Just to add to this, I also need to know what language is the "nnet"
package written in? Is it in pure R or is it a wrapper for a C
library. It really is performing very quickly, going through 200
epochs in seconds when it took "neural" minutes, and neural is
written in R.
It all may sound like a trivial question, but it is important as I
need to know all this for the analysis I'm doing for a paper.
On 22 Nov 2006, at 15:41, Wee-Jin Goh wrote:
> Greetings list,
> I've just swapped from the "neural" package to the "nnet" package and
> I've noticed that the training is orders of magnitude faster, and the
> results are way more accurate.
> This leads me to wonder, what training algorithm is "nnet" using? Is
> it a modification on the standard backpropagation? Or a completely
> different algorithm? I'm trying to account for the speed differences
> between neural and nnet, and the documentation on the nnet package is
> rather sparse on what training algorithm is used (either that, or I'm
> getting blind and missed it totally).
> Any help would be much appreciated.
> R-help@stat... mailing list
> https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guide http://www.R-project.org/posting- > guide.html
> and provide commented, minimal, self-contained, reproducible code.