Hello, There isn't one yet. But I will try to create it today. I will let you know
Cheers, Oleks On Apr 25, 2017 16:10, "francescoagati [via Smalltalk]" < ml+s1294792n494402...@n4.nabble.com> wrote: > Hi Oleks, > there is a mode for install neural network from metacello? > > 2017-04-25 13:00 GMT+02:00 Alexandre Bergel <[hidden email] > <http:///user/SendEmail.jtp?type=node&node=4944025&i=0>>: > >> Continue to push that topic Oleks. You are on the right track! >> >> Alexandre >> >> > On Apr 24, 2017, at 1:43 AM, Oleks <[hidden email] >> <http:///user/SendEmail.jtp?type=node&node=4944025&i=1>> wrote: >> > >> > Hello, >> > >> > Thanks a lot for your advice! It was very helpful and educating (for >> > example, I thought that we store biases in the weight matrix and >> prepend 1 >> > to input to make it faster, but now I see why it's actually slower that >> > way). >> > >> > I've implemented a multi-layer neural network as a linked list of layers >> > that propagate the input and error from one to another, similar to the >> Chain >> > of Responsibility pattern. Also, now I represent biases as separate >> vectors. >> > The LearningAlgorithm is a separate class with Backpropagation as its >> > subclass (though at this point the network can only learn through >> > backpropagation, but I'm planning to change that). I'm trying to figure >> out >> > how the activation and cost functions should be connected. For example, >> > cross-entropy works best with logistic sigmoid activation etc. I would >> like >> > to give the user a freedom to use whatever he wants (plug in whatever >> you >> > like and see what happens), but it can be very inefficient (because some >> > time-consuming parts of activation and cost derivatives cancel out each >> > other). >> > >> > Also, there is an interface for setting the learning rate for the whole >> > network, which can be used to choose the learning rate prior to >> learning, as >> > well as to change the learning rate after each iteration. I am planning >> to >> > implement some optimization algorithms that would automize the process >> of >> > choosing a learning rate (adagrad for example), but this would require >> a bit >> > different design (maybe I will implement the Optimizer, as you >> suggested). >> > >> > I'm attaching two images with UML diagrams, describing my current >> > implementation. Could you please tell me what you think about this >> design? >> > The first image is a class diagram that shows the whole architecture, >> and >> > the second one is a sequence diagram of backpropagation. >> > >> > mlnn.png <http://forum.world.st/file/n4943698/mlnn.png> >> > backprop.png <http://forum.world.st/file/n4943698/backprop.png> >> > >> > Sincerely yours, >> > Oleksandr >> > >> > >> > >> > -- >> > View this message in context: http://forum.world.st/Neural-N >> etworks-in-Pharo-tp4941271p4943698.html >> > Sent from the Pharo Smalltalk Users mailing list archive at Nabble.com. >> > >> >> -- >> _,.;:~^~:;._,.;:~^~:;._,.;:~^~:;._,.;:~^~:;._,.;: >> Alexandre Bergel http://www.bergel.eu >> ^~:;._,.;:~^~:;._,.;:~^~:;._,.;:~^~:;._,.;:~^~:;. >> >> >> >> >> > > > ------------------------------ > If you reply to this email, your message will be added to the discussion > below: > http://forum.world.st/Neural-Networks-in-Pharo-tp4941271p4944025.html > To unsubscribe from Neural Networks in Pharo, click here > <http://forum.world.st/template/NamlServlet.jtp?macro=unsubscribe_by_code&node=4941271&code=b2xrLnpheXRzZXZAZ21haWwuY29tfDQ5NDEyNzF8MTQxNTM2ODEzNw==> > . > NAML > <http://forum.world.st/template/NamlServlet.jtp?macro=macro_viewer&id=instant_html%21nabble%3Aemail.naml&base=nabble.naml.namespaces.BasicNamespace-nabble.view.web.template.NabbleNamespace-nabble.view.web.template.NodeNamespace&breadcrumbs=notify_subscribers%21nabble%3Aemail.naml-instant_emails%21nabble%3Aemail.naml-send_instant_email%21nabble%3Aemail.naml> > -- View this message in context: http://forum.world.st/Neural-Networks-in-Pharo-tp4941271p4944027.html Sent from the Pharo Smalltalk Users mailing list archive at Nabble.com.