now work all good with last version of polymath thanks :-) 2017-04-27 17:29 GMT+02:00 Oleks <olk.zayt...@gmail.com>:
> Hello, > > I have finally added a configuration to the NeuralNetwork project. Now you > can use this Metacello script to load it into your Pharo image: > > Metacello new > repository: 'http://smalltalkhub.com/mc/Oleks/NeuralNetwork/main'; > configuration: 'MLNeuralNetwork'; > version: #development; > load. > > Sorry for the delay > > Oleks > > On Tue, Apr 25, 2017 at 4:13 PM, francescoagati [via Smalltalk] <[hidden > email] <http:///user/SendEmail.jtp?type=node&node=4944473&i=0>> wrote: > >> thanks ;-) >> >> 2017-04-25 15:09 GMT+02:00 Oleks <[hidden email] >> <http:///user/SendEmail.jtp?type=node&node=4944028&i=0>>: >> >>> Hello, >>> >>> There isn't one yet. But I will try to create it today. I will let you >>> know >>> >>> Cheers, >>> Oleks >>> >>> On Apr 25, 2017 16:10, "francescoagati [via Smalltalk]" <[hidden email] >>> <http:///user/SendEmail.jtp?type=node&node=4944027&i=0>> wrote: >>> >>>> Hi Oleks, >>>> there is a mode for install neural network from metacello? >>>> >>>> 2017-04-25 13:00 GMT+02:00 Alexandre Bergel <[hidden email] >>>> <http:///user/SendEmail.jtp?type=node&node=4944025&i=0>>: >>>> >>>>> Continue to push that topic Oleks. You are on the right track! >>>>> >>>>> Alexandre >>>>> >>>>> > On Apr 24, 2017, at 1:43 AM, Oleks <[hidden email] >>>>> <http:///user/SendEmail.jtp?type=node&node=4944025&i=1>> wrote: >>>>> > >>>>> > Hello, >>>>> > >>>>> > Thanks a lot for your advice! It was very helpful and educating (for >>>>> > example, I thought that we store biases in the weight matrix and >>>>> prepend 1 >>>>> > to input to make it faster, but now I see why it's actually slower >>>>> that >>>>> > way). >>>>> > >>>>> > I've implemented a multi-layer neural network as a linked list of >>>>> layers >>>>> > that propagate the input and error from one to another, similar to >>>>> the Chain >>>>> > of Responsibility pattern. Also, now I represent biases as separate >>>>> vectors. >>>>> > The LearningAlgorithm is a separate class with Backpropagation as its >>>>> > subclass (though at this point the network can only learn through >>>>> > backpropagation, but I'm planning to change that). I'm trying to >>>>> figure out >>>>> > how the activation and cost functions should be connected. For >>>>> example, >>>>> > cross-entropy works best with logistic sigmoid activation etc. I >>>>> would like >>>>> > to give the user a freedom to use whatever he wants (plug in >>>>> whatever you >>>>> > like and see what happens), but it can be very inefficient (because >>>>> some >>>>> > time-consuming parts of activation and cost derivatives cancel out >>>>> each >>>>> > other). >>>>> > >>>>> > Also, there is an interface for setting the learning rate for the >>>>> whole >>>>> > network, which can be used to choose the learning rate prior to >>>>> learning, as >>>>> > well as to change the learning rate after each iteration. I am >>>>> planning to >>>>> > implement some optimization algorithms that would automize the >>>>> process of >>>>> > choosing a learning rate (adagrad for example), but this would >>>>> require a bit >>>>> > different design (maybe I will implement the Optimizer, as you >>>>> suggested). >>>>> > >>>>> > I'm attaching two images with UML diagrams, describing my current >>>>> > implementation. Could you please tell me what you think about this >>>>> design? >>>>> > The first image is a class diagram that shows the whole >>>>> architecture, and >>>>> > the second one is a sequence diagram of backpropagation. >>>>> > >>>>> > mlnn.png <http://forum.world.st/file/n4943698/mlnn.png> >>>>> > backprop.png <http://forum.world.st/file/n4943698/backprop.png> >>>>> > >>>>> > Sincerely yours, >>>>> > Oleksandr >>>>> > >>>>> > >>>>> > >>>>> > -- >>>>> > View this message in context: http://forum.world.st/Neural-N >>>>> etworks-in-Pharo-tp4941271p4943698.html >>>>> > Sent from the Pharo Smalltalk Users mailing list archive at >>>>> Nabble.com. >>>>> > >>>>> >>>>> -- >>>>> _,.;:~^~:;._,.;:~^~:;._,.;:~^~:;._,.;:~^~:;._,.;: >>>>> Alexandre Bergel http://www.bergel.eu >>>>> ^~:;._,.;:~^~:;._,.;:~^~:;._,.;:~^~:;._,.;:~^~:;. >>>>> >>>>> >>>>> >>>>> >>>>> >>>> >>>> >>>> ------------------------------ >>>> If you reply to this email, your message will be added to the >>>> discussion below: >>>> http://forum.world.st/Neural-Networks-in-Pharo-tp4941271p4944025.html >>>> To unsubscribe from Neural Networks in Pharo, click here. >>>> NAML >>>> <http://forum.world.st/template/NamlServlet.jtp?macro=macro_viewer&id=instant_html%21nabble%3Aemail.naml&base=nabble.naml.namespaces.BasicNamespace-nabble.view.web.template.NabbleNamespace-nabble.view.web.template.NodeNamespace&breadcrumbs=notify_subscribers%21nabble%3Aemail.naml-instant_emails%21nabble%3Aemail.naml-send_instant_email%21nabble%3Aemail.naml> >>>> >>> >>> ------------------------------ >>> View this message in context: Re: Neural Networks in Pharo >>> <http://forum.world.st/Neural-Networks-in-Pharo-tp4941271p4944027.html> >>> >>> Sent from the Pharo Smalltalk Users mailing list archive >>> <http://forum.world.st/Pharo-Smalltalk-Users-f1310670.html> at >>> Nabble.com. >>> >> >> >> >> ------------------------------ >> If you reply to this email, your message will be added to the discussion >> below: >> http://forum.world.st/Neural-Networks-in-Pharo-tp4941271p4944028.html >> To unsubscribe from Neural Networks in Pharo, click here. >> NAML >> <http://forum.world.st/template/NamlServlet.jtp?macro=macro_viewer&id=instant_html%21nabble%3Aemail.naml&base=nabble.naml.namespaces.BasicNamespace-nabble.view.web.template.NabbleNamespace-nabble.view.web.template.NodeNamespace&breadcrumbs=notify_subscribers%21nabble%3Aemail.naml-instant_emails%21nabble%3Aemail.naml-send_instant_email%21nabble%3Aemail.naml> >> > > > ------------------------------ > View this message in context: Re: Neural Networks in Pharo > <http://forum.world.st/Neural-Networks-in-Pharo-tp4941271p4944473.html> > Sent from the Pharo Smalltalk Users mailing list archive > <http://forum.world.st/Pharo-Smalltalk-Users-f1310670.html> at Nabble.com. >