thanks ;-)

2017-04-25 15:09 GMT+02:00 Oleks <olk.zayt...@gmail.com>:

> Hello,
>
> There isn't one yet. But I will try to create it today. I will let you know
>
> Cheers,
> Oleks
>
> On Apr 25, 2017 16:10, "francescoagati [via Smalltalk]" <[hidden email]
> <http:///user/SendEmail.jtp?type=node&node=4944027&i=0>> wrote:
>
>> Hi Oleks,
>> there is a mode for install neural network from metacello?
>>
>> 2017-04-25 13:00 GMT+02:00 Alexandre Bergel <[hidden email]
>> <http:///user/SendEmail.jtp?type=node&node=4944025&i=0>>:
>>
>>> Continue to push that topic Oleks. You are on the right track!
>>>
>>> Alexandre
>>>
>>> > On Apr 24, 2017, at 1:43 AM, Oleks <[hidden email]
>>> <http:///user/SendEmail.jtp?type=node&node=4944025&i=1>> wrote:
>>> >
>>> > Hello,
>>> >
>>> > Thanks a lot for your advice! It was very helpful and educating (for
>>> > example, I thought that we store biases in the weight matrix and
>>> prepend 1
>>> > to input to make it faster, but now I see why it's actually slower that
>>> > way).
>>> >
>>> > I've implemented a multi-layer neural network as a linked list of
>>> layers
>>> > that propagate the input and error from one to another, similar to the
>>> Chain
>>> > of Responsibility pattern. Also, now I represent biases as separate
>>> vectors.
>>> > The LearningAlgorithm is a separate class with Backpropagation as its
>>> > subclass (though at this point the network can only learn through
>>> > backpropagation, but I'm planning to change that). I'm trying to
>>> figure out
>>> > how the activation and cost functions should be connected. For example,
>>> > cross-entropy works best with logistic sigmoid activation etc. I would
>>> like
>>> > to give the user a freedom to use whatever he wants (plug in whatever
>>> you
>>> > like and see what happens), but it can be very inefficient (because
>>> some
>>> > time-consuming parts of activation and cost derivatives cancel out each
>>> > other).
>>> >
>>> > Also, there is an interface for setting the learning rate for the whole
>>> > network, which can be used to choose the learning rate prior to
>>> learning, as
>>> > well as to change the learning rate after each iteration. I am
>>> planning to
>>> > implement some optimization algorithms that would automize the process
>>> of
>>> > choosing a learning rate (adagrad for example), but this would require
>>> a bit
>>> > different design (maybe I will implement the Optimizer, as you
>>> suggested).
>>> >
>>> > I'm attaching two images with UML diagrams, describing my current
>>> > implementation. Could you please tell me what you think about this
>>> design?
>>> > The first image is a class diagram that shows the whole architecture,
>>> and
>>> > the second one is a sequence diagram of backpropagation.
>>> >
>>> > mlnn.png <http://forum.world.st/file/n4943698/mlnn.png>
>>> > backprop.png <http://forum.world.st/file/n4943698/backprop.png>
>>> >
>>> > Sincerely yours,
>>> > Oleksandr
>>> >
>>> >
>>> >
>>> > --
>>> > View this message in context: http://forum.world.st/Neural-N
>>> etworks-in-Pharo-tp4941271p4943698.html
>>> > Sent from the Pharo Smalltalk Users mailing list archive at Nabble.com.
>>> >
>>>
>>> --
>>> _,.;:~^~:;._,.;:~^~:;._,.;:~^~:;._,.;:~^~:;._,.;:
>>> Alexandre Bergel  http://www.bergel.eu
>>> ^~:;._,.;:~^~:;._,.;:~^~:;._,.;:~^~:;._,.;:~^~:;.
>>>
>>>
>>>
>>>
>>>
>>
>>
>> ------------------------------
>> If you reply to this email, your message will be added to the discussion
>> below:
>> http://forum.world.st/Neural-Networks-in-Pharo-tp4941271p4944025.html
>> To unsubscribe from Neural Networks in Pharo, click here.
>> NAML
>> <http://forum.world.st/template/NamlServlet.jtp?macro=macro_viewer&id=instant_html%21nabble%3Aemail.naml&base=nabble.naml.namespaces.BasicNamespace-nabble.view.web.template.NabbleNamespace-nabble.view.web.template.NodeNamespace&breadcrumbs=notify_subscribers%21nabble%3Aemail.naml-instant_emails%21nabble%3Aemail.naml-send_instant_email%21nabble%3Aemail.naml>
>>
>
> ------------------------------
> View this message in context: Re: Neural Networks in Pharo
> <http://forum.world.st/Neural-Networks-in-Pharo-tp4941271p4944027.html>
>
> Sent from the Pharo Smalltalk Users mailing list archive
> <http://forum.world.st/Pharo-Smalltalk-Users-f1310670.html> at Nabble.com.
>

Reply via email to