[erlang-questions] Machine Learning

lloyd@REDACTED lloyd@REDACTED
Sun Mar 6 00:12:31 CET 2016


Hello,

I can't claim anything but the shallowest machine learning chops beyond cursory exploration of the literature and what my data scientist son has taught me. But it seems like the first and foremost consideration in determining whether or not to proceed with a machine learning project in Erlang is careful consideration of the problem you're trying to solve.

My sense is that, as Alex points out, Erlang may work quite well for problems in the Machine Learning 101  domain; e.g. limited number of perceptron elements. How do we define limited? Exactly why it may be worthwhile to experiment. Certainly Gene Shor has done interesting work. Does anyone know of worthy follow-up?

That said, big-data machine learning isn't the only game in town. This piece throws down a significant challenge:

Neural modelling: Abstractions of the mind
http://www.nature.com/nature/journal/v531/n7592_supp/full/531S16a.html

All the best,

LRP

-----Original Message-----
From: "Alex Alvarez" <eajam@REDACTED>
Sent: Saturday, March 5, 2016 4:58pm
To: erlang-questions@REDACTED
Subject: Re: [erlang-questions] Machine Learning

_______________________________________________
erlang-questions mailing list
erlang-questions@REDACTED
http://erlang.org/mailman/listinfo/erlang-questions
Sorry for my late response to this topic, but I do believe Erlang is 
actually a great language for the ML and statistics space.  Take a basic 
feed-forward NN with back-propagation, for example.  What you'd normally 
have in terms of mathematical computation is mainly addition and 
multiplication.  You only need to put together a perceptron, which are 
only inputs (including a bias) multiplied by respective weights, you add 
them up and pass this value through a function like sigmoid or 
hyperbolic tangent and that's that. Back-propagation, as a way to adjust 
the weights during the training phase, doesn't require math-wise 
anything more complicated.  You combine the perceptrons for the hidden 
and output layers and you got yourself a NN.  In this configuration, 
deep learning will simply be two or more hidden layers, instead of one.  
The key thing to maximize the use of Erlang is certainly to distribute 
the load through processes, so each perception could be one individual 
process, for example.  Definitely, not rocket science.  Now, I concur 
that in some situations it might be advantageous to write a module, say 
for a perception, in C and make use of it within Erlang, but there is no 
reason why you couldn't start with Erlang and gradually move to that 
direction, if need be.

Cheers,
Alex


On 02/10/2016 06:31 AM, Jesper Louis Andersen wrote:
>
> On Wed, Feb 10, 2016 at 10:34 AM, Samuel <samuelrivas@REDACTED 
> <mailto:samuelrivas@REDACTED>> wrote:
>
>     I am not aware of existing ML or linear algrebra libraries in erlang
>     that can be used to quick start an ML project, but just piping into
>     tensorflow (or any other existing library/framework) isn't really
>     doing ML with erlang, is it? You can as well just use tensorflow
>     directly.
>
>
> The question is if this is a practical problem which needs solving or 
> it is for research. If you are researching how to construct, say, SVM 
> or NNs, then surely Erlang is a good vehicle. But in practice, there 
> is a number of things which makes Erlang unsuitable:
>
> * ML is often CPU bound. You don't want a bytecode interpreter to be a 
> limiting factor here. Even if the interpreter in Erlang is 
> state-of-the-art and highly optimized, it is not far fetched that a 
> FP-intensive program will be roughly a factor of 30 faster if compiled 
> in a lower level language.
>
> * GPUs are popular in ML models for a reason: they speed up the FP 
> computations by a factor of 3000 or more. This in itself should hint 
> you that you need something else than Erlang.
>
> * Erlangs word overhead per process and terms means a lower-level 
> model can pack many more entities in memory. This affects caching 
> behavior.
>
> Training of the model is often off-line and using the model is online 
> in the system. How you train your model is less important. This is why 
> I'd just outsource this problem to the libraries built and tuned for 
> it. It is like solving LinAlg problems but forgetting everything about 
> existing LAPACK and ATLAS routines in Fortran.
>
> A model, in Erlang, which could be viable is to use Erlang to produce 
> programs for lower level consumption by compilation. But these are 
> problems for which languages such as Haskell and OCaml dominates for a 
> reason: their type systems makes it far easier to pull off.
>
>
>
> -- 
> J.
>
>
> _______________________________________________
> erlang-questions mailing list
> erlang-questions@REDACTED
> http://erlang.org/mailman/listinfo/erlang-questions






More information about the erlang-questions mailing list