Pagini recente » Diferente pentru blog/idei-proaste intre reviziile 3 si 4 | Diferente pentru blog/lansare-mindcoding-2017 intre reviziile 9 si 8 | Diferente pentru blog/finala-algoritmiada-2015 intre reviziile 1 si 2 | Diferente pentru blog/idei-proaste intre reviziile 5 si 6 | Diferente pentru blog/deep-learning-alien-labs intre reviziile 5 si 6
Nu exista diferente intre titluri.
Diferente intre continut:
Here are some of my notes on deep learning:
Anecdotes:
*Anecdotes*:
* Google voice was unusable for me a few years ago, now it gets my bad English accent.
* DeepMind f*king solved Go. A game that 2 years ago was thought to be 10 years out of reach.
* Baidu is working on speech recognition for mandarin (lots of illiterate people with phones).
* Word embeddings give you natural language models that do away with the tweaks one used to encode all sorts of idiosyncrasies in the English language.
OpenAI CTO: "As Ilya likes to say, deep learning is a shallow field — it's actually relatively easy to pick up and start making contributions.
OpenAI CTO: "As Ilya likes to say, deep learning is a *shallow field* — it's actually relatively easy to pick up and start making contributions.
* short learning curve, deep learning techniques started to be effective recently (2010? 2012?)
* previous experience in the old techniques is not very relevant
* state of the art results significantly improving on the previous state of the art
(in real time object detection, speech recognition)
Interesting concepts:
Word2Vec - neat idea that maps words to points in n dimensional space. Then you can do algebra on the vector representation:
vec(“king”) – vec(“man”) + vec(“woman”) =~ vec(“queen”), or vec(“Montreal Canadiens”) – vec(“Montreal”) + vec(“Toronto”) resembles the vector for “Toronto Maple Leafs”
Mikolov, Tomas; Sutskever, Ilya; Chen, Kai; Corrado, Greg S.; Dean, Jeff (2013). Distributed representations of words and phrases and their compositionality.
Back propagation - (I see it as a dynamic programming technique that works well for the neural network setup to compute partial derivatives one needs when running gradient descent)
Convolutional neural networks - a convolutional layer reuses the same k weights instead of having k^2 weights between 2 layers (this concept makes sense in image input and makes algorithms much faster)
Rectified Linear Units f(x) = max(0, x) work much better than the historically used sigmoid and hyperbolic tangent functions
Dropout - some neurons are ignored with a set probability, inspired by how the neurons in the brain fire with some pro
*Interesting concepts*:
* *Word2Vec* - neat idea that maps words to points in n dimensional space. Then you can do algebra on the vector representation:
vec(“king”) – vec(“man”) + vec(“woman”) =~ vec(“queen”), or vec(“Montreal Canadiens”) – vec(“Montreal”) + vec(“Toronto”) resembles the vector for “Toronto Maple Leafs” (Mikolov, Tomas; Sutskever, Ilya; Chen, Kai; Corrado, Greg S.; Dean, Jeff (2013). Distributed representations of words and phrases and their compositionality)
* *Back propagation* - I see it as a dynamic programming technique that works well for the neural network setup to compute partial derivatives one needs when running gradient descent
* *Convolutional neural networks* - a convolutional layer reuses the same k weights instead of having k^2 weights between 2 layers (this concept makes sense in image input and makes algorithms much faster)
* *Rectified Linear Units* f(x) = max(0, x) work much better than the historically used sigmoid and hyperbolic tangent functions
* *Dropout* - some neurons are ignored with a set probability, inspired by how the neurons in the brain fire with some pro
Mircea Pasoi and Cristian Strat, after their successful stint at twitter, recently founded Alien Labs. They use deep learning to build intelligent chat bots for an office environment. This is an awesome opportunity to work together again. It's been 8 years since we last did. At Google, we worked on an ads inventory management problem using 'network flows':http://goo.gl/5SDRrM Our claim to fame is that we got help from Cliff Stein, the S in 'CLRS':https://mitpress.mit.edu/books/introduction-algorithms :). This is also an opportunity for me to jump on the deep learning train tackling real world problems.
Nu exista diferente intre securitate.
Topicul de forum nu a fost schimbat.