recentpopularlog in

pozorvlak : machinelearning   132

« earlier  
Dark knowledge
Use "soft targets" (temperature-smoothed average predictions from ensemble models) to train smaller summary models - they carry much of the information from the teacher models.

Random dropout can achieve a similar effect more cheaply.

Ensembles-of-specialists can perform well, but must be combined with care.
computers  machinelearning  deeplearning  ai  google 
8 days ago by pozorvlak
Out of shape? Why deep learning works differently than we thought
Current deep learning techniques for object recognition primarily rely on textures, not on object shapes.
computers  programming  ai  machinelearning  deeplearning  science  computervision 
13 days ago by pozorvlak
[1611.03530] Understanding deep learning requires rethinking generalization
Specifically, our experiments establish that state-of-the-art convolutional networks for image classification trained with stochastic gradient methods easily fit a random labeling of the training data. This phenomenon is qualitatively unaffected by explicit regularization, and occurs even if we replace the true images by completely unstructured random noise. We corroborate these experimental findings with a theoretical construction showing that simple depth two neural networks already have perfect finite sample expressivity as soon as the number of parameters exceeds the number of data points as it usually does in practice. 
computers  machinelearning  deeplearning  badscience 
7 weeks ago by pozorvlak
Togelius: Empiricism and the limits of gradient descent
Evolutionary algorithms might be able to learn things that gradient descent can't; a tortured analogy claiming (gradient descent:evolutionary algorithms)::(strict empiricism:Popperian hypothetico-deductivism).
machinelearning  deeplearning  ai  evolution  algorithms  philosophy 
june 2018 by pozorvlak
Generative Adversarial Networks (GANs) in 50 lines of code (PyTorch)
In 2014, Ian Goodfellow and his colleagues at the University of Montreal published a stunning paper introducing the world to GANs, or generative adversarial networks. Through an innovative…
ai  machinelearning  deeplearning  computers  python  programming 
march 2018 by pozorvlak
Copista: Training models for TensorFlow Mobile – Andrew G – Medium
For those who missed the first part Copista: Developing Neural Style Transfer application with TensorFlow Mobile, this blog is a software engineer take on Machine Learning. In this part, you can find…
android  computers  programming  machinelearning  deeplearning 
november 2017 by pozorvlak
Copista: Developing Neural Style Transfer application with TensorFlow Mobile
I should confess that I am not a Data Scientist :). It’s my take on Machine Learning as a Software Engineer. Its all started when I came across Pete Warden’s blog TensorFlow for Mobile Poets in the…
machinelearning  android  deeplearning  computers  programming 
november 2017 by pozorvlak
Zayd's Blog – Why is machine learning 'hard'?
In conclusion, fast and effective debugging is the skill that is most required for implementing modern day machine learning pipelines.
computers  programming  debugging  machinelearning  datascience 
march 2017 by pozorvlak
Building Jarvis
Mark Zuckerberg's home-automation AI system.
facebook  ai  machinelearning  music  language  lifeimitatessf 
january 2017 by pozorvlak
« earlier      
per page:    204080120160

Copy this bookmark:





to read