ImageNet Classification with Deep Convolutional Neural Networks
The 2012 breakthrough by Krizhevsky(of AlexNet fame), Sutskever(Co-Founder at OpenAI), and Hinton(Godfather of AI) with AlexNet revolutionized AI. By using deep convolutional neural networks and leveraging GPUs for training, they achieved insane accuracy on ImageNet dataset. This not only validated deep learning's potential but also introduced key innovations like ReLU activations, dropout, and data augmentation.
Alex Krizhevsky, Ilya Sutskever and Geoffrey E. Hinton
https://proceedings.neurips.cc/paper_files/paper/2012/file/c399862d3b9d6b76c8436e924a68c45b-Paper.pdf
Discover More
Curated from across
Data Scientists on
by Babel
Yubi
The Research paper that changed the world...
Anything happening in AI today can be traced back to this one brief moment in history... Share your favourite papers. GPT ftw :)
https://proceedings.neurips.cc/paper_files/paper/2017/file/3f5ee243547dee91fbd053c1c4a845aa-Paper.pdf
Data Scientists on
by Gooner7
Goldman Sachs
The Gods of AI made "Generative Adversarial Networks"
Please give this post +1 like, I am getting demoralized when I see bad posts get 100s of likes but a good post like this gets no likes whatsoever. If I don't get more than 100 likes then I will stop posting research papers here. Ian Goodfellow and his collaborators published a groundbreaking paper that revolutionized the field of artificial intelligence, especially in generative modeling. This is the first time the world was introduced to Generative Adversarial Networks (GANs), a revolutionary framework that pits two neural networks against each other in a game-theoretic scenario. Unlike traditional generative models, GANs consist of a generator and a discriminator. The generator creates data samples, while the discriminator evaluates them, distinguishing between real and generated data. This novel approach not only significantly enhanced the quality and realism of generated data but also provided a robust framework for training generative models in an unsupervised manner. Ian Goodfellow, along with his colleagues, not only pushed the boundaries of generative modeling but also set the stage for a plethora of applications across various domains, from image synthesis and enhancement to data augmentation and beyond. Their innovative work demonstrated the potential of adversarial training in neural networks, opening new avenues in both theoretical research and practical applications in AI.
We propose a new framework for estimating generative models via an adversarial process, in which we simultaneously train two models: a gener...
https://arxiv.org/pdf/1406.2661
Data Scientists on
by Gooner7
Goldman Sachs
Neural Machine Translation by Jointly Learning to Align and Translate
If you want more papers like these, drop a "+1" comment below. I will notify these people in DMs next time I upload a new paper. Bahdanau, Cho, and Bengio published a pivotal paper that reshaped the landscape of artificial intelligence, particularly in NLP. This is the first time the world was introduced to the attention mechanism, the most important thing in modern neural machine translation systems. Unlike traditional approaches that relied solely on fixed-length vector representations, the attention mechanism allowed models to dynamically focus on different parts of the input sequence during the translation process. This breakthrough not only significantly improved the accuracy of translation but also enabled the handling of longer sentences with greater fluency. Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Bengio not only innovated in the field of machine translation forward but also laid the foundation for attention-based architectures across various domains of deep learning. Their innovative approach demonstrated the power of neural networks to tackle complex sequence-to-sequence tasks and opened doors to a new era of natural language understanding and generation.
Dzmitry Bahdanau, KyungHyun Cho and Yoshua Bengio
https://arxiv.org/pdf/1409.0473
Data Scientists on
by Gooner7
Goldman Sachs
Playing Atari with Deep Reinforcement Learning
DeepMind's paper "Playing Atari with Deep Reinforcement Learning" (2013) marked a milestone in AI by demonstrating how deep Q-learning could learn to play Atari games directly from pixels. This breakthrough highlighted the potential of deep RL in complex decision-making tasks and laid the foundation for subsequent advancements in reinforcement learning algorithms.
https://arxiv.org/pdf/1312.5602