img

The Research paper that changed the world...

Anything happening in AI today can be traced back to this one brief moment in history... Share your favourite papers. GPT ftw :)

https://proceedings.neurips.cc/paper_files/paper/2017/file/3f5ee243547dee91fbd053c1c4a845aa-Paper.pdf

img
img

Dihaadi

Stealth

4 months ago

img

kpmg_guy

KPMG

4 months ago

img

TheOatmeal

Stealth

4 months ago

img

BanRakas

Amazon

4 months ago

img

Jackietrader

Google

4 months ago

img

OnlySurvey5

Amazon

4 months ago

img

inr

Oracle

4 months ago

img

Dihaadi

Stealth

4 months ago

img

placehldr

Linkedin

4 months ago

img

rat_racer

Globallogic

4 months ago

img

batman_on_leave

Zomato

4 months ago

img

FireHawk

Javis

4 months ago

img

WhiteLotus23

Stealth

4 months ago

Sign in to a Grapevine account for the full experience.

Discover More

Curated from across

img

Data Scientists on

by Gooner7

Goldman Sachs

This Research Paper changed my life forever.

It was one of the papers that was discussed in my interview at Goldman. I came to know about this research paper a few years back after consulting a friend doing an ML PhD at University of Maryland, College Park. The explanation of the paper: 1. Initialize the neural network with small random values typically (-0.1,0.1) to avoid symmetry issues. 2. Now get ready to do Forward propagation: you pass thetraining data through the multilayer perceptron and compute the output. For each neuron in the MLP, calculate the weighted sum of its inputs and apply the activation function. (my favourite is tanh for LSTM applications) 3. Now compute the loss using a loss function like mean squared error, between output computed and the actual value. 4. Now get ready to do backpropagation, where you need to calculate the gradient of the loss function with respect to each weight by propagating the error backward through the network. 5. So, compute partial derivatives of the loss with respect to each weight, starting from the output layer and moving back to the input layer. 6. Here is the fun part: update the weights using the gradients obtained from the backward pass. here people usually use adam optimizer, which allows for accelerated stochastic gradient descent. Fun trivia: Adam stands for "Adaptive Moment Estimation". 7. Now repeat the forward and backward propagation process for numerous tries until theperformance of the model stabilizes.

https://www.iro.umontreal.ca/~vincentp/ift3395/lectures/backprop_old.pdf

img
img

Data Scientists on

by Welt

Remote

Recent research on Robotics and gen ai

Octo: Generalist Robot Policy (Gen AI in Robotics) Imagine a GPT for robots! That's what this paper is aiming for—a generalist robot that can perform a wide range of tasks. Right now, we have robots designed for specific, narrow tasks. This research is a step towards creating robots that can handle anything you throw at them What’s a Generalist Robot Policy? In simple terms, a policy is a neural network that decides what action a robot should take. It takes in observations (like text and images) and outputs control units. Types of Control Units 1. Joint Control: The model specifies exactly what each joint should do (like the angle of an elbow). 2. End Effector: This is a fancy name for the gripper. The model outputs movement and rotation vectors for it. The Process 1. Input: Task tokens (text), observation tokens (images), and readout tokens (a summary of task and observation tokens). 2. Output: Control units from the action head (a diffusion model). Challenges - Speed: Transformers are powerful but can be slow due to processing limitations. MLPs (Multi-Layer Perceptrons) can be faster but lack the generalization needed for different robots. - Data: Currently, the available dataset is limited. However, with access to more data, we can expect these policies to become much more efficient. In essence, this paper is pushing the boundaries in the field of robotics, aiming to create versatile robots capable of performing a wide array of tasks. link to the paper: https://arxiv.org/pdf/2405.12213 If you want me dive deeper,do let me know

img

Data Scientists on

by Gooner7

Goldman Sachs

The Gods of AI made "Generative Adversarial Networks"

Please give this post +1 like, I am getting demoralized when I see bad posts get 100s of likes but a good post like this gets no likes whatsoever. If I don't get more than 100 likes then I will stop posting research papers here. Ian Goodfellow and his collaborators published a groundbreaking paper that revolutionized the field of artificial intelligence, especially in generative modeling. This is the first time the world was introduced to Generative Adversarial Networks (GANs), a revolutionary framework that pits two neural networks against each other in a game-theoretic scenario. Unlike traditional generative models, GANs consist of a generator and a discriminator. The generator creates data samples, while the discriminator evaluates them, distinguishing between real and generated data. This novel approach not only significantly enhanced the quality and realism of generated data but also provided a robust framework for training generative models in an unsupervised manner. Ian Goodfellow, along with his colleagues, not only pushed the boundaries of generative modeling but also set the stage for a plethora of applications across various domains, from image synthesis and enhancement to data augmentation and beyond. Their innovative work demonstrated the potential of adversarial training in neural networks, opening new avenues in both theoretical research and practical applications in AI.

We propose a new framework for estimating generative models via an adversarial process, in which we simultaneously train two models: a gener...

https://arxiv.org/pdf/1406.2661

img