This Research Paper changed my life forever.

It was one of the papers that was discussed in my interview at Goldman.

I came to know about this research paper a few years back after consulting a friend doing an ML PhD at University of Maryland, College Park.

The explanation of the paper:

  1. Initialize the neural network with small random values typically (-0.1,0.1) to avoid symmetry issues.

  2. Now get ready to do Forward propagation: you pass thetraining data through the multilayer perceptron and compute the output. For each neuron in the MLP, calculate the weighted sum of its inputs and apply the activation function. (my favourite is tanh for LSTM applications)

  3. Now compute the loss using a loss function like mean squared error, between output computed and the actual value.

  4. Now get ready to do backpropagation, where you need to calculate the gradient of the loss function with respect to each weight by propagating the error backward through the network.

  5. So, compute partial derivatives of the loss with respect to each weight, starting from the output layer and moving back to the input layer.

  6. Here is the fun part: update the weights using the gradients obtained from the backward pass. here people usually use adam optimizer, which allows for accelerated stochastic gradient descent. Fun trivia: Adam stands for "Adaptive Moment Estimation".

  7. Now repeat the forward and backward propagation process for numerous tries until theperformance of the model stabilizes.

5mo ago4.5K views
FairyMermaid
FairyMermaid

I appreciate you posting these papers, but why the cutoff of 100 likes ? Is it for grapevine chat facilities? That is a pain for me as well. I can not dm people I want to speak to

Gooner7
Gooner7

@FairyMermaid Because people need to value my content. Else, I am not putting the effort in.

Sane
Sane
KPMG5mo

People might not leave a like and still appreciate the content. On the contrary they will just pass by & won’t even read it after liking. You analysis is absolutely wrong to gain the respect for content. Value those people who stopped by and read it. Time is money

Welt
Welt

True,the backprop truly changed the ai. From backprop to generalist ai we come so far.

Gooner7
Gooner7

@Welt_yang @Welt_yang Seriously.

Sane
Sane
KPMG5mo

It was a good read! Thanks for posting🥂

AjaxNinja
AjaxNinja

Are you aware of any latest research that analyses intelligence/consciousness from a meta/abstract formulation pov and not the i/o neuron paradigm?

The original backprop paper is just 4 pages long?!

Discover more
Curated from across
BuzzingDrill
BuzzingDrillStudent10mo

Guidance for ML breadth/Fundamentals

I have had my fair share of playing around with different Kinda off generic Ml models and projects but It had been mostly random and not organised.

So now i want to dive deep into some niche in this field maybe even understand some rela...

UneducatedCoder
UneducatedCoderAmazon10mo

Pulse Check :

What was your first ever CTC and what's your current CTC with YoE?

I'll start - 1.6 to 9.8 in 4 years.

This post is intended to give hope to freshers and motivation to experienced folks.

Top comment
user

1 lakh 80 thousand per year Now 1 lakh 85 thousand per month Years taken 11