Welcome to Incels.is - Involuntary Celibate Forum

Welcome! This is a forum for involuntary celibates: people who lack a significant other. Are you lonely and wish you had someone in your life? You're not alone! Join our forum and talk to people just like you.

Implemented dropout, starting convnets

G

gstvtrp

Banned
-
Joined
Nov 7, 2017
Posts
2,281
My Deeplearning skills are growing every day.
 
Nice. Do you prefer deep learning over machine learning?
 
Nice. Do you prefer deep learning over machine learning?

kinda yeah even though DL is a subset of ML. I believe in connectionist approaches overall. I don't think there is any other path to AGI.
 
I spent like 3 hours last week trying to understand back propagation lol
 
I spent like 3 hours last week trying to understand back propagation lol

You get used to it. Try watching the 3blue1brown videos on neural networks.

Basically, we have a loss function right? We want to minimize the loss. If we imagine the loss as a hilly terrain, then we want to find the low points in the terrain (local minima). To find the direction we should move in, we need to know the gradient of the loss with respect to the parameters. This is where the chain rule comes in. As we backprop we get the gradients for each layer.

Then with SGD, you simply update the parameters W of a layer by W := W - learning_rate * dW.
 
You get used to it. Try watching the 3blue1brown videos on neural networks.

Basically, we have a loss function right? We want to minimize the loss. If we imagine the loss as a hilly terrain, then we want to find the low points in the terrain (local minima). To find the direction we should move in, we need to know the gradient of the loss with respect to the parameters. This is where the chain rule comes in. As we backprop we get the gradients for each layer.

Then with SGD, you simply update the parameters W of a layer by W := W - learning_rate * dW.

I used to watch everything from 3blue1brown, he puts a lot of effort.

I understood it a little better after his videos and a few from udacity, I guess I had a few problems with the backprop calc
 
I used to watch everything from 3blue1brown, he puts a lot of effort.

I understood it a little better after his videos and a few from udacity, I guess I had a few problems with the backprop calc

Keep at it man. When I first tried learning backprop, I was like wtf? Now it makes complete sense to me but it took a while. You should try the Coursera course by Andrew Ng. Even in that course Ng says sometimes he felt like he didn't understand backprop. Of course by now he probably does understand it completely. I think this is a very common thing among beginner ML practitioners.
 
LOL I just read the title, I meant to say implemented DROPOUT lol. Of course I've implemented backprop before.

First time I've done Dropout. JFL.
 
Keep at it man. When I first tried learning backprop, I was like wtf? Now it makes complete sense to me but it took a while. You should try the Coursera course by Andrew Ng. Even in that course Ng says sometimes he felt like he didn't understand backprop. Of course by now he probably does understand it completely. I think this is a very common thing among beginner ML practitioners.
yup, was planning to after I refreshed on linear algebra
 
tfw not high iq programmer. tfw rarted laborer
Peep
 
I spent like 3 hours last week trying to understand back propagation lol

My professor made me memorize the entire equation. I did, but now I've forgotten all of it.

Nice. Do you prefer deep learning over machine learning?

Deep learning is the best approach as a whole but sucks to program/work with. Machine learning makes for more enjoyable programming since you are the one doing the thinking and finding patterns, not your computer. I think deep learning will be what most programs use in the future which is unfortunate imo.
 
fuarkk, I implemented forward and backward passes for a convolutional layer. That was pretty tough. I think that's one of the tougher things in the whole class.
 
fuarrr, I'm getting even further. Implemented a 3-layer convnet and also spatial batchnorm. Almost done with assignment 2!

 
A technique where some neurons are probabilistically turned off during training time. This has the effect of helping a neural network generalize better.
i still dont understand kek
 

Similar threads

AutistSupremacist
Replies
6
Views
181
Grodd
Grodd
SuperKanga.Belgrade
2
Replies
90
Views
954
SuperKanga.Belgrade
SuperKanga.Belgrade
TheDragon
Replies
21
Views
444
PolskiKartofel
PolskiKartofel
Friezacel
Replies
6
Views
144
Grodd
Grodd
Left4DeadKoala
Replies
28
Views
519
based_meme
B

Users who are viewing this thread

shape1
shape2
shape3
shape4
shape5
shape6
Back
Top