Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Intiallizing the weight matrix #3

Closed
vanrajbrown opened this issue Feb 5, 2019 · 4 comments
Closed

Intiallizing the weight matrix #3

vanrajbrown opened this issue Feb 5, 2019 · 4 comments

Comments

@vanrajbrown
Copy link

How are you initializing weight matrix in multiclass_classification_gpu.py line number 83. Although when avoiding vanishing and exploding gradient, we need to use np.random.randn(size_l, size_l-1). Can you please explain the numbers you are using - tf.truncated_normal([11,11,3,96], stddev=0.01).
Thanks

@vanrajbrown
Copy link
Author

Okay, I got the reason as to why are you using these number, these are from AlexNet Architecture. But do you have any write up written on this, as to why are you using what you are on each step. That would be so helpful, I have lot more questions, will it be fine if I post them under the issues page?

@MuhammedBuyukkinaci
Copy link
Owner

Sorry for being late. I tried to implement AlexNet Paper. The hyperparameters used in the model are from that paper. You are highly welcome to ask your questions here. I will try to answer them.

@vanrajbrown
Copy link
Author

Thanks, I Figured out the Alex net Architecture. I have another question, why the value of Bias constant varies either 0 or 1. How do you decide which layer bias value as 1 or 0 ?

@MuhammedBuyukkinaci
Copy link
Owner

I just got this information from CS224 of Andrej Karpathy on youtube.com .

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants