Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

some typos in ch 2 #7

Open
murphyk opened this issue Jan 5, 2023 · 2 comments
Open

some typos in ch 2 #7

murphyk opened this issue Jan 5, 2023 · 2 comments

Comments

@murphyk
Copy link

murphyk commented Jan 5, 2023

-p47 you write W_{d,c} instead of W^{d,c}

  • p46 Your comment about Y_pred.ravel() could have been made earlier on p40 where it was first introduced
@murphyk
Copy link
Author

murphyk commented Jan 5, 2023

Also on p56-p57 there is an inconsistency between y-hat meaning logits (colored equation for loss(sm(yhat), y)) and y-hat meaning probabilities (code snippet, y_hat = F.softmax(logits)). Maybe call the latter p_hat?

@yebangyu
Copy link

yebangyu commented May 25, 2023

Wow,murphyk is also reading this book ? so cool

I am confused about the notation, too. on P56 there is a:

loss(yhat,y) = -log(sm(yhat), y))

accordinng to the text, yhat is a vector and y is an index. so yhat is a vector and y is a scalar ?

it is so weird to say a loss between a vector and a scalar because loss often means to the things which are of same level?

And, what does sm(yhat), y) mean ? the prob of yth element ?

if it is , this is confusing. what does yth mean?

if not, there is an inconsistency (since On P53, sm(x)i means the prob of ith element )

@EdwardRaff

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants