Skip to content

why my GPT2 output(logits) different from the book #631

Closed Answered by casinca
Jessen-Li asked this question in Q&A

You must be logged in to vote

I believe it is to be expected, it's the same scheme as #607.
I have the same output as yours on Win and the same output from the notebook + p.120 book on Mac (Sebastian is on Mac).

The dummy model, just as GPT-2 used nn.Dropout() which explains the discrepancy that Sebastian and @d-kleine talked about in #607 earlier.

Replies: 1 comment

You must be logged in to vote
0 replies
Answer selected by Jessen-Li
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants