-
In chapter 4.6, on page 120, it is written that |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 4 replies
-
Hmm interesting. My GPT model also deviates in trainable parameters from the book, however, I did not do an exact 1:1 copy if I remember correctly. My gpt2-small has 163,059,793 trainable parameters. I suggest you install ´torch-summary´ from pip and run your model both through My model config is Code outputprint(model):
summary(model):
|
Beta Was this translation helpful? Give feedback.
Thanks for the comments! It's interesting and weird that the numbers may be different. I suspect this is maybe due to some minor code difference that is easy to overlook.
@Jessen-Li could you try to run the following standalone code and see what you get? If it is 163,009,536 then there's perhaps some small discrepancy in your code. If not, we can investigate further.