- Notifications
You must be signed in to change notification settings - Fork11.9k
-
In ch06, I have noticed that generate_text_simple with text2 in the pretrained model, i.e. before finetuning, simply just repeats indefinitely the input text2 many times over word for word. |
BetaWas this translation helpful?Give feedback.
All reactions
Replies: 1 comment
-
Hi there, this is normal behavior for several pretrained LLMs that haven't undergone finetuning, yet. Especially with smaller LLMs. Not sure why this happens exactly (maybe an artifact of repetitive structures in the training data or the LLM isn't good at longer contexts) |
BetaWas this translation helpful?Give feedback.
All reactions
0 replies
Sign up for freeto join this conversation on GitHub. Already have an account?Sign in to comment
