-
In-context Learning and Induction Heads: This paper talks about an emerging phenomena in training transformers called induction heads. As the naming indicates, an induction heads refers to a special type of attention heads that is able to do "induction" reasoning.
-
SUCCESSOR HEADS: RECURRING, INTERPRETABLE ATTENTION HEADS IN THE WILD
-
Transformers generalize differently from information stored in context vs in weights
-
Rethinking the Role of Demonstrations: What Makes In-Context Learning Work?
-
[https://arxiv.org/abs/2402.11004](The Evolution of Statistical Induction Heads: In-Context Learning Markov Chains)
-
[https://www.lesswrong.com/posts/j6s9H9SHrEhEfuJnq/causal-scrubbing-results-on-induction-heads](Causal scrubbing: results on induction heads)
-
Notifications
You must be signed in to change notification settings - Fork 0
Alialsaeedi1/In-context-learning-secrets
Folders and files
Name | Name | Last commit message | Last commit date | |
---|---|---|---|---|
Repository files navigation
About
No description, website, or topics provided.
Resources
Stars
Watchers
Forks
Releases
No releases published
Packages 0
No packages published