Discussion about this post

User's avatar
Brian Dalton's avatar

This is very helpful. Repurposing errors as new examples is brilliant!

Two questions: Is role playing ("You are a xxx AI)" really useful? What about using first person ("I am a xxx AI")?

Expand full comment
Jerry Ding's avatar

Really enjoying your blog so far, very few people study the nuances of prompting like you do.

For making an LLM learn new knowledge that can be applied in diverse tasks, how do you think prompting compares vs. other methods like continued pretraining in terms of response quality? Do you think in-context learning with the prompt achieve the highest quality of learning, and other methods are mainly for cost saving / large scale? Or do you think prompting is a suboptimal way of making an LLM learn new information, even if the information can fit in the context window?

These days, the context window can fit millions of tokens and prompt caching can make them quite cheap, but I'm not sure it's possible to teach an LLM new stuff like a human, simply by patiently explaining how to do things, providing feedback when it makes mistakes, and keeping the full history in the prompt.

Also, what are your thoughts on "infinite context window" algorithms (i.e. linear complexity LLMs)? Will they eventually dominate and replace the finite size attention we use now?

Expand full comment
1 more comment...

No posts