Cedric Yau(@ctyau) 's Twitter Profile Photo

This talk Vishal Misra was an eye-opener to how few shot prompting works under the hood to change output token probability distributions. I forgot LLMs don't produce tokens directly. They produce embeddings from which nearest neighbors tokens are chosen.

youtu.be/sLodkyHlQhY

This talk @vishalmisra was an eye-opener to how few shot prompting works under the hood to change output token probability distributions. I forgot LLMs don't produce tokens directly. They produce embeddings from which nearest neighbors tokens are chosen.

youtu.be/sLodkyHlQhY
account_circle
martin_casado(@martin_casado) 's Twitter Profile Photo

tl;dr LLMS as Bayesian learning : Given a prompt, looks for something close in training set, then uses prompt for new evidence. Then computes a posterior using this new evidence. This posterior distribution is what is used to generate the new text. (Vishal Misra)

It's so crazy…

tl;dr LLMS as Bayesian learning : Given a prompt, looks for something close in training set, then uses prompt for new evidence. Then computes a posterior using this new evidence. This posterior distribution is what is used to generate the new text.  (@vishalmisra)

It's so crazy…
account_circle