What Happens When an LLM "Thinks": Tokens, Logits, and Sampling
You send a prompt. The model "thinks" for a few seconds. And generates a response that seems intelligent.
What happened in that time?
If your way of implementing systems with LLMs is based on using a high-level framework plus OpenAI APIs and derivati...
luisciber.com20 min read