I kept a running list of every LLM term that actually matters for production, cleaned it up and open sourced it
Mirrored from r/LocalLLaMA for archival readability. Support the source by reading on the original site.
| Been building with LLMs for a while and kept hitting terms where the standard definition was useless for making engineering decisions. So I kept a personal doc, eventually it hit 30+ terms across inference, retrieval, agents, training, and prompting. Each entry has the plain-English definition plus the production implication, the thing that actually affects your architecture or debugging. Cleaned it up, built a small interactive UI with search and category filtering, and put it on GitHub. Not trying to compete with papers or courses, it's more of a field reference for when you're mid-build and need the practical version of a term fast. Would genuinely appreciate corrections or additions. The bar I set for new terms: does the definition help someone make a better engineering decision? [link] [comments] |
More from r/LocalLLaMA
-
Dynamically allocating compute budget to hard set of problems and evolving the sections with Qwen-35B-A3B gets you near GPT-5.4-xHigh on HLE
May 15
-
Orthrus-Qwen3-8B : up to 7.8×tokens/forward on Qwen3-8B, frozen backbone, provably identical output distribution
May 15
-
I built a self-hosted open-source MCP server that gives any local LLM real financial data — SEC filings, 13F, insider & congressional trades, short data, FRED
May 15
-
Qwen 3.6 27B: IQ3XXS KV Q8 vs Q4XL KV Q4 (262K context)
May 15
Discussion (0)
Sign in to join the discussion. Free account, 30 seconds — email code or GitHub.
Sign in →No comments yet. Sign in and be the first to say something.