r/LocalLLaMA
42 articles archived · Visit source ↗ · RSS
-
r/LocalLLaMA community 4h ago
Side Projects.
​ Little something I put together to play with for larger contexts than my 9070xt. 8700k, dual P100's, 16gb DDR4, 32gb Optane, Samsung sata SSD. Nothing too fancy. Anyone else do a recent build? How's it working out?   submitted by   /u/apollo_mg [link]  …
15 -
-
-
r/LocalLLaMA community 6h ago
Efficient pretraining with token superposition by Nous Research
  submitted by   /u/de4dee [link]   [comments]
14 -
-
-
-
r/LocalLLaMA community 7h ago
Sipeed's K3 RISC-V SBCs can run 30B-parameter LLMs 60 TOPS (INT4), Supports BF16/FP16/INT4
https://wccftech.com/sipeed-crams-32gb-lpddr5-60-tops-npu-compact-risc-v-board-hits-15-tokens-s-ai-llms/   submitted by   /u/MundanePercentage674 [link]   [comments]
19 -
r/LocalLLaMA community 9h ago
qwen3.6 just stops
https://preview.redd.it/74cj1xu9pw0h1.png?width=1229&format=png&auto=webp&s=3ae999cc3530ecb4eccf70e25f1a9eb2aa3f2d7b Sometimes qwen 3.6 just stops at the middle of a task, is there a way to avoid it? This is qwen-code CLI, but also happens on opencode. Running with vLLM with…
17 -
-
-
-
-
r/LocalLLaMA community 15h ago
Does THINKING MODE significantly improve translation?
Between a solid model from Qwen or Gemma 4, when translating a text, does "thinking mode" significantly boost the quality of the translation, or is the difference negligible?   submitted by   /u/Sostrene_Blue [link]   [comments]
27 -
r/LocalLLaMA community 21h ago
AntAngelMed - 100a6b Healthcare LLM
  submitted by   /u/Zc5Gwu [link]   [comments]
38 -
-
-
-
r/LocalLLaMA community 1d ago
Dad why is my sisters name Lora?
  submitted by   /u/rwitz4 [link]   [comments]
35 -
-
-
-
-
-
-
r/LocalLLaMA community 1d ago
examples : add llama-eval by ggerganov · Pull Request #21152 · ggml-org/llama.cpp
now you can evaluate your models at home, sounds like a perfect tool to compare quants and finetunes Datasets: AIME, AIME2025, GSM8K, GPQA   submitted by   /u/jacek2023 [link]   [comments]
15