[R] Which LLMs are actually best for bleeding-edge Linux/ML debugging workflows in 2026? [R]
Mirrored from r/MachineLearning for archival readability. Support the source by reading on the original site.
I’m trying to optimize an AI workflow for bleeding-edge Linux/ML debugging (Arch/CachyOS, CUDA, Python, unsloth, etc.).
Current stack:
- Claude = deep reasoning/mastermind
- Gemini 3.1 Pro = execution/logistics
- Perplexity = retrieval
Main problem: Gemini often gives high-friction or impractical fixes and degrades badly in long troubleshooting sessions. Example: suggested a long Podman workflow for an unsloth/Python issue where micromamba solved it much faster.
I also have access to hosted open models:
- Qwen 3 Coder 30B
- Qwen 3.5 122B
- Mistral Large 675B
- DeepSeek R1 Distill 70B
etc.
Question:
For people doing real-world Linux/ML/debugging workflows (not benchmarks), what currently works best as the “execution/logistics” model with strong web/recent-ecosystem awareness?
I care more about:
- practical fixes
- low friction
- stable long sessions
- debugging quality
than benchmark scores.
[link] [comments]
Discussion (0)
Sign in to join the discussion. Free account, 30 seconds — email code or GitHub.
Sign in →No comments yet. Sign in and be the first to say something.