Dynamically allocating compute budget to hard set of problems and evolving the sections with Qwen-35B-A3B gets you near GPT-5.4-xHigh on HLE
Mirrored from r/LocalLLaMA for archival readability. Support the source by reading on the original site.
| submitted by /u/Ryoiki-Tokuiten [link] [comments] |
More from r/LocalLLaMA
-
Finding the 4x 3090 Sweet Spot
May 15
-
RAG on Snapdragon X2 Laptop, 200K documents.
May 15
-
Orthrus-Qwen3-8B : up to 7.8×tokens/forward on Qwen3-8B, frozen backbone, provably identical output distribution
May 15
-
I built a self-hosted open-source MCP server that gives any local LLM real financial data — SEC filings, 13F, insider & congressional trades, short data, FRED
May 15
Discussion (0)
Sign in to join the discussion. Free account, 30 seconds — email code or GitHub.
Sign in →No comments yet. Sign in and be the first to say something.