|
2 | 2 | <feed xmlns="http://www.w3.org/2005/Atom"> |
3 | 3 | <id>/r/ollama/.rss</id> |
4 | 4 | <title>ollama</title> |
5 | | - <updated>2026-03-28T18:17:52+00:00</updated> |
| 5 | + <updated>2026-03-28T18:47:22+00:00</updated> |
6 | 6 | <link href="https://old.reddit.com/r/ollama/" rel="alternate"/> |
7 | 7 | <generator uri="https://lkiesow.github.io/python-feedgen" version="1.0.0">python-feedgen</generator> |
8 | 8 | <icon>https://www.redditstatic.com/icon.png/</icon> |
9 | 9 | <subtitle>Atom feed for r/ollama</subtitle> |
10 | | - <entry> |
11 | | - <id>t3_1s489fo</id> |
12 | | - <title>MiroThinker 1.7 mini: 3B active params, beats GPT 5 on multiple benchmarks, weights on HuggingFace</title> |
13 | | - <updated>2026-03-26T14:00:08+00:00</updated> |
14 | | - <author> |
15 | | - <name>/u/Middle-Wafer4480</name> |
16 | | - <uri>https://old.reddit.com/user/Middle-Wafer4480</uri> |
17 | | - </author> |
18 | | - <content type="html"><!-- SC_OFF --><div class="md"><p>Been following the MiroThinker project since v1.0 and wanted to share the latest release since the open source models are genuinely impressive for their size.</p> <h1>The tldr</h1> <p>MiroMind just dropped <strong>MiroThinker 1.7</strong> and <strong>MiroThinker 1.7 mini</strong> as open source models. The mini variant uses only 3B activated parameters (it's a MoE architecture based on Qwen3) and punches way above its weight on research and reasoning benchmarks.</p> <h1>Why this matters for local runners</h1> <p>The 1.7 mini model with 3B active params is small enough to run on consumer hardware. Weights are already on HuggingFace: <a href="https://huggingface.co/miromind-ai/MiroThinker-1.7">miromind-ai/MiroThinker-1.7</a></p> <p>If anyone has already converted this to GGUF or created an Ollama modelfile, please drop it in the comments. The base architecture is Qwen3 MoE so the conversion path should be straightforward.</p> <h1>Benchmarks that caught my eye</h1> <p>Here's where the mini model (3B active) lands compared to some big names:</p> <table><thead> <tr> <th align="left">Benchmark</th> <th align="left">MiroThinker 1.7 mini</th> <th align="left">GPT 5</th> <th align="left">DeepSeek V3.2</th> <th align="left">Gemini 3 Pro</th> </tr> </thead><tbody> <tr> <td align="left">BrowseComp ZH</td> <td align="left">72.3</td> <td align="left">65.0</td> <td align="left">65.0</td> <td align="left">66.8</td> </tr> <tr> <td align="left">GAIA</td> <td align="left">80.3</td> <td align="left">76.4</td> <td align="left">—</td> <td align="left">—</td> </tr> <tr> <td align="left">xbench DeepResearch</td> <td align="left">57.2</td> <td align="left">75.0</td> <td align="left">—</td> <td align="left">53.0</td> </tr> <tr> <td align="left">FinSearchComp</td> <td align="left">62.6</td> <td align="left">73.8</td> <td align="left">—</td> <td align="left">52.7</td> </tr> </tbody></table> <p>The mini model beating GPT 5 on BrowseComp ZH and GAIA while running at a fraction of the compute is wild. The full 1.7 model scores even higher across the board.</p> <p>The bigger sibling, MiroThinker H1, hits 88.2 on BrowseComp (vs Gemini 3.1 Pro at 85.9 and Claude 4.6 Opus at 84.0) but that one is hosted only, not open source.</p> <h1>What makes it different from a regular chat model</h1> <p>This isn't just another instruct model. It's trained specifically as a research agent with a four stage pipeline: mid training for planning and reasoning fundamentals, SFT on expert trajectories, DPO for preference alignment, then RL with GRPO in live environments. The mid training stage is the interesting part; they train the model on isolated agentic &quot;atoms&quot; (planning from scratch, reasoning given partial context, summarizing partial observations) rather than just full trajectories. This apparently makes each reasoning step more reliable so the model needs fewer total steps to solve problems.</p> <p>In their ablations, the 1.7 mini achieved 16.7% better performance with about 43% fewer interaction rounds compared to v1.5 at the same parameter count.</p> <h1>The agentic setup caveat</h1> <p>Full disclosure: the benchmark numbers above come from running the model inside their agentic framework (<a href="https://github.com/MiroMindAI/MiroThinker">MiroThinker on GitHub</a>) which includes web search, code sandboxes, and file transfer tools. So you won't replicate these exact scores just running the raw model through Ollama for chat. But the underlying model capabilities (planning, multi step reasoning, tool call formatting) are all baked into the weights, so it should still be a strong reasoning model for local use even without the full agent stack.</p> <p>For those who want the full agent experience locally, their framework is open source and you could potentially wire it up with a local inference backend.</p> <h1>Links</h1> <ul> <li>Model weights: <a href="https://huggingface.co/miromind-ai/MiroThinker-1.7">HuggingFace</a></li> <li>Agent framework: <a href="https://github.com/MiroMindAI/MiroThinker">GitHub</a></li> <li>General agent framework: <a href="https://github.com/MiroMindAI/MiroFlow">MiroFlow GitHub</a></li> <li>Full technical report has all the details on training pipeline and benchmarks</li> </ul> <p>Would love to hear from anyone who gets this running through Ollama. Curious how it performs as a general reasoning model outside the agentic setup, and what kind of VRAM usage people are seeing with different quantizations.</p> </div><!-- SC_ON --> &#32; submitted by &#32; <a href="https://old.reddit.com/user/Middle-Wafer4480"> /u/Middle-Wafer4480 </a> <br /> <span><a href="https://old.reddit.com/r/ollama/comments/1s489fo/mirothinker_17_mini_3b_active_params_beats_gpt_5/">[link]</a></span> &#32; <span><a href="https://old.reddit.com/r/ollama/comments/1s489fo/mirothinker_17_mini_3b_active_params_beats_gpt_5/">[comments]</a></span></content> |
19 | | - <link href="https://old.reddit.com/r/ollama/comments/1s489fo/mirothinker_17_mini_3b_active_params_beats_gpt_5/"/> |
20 | | - <category term="ollama" label="r/ollama"/> |
21 | | - <published>2026-03-26T14:00:08+00:00</published> |
22 | | - </entry> |
23 | 10 | <entry> |
24 | 11 | <id>t3_1s50o0p</id> |
25 | 12 | <title>Added branching + switch logic to my Ollama-based AI workflow builder (v0.7.0)</title> |
|
280 | 267 | <category term="ollama" label="r/ollama"/> |
281 | 268 | <published>2026-03-27T21:03:11+00:00</published> |
282 | 269 | </entry> |
| 270 | + <entry> |
| 271 | + <id>t3_1s687pc</id> |
| 272 | + <title>ROCm Support Dropped for RDNA1/2?</title> |
| 273 | + <updated>2026-03-28T18:34:25+00:00</updated> |
| 274 | + <author> |
| 275 | + <name>/u/Tyr_Kukulkan</name> |
| 276 | + <uri>https://old.reddit.com/user/Tyr_Kukulkan</uri> |
| 277 | + </author> |
| 278 | + <content type="html"><!-- SC_OFF --><div class="md"><p>I updated Ollama and OpenWebUI on my homelab a few weeks ago only to discover inference was no longer running on my old Vega 64. I replaced it with an RX 5700 XT I had lying around and everything immediately worked again.</p> <p>I looked up that problem and it appeared as though ROCm support for the Vega series GPUs had been dropped with a ROCm update. Thanks AMD!</p> <p>Now a few weeks later and my RX 5700 XT Is showing exactly the same signs and isn't being used. I went onto the AMD ROCm documentation and only RDNA3/4 cards are showing as supported from this month onwards. Well, fuck... Thanks again AMD!</p> <p>I checked which cards are supported and only cards 7700 and up for RDNA3 and all current RDNA4 (9060 and up).</p> <p>I'm still quite new to this so when I tried to force support by editing the relevant file, ollama then wouldn't run. Removing the HSA override line allowed it to run again.</p> <p>I don't have enough RAM in my main PC to run larger models, which is what I was doing with my homelab, as that has 128GB of RAM.</p> <p>Am I just up shit creek without a paddle with my homelab setup? I was going to get a second hand 6800 16GB (or two) fairly cheap to drop in originally but glad I didn't. I'm put off getting a 7900 XTX second hand in case support for that is also dropped soon.</p> <p>I didn't want to spend loads on my homelab to just mess around with my VMs and LLMs.</p> <p>Any help or advice anyone can offer would be appreciated.</p> </div><!-- SC_ON --> &#32; submitted by &#32; <a href="https://old.reddit.com/user/Tyr_Kukulkan"> /u/Tyr_Kukulkan </a> <br /> <span><a href="https://old.reddit.com/r/ollama/comments/1s687pc/rocm_support_dropped_for_rdna12/">[link]</a></span> &#32; <span><a href="https://old.reddit.com/r/ollama/comments/1s687pc/rocm_support_dropped_for_rdna12/">[comments]</a></span></content> |
| 279 | + <link href="https://old.reddit.com/r/ollama/comments/1s687pc/rocm_support_dropped_for_rdna12/"/> |
| 280 | + <category term="ollama" label="r/ollama"/> |
| 281 | + <published>2026-03-28T18:34:25+00:00</published> |
| 282 | + </entry> |
283 | 283 | <entry> |
284 | 284 | <id>t3_1s63amy</id> |
285 | 285 | <title>My first MAUI app integrating Ollama: A private AI Database Auditor for SQL/Postgres/MySQL.</title> |
|
0 commit comments