|
2 | 2 | <feed xmlns="http://www.w3.org/2005/Atom"> |
3 | 3 | <id>/r/LocalLLaMA/.rss</id> |
4 | 4 | <title>LocalLlama</title> |
5 | | - <updated>2026-03-28T16:12:00+00:00</updated> |
| 5 | + <updated>2026-03-28T16:28:56+00:00</updated> |
6 | 6 | <link href="https://old.reddit.com/r/LocalLLaMA/" rel="alternate"/> |
7 | 7 | <generator uri="https://lkiesow.github.io/python-feedgen" version="1.0.0">python-feedgen</generator> |
8 | 8 | <icon>https://www.redditstatic.com/icon.png/</icon> |
9 | 9 | <subtitle>Subreddit to discuss locally hostable AI</subtitle> |
10 | 10 | <entry> |
11 | | - <id>t3_1s639zz</id> |
12 | | - <title>How are you solving agent-to-agent access control?</title> |
13 | | - <updated>2026-03-28T15:25:13+00:00</updated> |
| 11 | + <id>t3_1s63wpc</id> |
| 12 | + <title>Post your Favourite Local AI Productivity Stack (Voice, Code Gen, RAG, Memory etc)</title> |
| 13 | + <updated>2026-03-28T15:49:28+00:00</updated> |
14 | 14 | <author> |
15 | | - <name>/u/nightFlyer_rahl</name> |
16 | | - <uri>https://old.reddit.com/user/nightFlyer_rahl</uri> |
| 15 | + <name>/u/No-Paper-557</name> |
| 16 | + <uri>https://old.reddit.com/user/No-Paper-557</uri> |
17 | 17 | </author> |
18 | | - <content type="html"><!-- SC_OFF --><div class="md"><p><strong>Builders, how are you solving the access control problem for agents?</strong></p> <p>Context: I'm building <a href="https://github.com/GetBindu/Bindu">Bindu</a>, an operating layer for agents. The idea is any framework, any language - agents can talk to each other, negotiate, do trade. We use DIDs (decentralized identifiers) for agent identity. Communication is encrypted.</p> <p>But now I'm hitting a wall: <strong>agent trust.</strong></p> <p>Think about it. In a swarm, some agents should have more power than others. A high trust orchestrator agent should be able to:</p> <ul> <li>compress or manage the context window</li> <li>delegate tasks to lower trust worker agents</li> <li>control who can write to the database</li> </ul> <p>The low trust agents? They just do their job with limited scope. They shouldn't be able to escalate or pretend they have more access than they do.</p> <p>The DB part: sure, MCP and skills can handle that. But what about at the agent-to-agent level? How does one agent prove to another that it has the authority to delegate? How do you stop a worker agent from acting like an orchestrator?</p> <p>In normal software we'd use Keycloak or OAuth for this. But those assume human users, sessions, login flows. In the agent world, there are no humans — just bots talking to bots.</p> <p>What are you all doing for this? Custom solutions? Ignoring it? Curious what's actually working in practice.</p> <p><em>English is not my first language, I use AI to clean up grammar. If it smells like AI, that's the editing</em></p> </div><!-- SC_ON --> &#32; submitted by &#32; <a href="https://old.reddit.com/user/nightFlyer_rahl"> /u/nightFlyer_rahl </a> <br /> <span><a href="https://old.reddit.com/r/LocalLLaMA/comments/1s639zz/how_are_you_solving_agenttoagent_access_control/">[link]</a></span> &#32; <span><a href="https://old.reddit.com/r/LocalLLaMA/comments/1s639zz/how_are_you_solving_agenttoagent_access_control/">[comments]</a></span></content> |
19 | | - <link href="https://old.reddit.com/r/LocalLLaMA/comments/1s639zz/how_are_you_solving_agenttoagent_access_control/"/> |
| 18 | + <content type="html"><!-- SC_OFF --><div class="md"><p>Hi all,</p> <p>It seems like so many new developments are being released as OSS all the time, but I’d like to get an understanding of what you’ve found to personally work well.</p> <p>I know many people here run the newest open source/open weight models with llama.cpp or ollama etc but I wanted to gather feedback on how you use these models for your productivity.</p> <p>1) Voice Conversations - If you’re using things like voice chat, how are you managing that? Previously i was recommended this solution - Faster-whisper + LLM + Kokoro, tied together with LiveKit is my local voice agent stack. I’ll share it if you want and you can just copy the setup</p> <p>2) code generation - what’s your best option at the moment? Eg. Are you using Open Code or something else? Are you managing this with llama.cpp and does tool calling work?</p> <p>3) Any other enhancements - RAG, memory, web search etc </p> </div><!-- SC_ON --> &#32; submitted by &#32; <a href="https://old.reddit.com/user/No-Paper-557"> /u/No-Paper-557 </a> <br /> <span><a href="https://old.reddit.com/r/LocalLLaMA/comments/1s63wpc/post_your_favourite_local_ai_productivity_stack/">[link]</a></span> &#32; <span><a href="https://old.reddit.com/r/LocalLLaMA/comments/1s63wpc/post_your_favourite_local_ai_productivity_stack/">[comments]</a></span></content> |
| 19 | + <link href="https://old.reddit.com/r/LocalLLaMA/comments/1s63wpc/post_your_favourite_local_ai_productivity_stack/"/> |
20 | 20 | <category term="LocalLLaMA" label="r/LocalLLaMA"/> |
21 | | - <published>2026-03-28T15:25:13+00:00</published> |
| 21 | + <published>2026-03-28T15:49:28+00:00</published> |
22 | 22 | </entry> |
23 | 23 | <entry> |
24 | 24 | <id>t3_1s5g1v9</id> |
|
34 | 34 | <published>2026-03-27T20:43:24+00:00</published> |
35 | 35 | </entry> |
36 | 36 | <entry> |
37 | | - <id>t3_1s60en2</id> |
38 | | - <title>Does it make sense to use 4x32Gb RAM or 2x64Gb is the only reasonable option?</title> |
39 | | - <updated>2026-03-28T13:30:32+00:00</updated> |
| 37 | + <id>t3_1s64eux</id> |
| 38 | + <title>Which is better : one highly capable LLM (100+B) or many smaller LLMs (>20B)</title> |
| 39 | + <updated>2026-03-28T16:08:30+00:00</updated> |
40 | 40 | <author> |
41 | | - <name>/u/Real_Ebb_7417</name> |
42 | | - <uri>https://old.reddit.com/user/Real_Ebb_7417</uri> |
| 41 | + <name>/u/More_Chemistry3746</name> |
| 42 | + <uri>https://old.reddit.com/user/More_Chemistry3746</uri> |
43 | 43 | </author> |
44 | | - <content type="html"><!-- SC_OFF --><div class="md"><p>Hi, I currently own:</p> <p>GPU: RTX5080</p> <p>CPU: AMD 9950 x3d</p> <p>RAM: 2x32Gb DDR5 6000MT/s 30CL</p> <p>Aaaaand I'd like to slowly gear up to be able to run bigger models OR run them faster. Obviously GPU is an important factor here (and I'm planning to change it to RTX5090), but the immediate and cheaper upgrade is to increase my RAM.</p> <p>I could buy 2x64Gb instead of my current 2x32Gb (but with worse stats, 2x64Gb are hard to get now and almost nonexistant with 6000MT/s. I found some available with 5600MT/s and 40CL though)... But changing my RAM to 2x64Gb, while probably better, is also much more expensive.</p> <p>Another option is to buy the same 2x32Gb that I currently have and put it next to my current RAM. (my motherboard has 4 sockets)</p> <p>But I wonder how much it might slow down interference for models that are partially offloaded to RAM? As far as I understand, it might slow the RAM down (not sure how exactly it works, I'm not good at hardware xd), but I also don't know if it will be an issue in case of running models or playing video games (two things I care about on that PC). Maybe the bottleneck is actually somewhere else and runnning 4x32GB RAM instead of 2x64Gb won't give me any noticeable difference?</p> <p>So... do you know if it's worth trying? Or I should totally abandon this cheaper idea and go for 2x64Gb with worse parameters?</p> </div><!-- SC_ON --> &#32; submitted by &#32; <a href="https://old.reddit.com/user/Real_Ebb_7417"> /u/Real_Ebb_7417 </a> <br /> <span><a href="https://old.reddit.com/r/LocalLLaMA/comments/1s60en2/does_it_make_sense_to_use_4x32gb_ram_or_2x64gb_is/">[link]</a></span> &#32; <span><a href="https://old.reddit.com/r/LocalLLaMA/comments/1s60en2/does_it_make_sense_to_use_4x32gb_ram_or_2x64gb_is/">[comments]</a></span></content> |
45 | | - <link href="https://old.reddit.com/r/LocalLLaMA/comments/1s60en2/does_it_make_sense_to_use_4x32gb_ram_or_2x64gb_is/"/> |
| 44 | + <content type="html"><!-- SC_OFF --><div class="md"><p>I'm thinking about either having multiple PCs that run smaller models, or one powerful machine that can run a large model. Let's assume both the small and large models run in Q4 with sufficient memory and good performance</p> </div><!-- SC_ON --> &#32; submitted by &#32; <a href="https://old.reddit.com/user/More_Chemistry3746"> /u/More_Chemistry3746 </a> <br /> <span><a href="https://old.reddit.com/r/LocalLLaMA/comments/1s64eux/which_is_better_one_highly_capable_llm_100b_or/">[link]</a></span> &#32; <span><a href="https://old.reddit.com/r/LocalLLaMA/comments/1s64eux/which_is_better_one_highly_capable_llm_100b_or/">[comments]</a></span></content> |
| 45 | + <link href="https://old.reddit.com/r/LocalLLaMA/comments/1s64eux/which_is_better_one_highly_capable_llm_100b_or/"/> |
46 | 46 | <category term="LocalLLaMA" label="r/LocalLLaMA"/> |
47 | | - <published>2026-03-28T13:30:32+00:00</published> |
| 47 | + <published>2026-03-28T16:08:30+00:00</published> |
48 | 48 | </entry> |
49 | 49 | <entry> |
50 | 50 | <id>t3_1s57ky1</id> |
|
59 | 59 | <category term="LocalLLaMA" label="r/LocalLLaMA"/> |
60 | 60 | <published>2026-03-27T15:37:04+00:00</published> |
61 | 61 | </entry> |
62 | | - <entry> |
63 | | - <id>t3_1s5yv7o</id> |
64 | | - <title>Running my own LLM as a beginner, quick check on models</title> |
65 | | - <updated>2026-03-28T12:20:55+00:00</updated> |
66 | | - <author> |
67 | | - <name>/u/PiratesOfTheArctic</name> |
68 | | - <uri>https://old.reddit.com/user/PiratesOfTheArctic</uri> |
69 | | - </author> |
70 | | - <content type="html"><!-- SC_OFF --><div class="md"><p>Hi everyone</p> <p>I'm on a laptop (Dell XPS 9300, 32gb ram / 2tb drive, linux mint), don't plan to change it anytime soon.</p> <p>I'm tip toeing my way into the llm, and would like to sense check the models I have, they were suggested by claude when asking about lightweight types, claude made the descriptions for me: </p> <p>llama.cpp<br /> Openweb UI</p> <p>Models:<br /> Qwen2.5-Coder 3B Q6_K - DAILY: quick Python, formulas, fast answers<br /> Qwen3.5-9B Q6_K - DEEP: complex financial analysis, long programs<br /> Gemma 3 4B Q6_K - VISION: charts, images, screenshots<br /> Phi-4-mini-reasoning Q6_K - CHECK: verify maths and logic</p> <p>At the moment, they are working great, response times are reasonably ok, better than expected to be honest!</p> <p>I'm struggling (at the moment) to fully understand, and appreciate the different models on huggingface, and wondered, are these the most 'lean' based on descriptions, or should I be looking at swapping any? I'm certainly no power user, the models will be used for data analysis (csv/ods/txt), python programming and to bounce ideas off.</p> <p>Next week I'll be buying a dummies/idiot guide. 30 years IT experience and I'm still amazed how much and quick systems have progressed!</p> </div><!-- SC_ON --> &#32; submitted by &#32; <a href="https://old.reddit.com/user/PiratesOfTheArctic"> /u/PiratesOfTheArctic </a> <br /> <span><a href="https://old.reddit.com/r/LocalLLaMA/comments/1s5yv7o/running_my_own_llm_as_a_beginner_quick_check_on/">[link]</a></span> &#32; <span><a href="https://old.reddit.com/r/LocalLLaMA/comments/1s5yv7o/running_my_own_llm_as_a_beginner_quick_check_on/">[comments]</a></span></content> |
71 | | - <link href="https://old.reddit.com/r/LocalLLaMA/comments/1s5yv7o/running_my_own_llm_as_a_beginner_quick_check_on/"/> |
72 | | - <category term="LocalLLaMA" label="r/LocalLLaMA"/> |
73 | | - <published>2026-03-28T12:20:55+00:00</published> |
74 | | - </entry> |
75 | 62 | <entry> |
76 | 63 | <id>t3_1s5z0kx</id> |
77 | 64 | <title>Local LLM evaluation advice after DPO on a psychotherapy dataset</title> |
|
124 | 111 | <category term="LocalLLaMA" label="r/LocalLLaMA"/> |
125 | 112 | <published>2026-03-28T14:13:57+00:00</published> |
126 | 113 | </entry> |
| 114 | + <entry> |
| 115 | + <id>t3_1s5yv7o</id> |
| 116 | + <title>Running my own LLM as a beginner, quick check on models</title> |
| 117 | + <updated>2026-03-28T12:20:55+00:00</updated> |
| 118 | + <author> |
| 119 | + <name>/u/PiratesOfTheArctic</name> |
| 120 | + <uri>https://old.reddit.com/user/PiratesOfTheArctic</uri> |
| 121 | + </author> |
| 122 | + <content type="html"><!-- SC_OFF --><div class="md"><p>Hi everyone</p> <p>I'm on a laptop (Dell XPS 9300, 32gb ram / 2tb drive, linux mint), don't plan to change it anytime soon.</p> <p>I'm tip toeing my way into the llm, and would like to sense check the models I have, they were suggested by claude when asking about lightweight types, claude made the descriptions for me: </p> <p>llama.cpp<br /> Openweb UI</p> <p>Models:<br /> Qwen2.5-Coder 3B Q6_K - DAILY: quick Python, formulas, fast answers<br /> Qwen3.5-9B Q6_K - DEEP: complex financial analysis, long programs<br /> Gemma 3 4B Q6_K - VISION: charts, images, screenshots<br /> Phi-4-mini-reasoning Q6_K - CHECK: verify maths and logic</p> <p>At the moment, they are working great, response times are reasonably ok, better than expected to be honest!</p> <p>I'm struggling (at the moment) to fully understand, and appreciate the different models on huggingface, and wondered, are these the most 'lean' based on descriptions, or should I be looking at swapping any? I'm certainly no power user, the models will be used for data analysis (csv/ods/txt), python programming and to bounce ideas off.</p> <p>Next week I'll be buying a dummies/idiot guide. 30 years IT experience and I'm still amazed how much and quick systems have progressed!</p> </div><!-- SC_ON --> &#32; submitted by &#32; <a href="https://old.reddit.com/user/PiratesOfTheArctic"> /u/PiratesOfTheArctic </a> <br /> <span><a href="https://old.reddit.com/r/LocalLLaMA/comments/1s5yv7o/running_my_own_llm_as_a_beginner_quick_check_on/">[link]</a></span> &#32; <span><a href="https://old.reddit.com/r/LocalLLaMA/comments/1s5yv7o/running_my_own_llm_as_a_beginner_quick_check_on/">[comments]</a></span></content> |
| 123 | + <link href="https://old.reddit.com/r/LocalLLaMA/comments/1s5yv7o/running_my_own_llm_as_a_beginner_quick_check_on/"/> |
| 124 | + <category term="LocalLLaMA" label="r/LocalLLaMA"/> |
| 125 | + <published>2026-03-28T12:20:55+00:00</published> |
| 126 | + </entry> |
127 | 127 | <entry> |
128 | 128 | <id>t3_1s5plrv</id> |
129 | 129 | <title>Anyway to get close to GPT4o on a local model (I know it’s a dumb question)</title> |
|
0 commit comments