Skip to content
This repository was archived by the owner on Aug 8, 2025. It is now read-only.

Commit 9cef4b3

Browse files
committed
[dump]: data dump at 2025.06.29 18.55 GMT+6;
1 parent 2b6a121 commit 9cef4b3

File tree

93 files changed

+89646
-3094
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

93 files changed

+89646
-3094
lines changed

src/dump/1lm4s6i/index.json

Lines changed: 872 additions & 0 deletions
Large diffs are not rendered by default.

src/dump/1lm4tno/index.json

Lines changed: 77 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -152,7 +152,7 @@
152152
"report_reasons": null,
153153
"author": "HeroesDieYoung0",
154154
"discussion_type": null,
155-
"num_comments": 3,
155+
"num_comments": 5,
156156
"send_replies": true,
157157
"media": null,
158158
"contest_mode": false,
@@ -161,7 +161,7 @@
161161
"permalink": "/r/LocalLLaMA/comments/1lm4tno/build_advice_question_for_repurposing_spare_gpus/",
162162
"stickied": false,
163163
"url": "https://www.reddit.com/r/LocalLLaMA/comments/1lm4tno/build_advice_question_for_repurposing_spare_gpus/",
164-
"subreddit_subscribers": 492232,
164+
"subreddit_subscribers": 492572,
165165
"created_utc": 1751057588,
166166
"num_crossposts": 0,
167167
"mod_reports": [],
@@ -404,6 +404,81 @@
404404
"num_reports": null,
405405
"ups": 1
406406
}
407+
},
408+
{
409+
"kind": "t1",
410+
"data": {
411+
"subreddit_id": "t5_81eyvm",
412+
"approved_at_utc": null,
413+
"author_is_blocked": false,
414+
"comment_type": null,
415+
"awarders": [],
416+
"mod_reason_by": null,
417+
"banned_by": null,
418+
"author_flair_type": "text",
419+
"total_awards_received": 0,
420+
"subreddit": "LocalLLaMA",
421+
"author_flair_template_id": null,
422+
"likes": null,
423+
"replies": "",
424+
"user_reports": [],
425+
"saved": false,
426+
"id": "n08vp0q",
427+
"banned_at_utc": null,
428+
"mod_reason_title": null,
429+
"gilded": 0,
430+
"archived": false,
431+
"collapsed_reason_code": null,
432+
"no_follow": true,
433+
"author": "Marksta",
434+
"can_mod_post": false,
435+
"created_utc": 1751121027,
436+
"send_replies": true,
437+
"parent_id": "t3_1lm4tno",
438+
"score": 1,
439+
"author_fullname": "t2_559a1",
440+
"approved_by": null,
441+
"mod_note": null,
442+
"all_awardings": [],
443+
"collapsed": false,
444+
"body": "Hook them up bud, that system is going to rip with even just the 2x 3080. 60gb vram is going to be really nice. No problems using them all, even on 1x.",
445+
"edited": false,
446+
"top_awarded_type": null,
447+
"author_flair_css_class": null,
448+
"name": "t1_n08vp0q",
449+
"is_submitter": false,
450+
"downs": 0,
451+
"author_flair_richtext": [],
452+
"author_patreon_flair": false,
453+
"body_html": "<div class=\"md\"><p>Hook them up bud, that system is going to rip with even just the 2x 3080. 60gb vram is going to be really nice. No problems using them all, even on 1x.</p>\n</div>",
454+
"removal_reason": null,
455+
"collapsed_reason": null,
456+
"distinguished": null,
457+
"associated_award": null,
458+
"stickied": false,
459+
"author_premium": false,
460+
"can_gild": false,
461+
"gildings": {},
462+
"unrepliable_reason": null,
463+
"author_flair_text_color": null,
464+
"score_hidden": false,
465+
"permalink": "/r/LocalLLaMA/comments/1lm4tno/build_advice_question_for_repurposing_spare_gpus/n08vp0q/",
466+
"subreddit_type": "public",
467+
"locked": false,
468+
"report_reasons": null,
469+
"created": 1751121027,
470+
"author_flair_text": null,
471+
"treatment_tags": [],
472+
"link_id": "t3_1lm4tno",
473+
"subreddit_name_prefixed": "r/LocalLLaMA",
474+
"controversiality": 0,
475+
"depth": 0,
476+
"author_flair_background_color": null,
477+
"collapsed_because_crowd_control": null,
478+
"mod_reports": [],
479+
"num_reports": null,
480+
"ups": 1
481+
}
407482
}
408483
],
409484
"before": null

src/dump/1lm58q1/index.json

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -39,7 +39,7 @@
3939
"upvote_ratio": 1,
4040
"author_flair_background_color": null,
4141
"subreddit_type": "public",
42-
"ups": 1,
42+
"ups": 2,
4343
"total_awards_received": 0,
4444
"media_embed": {},
4545
"thumbnail_width": null,
@@ -53,7 +53,7 @@
5353
"secure_media_embed": {},
5454
"link_flair_text": "Question | Help",
5555
"can_mod_post": false,
56-
"score": 1,
56+
"score": 2,
5757
"approved_by": null,
5858
"is_created_from_ads_ui": false,
5959
"author_premium": false,
@@ -79,7 +79,7 @@
7979
"banned_at_utc": null,
8080
"view_count": null,
8181
"archived": false,
82-
"no_follow": true,
82+
"no_follow": false,
8383
"is_crosspostable": false,
8484
"pinned": false,
8585
"over_18": false,
@@ -116,7 +116,7 @@
116116
"permalink": "/r/LocalLLaMA/comments/1lm58q1/problems_on_rvc_webui_creating_new_vocal_model/",
117117
"stickied": false,
118118
"url": "https://www.reddit.com/r/LocalLLaMA/comments/1lm58q1/problems_on_rvc_webui_creating_new_vocal_model/",
119-
"subreddit_subscribers": 492233,
119+
"subreddit_subscribers": 492572,
120120
"created_utc": 1751058659,
121121
"num_crossposts": 0,
122122
"mod_reports": [],

src/dump/1lm5a05/index.json

Lines changed: 3442 additions & 0 deletions
Large diffs are not rendered by default.

src/dump/1lm5muh/index.json

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -39,7 +39,7 @@
3939
"upvote_ratio": 1,
4040
"author_flair_background_color": null,
4141
"subreddit_type": "public",
42-
"ups": 3,
42+
"ups": 5,
4343
"total_awards_received": 0,
4444
"media_embed": {},
4545
"thumbnail_width": null,
@@ -53,7 +53,7 @@
5353
"secure_media_embed": {},
5454
"link_flair_text": "Question | Help",
5555
"can_mod_post": false,
56-
"score": 3,
56+
"score": 5,
5757
"approved_by": null,
5858
"is_created_from_ads_ui": false,
5959
"author_premium": false,
@@ -116,7 +116,7 @@
116116
"permalink": "/r/LocalLLaMA/comments/1lm5muh/what_is_your_favorite_opensource_image_embedding/",
117117
"stickied": false,
118118
"url": "https://www.reddit.com/r/LocalLLaMA/comments/1lm5muh/what_is_your_favorite_opensource_image_embedding/",
119-
"subreddit_subscribers": 492232,
119+
"subreddit_subscribers": 492572,
120120
"created_utc": 1751059680,
121121
"num_crossposts": 0,
122122
"mod_reports": [],
@@ -165,7 +165,7 @@
165165
"created_utc": 1751077155,
166166
"send_replies": true,
167167
"parent_id": "t3_1lm5muh",
168-
"score": 1,
168+
"score": 2,
169169
"author_fullname": "t2_26yp67qq",
170170
"approved_by": null,
171171
"mod_note": null,
@@ -207,7 +207,7 @@
207207
"collapsed_because_crowd_control": null,
208208
"mod_reports": [],
209209
"num_reports": null,
210-
"ups": 1
210+
"ups": 2
211211
}
212212
}
213213
],

src/dump/1lm66fy/index.json

Lines changed: 15 additions & 15 deletions
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@
1212
"data": {
1313
"approved_at_utc": null,
1414
"subreddit": "LocalLLaMA",
15-
"selftext": "Hey everyone. So I'll keep it short. I've written a Claude Code \"clone\", [mcp-agent](https://github.com/amranu/mcp-agent) which allows tool use for arbitrary LLMs (though they have to support tool use, I'm not using any templating). Currently it has tested support for Deepseek, Gemini, OpenAI and Anthropic APIs but I want it to work with ollama. Main problem is I don't have a setup that can work with ollama (I have an old AMD card, no nvidia). So I need someone to test out the ollama support I've added and see if it works.\n\nmcp-agent exposes all the tools Claude Code has, along with arbitrary subagent support. It also has an mcp server, similar to Zen MCP to allow any LLM to talk to any other LLM you have configured. Except unlike Zen MCP, the LLMs have access to tools.\n\nAnyone willing to help me out and test ollama support would be greatly appreciated!",
15+
"selftext": "Hey everyone. So I'll keep it short. I've written a Claude Code \"clone\", [cli-agent](https://github.com/amranu/cli-agent) which allows tool use for arbitrary LLMs (though they have to support tool use, I'm not using any templating). Currently it has tested support for Deepseek, Gemini, OpenAI and Anthropic APIs but I want it to work with ollama. Main problem is I don't have a setup that can work with ollama (I have an old AMD card, no nvidia). So I need someone to test out the ollama support I've added and see if it works.\n\nmcp-agent exposes all the tools Claude Code has, along with arbitrary subagent support. It also has an mcp server, similar to Zen MCP to allow any LLM to talk to any other LLM you have configured. Except unlike Zen MCP, the LLMs have access to tools.\n\nAnyone willing to help me out and test ollama support would be greatly appreciated!",
1616
"user_reports": [],
1717
"saved": false,
1818
"mod_reason_title": null,
@@ -36,10 +36,10 @@
3636
"name": "t3_1lm66fy",
3737
"quarantine": false,
3838
"link_flair_text_color": "light",
39-
"upvote_ratio": 0.5,
39+
"upvote_ratio": 0.6,
4040
"author_flair_background_color": null,
4141
"subreddit_type": "public",
42-
"ups": 0,
42+
"ups": 1,
4343
"total_awards_received": 0,
4444
"media_embed": {},
4545
"thumbnail_width": null,
@@ -53,12 +53,12 @@
5353
"secure_media_embed": {},
5454
"link_flair_text": "Other",
5555
"can_mod_post": false,
56-
"score": 0,
56+
"score": 1,
5757
"approved_by": null,
5858
"is_created_from_ads_ui": false,
5959
"author_premium": false,
6060
"thumbnail": "self",
61-
"edited": false,
61+
"edited": 1751140914,
6262
"author_flair_css_class": null,
6363
"author_flair_richtext": [],
6464
"gildings": {},
@@ -74,7 +74,7 @@
7474
"author_flair_type": "text",
7575
"domain": "self.LocalLLaMA",
7676
"allow_live_comments": false,
77-
"selftext_html": "<!-- SC_OFF --><div class=\"md\"><p>Hey everyone. So I'll keep it short. I've written a Claude Code "clone", <a href=\"https://github.com/amranu/mcp-agent\">mcp-agent</a> which allows tool use for arbitrary LLMs (though they have to support tool use, I'm not using any templating). Currently it has tested support for Deepseek, Gemini, OpenAI and Anthropic APIs but I want it to work with ollama. Main problem is I don't have a setup that can work with ollama (I have an old AMD card, no nvidia). So I need someone to test out the ollama support I've added and see if it works.</p>\n\n<p>mcp-agent exposes all the tools Claude Code has, along with arbitrary subagent support. It also has an mcp server, similar to Zen MCP to allow any LLM to talk to any other LLM you have configured. Except unlike Zen MCP, the LLMs have access to tools.</p>\n\n<p>Anyone willing to help me out and test ollama support would be greatly appreciated!</p>\n</div><!-- SC_ON -->",
77+
"selftext_html": "<!-- SC_OFF --><div class=\"md\"><p>Hey everyone. So I'll keep it short. I've written a Claude Code "clone", <a href=\"https://github.com/amranu/cli-agent\">cli-agent</a> which allows tool use for arbitrary LLMs (though they have to support tool use, I'm not using any templating). Currently it has tested support for Deepseek, Gemini, OpenAI and Anthropic APIs but I want it to work with ollama. Main problem is I don't have a setup that can work with ollama (I have an old AMD card, no nvidia). So I need someone to test out the ollama support I've added and see if it works.</p>\n\n<p>mcp-agent exposes all the tools Claude Code has, along with arbitrary subagent support. It also has an mcp server, similar to Zen MCP to allow any LLM to talk to any other LLM you have configured. Except unlike Zen MCP, the LLMs have access to tools.</p>\n\n<p>Anyone willing to help me out and test ollama support would be greatly appreciated!</p>\n</div><!-- SC_ON -->",
7878
"likes": null,
7979
"suggested_sort": null,
8080
"banned_at_utc": null,
@@ -88,44 +88,44 @@
8888
"images": [
8989
{
9090
"source": {
91-
"url": "https://external-preview.redd.it/JjssSrYiQ_fO-pBkgXAqjNKlDyH24UlZePYPK2yw1po.png?auto=webp&s=0cad53d5b002b8bbcab582e456d0ad6d4580749e",
91+
"url": "https://external-preview.redd.it/ky4q-HJ1F3S2UdCuaGkloWjj4Ru8GaNbo0jpnr086rM.png?auto=webp&s=73d3e4c6457f8e6bbd5981b68295bab611fbfe97",
9292
"width": 1200,
9393
"height": 600
9494
},
9595
"resolutions": [
9696
{
97-
"url": "https://external-preview.redd.it/JjssSrYiQ_fO-pBkgXAqjNKlDyH24UlZePYPK2yw1po.png?width=108&crop=smart&auto=webp&s=d148b11f5d049c242e0b8ccebd09c595f24561ee",
97+
"url": "https://external-preview.redd.it/ky4q-HJ1F3S2UdCuaGkloWjj4Ru8GaNbo0jpnr086rM.png?width=108&crop=smart&auto=webp&s=3b376e1aa1902b7556bb3536cbf55124d2711777",
9898
"width": 108,
9999
"height": 54
100100
},
101101
{
102-
"url": "https://external-preview.redd.it/JjssSrYiQ_fO-pBkgXAqjNKlDyH24UlZePYPK2yw1po.png?width=216&crop=smart&auto=webp&s=0ca3894115999ddafafce2ec6ba026dc9ca52c61",
102+
"url": "https://external-preview.redd.it/ky4q-HJ1F3S2UdCuaGkloWjj4Ru8GaNbo0jpnr086rM.png?width=216&crop=smart&auto=webp&s=6dfe7ffd40bbab26c0485457d0a2fff177288a62",
103103
"width": 216,
104104
"height": 108
105105
},
106106
{
107-
"url": "https://external-preview.redd.it/JjssSrYiQ_fO-pBkgXAqjNKlDyH24UlZePYPK2yw1po.png?width=320&crop=smart&auto=webp&s=86f14a97de86022f331b9aaf11ab54ebf6a68c9a",
107+
"url": "https://external-preview.redd.it/ky4q-HJ1F3S2UdCuaGkloWjj4Ru8GaNbo0jpnr086rM.png?width=320&crop=smart&auto=webp&s=1337de2e84c021178d03af21aca3ecc6af91df4e",
108108
"width": 320,
109109
"height": 160
110110
},
111111
{
112-
"url": "https://external-preview.redd.it/JjssSrYiQ_fO-pBkgXAqjNKlDyH24UlZePYPK2yw1po.png?width=640&crop=smart&auto=webp&s=ac89bad57a49c1d66a2e32f964a34fc3d08a73d9",
112+
"url": "https://external-preview.redd.it/ky4q-HJ1F3S2UdCuaGkloWjj4Ru8GaNbo0jpnr086rM.png?width=640&crop=smart&auto=webp&s=1bf7ee1161ce602fb904e807403f07417ec4d842",
113113
"width": 640,
114114
"height": 320
115115
},
116116
{
117-
"url": "https://external-preview.redd.it/JjssSrYiQ_fO-pBkgXAqjNKlDyH24UlZePYPK2yw1po.png?width=960&crop=smart&auto=webp&s=f64ad6b039fd8cb823a5e4c066497c4da38af452",
117+
"url": "https://external-preview.redd.it/ky4q-HJ1F3S2UdCuaGkloWjj4Ru8GaNbo0jpnr086rM.png?width=960&crop=smart&auto=webp&s=f17963fe73bb28a9dad1f18d1e3455c063524ece",
118118
"width": 960,
119119
"height": 480
120120
},
121121
{
122-
"url": "https://external-preview.redd.it/JjssSrYiQ_fO-pBkgXAqjNKlDyH24UlZePYPK2yw1po.png?width=1080&crop=smart&auto=webp&s=da184f72071d4b12f05e10eae7fe7e6861d9c58b",
122+
"url": "https://external-preview.redd.it/ky4q-HJ1F3S2UdCuaGkloWjj4Ru8GaNbo0jpnr086rM.png?width=1080&crop=smart&auto=webp&s=b3c52c8d7e5cfd23bac748411d52d2f33ea125f0",
123123
"width": 1080,
124124
"height": 540
125125
}
126126
],
127127
"variants": {},
128-
"id": "JjssSrYiQ_fO-pBkgXAqjNKlDyH24UlZePYPK2yw1po"
128+
"id": "ky4q-HJ1F3S2UdCuaGkloWjj4Ru8GaNbo0jpnr086rM"
129129
}
130130
],
131131
"enabled": false
@@ -163,7 +163,7 @@
163163
"permalink": "/r/LocalLLaMA/comments/1lm66fy/i_need_help_testing_my_agentic_wrapper_for_llms/",
164164
"stickied": false,
165165
"url": "https://www.reddit.com/r/LocalLLaMA/comments/1lm66fy/i_need_help_testing_my_agentic_wrapper_for_llms/",
166-
"subreddit_subscribers": 492233,
166+
"subreddit_subscribers": 492572,
167167
"created_utc": 1751061087,
168168
"num_crossposts": 0,
169169
"mod_reports": [],

src/dump/1lm76gk/index.json

Lines changed: 8 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -36,9 +36,9 @@
3636
"name": "t3_1lm76gk",
3737
"quarantine": false,
3838
"link_flair_text_color": "light",
39-
"upvote_ratio": 0.96,
39+
"upvote_ratio": 0.97,
4040
"author_flair_background_color": null,
41-
"ups": 56,
41+
"ups": 81,
4242
"total_awards_received": 0,
4343
"media_embed": {},
4444
"thumbnail_width": 140,
@@ -52,7 +52,7 @@
5252
"secure_media_embed": {},
5353
"link_flair_text": "Resources",
5454
"can_mod_post": false,
55-
"score": 56,
55+
"score": 81,
5656
"approved_by": null,
5757
"is_created_from_ads_ui": false,
5858
"author_premium": false,
@@ -164,7 +164,7 @@
164164
"permalink": "/r/LocalLLaMA/comments/1lm76gk/hugging_face_releases_a_50_page_report_on_how/",
165165
"stickied": false,
166166
"url": "https://i.redd.it/ixin9dvyqj9f1.png",
167-
"subreddit_subscribers": 492232,
167+
"subreddit_subscribers": 492572,
168168
"created_utc": 1751063663,
169169
"num_crossposts": 0,
170170
"mod_reports": [],
@@ -213,7 +213,7 @@
213213
"created_utc": 1751063748,
214214
"send_replies": true,
215215
"parent_id": "t3_1lm76gk",
216-
"score": 3,
216+
"score": 5,
217217
"author_fullname": "t2_911kk5e7",
218218
"approved_by": null,
219219
"mod_note": null,
@@ -255,7 +255,7 @@
255255
"collapsed_because_crowd_control": null,
256256
"mod_reports": [],
257257
"num_reports": null,
258-
"ups": 3
258+
"ups": 5
259259
}
260260
},
261261
{
@@ -288,7 +288,7 @@
288288
"created_utc": 1751074087,
289289
"send_replies": true,
290290
"parent_id": "t3_1lm76gk",
291-
"score": 1,
291+
"score": 0,
292292
"author_fullname": "t2_16kjuck66n",
293293
"approved_by": null,
294294
"mod_note": null,
@@ -330,7 +330,7 @@
330330
"collapsed_because_crowd_control": null,
331331
"mod_reports": [],
332332
"num_reports": null,
333-
"ups": 1
333+
"ups": 0
334334
}
335335
}
336336
],

src/dump/1lm76yz/index.json

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -116,7 +116,7 @@
116116
"permalink": "/r/LocalLLaMA/comments/1lm76yz/hubert_checkpoint_hubertsoft0d54a1f4pt_for_sovits/",
117117
"stickied": false,
118118
"url": "https://www.reddit.com/r/LocalLLaMA/comments/1lm76yz/hubert_checkpoint_hubertsoft0d54a1f4pt_for_sovits/",
119-
"subreddit_subscribers": 492233,
119+
"subreddit_subscribers": 492572,
120120
"created_utc": 1751063702,
121121
"num_crossposts": 0,
122122
"mod_reports": [],

src/dump/1lm7dox/index.json

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -36,7 +36,7 @@
3636
"name": "t3_1lm7dox",
3737
"quarantine": false,
3838
"link_flair_text_color": "dark",
39-
"upvote_ratio": 0.29,
39+
"upvote_ratio": 0.36,
4040
"author_flair_background_color": null,
4141
"subreddit_type": "public",
4242
"ups": 0,
@@ -116,7 +116,7 @@
116116
"permalink": "/r/LocalLLaMA/comments/1lm7dox/computing_power_to_locally_run_a_model_equivalent/",
117117
"stickied": false,
118118
"url": "https://www.reddit.com/r/LocalLLaMA/comments/1lm7dox/computing_power_to_locally_run_a_model_equivalent/",
119-
"subreddit_subscribers": 492233,
119+
"subreddit_subscribers": 492572,
120120
"created_utc": 1751064196,
121121
"num_crossposts": 0,
122122
"mod_reports": [],

0 commit comments

Comments
 (0)