|
4 | 4 | "data": { |
5 | 5 | "after": null, |
6 | 6 | "dist": 1, |
7 | | - "modhash": "dy0634ws1m6eb4c30dd5fb8bd868b0a672db766f5fbe81224b", |
| 7 | + "modhash": "", |
8 | 8 | "geo_filter": "", |
9 | 9 | "children": [ |
10 | 10 | { |
11 | 11 | "kind": "t3", |
12 | 12 | "data": { |
13 | 13 | "approved_at_utc": null, |
14 | 14 | "subreddit": "LocalLLaMA", |
15 | | - "selftext": "# Hey everyone! 👋\n\nI’ve built an **open snapshot** of r/LocalLLaMA to help preserve its discussions, discoveries, and resources for all of us - *especially given how shaky things can get with subs these days*.\n\n👉 **Dive into the live archive:** [https://maifeeulasad.github.io/LocalLLaMA/](https://maifeeulasad.github.io/LocalLLaMA/)\n\nThis little bot quietly **fetches and saves new posts every hour**, so all the local LLM experiments, model links, feedback threads, and tips stay safe and easy to explore - *now and in the future*.\n\nI stitched this together with **React, Ant Design, Node.js**, and a pinch of automation magic. It runs on its own, taking snapshots and refreshing the archive around the clock.\n\n💡 *Don’t trust me?* Fork it. Run it yourself. The goal is simple: **keep the knowledge free**.\n\n**⚡ NB:** Right now, this only auto-archives **new posts**. I’d love help figuring out how to **scrape and backfill older gems** too - so we can lock in the full history. *This needs everyone’s ideas and support!*\n\nIf this helps you, please **star the repo** ❤️ [https://github.com/maifeeulasad/LocalLLaMA](https://github.com/maifeeulasad/LocalLLaMA) \\- issues, ideas, and pull requests are all welcome!\n\nI’ve learned so much from this sub - this is just my small way of giving back. Let’s keep **open models** and **shared knowledge** alive and accessible, no matter what. 🌍✨", |
| 15 | + "selftext": "[removed]", |
16 | 16 | "user_reports": [], |
17 | 17 | "saved": false, |
18 | 18 | "mod_reason_title": null, |
|
41 | 41 | "subreddit_type": "public", |
42 | 42 | "ups": 1, |
43 | 43 | "total_awards_received": 0, |
44 | | - "media_embed": { |
45 | | - |
46 | | - }, |
| 44 | + "media_embed": {}, |
47 | 45 | "thumbnail_width": null, |
48 | 46 | "author_flair_template_id": "50c36eba-fdca-11ee-9735-92a88d7e3b87", |
49 | 47 | "is_original_content": false, |
|
52 | 50 | "is_reddit_media_domain": false, |
53 | 51 | "is_meta": false, |
54 | 52 | "category": null, |
55 | | - "secure_media_embed": { |
56 | | - |
57 | | - }, |
| 53 | + "secure_media_embed": {}, |
58 | 54 | "link_flair_text": "Discussion", |
59 | 55 | "can_mod_post": false, |
60 | 56 | "score": 1, |
61 | 57 | "approved_by": null, |
62 | 58 | "is_created_from_ads_ui": false, |
63 | 59 | "author_premium": false, |
64 | | - "thumbnail": "self", |
| 60 | + "thumbnail": "default", |
65 | 61 | "edited": false, |
66 | 62 | "author_flair_css_class": null, |
67 | 63 | "author_flair_richtext": [ |
|
70 | 66 | "t": "Ollama" |
71 | 67 | } |
72 | 68 | ], |
73 | | - "gildings": { |
74 | | - |
75 | | - }, |
| 69 | + "gildings": {}, |
76 | 70 | "content_categories": null, |
77 | 71 | "is_self": true, |
78 | 72 | "mod_note": null, |
|
84 | 78 | "author_flair_type": "richtext", |
85 | 79 | "domain": "self.LocalLLaMA", |
86 | 80 | "allow_live_comments": false, |
87 | | - "selftext_html": "<!-- SC_OFF --><div class=\"md\"><h1>Hey everyone! 👋</h1>\n\n<p>I’ve built an <strong>open snapshot</strong> of <a href=\"/r/LocalLLaMA\">r/LocalLLaMA</a> to help preserve its discussions, discoveries, and resources for all of us - <em>especially given how shaky things can get with subs these days</em>.</p>\n\n<p>👉 <strong>Dive into the live archive:</strong> <a href=\"https://maifeeulasad.github.io/LocalLLaMA/\">https://maifeeulasad.github.io/LocalLLaMA/</a></p>\n\n<p>This little bot quietly <strong>fetches and saves new posts every hour</strong>, so all the local LLM experiments, model links, feedback threads, and tips stay safe and easy to explore - <em>now and in the future</em>.</p>\n\n<p>I stitched this together with <strong>React, Ant Design, Node.js</strong>, and a pinch of automation magic. It runs on its own, taking snapshots and refreshing the archive around the clock.</p>\n\n<p>💡 <em>Don’t trust me?</em> Fork it. Run it yourself. The goal is simple: <strong>keep the knowledge free</strong>.</p>\n\n<p><strong>⚡ NB:</strong> Right now, this only auto-archives <strong>new posts</strong>. I’d love help figuring out how to <strong>scrape and backfill older gems</strong> too - so we can lock in the full history. <em>This needs everyone’s ideas and support!</em></p>\n\n<p>If this helps you, please <strong>star the repo</strong> ❤️ <a href=\"https://github.com/maifeeulasad/LocalLLaMA\">https://github.com/maifeeulasad/LocalLLaMA</a> - issues, ideas, and pull requests are all welcome!</p>\n\n<p>I’ve learned so much from this sub - this is just my small way of giving back. Let’s keep <strong>open models</strong> and <strong>shared knowledge</strong> alive and accessible, no matter what. 🌍✨</p>\n</div><!-- SC_ON -->", |
88 | | - "likes": true, |
| 81 | + "selftext_html": "<!-- SC_OFF --><div class=\"md\"><p>[removed]</p>\n</div><!-- SC_ON -->", |
| 82 | + "likes": null, |
89 | 83 | "suggested_sort": null, |
90 | 84 | "banned_at_utc": null, |
91 | 85 | "view_count": null, |
92 | 86 | "archived": false, |
93 | | - "no_follow": false, |
| 87 | + "no_follow": true, |
94 | 88 | "is_crosspostable": false, |
95 | 89 | "pinned": false, |
96 | 90 | "over_18": false, |
|
103 | 97 | "locked": false, |
104 | 98 | "author_flair_text": "Ollama", |
105 | 99 | "treatment_tags": [], |
106 | | - "rte_mode": "richtext", |
107 | 100 | "visited": false, |
108 | 101 | "removed_by": null, |
109 | 102 | "num_reports": null, |
|
128 | 121 | "permalink": "/r/LocalLLaMA/comments/1lmjg3p/helping_archive_rlocalllama_for_everyone_for_open/", |
129 | 122 | "stickied": false, |
130 | 123 | "url": "https://www.reddit.com/r/LocalLLaMA/comments/1lmjg3p/helping_archive_rlocalllama_for_everyone_for_open/", |
131 | | - "subreddit_subscribers": 492256, |
| 124 | + "subreddit_subscribers": 492572, |
132 | 125 | "created_utc": 1751105890, |
133 | 126 | "num_crossposts": 0, |
134 | 127 | "mod_reports": [], |
|
144 | 137 | "data": { |
145 | 138 | "after": null, |
146 | 139 | "dist": null, |
147 | | - "modhash": "dy0634ws1m6eb4c30dd5fb8bd868b0a672db766f5fbe81224b", |
| 140 | + "modhash": "", |
148 | 141 | "geo_filter": "", |
149 | 142 | "children": [], |
150 | 143 | "before": null |
|
0 commit comments