|
36 | 36 | "name": "t3_1lojd3e", |
37 | 37 | "quarantine": false, |
38 | 38 | "link_flair_text_color": "dark", |
39 | | - "upvote_ratio": 0.85, |
| 39 | + "upvote_ratio": 0.84, |
40 | 40 | "author_flair_background_color": null, |
41 | 41 | "subreddit_type": "public", |
42 | | - "ups": 10, |
| 42 | + "ups": 9, |
43 | 43 | "total_awards_received": 0, |
44 | 44 | "media_embed": {}, |
45 | 45 | "thumbnail_width": null, |
|
53 | 53 | "secure_media_embed": {}, |
54 | 54 | "link_flair_text": "Question | Help", |
55 | 55 | "can_mod_post": false, |
56 | | - "score": 10, |
| 56 | + "score": 9, |
57 | 57 | "approved_by": null, |
58 | 58 | "is_created_from_ads_ui": false, |
59 | 59 | "author_premium": false, |
|
107 | 107 | "report_reasons": null, |
108 | 108 | "author": "el_pr3sid3nt3", |
109 | 109 | "discussion_type": null, |
110 | | - "num_comments": 6, |
| 110 | + "num_comments": 8, |
111 | 111 | "send_replies": true, |
112 | 112 | "media": null, |
113 | 113 | "contest_mode": false, |
|
116 | 116 | "permalink": "/r/LocalLLaMA/comments/1lojd3e/gemma3n_vram_usage/", |
117 | 117 | "stickied": false, |
118 | 118 | "url": "https://www.reddit.com/r/LocalLLaMA/comments/1lojd3e/gemma3n_vram_usage/", |
119 | | - "subreddit_subscribers": 493242, |
| 119 | + "subreddit_subscribers": 493458, |
120 | 120 | "created_utc": 1751317817, |
121 | 121 | "num_crossposts": 0, |
122 | 122 | "mod_reports": [], |
|
219 | 219 | "subreddit": "LocalLLaMA", |
220 | 220 | "author_flair_template_id": null, |
221 | 221 | "likes": null, |
222 | | - "replies": "", |
| 222 | + "replies": { |
| 223 | + "kind": "Listing", |
| 224 | + "data": { |
| 225 | + "after": null, |
| 226 | + "dist": null, |
| 227 | + "modhash": "", |
| 228 | + "geo_filter": "", |
| 229 | + "children": [ |
| 230 | + { |
| 231 | + "kind": "t1", |
| 232 | + "data": { |
| 233 | + "subreddit_id": "t5_81eyvm", |
| 234 | + "approved_at_utc": null, |
| 235 | + "author_is_blocked": false, |
| 236 | + "comment_type": null, |
| 237 | + "awarders": [], |
| 238 | + "mod_reason_by": null, |
| 239 | + "banned_by": null, |
| 240 | + "author_flair_type": "text", |
| 241 | + "total_awards_received": 0, |
| 242 | + "subreddit": "LocalLLaMA", |
| 243 | + "author_flair_template_id": null, |
| 244 | + "distinguished": null, |
| 245 | + "likes": null, |
| 246 | + "replies": "", |
| 247 | + "user_reports": [], |
| 248 | + "saved": false, |
| 249 | + "id": "n0sbt5a", |
| 250 | + "banned_at_utc": null, |
| 251 | + "mod_reason_title": null, |
| 252 | + "gilded": 0, |
| 253 | + "archived": false, |
| 254 | + "collapsed_reason_code": null, |
| 255 | + "no_follow": true, |
| 256 | + "author": "vk3r", |
| 257 | + "can_mod_post": false, |
| 258 | + "send_replies": true, |
| 259 | + "parent_id": "t1_n0qd1my", |
| 260 | + "score": 1, |
| 261 | + "author_fullname": "t2_hyklw8a", |
| 262 | + "removal_reason": null, |
| 263 | + "approved_by": null, |
| 264 | + "mod_note": null, |
| 265 | + "all_awardings": [], |
| 266 | + "body": "I think you don't understand enough of the subject. That the model mentions that it occupies 2-3GB is an approximate weight that will depend on the architecture for which it was made, the tools that are occupying, the context that it has and the quantization occupied.\n\nIt is never exact.\n\nAbout the quantizations, search in Hugginface and depending on the tool you use to build the model, you can find one quantized by someone. Unsloth and Bartwski are known for their work.", |
| 267 | + "edited": false, |
| 268 | + "author_flair_css_class": null, |
| 269 | + "name": "t1_n0sbt5a", |
| 270 | + "is_submitter": false, |
| 271 | + "downs": 0, |
| 272 | + "author_flair_richtext": [], |
| 273 | + "author_patreon_flair": false, |
| 274 | + "body_html": "<div class=\"md\"><p>I think you don&#39;t understand enough of the subject. That the model mentions that it occupies 2-3GB is an approximate weight that will depend on the architecture for which it was made, the tools that are occupying, the context that it has and the quantization occupied.</p>\n\n<p>It is never exact.</p>\n\n<p>About the quantizations, search in Hugginface and depending on the tool you use to build the model, you can find one quantized by someone. Unsloth and Bartwski are known for their work.</p>\n</div>", |
| 275 | + "gildings": {}, |
| 276 | + "collapsed_reason": null, |
| 277 | + "link_id": "t3_1lojd3e", |
| 278 | + "associated_award": null, |
| 279 | + "stickied": false, |
| 280 | + "author_premium": false, |
| 281 | + "can_gild": false, |
| 282 | + "top_awarded_type": null, |
| 283 | + "unrepliable_reason": null, |
| 284 | + "author_flair_text_color": null, |
| 285 | + "treatment_tags": [], |
| 286 | + "score_hidden": false, |
| 287 | + "permalink": "/r/LocalLLaMA/comments/1lojd3e/gemma3n_vram_usage/n0sbt5a/", |
| 288 | + "subreddit_type": "public", |
| 289 | + "locked": false, |
| 290 | + "report_reasons": null, |
| 291 | + "created": 1751388313, |
| 292 | + "author_flair_text": null, |
| 293 | + "collapsed": false, |
| 294 | + "created_utc": 1751388313, |
| 295 | + "subreddit_name_prefixed": "r/LocalLLaMA", |
| 296 | + "controversiality": 0, |
| 297 | + "depth": 4, |
| 298 | + "author_flair_background_color": null, |
| 299 | + "collapsed_because_crowd_control": null, |
| 300 | + "mod_reports": [], |
| 301 | + "num_reports": null, |
| 302 | + "ups": 1 |
| 303 | + } |
| 304 | + } |
| 305 | + ], |
| 306 | + "before": null |
| 307 | + } |
| 308 | + }, |
223 | 309 | "user_reports": [], |
224 | 310 | "saved": false, |
225 | 311 | "id": "n0qd1my", |
|
233 | 319 | "can_mod_post": false, |
234 | 320 | "send_replies": true, |
235 | 321 | "parent_id": "t1_n0oc69x", |
236 | | - "score": 1, |
| 322 | + "score": 0, |
237 | 323 | "author_fullname": "t2_6hmjiu4n", |
238 | 324 | "removal_reason": null, |
239 | 325 | "approved_by": null, |
|
276 | 362 | "collapsed_because_crowd_control": null, |
277 | 363 | "mod_reports": [], |
278 | 364 | "num_reports": null, |
279 | | - "ups": 1 |
| 365 | + "ups": 0 |
280 | 366 | } |
281 | 367 | } |
282 | 368 | ], |
|
423 | 509 | "created_utc": 1751325210, |
424 | 510 | "send_replies": true, |
425 | 511 | "parent_id": "t3_1lojd3e", |
426 | | - "score": 5, |
| 512 | + "score": 6, |
427 | 513 | "author_fullname": "t2_hyklw8a", |
428 | 514 | "approved_by": null, |
429 | 515 | "mod_note": null, |
|
465 | 551 | "collapsed_because_crowd_control": null, |
466 | 552 | "mod_reports": [], |
467 | 553 | "num_reports": null, |
468 | | - "ups": 5 |
| 554 | + "ups": 6 |
| 555 | + } |
| 556 | + }, |
| 557 | + { |
| 558 | + "kind": "t1", |
| 559 | + "data": { |
| 560 | + "subreddit_id": "t5_81eyvm", |
| 561 | + "approved_at_utc": null, |
| 562 | + "author_is_blocked": false, |
| 563 | + "comment_type": null, |
| 564 | + "awarders": [], |
| 565 | + "mod_reason_by": null, |
| 566 | + "banned_by": null, |
| 567 | + "author_flair_type": "text", |
| 568 | + "total_awards_received": 0, |
| 569 | + "subreddit": "LocalLLaMA", |
| 570 | + "author_flair_template_id": null, |
| 571 | + "likes": null, |
| 572 | + "replies": "", |
| 573 | + "user_reports": [], |
| 574 | + "saved": false, |
| 575 | + "id": "n0rasvn", |
| 576 | + "banned_at_utc": null, |
| 577 | + "mod_reason_title": null, |
| 578 | + "gilded": 0, |
| 579 | + "archived": false, |
| 580 | + "collapsed_reason_code": null, |
| 581 | + "no_follow": true, |
| 582 | + "author": "sciencewarrior", |
| 583 | + "can_mod_post": false, |
| 584 | + "created_utc": 1751377674, |
| 585 | + "send_replies": true, |
| 586 | + "parent_id": "t3_1lojd3e", |
| 587 | + "score": 4, |
| 588 | + "author_fullname": "t2_4feaa", |
| 589 | + "approved_by": null, |
| 590 | + "mod_note": null, |
| 591 | + "all_awardings": [], |
| 592 | + "collapsed": false, |
| 593 | + "body": "From what their model cards suggest, the software needs to support their architecture to make it work. Make sure you are running the latest version of llama.cpp. This tutorial should be handy: https://docs.unsloth.ai/basics/gemma-3n-how-to-run-and-fine-tune", |
| 594 | + "edited": 1751404517, |
| 595 | + "top_awarded_type": null, |
| 596 | + "author_flair_css_class": null, |
| 597 | + "name": "t1_n0rasvn", |
| 598 | + "is_submitter": false, |
| 599 | + "downs": 0, |
| 600 | + "author_flair_richtext": [], |
| 601 | + "author_patreon_flair": false, |
| 602 | + "body_html": "<div class=\"md\"><p>From what their model cards suggest, the software needs to support their architecture to make it work. Make sure you are running the latest version of llama.cpp. This tutorial should be handy: <a href=\"https://docs.unsloth.ai/basics/gemma-3n-how-to-run-and-fine-tune\">https://docs.unsloth.ai/basics/gemma-3n-how-to-run-and-fine-tune</a></p>\n</div>", |
| 603 | + "removal_reason": null, |
| 604 | + "collapsed_reason": null, |
| 605 | + "distinguished": null, |
| 606 | + "associated_award": null, |
| 607 | + "stickied": false, |
| 608 | + "author_premium": false, |
| 609 | + "can_gild": false, |
| 610 | + "gildings": {}, |
| 611 | + "unrepliable_reason": null, |
| 612 | + "author_flair_text_color": null, |
| 613 | + "score_hidden": false, |
| 614 | + "permalink": "/r/LocalLLaMA/comments/1lojd3e/gemma3n_vram_usage/n0rasvn/", |
| 615 | + "subreddit_type": "public", |
| 616 | + "locked": false, |
| 617 | + "report_reasons": null, |
| 618 | + "created": 1751377674, |
| 619 | + "author_flair_text": null, |
| 620 | + "treatment_tags": [], |
| 621 | + "link_id": "t3_1lojd3e", |
| 622 | + "subreddit_name_prefixed": "r/LocalLLaMA", |
| 623 | + "controversiality": 0, |
| 624 | + "depth": 0, |
| 625 | + "author_flair_background_color": null, |
| 626 | + "collapsed_because_crowd_control": null, |
| 627 | + "mod_reports": [], |
| 628 | + "num_reports": null, |
| 629 | + "ups": 4 |
469 | 630 | } |
470 | 631 | }, |
471 | 632 | { |
|
0 commit comments