|
13043 | 13043 | - filename: Delta-Vector_Austral-24B-Winton-Q4_K_M.gguf |
13044 | 13044 | sha256: feb76e0158d1ebba1809de89d01671b86037f768ebd5f6fb165885ae6338b1b7 |
13045 | 13045 | uri: huggingface://bartowski/Delta-Vector_Austral-24B-Winton-GGUF/Delta-Vector_Austral-24B-Winton-Q4_K_M.gguf |
| 13046 | +- !!merge <<: *mistral03 |
| 13047 | + name: "mistral-small-3.2-46b-the-brilliant-raconteur-ii-instruct-2506" |
| 13048 | + url: "github:mudler/LocalAI/gallery/chatml.yaml@master" |
| 13049 | + icon: https://huggingface.co/DavidAU/Mistral-Small-3.2-46B-The-Brilliant-Raconteur-II-Instruct-2506/resolve/main/mistral-2506.jpg |
| 13050 | + urls: |
| 13051 | + - https://huggingface.co/DavidAU/Mistral-Small-3.2-46B-The-Brilliant-Raconteur-II-Instruct-2506 |
| 13052 | + - https://huggingface.co/mradermacher/Mistral-Small-3.2-46B-The-Brilliant-Raconteur-II-Instruct-2506-GGUF |
| 13053 | + description: | |
| 13054 | + WARNING: MADNESS - UN HINGED and... NSFW. Vivid prose. INTENSE. Visceral Details. Violence. HORROR. GORE. Swearing. UNCENSORED... humor, romance, fun. |
| 13055 | + Mistral-Small-3.2-46B-The-Brilliant-Raconteur-II-Instruct-2506 |
| 13056 | + |
| 13057 | + This repo contains the full precision source code, in "safe tensors" format to generate GGUFs, GPTQ, EXL2, AWQ, HQQ and other formats. The source code can also be used directly. |
| 13058 | + |
| 13059 | + ABOUT: |
| 13060 | + |
| 13061 | + A stronger, more creative Mistral (Mistral-Small-3.2-24B-Instruct-2506) extended to 79 layers, 46B parameters with Brainstorm 40x by DavidAU (details at very bottom of the page). This is version II, which has a jump in detail, and raw emotion relative to version 1. |
| 13062 | + |
| 13063 | + This model pushes Mistral's Instruct 2506 to the limit: |
| 13064 | + |
| 13065 | + Regens will be very different, even with same prompt / settings. |
| 13066 | + Output generation will vary vastly on each generation. |
| 13067 | + Reasoning will be changed, and often shorter. |
| 13068 | + Prose, creativity, word choice, and general "flow" are improved. |
| 13069 | + Several system prompts below help push this model even further. |
| 13070 | + Model is partly de-censored / abliterated. Most Mistrals are more uncensored that most other models too. |
| 13071 | + This model can also be used for coding too; even at low quants. |
| 13072 | + Model can be used for all use cases too. |
| 13073 | + |
| 13074 | + As this is an instruct model, this model thrives on instructions - both in the system prompt and/or the prompt itself. |
| 13075 | + |
| 13076 | + One example below with 3 generations using Q4_K_S. |
| 13077 | + |
| 13078 | + Second example below with 2 generations using Q4_K_S. |
| 13079 | + |
| 13080 | + Quick Details: |
| 13081 | + |
| 13082 | + Model is 128k context, Jinja template (embedded) OR Chatml Template. |
| 13083 | + Reasoning can be turned on/off (see system prompts below) and is OFF by default. |
| 13084 | + Temp range .1 to 1 suggested, with 1-2 for enhanced creative. Above temp 2, is strong but can be very different. |
| 13085 | + Rep pen range: 1 (off) or very light 1.01, 1.02 to 1.05. (model is sensitive to rep pen - this affects reasoning / generation length.) |
| 13086 | + For creative/brainstorming use: suggest 2-5 generations due to variations caused by Brainstorm. |
| 13087 | + |
| 13088 | + Observations: |
| 13089 | + |
| 13090 | + Sometimes using Chatml (or Alpaca / others ) template (VS Jinja) will result in stronger creative generation. |
| 13091 | + Model can be operated with NO system prompt; however a system prompt will enhance generation. |
| 13092 | + Longer prompts, that more detailed, with more instructions will result in much stronger generations. |
| 13093 | + For prose directives: You may need to add directions, because the model may follow your instructions too closely. IE: "use short sentences" vs "use short sentences sparsely". |
| 13094 | + Reasoning (on) can lead to better creative generation, however sometimes generation with reasoning off is better. |
| 13095 | + Rep pen of up to 1.05 may be needed on quants Q2k/q3ks for some prompts to address "low bit" issues. |
| 13096 | + |
| 13097 | + Detailed settings, system prompts, how to and examples below. |
| 13098 | + |
| 13099 | + NOTES: |
| 13100 | + |
| 13101 | + Image generation should also be possible with this model, just like the base model. Brainstorm was not applied to the image generation systems of the model... yet. |
| 13102 | + |
| 13103 | + This is Version II and subject to change / revision. |
| 13104 | + |
| 13105 | + This model is a slightly different version of: |
| 13106 | + |
| 13107 | + https://huggingface.co/DavidAU/Mistral-Small-3.2-46B-The-Brilliant-Raconteur-Instruct-2506 |
| 13108 | + overrides: |
| 13109 | + parameters: |
| 13110 | + model: Mistral-Small-3.2-46B-The-Brilliant-Raconteur-II-Instruct-2506.Q4_K_M.gguf |
| 13111 | + files: |
| 13112 | + - filename: Mistral-Small-3.2-46B-The-Brilliant-Raconteur-II-Instruct-2506.Q4_K_M.gguf |
| 13113 | + sha256: 5c8b6f21ae4f671880fafe60001f30f4c639a680e257701e474777cfcf00f8f6 |
| 13114 | + uri: huggingface://mradermacher/Mistral-Small-3.2-46B-The-Brilliant-Raconteur-II-Instruct-2506-GGUF/Mistral-Small-3.2-46B-The-Brilliant-Raconteur-II-Instruct-2506.Q4_K_M.gguf |
13046 | 13115 | - &mudler |
13047 | 13116 | url: "github:mudler/LocalAI/gallery/mudler.yaml@master" ### START mudler's LocalAI specific-models |
13048 | 13117 | name: "LocalAI-llama3-8b-function-call-v0.2" |
|
0 commit comments