Skip to content

Commit cb87d33

Browse files
authored
chore(model gallery): add sicariussicariistuff_impish_llama_4b (#5799)
Signed-off-by: Ettore Di Giacinto <[email protected]>
1 parent 6dfc962 commit cb87d33

File tree

1 file changed

+26
-0
lines changed

1 file changed

+26
-0
lines changed

gallery/index.yaml

Lines changed: 26 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -11030,6 +11030,32 @@
1103011030
- filename: ockerman0_AnubisLemonade-70B-v1-Q4_K_M.gguf
1103111031
sha256: 44a06924a131fafde604a6c4e2f9f5209b9e79452b2211c9dbb0b14a1e177c43
1103211032
uri: huggingface://bartowski/ockerman0_AnubisLemonade-70B-v1-GGUF/ockerman0_AnubisLemonade-70B-v1-Q4_K_M.gguf
11033+
- !!merge <<: *llama31
11034+
name: "sicariussicariistuff_impish_llama_4b"
11035+
icon: https://huggingface.co/SicariusSicariiStuff/Impish_LLAMA_4B/resolve/main/Images/Impish_LLAMA_4B.png
11036+
urls:
11037+
- https://huggingface.co/SicariusSicariiStuff/Impish_LLAMA_4B
11038+
- https://huggingface.co/bartowski/SicariusSicariiStuff_Impish_LLAMA_4B-GGUF
11039+
description: |
11040+
5th of May, 2025, Impish_LLAMA_4B.
11041+
11042+
Almost a year ago, I created Impish_LLAMA_3B, the first fully coherent 3B roleplay model at the time. It was quickly adopted by some platforms, as well as one of the go-to models for mobile. After some time, I made Fiendish_LLAMA_3B and insisted it was not an upgrade, but a different flavor (which was indeed the case, as a different dataset was used to tune it).
11043+
11044+
Impish_LLAMA_4B, however, is an upgrade, a big one. I've had over a dozen 4B candidates, but none of them were 'worthy' of the Impish badge. This model has superior responsiveness and context awareness, and is able to pull off very coherent adventures. It even comes with some additional assistant capabilities too. Of course, while it is exceptionally competent for its size, it is still 4B. Manage expectations and all that. I, however, am very much pleased with it. It took several tries to pull off just right. Total tokens trained: about 400m (due to being a generalist model, lots of tokens went there, despite the emphasis on roleplay & adventure).
11045+
11046+
This took more effort than I thought it would. Because of course it would. This is mainly due to me refusing to release a model only 'slightly better' than my two 3B models mentioned above. Because "what would be the point" in that? The reason I included so many tokens for this tune is that small models are especially sensitive to many factors, including the percentage of moisture in the air and how many times I ran nvidia-smi since the system last started.
11047+
11048+
It's no secret that roleplay/creative writing models can reduce a model's general intelligence (any tune and RL risk this, but roleplay models are especially 'fragile'). Therefore, additional tokens of general assistant data were needed in my opinion, and indeed seemed to help a lot with retaining intelligence.
11049+
11050+
This model is also 'built a bit different', literally, as it is based on nVidia's prune; it does not 'behave' like a typical 8B, from my own subjective impression. This helped a lot with keeping it smart at such size.
11051+
To be honest, my 'job' here in open source is 'done' at this point. I've achieved everything I wanted to do here, and then some.
11052+
overrides:
11053+
parameters:
11054+
model: SicariusSicariiStuff_Impish_LLAMA_4B-Q4_K_M.gguf
11055+
files:
11056+
- filename: SicariusSicariiStuff_Impish_LLAMA_4B-Q4_K_M.gguf
11057+
sha256: 84d14bf15e198465336220532cb0fbcbdad81b33f1ab6748551218ee432208f6
11058+
uri: huggingface://bartowski/SicariusSicariiStuff_Impish_LLAMA_4B-GGUF/SicariusSicariiStuff_Impish_LLAMA_4B-Q4_K_M.gguf
1103311059
- &deepseek
1103411060
url: "github:mudler/LocalAI/gallery/deepseek.yaml@master" ## Deepseek
1103511061
name: "deepseek-coder-v2-lite-instruct"

0 commit comments

Comments
 (0)