๐ GEMBRAIN-31B ๐ง
๐ง About The Model
Gembrain-31B is a synthesis of several models, including Gemsicle-31B as important ingredient. The goal of this release was to stabilize and improve the initial Gemsicle-31B, but also to enhance its logical and lateral thinking, both with and without reasoning.
It's build to create the most unhinged narratives and construct image prompts about anything accordingly to a given structure with high precision.
Expect creative swipe variance, unique and non-robotic prose, and sharper instruction adherence.
๐๏ธ Samplers
| Temperature | 1.0 |
| Top-K | 0 |
| Top-P | 0.95 |
| Min-P | 0.03 |
| DRY Multiplier | 0.8 |
| DRY Base | 1.75 |
| DRY Allowed Length | 10 |
| Optional: Adaptive-P Target | 0.6 |
| Optional: Adaptive-P Decay | 0.5 |
๐ฎ Prompt Format
Please refer to the original google/gemma-4-31b-it for the correct chat template.
Let your frontend handle the chat template if possible (e.g., Chat Completion in SillyTavern).
For Reasoning: Add <|think|> at the very beginning of the system prompt. Thinking happens between <|channel>thought\n and
<channel|> tags.
<|turn>system
<|think|>
You are a helpful assistant<turn|>
<|turn>user
Hello<turn|>
<|turn>model
Hi there<turn|>
<|turn>user
How are you?<turn|>
<|turn>model
๐งช Merge Details
This model was systematically created through a five-stage process of priming models for their given purpose and merging the results:
Phase 01: breadcrumbs_ties
Gemopus X MeroMero
models:
- model: ./G4-MeroMero-31B
- model: ./G4-Gemopus-4-31B-it
merge_method: breadcrumbs_ties
base_model: ./G4-31B-it
parameters:
density: 0.85
weight: 0.5
int8_mask: true
dtype: bfloat16
Phase 02: slerp
GarnetV2 X Musica-v1
models:
- model: ./G4-Gemma4-GarnetV2-31B
- model: ./G4-31B-Musica-v1
merge_method: slerp
base_model: ./G4-Gemma4-GarnetV2-31B
parameters:
t:
- value: 0.6
dtype: bfloat16
Phase 03: della_linear
Gemsicle X Gemma-4-31B-it-heretic-ara
models:
- model: ./Gemsicle-31B
parameters:
weight: 1.0
- model: ./G4-gemma-4-31b-it-heretic-ara
parameters:
weight: 0.75
density: 0.65
merge_method: della_linear
base_model: ./G4-31B-it
parameters:
weight: 1.0
normalize: false
epsilon: 0.05
lambda: 1.0
dtype: bfloat16
Phase 04: model_stock
Phase 01 X Phase 02 X Phase 03
models:
- model: ./phase01_breadcrumbs_ties
- model: ./phase02_slerp
merge_method: model_stock
base_model: ./phase03_della_linear
dtype: bfloat16
tokenizer_source: "base"
Phase 05: arcee_fusion
Gemsicle X Phase 04
models:
- model: ./Gemsicle-31B
- model: ./phase04_model_stock
merge_method: arcee_fusion
base_model: ./Gemsicle-31B
dtype: bfloat16
tokenizer_source: "base"
๐ Credits & Honors
- Downloads last month
- 28