π Cthulhu-24B-v1
This is a creative, uncensored merge of pre-trained language models created using mergekit.
- Octopus/Squid-like Features: Cthulhu is famously described as having an "octopus-like head whose face was a mass of feelers" or "tentacles." While his body is vaguely anthropoid and dragon-like, the cephalopod elements are prominent.
- Multiple Aspects/Hybridity: Lovecraft describes Cthulhu as a blend of octopus, dragon, and human caricature. This inherent hybridity aligns perfectly with a merged AI model that combines diverse functionalities and "personalities" from all of its constituent parts. Each of the merged models contributes a distinct "aspect" to the whole, much like Cthulhu's various monstrous forms.
- Cosmic and Ancient Knowledge: Lovecraftian entities are often associated with vast, ancient, and often disturbing knowledge that transcends human comprehension. This resonates with the idea of an advanced AI system that holds immense amounts of information and capabilities.
- Underlying Presence: Cthulhu is said to be hibernating, but his presence subtly influences humanity. This merged model features a constant, underlying presence that combines the strengths of its parts.
- Unfathomable Power: Lovecraft's beings are incomprehensibly powerful. This merge aims for a similar sense of enhanced capability. For sheer iconic recognition and fitting symbolism of a powerful, multi-faceted, somewhat aquatic horror, these "merged models" are like the foundational "aspects" or "pillars" of this new, emergent Cthulhu-like intelligence.
Format
</s>[INST] [/INST]
Merge Details
Merge Method
This model was merged using the DARE_TIES merge method using PocketDoc/Dans-PersonalityEngine-V1.3.0-24b as a base.
Models Merged
The following models were included in the merge:
- PocketDoc/Dans-PersonalityEngine-V1.3.0-24b
- TheDrummer/Cydonia-24B-v3.1
- Gryphe/Codex-24B-Small-3.2
- LatitudeGames/Harbinger-24B
- Doctor-Shotgun/MS3.2-24B-Magnum-Diamond
- aixonlab/Eurydice-24b-v3.5
- SicariusSicariiStuff/Impish_Magic_24B
- ReadyArt/MS3.2-The-Omega-Directive-24B-Unslop-v2.0
Configuration
The following YAML configuration was used to produce this model:
base_model: PocketDoc/Dans-PersonalityEngine-V1.3.0-24b
merge_method: dare_ties
dtype: bfloat16
models:
- model: PocketDoc/Dans-PersonalityEngine-V1.3.0-24b
parameters:
density: 0.5
weight: 0.125
- model: TheDrummer/Cydonia-24B-v3.1
parameters:
density: 0.5
weight: 0.125
- model: Gryphe/Codex-24B-Small-3.2
parameters:
density: 0.5
weight: 0.125
- model: LatitudeGames/Harbinger-24B
parameters:
density: 0.5
weight: 0.125
- model: Doctor-Shotgun/MS3.2-24B-Magnum-Diamond
parameters:
density: 0.5
weight: 0.125
- model: aixonlab/Eurydice-24b-v3.5
parameters:
density: 0.5
weight: 0.125
- model: SicariusSicariiStuff/Impish_Magic_24B
parameters:
density: 0.5
weight: 0.125
- model: ReadyArt/MS3.2-The-Omega-Directive-24B-Unslop-v2.0
parameters:
density: 0.5
weight: 0.125
tokenizer:
source: union
chat_template: auto
- Downloads last month
- 34
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
π
Ask for provider support
Model tree for Fentible/Cthulhu-24B-v1
Merge model
this model