Quazim0t0/RZA-14B-sce

Quazim0t0/RZA-14B-sce is a merge of the following models: RZA & GZA Models provided me data to which model ranked to influence the merge. RZA blended somewhere between two models with different variations to answers questions in different formats. While GZA continued to use one type of formatting. https://huggingface.co/Quazim0t0/RZA-14B-sce

🧩 Configuration

models:
  # Pivot model
  - model: Quazim0t0/ThinkPhi1.1-Tensors
  # Target models
  - model: Quazim0t0/Phi4.Turn.R1Distill.16bit
  - model: Pinkstack/SuperThoughts-CoT-14B-16k-o1-QwQ
  - model: prithivMLmods/Phi-4-QwQ
merge_method: sce
base_model: Quazim0t0/ThinkPhi1.1-Tensors
parameters:
  select_topk: 1.0
dtype: bfloat16

πŸ’» Usage

!pip install -qU transformers accelerate

from transformers import AutoTokenizer
import transformers
import torch

model = "Quazim0t0/RZA-14B-sce"
messages = [{"role": "user", "content": "What is a large language model?"}]

tokenizer = AutoTokenizer.from_pretrained(model)
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
pipeline = transformers.pipeline(
    "text-generation",
    model=model,
    torch_dtype=torch.float16,
    device_map="auto",
)

outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
print(outputs[0]["generated_text"])
Downloads last month
11
GGUF
Model size
14.7B params
Architecture
llama

4-bit

Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.

Collection including Quazim0t0/RZA-GGUF