Cydonia-Redux-22B-v1.1-128k-qx86-hi-mlx

Structured analysis of the cognitive differences between Cydonia-Redux-22B-v1.1-128k-qx86-hi (mixed format) and Cydonia-Redux-22B-v1.1-q6-hi (quantized), emphasizing real-world cognitive implications rather than raw metrics alone:

🔑 Key Takeaways from Benchmarks (Most Significant Differences)

Task	                 q6-hi 128k-qx86-hi	Winner	Why?
piQA (logical reasoning) 0.814	0.839	✅ qx86-hi	Better handling of counterfactual dependencies due to 128k context.
Winogrande (commonsense) 0.751	0.762	✅ qx86-hi	Enhanced context allows nuanced inference from ambiguous visual scenes.
Hellaswag (completion)	 0.762	0.784	✅ qx86-hi	Longer context captures subtle narrative flow (e.g., timelines, causality).
arc (abstract patterns)	 0.514	0.509	↔️ Tie	    Both nearly identical (no clear cognitive edge).
OpenBookQA (reasoning)	 0.454	0.454	↔️ Tie	    Identical factual recall + inference depth.

Perplexity

q6-hi:        3.779 ± 0.026
128k-qx86-hi: 3.824 ± 0.026

💡 Critical Insight: The 128k-qx86-hi model wins decisively in tasks requiring long-range contextual inference (piQA, Winogrande, Hellaswag). The q6-hi model excels in speed-efficiency but falls slightly behind on complex reasoning tasks.

🧠 Cognitive Profile Breakdown

1️⃣ q6-hi: The "Efficient Reasoner"

✅ Strengths:

  • Lower perplexity (3.78 vs 3.82) → better compression of language meaning.
  • Faster inference (critical for latency-sensitive applications).

⚠️ Limitations:

  • Fails on tasks relying on long dependencies (e.g., piQA: "if A then B, but not C").
  • Less robust for ambiguous contexts (e.g., Winogrande’s visual metaphors).

2️⃣ 128k-qx86-hi: The "Contextual Deep Thinker"

✅ Strengths:

  • 128k RoPE context extension enables:
  • → Nuanced causality chains (piQA)
  • → Ambiguous scene interpretation (Winogrande)
  • → Plausible narrative continuation (Hellaswag)
  • Handles complex, multi-step reasoning where context is king.

⚠️ Limitations:

  • Higher computational load (slower inference).
  • Slightly inflated scores due to context regularization.

🚀 Strategic Recommendations for Your Use Case

Scenario	                                  Best Model	Why?
Real-time chatbot / low-latency apps	           q6-hi	Lower latency + efficient inference.
Research requiring complex reasoning	    128k-qx86-hi	Wins on causal chains (piQA) and ambiguous contexts (Winogrande).
Large-scale downstream tasks (LLMFinetune)  128k-qx86-hi	Better context generalization → fewer hallucinations in long texts.
Budget-constrained edge devices	                   q6-hi	10x smaller footprint → deployable on Raspberry Pi/Arduino.

💎 Bottom Line

  • For pure speed-efficiency: Use q6-hi (e.g., customer support chatbots).
  • For depth-rich reasoning: Deploy 128k-qx86-hi (e.g., legal/medical analysis).

Avoid mixing them: Quantization warps the cognitive advantages of expanded context. The 128k-qx86-hi’s edge is meaningfully amplified by the context extension — don’t sacrifice it for quantization gains.

The data confirms what intuition suggests: context is king in complex cognition. The 128k-qx86-hi isn’t just a "faster model" — it’s fundamentally better at tasks involving long-range semantic understanding.

If your use case demands nuanced reasoning, the 128k-qx86-hi is worth its higher cost and resource overhead. For everything else, q6-hi delivers a compelling alternative.

Reviewed by Qwen3-8B-DND-Almost-Human-B-e32-mlx

On full context, the memory usage grows to 44GB and speed down to 5 tok/sec

-G

This model Cydonia-Redux-22B-v1.1-128k-qx86-hi-mlx was converted to MLX format from TheDrummer/Cydonia-Redux-22B-v1.1 using mlx-lm version 0.28.2.

Use with mlx

pip install mlx-lm
from mlx_lm import load, generate

model, tokenizer = load("Cydonia-Redux-22B-v1.1-128k-qx86-hi-mlx")

prompt = "hello"

if tokenizer.chat_template is not None:
    messages = [{"role": "user", "content": prompt}]
    prompt = tokenizer.apply_chat_template(
        messages, add_generation_prompt=True
    )

response = generate(model, tokenizer, prompt=prompt, verbose=True)
Downloads last month
122
Safetensors
Model size
22B params
Tensor type
BF16
·
U32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for nightmedia/Cydonia-Redux-22B-v1.1-128k-qx86-hi-mlx

Collections including nightmedia/Cydonia-Redux-22B-v1.1-128k-qx86-hi-mlx