Support this work: donate.sybilsolutions.ai

REAP surfaces: GLM | MiniMax | Qwen | Gemma | Paper | Code | PR17 | Cerebras Collection

Qwen3.5-264B-REAP

  • Repository: 0xSero/Qwen3.5-264B-REAP
  • Base model: Qwen/Qwen3.5-397B-A17B
  • Artifact kind: pruned
  • Compression ratio: 34%
  • Prune metric: reap

Details

  • Maintainer: 0xSero
  • Organization: Sybil Solutions
  • Project: REAP PR17
  • Hub owner: 0xSero
  • Summary: BF16 REAP-pruned Qwen3.5-397B-A17B with 176 of 512 experts removed per MoE layer, retaining 336 experts per layer, for an estimated 264B total parameters.

Provenance

  • Observer state: /home/ubuntu/qwen397-full/observer-calibv1/qwen397-pr17-calibv1-23k-16k-observer-state.raw.pt
  • Detail state: /home/ubuntu/qwen397-full/observer-calibv1/qwen397-pr17-calibv1-23k-16k-detail-state.raw.pt

Benchmarks

No benchmark summary was found.

Custom Stress

No custom stress summary was found.

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("0xSero/Qwen3.5-264B-REAP", trust_remote_code=True)
tokenizer = AutoTokenizer.from_pretrained("0xSero/Qwen3.5-264B-REAP", trust_remote_code=True)

Sponsors

Thank you for the kind sponsors, wouldn't be possible without them:

  • Nvidia
  • TNG Technology
  • Lambda
  • Prime Intellect
  • HotAisle
Downloads last month
1,109
Safetensors
Model size
263B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for 0xSero/Qwen3.5-264B-REAP

Finetuned
(28)
this model
Quantizations
1 model

Datasets used to train 0xSero/Qwen3.5-264B-REAP

Space using 0xSero/Qwen3.5-264B-REAP 1

Paper for 0xSero/Qwen3.5-264B-REAP