Hugging Face
Models
Datasets
Spaces
Buckets
new
Docs
Enterprise
Pricing
Log In
Sign Up
3107.2
TFLOPS
48
6
414
seawolf
PRO
seawolf2357
Follow
IronNeuron's profile picture
jbl2244's profile picture
basil121ert's profile picture
349 followers
·
230 following
AI & ML interests
None yet
Recent Activity
published
a Space
1 day ago
Heartsync/MARL-mobile-backup
reacted
to
SeaWolf-AI
's
post
with 🔥
4 days ago
Darwin-TTS: 3% of an LLM's Brain Makes TTS Speak with Emotion — Zero Training We blended 3% of Qwen3-1.7B (LLM) FFN weights into Qwen3-TTS-1.7B's talker module. The result: emotionally enhanced speech synthesis — with zero training, zero data, and zero GPU hours. Try the Demo: https://huggingface.co/spaces/FINAL-Bench/Darwin-TTS-1.7B-Cross Model Weights: https://huggingface.co/FINAL-Bench/Darwin-TTS-1.7B-Cross Full Research Article: https://huggingface.co/blog/FINAL-Bench/darwin-tts Qwen3-1.7B (LLM) and Qwen3-TTS-1.7B's talker share 100% identical architecture — same hidden_size (2048), same layers (28), same heads (16). This enabled pure 1:1 weight blending across 84 FFN tensors with a single lerp operation. At 3% blend, emotion appears. At 5%, emotion intensifies. At 10%, the model breaks — producing 655-second outputs for a 3-second sentence, because the LLM's "keep generating" pattern overwhelms the TTS stop signal. To our knowledge, this is the first training-free cross-modal weight transfer between an LLM and a TTS model. Prior work either requires adapter training (SmolTolk, 2025), fine-tuning (CSLM, 2025), or massive end-to-end compute (GPT-4o). Darwin-TTS achieves cross-modal capability transfer in under 2 minutes on CPU. The key insight: TTS models with LLM backbones already "think" in language. We're just restoring 3% of the original LLM's language understanding patterns — particularly those related to emotional semantics and prosody planning. The code is three lines: load the model, load the LLM FFN, call p.lerp_(llm_weight, 0.03). creators of the Darwin Evolutionary Merge Framework. Darwin LLM V7 achieved GPQA Diamond 86.9% (HF Benchmark #3) through CMA-ES optimized FFN crossbreeding. Darwin-TTS extends this principle from LLM-to-LLM merging into cross-modal LLM-to-TTS transfer. Apache 2.0.
upvoted
an
article
4 days ago
Darwin-TTS: We Gave a TTS Model 3% of an LLM's Brain — It Started Showing Emotion
View all activity
Organizations
seawolf2357
's models
39
Sort: Recently updated
seawolf2357/Meta-Llama-3-8B-Instruct-Q4_K_M-GGUF
Text Generation
•
8B
•
Updated
May 14, 2024
•
32
seawolf2357/cerberus-llama-3-8b-instruct-gguf
Text Generation
•
8B
•
Updated
May 14, 2024
•
23
seawolf2357/kollm8
Text Generation
•
7B
•
Updated
May 12, 2024
seawolf2357/kollm7
Text Generation
•
29B
•
Updated
May 12, 2024
•
1
seawolf2357/kollm3
Text Generation
•
7B
•
Updated
May 12, 2024
seawolf2357/mergetest2
Text Generation
•
7B
•
Updated
May 12, 2024
•
1
seawolf2357/llama7b-merge
Text Generation
•
7B
•
Updated
May 12, 2024
•
2
seawolf2357/c4ai-command-r-plus-Q4_K_M-GGUF
Updated
Apr 28, 2024
seawolf2357/Phi-3-mini-128k-instruct-Q4_K_M-GGUF
Text Generation
•
4B
•
Updated
Apr 28, 2024
•
92
Previous
1
2
Next