NoesisLab advances machine learning research in deep contemplation and reflective reasoning to enable more profound and self-aware artificial intelligence.
Arcade-3B — SmolReasoner NoesisLab/Arcade-3B Arcade-3B is a 3B instruction-following and reasoning model built on SmolLM3-3B. It is the public release from the ARCADE project at NoesisLab, which investigates the State–Constraint Orthogonality Hypothesis: standard Transformer hidden states conflate factual content and reasoning structure in the same subspace, and explicitly decoupling them improves generalization.
We deleted the Embedding Layer -- INTRO Our Collins-Embedding-3M NoesisLab/Collins-Embedding-3M Most "small" models are just giant vocab tables in a trench coat. Collins-3M changes that. By using 2-Universal Hashing and Chernoff-bound noise suppression, we’ve collapsed the embedding space into a fixed O(1) hash-map. * STSB: 0.7114 (Beating many 100M+ models) * Size: 3M (Edge-ready, IoT-ready) * Tech: Randomized Sign-Hashing + RoPE positional injection. Built by NoesisLab
🔥 UPGRADE in Kai: 30B Scaling! 🔥 NoesisLab/Kai-30B-Instruct NoesisLab/Kai-30B-Instruct We are incredibly excited to announce that the Kai-30B-Instruct model and its official Space are now LIVE! 🚀 If you've been following the journey from Kai-0.35B to Kai-3B, you know we're rethinking how models reason. Tired of verbose, slow Chain-of-Thought (CoT) outputs that flood your screen with self-talk? So are we. Kai-30B-Instruct scales up our Adaptive Dual-Search Distillation (ADS) framework. By bridging classical A* heuristic search with continuous gradient descent , we use an information-theoretic log-barrier to physically prune high-entropy reasoning paths during training. The result? Pure implicit reasoning. The model executes structured logic, arithmetic carries, and branch selections as a reflex in a single forward pass—no external scaffolding required. At 3B, we observed a phase transition where the model achieved "logical crystallization". Now, at 30B, we are giving the ADS regularizer the massive representational capacity it needs to tackle higher-order symbolic abstractions and complex reasoning tasks. 🧪 Test Kai yourself in our new Space: NoesisLab/Kai-30B-Instruct 📦 Model Weights: NoesisLab/Kai-30B-Instruct Bring your hardest math, logic, and coding benchmarks. We invite the community to stress-test the limits of the penalty wall! 🧱💥