β€’4 min readβ€’from Machine Learning

[R] Structure Over Scale: Memory-First Reasoning and Depth-Pruned Efficiency in Magnus and Seed Architecture Auto-Discovery

[R] Structure Over Scale: Memory-First Reasoning and Depth-Pruned Efficiency in Magnus and Seed Architecture Auto-Discovery
[R] Structure Over Scale: Memory-First Reasoning and Depth-Pruned Efficiency in Magnus and Seed Architecture Auto-Discovery
Dataset Model Acc F1 Ξ” vs Log Ξ” vs Static Avg Params Peak Params Steps Infer ms Size
Banking77-20 Logistic TF-IDF 92.37% 0.9230 +0.00pp +0.76pp 64,940 64,940 0.00M 0.473 1.000x
Static Seed 91.61% 0.9164 -0.76pp +0.00pp 52,052 52,052 94.56M 0.264 0.801x
Dynamic Seed Distill 93.53% 0.9357 +1.17pp +1.92pp 12,648 16,881 70.46M 0.232 0.195x
CLINC150 | Logistic TF-IDF | 97.00% | 0.9701 | +0.00pp | +1.78pp | 41,020 | 41,020 | 0.00M | 0.000 | 1.000x | Static Seed | 95.22% | 0.9521 | -1.78pp | +0.00pp | 52,052 | 52,052 | 66.80M | 0.302 | 1.269x | Dynamic Seed | 94.78% | 0.9485 | -2.22pp | -0.44pp | 10,092 | 10,136 | 28.41M | 0.324 | 0.246x | Dynamic Seed Distill | 95.44% | 0.9544 | -1.56pp | +0.22pp | 9,956 | 9,956 | 32.69M | 0.255 | 0.243x HWU64 | Logistic TF-IDF | 87.94% | 0.8725 | +0.00pp | +0.81pp | 42,260 | 42,260 | 0.00M | 0.000 | 1.000x | Static Seed | 87.13% | 0.8674 | -0.81pp | +0.00pp | 52,052 | 52,052 | 146.61M | 0.300 | 1.232x | Dynamic Seed | 86.63% | 0.8595 | -1.31pp | -0.50pp | 12,573 | 17,565 | 62.54M | 0.334 | 0.297x | Dynamic Seed Distill | 87.23% | 0.8686 | -0.71pp | +0.10pp | 13,117 | 17,575 | 62.86M | 0.340 | 0.310x MASSIVE-20 | Logistic TF-IDF | 86.06% | 0.7324 | +0.00pp | -1.92pp | 74,760 | 74,760 | 0.00M | 0.000 | 1.000x | Static Seed | 87.98% | 0.8411 | +1.92pp | +0.00pp | 52,052 | 52,052 | 129.26M | 0.247 | 0.696x | Dynamic Seed | 86.94% | 0.7364 | +0.88pp | -1.04pp | 11,595 | 17,565 | 47.62M | 0.257 | 0.155x | Dynamic Seed Distill | 86.45% | 0.7380 | +0.39pp | -1.53pp | 11,851 | 19,263 | 51.90M | 0.442 | 0.159x 

TL;DR:
I built a system that finds much smaller models that stay competitive β€” and sometimes outperform larger baselines.

Built a small experiment around Seed (architecture discovery).

Instead of training bigger models, Seed:

  • generates candidate architectures
  • evaluates them
  • keeps the smallest ones that still perform well

Tested across 4 datasets:

  • Banking77
  • CLINC150
  • HWU64
  • MASSIVE

🧠 Key result (Banking77)

  • Logistic TF-IDF: 92.37%
  • Dynamic Seed (distilled): 93.53%

πŸ‘‰ Higher accuracy + ~5x smaller (12.6k vs 64.9k params)

πŸ“Š Other results

  • MASSIVE β†’ quality + size wins
  • CLINC150 / HWU64 β†’ not always higher accuracy but ~4–5x smaller models with competitive performance

πŸ”₯ What actually matters (not just accuracy)

If you only look at accuracy β†’ mixed

If you include:

  • model size
  • training compute
  • inference latency

πŸ‘‰ this becomes a much stronger result

🧠 Takeaway

Traditional ML:
πŸ‘‰ scale model size and hope

Seed:
πŸ‘‰ search for better structure

Smaller models can compete with larger ones
if you find the right architecture

Not AGI
Not β€œwe solved NLU”

But a real signal that:

πŸ‘‰ structure > scale

Smaller models can compete with larger ones β€” if you find the right structure

Key takeaway:

  • Text β†’ clear wins (better + smaller)
  • Sensor β†’ huge efficiency gains
  • Vision β†’ compact tradeoff
  • Audio β†’ failed due to weak representation

So, this isn’t just about accuracy, it’s about moving the efficiency frontier.

submitted by /u/califalcon
[link] [comments]

Want to read more?

Check out the full article on the original site

View original article→

Tagged with

#financial modeling with spreadsheets
#rows.com
#natural language processing for spreadsheets
#generative AI for data analysis
#large dataset processing
#Excel alternatives for data analysis
#real-time data collaboration
#no-code spreadsheet solutions
#real-time collaboration
#big data performance
#Logistic TF-IDF
#Memory-First Reasoning
#Architecture Auto-Discovery
#Accuracy
#Efficiency Frontier
#Structure Over Scale
#Depth-Pruned Efficiency
#Model Size
#Parameters
#Competitive Performance