Skip to content

Add TECS-L: Dense-to-MoE Compression via Golden Zone Sparsity#61

Open
dancinlife wants to merge 1 commit intoHuangOwen:mainfrom
dancinlife:add-tecs-l-golden-moe
Open

Add TECS-L: Dense-to-MoE Compression via Golden Zone Sparsity#61
dancinlife wants to merge 1 commit intoHuangOwen:mainfrom
dancinlife:add-tecs-l-golden-moe

Conversation

@dancinlife
Copy link
Copy Markdown

Summary

  • Adds TECS-L (Golden MoE) to the Pruning and Sparsity section
  • TECS-L is a dense-to-sparse MoE conversion framework that mathematically proves the optimal expert activation ratio (Inhibition I ≈ 1/e ≈ 0.368), achieving 62.5% parameter efficiency with structured sparsity
  • GitHub: https://github.com/need-singularity/TECS-L

Details

TECS-L converts dense LLMs into Mixture-of-Experts models using a "Golden Zone" sparsity criterion (I ∈ [0.213, 0.500]). Applied to Mistral 7B, it reduces active parameters to 62.5% while maintaining model quality through Boltzmann-routed expert selection.

🤖 Generated with Claude Code

TECS-L (Golden MoE) is a dense-to-sparse MoE conversion framework that
mathematically proves the optimal expert activation ratio (I≈1/e),
achieving 62.5% parameter efficiency with structured sparsity.

Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant