TacoSkill LAB
TacoSkill LAB
HomeSkillHubCreatePlaygroundSkillKit
© 2026 TacoSkill LAB
AboutPrivacyTerms
  1. Home
  2. /
  3. SkillHub
  4. /
  5. model-pruning
Improve

model-pruning

8.1

by davila7

75Favorites
403Upvotes
0Downvotes

Reduce LLM size and accelerate inference using pruning techniques like Wanda and SparseGPT. Use when compressing models without retraining, achieving 50% sparsity with minimal accuracy loss, or enabling faster inference on hardware accelerators. Covers unstructured pruning, structured pruning, N:M sparsity, magnitude pruning, and one-shot methods.

model-compression

8.1

Rating

0

Installs

Machine Learning

Category

Quick Review

Excellent skill covering model pruning techniques with clear, actionable guidance. The description accurately reflects comprehensive capabilities including Wanda, SparseGPT, and N:M sparsity methods. Task knowledge is strong with working code examples, complete implementation patterns, and production-ready pipelines. Structure is well-organized with logical flow from quick start to advanced strategies, though some sections are verbose. Novelty is moderate-to-good: while pruning concepts are established, implementing one-shot methods like Wanda and SparseGPT with proper calibration is non-trivial for a CLI agent and would require significant token expenditure to derive independently. The skill meaningfully reduces complexity by packaging activation-aware pruning logic, layer-wise strategies, and hardware-optimized N:M patterns. Performance benchmarks and best practices add substantial practical value. Minor improvements could include more concise core sections and clearer delineation between beginner and advanced content.

LLM Signals

Description coverage9
Task knowledge9
Structure8
Novelty7

GitHub Signals

18,073
1,635
132
71
Last commit 0 days ago

Publisher

davila7

davila7

Skill Author

Related Skills

ml-pipelinesparse-autoencoder-traininghuggingface-accelerate

Loading SKILL.md…

Try onlineView on GitHub

Publisher

davila7 avatar
davila7

Skill Author

Related Skills

ml-pipeline

Jeffallan

6.4

sparse-autoencoder-training

zechenzhangAGI

7.6

huggingface-accelerate

zechenzhangAGI

7.6

moe-training

zechenzhangAGI

7.6
Try online