Moatasemmofadal/ssd
52
Accelerate LLM inference by running speculative decoding in parallel, improving speed without sacrificing exactness or output quality.
What's novel
Accelerate LLM inference by running speculative decoding in parallel, improving speed without sacrificing exactness or output quality.
Score Breakdown
Innovation
4 (25%)
Craft
39 (35%)
Traction
11 (15%)
Scope
61 (25%)
Signal breakdown
Innovation
Not Fork+1
Code Novelty+0
Unique Niche+1
Concept Novelty+2
Craft
Ci-3
Tests-5
Polish+0
Releases-2
Has License+5
Code Quality+12
Readme Quality+15
Recent Activity+7
Structure Quality+5
Commit Consistency+0
Has Dependency Mgmt+5
Traction
Forks+0
Stars+6
Hn Points+0
Watchers+3
Early Traction+0
Devto Reactions+0
Community Contribs+2
Scope
Commits+5
Languages+3
Subsystems+10
Bloat Penalty+0
Completeness+7
Contributors+6
Authored Files+12
Readme Code Match+3
Architecture Depth+7
Implementation Depth+8
Evidence
Commits
16
Contributors
2
Files
54
Active weeks
1
TestsCI/CDREADMELicenseContributing
Repository
Language
Python
Stars
1
Forks
0
License
MIT