alozowski
Β·
AI & ML interests
NLP in all aspects
Recent Activity
reacted
to
Molbap's
post
with β€οΈ
28 days ago
π New blog: Maintain the unmaintainable β 1M+ Python LOC, 400+ models
How do you stop a million-line library built by thousands of contributors from collapsing under its own weight?
At π€ Transformers, we do it with explicit software-engineering tenets, principles that make the codebase hackable at scale.
π Inside the post:
β One Model, One File: readability first β you can still open a modeling file and see the full logic, top to bottom.
β Modular Transformers: visible inheritance that cuts maintenance cost by ~15Γ while keeping models readable.
β Config-Driven Performance: FlashAttention, tensor parallelism, and attention scheduling are config-level features, not rewrites.
Written with @lysandre,@pcuenq and @yonigozlan, this is a deep dive into how Transformers stays fast, open, and maintainable.
Read it here β https://huggingface.co/spaces/transformers-community/Transformers-tenets
reacted
to
Molbap's
post
with π₯
28 days ago
π New blog: Maintain the unmaintainable β 1M+ Python LOC, 400+ models
How do you stop a million-line library built by thousands of contributors from collapsing under its own weight?
At π€ Transformers, we do it with explicit software-engineering tenets, principles that make the codebase hackable at scale.
π Inside the post:
β One Model, One File: readability first β you can still open a modeling file and see the full logic, top to bottom.
β Modular Transformers: visible inheritance that cuts maintenance cost by ~15Γ while keeping models readable.
β Config-Driven Performance: FlashAttention, tensor parallelism, and attention scheduling are config-level features, not rewrites.
Written with @lysandre,@pcuenq and @yonigozlan, this is a deep dive into how Transformers stays fast, open, and maintainable.
Read it here β https://huggingface.co/spaces/transformers-community/Transformers-tenets
View all activity
Organizations
-
-
-
-
-
-
-
-
-
-
-
view article
COβ Emissions and Models Performance: Insights from the Open LLM Leaderboard