It should just be taking you to the Strawberry homepage. I guess they are tracking clicks.
Kenneth Hamilton PRO
ZennyKenny
AI & ML interests
Building and enablement @ montebello.ai
Certified vibe coder
Recent Activity
new activity
9 days ago
huggingface/InferenceSupport:ZennyKenny/oss-20b-prereform-to-modern-ru-merged
Organizations
replied to
their
post
6 days ago
posted
an
update
6 days ago
Post
278
Has anyone tried Strawberry Browser? https://strawberrybrowser.com/?ref_id=8D41NQCY7
😇 Shamelessly sharing my referral link here to move up in the waitlist line. Help me out, give it a click.
😇 Shamelessly sharing my referral link here to move up in the waitlist line. Help me out, give it a click.
replied to
their
post
13 days ago
Subbed, we'll make it through this traumatic experience together.
posted
an
update
16 days ago
Post
208
🔥 BigCodeArena: Unveiling More Reliable Human Preferences in Code Generation via Execution from
bigcode
is now available on Hugging Face!
👉 Check out the paper and please drop an upvote if you like the work BigCodeArena: Unveiling More Reliable Human Preferences in Code Generation via Execution (2510.08697)
👉 Check out the paper and please drop an upvote if you like the work BigCodeArena: Unveiling More Reliable Human Preferences in Code Generation via Execution (2510.08697)
posted
an
update
26 days ago
Post
1235
🥊 Big Code Arena is live!
bigcode/arena
💡
bigcode
is an open scientific collaboration working on responsible training of large language models for coding applications.
👉 The Arena ranks LLMs based on their ability to support natural language vibe coding requests in a competitive format, based on feedback from human reviewers.
🧠 It was a pleasure to contribute to this project led by @terryyz and appear as an additional contributor in the Big Code Arena paper.
💡
👉 The Arena ranks LLMs based on their ability to support natural language vibe coding requests in a competitive format, based on feedback from human reviewers.
🧠 It was a pleasure to contribute to this project led by @terryyz and appear as an additional contributor in the Big Code Arena paper.
reacted to
giadap's
post with 👍
about 1 month ago
Post
10839
One of the hardest challenges in AI safety is finding the right balance: how do we protect people from harm without undermining their agency? This tension is especially visible in conversational systems, where safeguards can sometimes feel more paternalistic than supportive.
In my latest piece for Hugging Face, I argue that open source and community-driven approaches offer a promising (though not exclusive) way forward.
✨ Transparency can make safety mechanisms into learning opportunities.
✨ Collaboration with diverse communities makes safeguards more relevant across contexts.
✨ Iteration in the open lets protections evolve rather than freeze into rigid, one-size-fits-all rules.
Of course, this isn’t a silver bullet. Top-down safety measures will still be necessary in some cases. But if we only rely on corporate control, we risk building systems that are safe at the expense of trust and autonomy.
Read the blog post here: https://huggingface.co/blog/giadap/preserving-agency
In my latest piece for Hugging Face, I argue that open source and community-driven approaches offer a promising (though not exclusive) way forward.
✨ Transparency can make safety mechanisms into learning opportunities.
✨ Collaboration with diverse communities makes safeguards more relevant across contexts.
✨ Iteration in the open lets protections evolve rather than freeze into rigid, one-size-fits-all rules.
Of course, this isn’t a silver bullet. Top-down safety measures will still be necessary in some cases. But if we only rely on corporate control, we risk building systems that are safe at the expense of trust and autonomy.
Read the blog post here: https://huggingface.co/blog/giadap/preserving-agency
posted
an
update
about 1 month ago
Post
8887
🖤 Probably one of my favorite projects that I've worked on so far, introducing Новояз (Novoyaz).
🛠 One of the first acts of the Bolshevik government after the Russian Revolution was the reform and standardization of the Russian language, which at the time had a non-standard and challenging orthography.
📚 Upon its reform the government launched a nationwide campaign called Ликбез (Likbez), which sought to improve literacy in the country (by the way, it worked, bringing the national literacy rate from <20% in the 1920s to >80% by the 1930s).
‼ While this is a remarkable result that should absolutely be celebrated, it's one that has left behind literally hundreds of thousands if not millions of artifacts using pre-reform Russian orthography.
😓 Researchers and historians are working tirelessly to translate these artifacts to modern Russian so that they may be archived and studied but many have told me that. they are doing this BY HAND (!).
💡 I thought, well this is a perfect use case for OCR and a fine-tuned LLM to step in and help to aid in this important work!
🌏 Introducing НОВОЯЗ (NOVOYAZ)! Powered by ChatDOC/OCRFlux-3B and ZennyKenny/oss-20b-prereform-to-modern-ru-merged, researchers can now convert images of their pre-reform documents to modern Russian orthography using the power of open-source AI!
Check it out and drop a like to support more real-world use cases for open source AI outside of traditional tech-centric domains!
ZennyKenny/Novoyaz
🛠 One of the first acts of the Bolshevik government after the Russian Revolution was the reform and standardization of the Russian language, which at the time had a non-standard and challenging orthography.
📚 Upon its reform the government launched a nationwide campaign called Ликбез (Likbez), which sought to improve literacy in the country (by the way, it worked, bringing the national literacy rate from <20% in the 1920s to >80% by the 1930s).
‼ While this is a remarkable result that should absolutely be celebrated, it's one that has left behind literally hundreds of thousands if not millions of artifacts using pre-reform Russian orthography.
😓 Researchers and historians are working tirelessly to translate these artifacts to modern Russian so that they may be archived and studied but many have told me that. they are doing this BY HAND (!).
💡 I thought, well this is a perfect use case for OCR and a fine-tuned LLM to step in and help to aid in this important work!
🌏 Introducing НОВОЯЗ (NOVOYAZ)! Powered by ChatDOC/OCRFlux-3B and ZennyKenny/oss-20b-prereform-to-modern-ru-merged, researchers can now convert images of their pre-reform documents to modern Russian orthography using the power of open-source AI!
Check it out and drop a like to support more real-world use cases for open source AI outside of traditional tech-centric domains!
ZennyKenny/Novoyaz
posted
an
update
about 1 month ago
Post
555
🔒 Like a lot of other AI builders, I have some anxiety about the emerging surveillance-capitalist paradigm emerging in the AI space.
👉 Of course-- this kind of thing isn't completely new and has been going on for decades, but the difference is the stronger immersion of AI tools into our daily lives (compared to something like a search engine or social network).
❕ That's why I was really excited to come across Lumo: https://lumo.proton.me/u/1/
❕ Lumo is created by
ProtonPrivacy
and offers privacy-first features that make sure that what you do with you AI assistant is your business.
❕ I already trust Proton with my other business apps and I've never been disappointed, plus the Lumo architecture is really fantastic, dynamically routing each query to the most appropriate model for the request.
🔥 Really awesome stuff Proton, thank you as always.
👉 Of course-- this kind of thing isn't completely new and has been going on for decades, but the difference is the stronger immersion of AI tools into our daily lives (compared to something like a search engine or social network).
❕ That's why I was really excited to come across Lumo: https://lumo.proton.me/u/1/
❕ Lumo is created by
❕ I already trust Proton with my other business apps and I've never been disappointed, plus the Lumo architecture is really fantastic, dynamically routing each query to the most appropriate model for the request.
🔥 Really awesome stuff Proton, thank you as always.
posted
an
update
about 1 month ago
Post
2376
The reactions to
mostlyai/synthetic-sdk-demo have been incredible! 🔥
Some users wrote that they were having performance issues on larger datasets, so I've capped the Space's input to 5000 rows and 10 columns, but you can always use the open source SDK that powers the space any time you want on datasets of arbitrary size and shape!
Check it out: https://github.com/mostly-ai/mostlyai 👈
Some users wrote that they were having performance issues on larger datasets, so I've capped the Space's input to 5000 rows and 10 columns, but you can always use the open source SDK that powers the space any time you want on datasets of arbitrary size and shape!
Check it out: https://github.com/mostly-ai/mostlyai 👈
posted
an
update
about 1 month ago
Post
2636
The open source Synthetic Data SDK from MOSTLY AI:
mostlyai
offers the ability to generate realistic, privacy-safe synthetic data with just a few lines of Python.
Try it out yourself in a No Code UI in the SDK Demo Space: mostlyai/synthetic-sdk-demo
Try it out yourself in a No Code UI in the SDK Demo Space: mostlyai/synthetic-sdk-demo
posted
an
update
3 months ago
Post
2595
It's just a matter of time before all the data leakage and data scraping associated with building, training, and using AI results in some kind of major scandal.
That's why I think this paper by @spintronic is so important: Privacy-Preserving Tabular Synthetic Data Generation Using TabularARGN (2508.06647)
Glad to know that there are already researchers looking to mitigate and address this risk before the s**t hits the fan.
That's why I think this paper by @spintronic is so important: Privacy-Preserving Tabular Synthetic Data Generation Using TabularARGN (2508.06647)
Glad to know that there are already researchers looking to mitigate and address this risk before the s**t hits the fan.
reacted to
ProCreations's
post with 🚀
5 months ago
reacted to
hesamation's
post with 🚀
6 months ago
Post
3578
60+ Generative AI projects for your resume. grind this GitHub repo if you want to level up:
> LLM fine-tuning and applications
> advanced RAG apps
> Agentic AI projects
> MCP and A2A (new)
GitHub: https://github.com/aishwaryanr/awesome-generative-ai-guide/blob/main/resources/60_ai_projects.md
> LLM fine-tuning and applications
> advanced RAG apps
> Agentic AI projects
> MCP and A2A (new)
GitHub: https://github.com/aishwaryanr/awesome-generative-ai-guide/blob/main/resources/60_ai_projects.md
reacted to
jeffboudier's
post with 🚀
6 months ago
Post
2608
Transcribing 1 hour of audio for less than $0.01 🤯
@mfuntowicz cooked with 8x faster Whisper speech recognition - whisper-large-v3-turbo transcribes at 100x real time on a $0.80/hr L4 GPU!
How they did it: https://huggingface.co/blog/fast-whisper-endpoints
1-click deploy with HF Inference Endpoints: https://endpoints.huggingface.co/new?repository=openai%2Fwhisper-large-v3-turbo&vendor=aws®ion=us-east&accelerator=gpu&instance_id=aws-us-east-1-nvidia-l4-x1&task=automatic-speech-recognition&no_suggested_compute=true
@mfuntowicz cooked with 8x faster Whisper speech recognition - whisper-large-v3-turbo transcribes at 100x real time on a $0.80/hr L4 GPU!
How they did it: https://huggingface.co/blog/fast-whisper-endpoints
1-click deploy with HF Inference Endpoints: https://endpoints.huggingface.co/new?repository=openai%2Fwhisper-large-v3-turbo&vendor=aws®ion=us-east&accelerator=gpu&instance_id=aws-us-east-1-nvidia-l4-x1&task=automatic-speech-recognition&no_suggested_compute=true
reacted to
AdinaY's
post with 🔥
6 months ago
Post
2442
Bytedance is on fire this week 🔥🔥🔥
They released Seed1.5-VL, A vision-language model for general-purpose multimodal reasoning.
It’s not open-source, but the paper and demo are available here👇
✨ Seed1.5-VL Technical Report (2505.07062)
✨ ByteDance-Seed/Seed1.5-VL
They released Seed1.5-VL, A vision-language model for general-purpose multimodal reasoning.
It’s not open-source, but the paper and demo are available here👇
✨ Seed1.5-VL Technical Report (2505.07062)
✨ ByteDance-Seed/Seed1.5-VL
replied to
as-cle-bert's
post
6 months ago
Whoa. Reliable open-sourced crawling software is a big win. I'll take it for a spin but I'm optimistic as this is the kind of thing I (and every other AI builder) has been building for years to avoid paying FireCrawl.
reacted to
onekq's
post with 🔥
6 months ago
Post
3303
This time Gemini is very quick with API support on its 2.5 pro May release. The performance is impressive too, now it is among top contenders like o4, R1, and Claude.
onekq-ai/WebApp1K-models-leaderboard
onekq-ai/WebApp1K-models-leaderboard
reacted to
wolfram's
post with 🔥
6 months ago
Post
7670
Finally finished my extensive **Qwen 3 evaluations** across a range of formats and quantisations, focusing on **MMLU-Pro** (Computer Science).
A few take-aways stood out - especially for those interested in local deployment and performance trade-offs:
1️⃣ **Qwen3-235B-A22B** (via Fireworks API) tops the table at **83.66%** with ~55 tok/s.
2️⃣ But the **30B-A3B Unsloth** quant delivered **82.20%** while running locally at ~45 tok/s and with zero API spend.
3️⃣ The same Unsloth build is ~5x faster than Qwen's **Qwen3-32B**, which scores **82.20%** as well yet crawls at <10 tok/s.
4️⃣ On Apple silicon, the **30B MLX** port hits **79.51%** while sustaining ~64 tok/s - arguably today's best speed/quality trade-off for Mac setups.
5️⃣ The **0.6B** micro-model races above 180 tok/s but tops out at **37.56%** - that's why it's not even on the graph (50 % performance cut-off).
All local runs were done with LM Studio on an M4 MacBook Pro, using Qwen's official recommended settings.
**Conclusion:** Quantised 30B models now get you ~98 % of frontier-class accuracy - at a fraction of the latency, cost, and energy. For most local RAG or agent workloads, they're not just good enough - they're the new default.
Well done, Qwen - you really whipped the llama's ass! And to OpenAI: for your upcoming open model, please make it MoE, with toggleable reasoning, and release it in many sizes. *This* is the future!
A few take-aways stood out - especially for those interested in local deployment and performance trade-offs:
1️⃣ **Qwen3-235B-A22B** (via Fireworks API) tops the table at **83.66%** with ~55 tok/s.
2️⃣ But the **30B-A3B Unsloth** quant delivered **82.20%** while running locally at ~45 tok/s and with zero API spend.
3️⃣ The same Unsloth build is ~5x faster than Qwen's **Qwen3-32B**, which scores **82.20%** as well yet crawls at <10 tok/s.
4️⃣ On Apple silicon, the **30B MLX** port hits **79.51%** while sustaining ~64 tok/s - arguably today's best speed/quality trade-off for Mac setups.
5️⃣ The **0.6B** micro-model races above 180 tok/s but tops out at **37.56%** - that's why it's not even on the graph (50 % performance cut-off).
All local runs were done with LM Studio on an M4 MacBook Pro, using Qwen's official recommended settings.
**Conclusion:** Quantised 30B models now get you ~98 % of frontier-class accuracy - at a fraction of the latency, cost, and energy. For most local RAG or agent workloads, they're not just good enough - they're the new default.
Well done, Qwen - you really whipped the llama's ass! And to OpenAI: for your upcoming open model, please make it MoE, with toggleable reasoning, and release it in many sizes. *This* is the future!
posted
an
update
6 months ago
Post
952
Community! 💡💡💡
It's the last day to submit your datasets for the Reasoning Datasets Competition: https://www.bespokelabs.ai/blog/reasoning-datasets-competition
Here are my submissions:
- ZennyKenny/synthetic_vc_financial_decisions_reasoning_dataset
- ZennyKenny/cosa-benchmark-dataset
- ZennyKenny/tactical-military-reasoning-v.1.0
- ZennyKenny/tron-dataset-v.1.0
Have a look and drop a ❤️ or comment! Check out the entire collection of submissions here: https://huggingface.co/datasets?other=reasoning-datasets-competition
It's the last day to submit your datasets for the Reasoning Datasets Competition: https://www.bespokelabs.ai/blog/reasoning-datasets-competition
Here are my submissions:
- ZennyKenny/synthetic_vc_financial_decisions_reasoning_dataset
- ZennyKenny/cosa-benchmark-dataset
- ZennyKenny/tactical-military-reasoning-v.1.0
- ZennyKenny/tron-dataset-v.1.0
Have a look and drop a ❤️ or comment! Check out the entire collection of submissions here: https://huggingface.co/datasets?other=reasoning-datasets-competition