Friedrich Marty

Smorty100

AI & ML interests

I'm most interested in content rerouting between LLM and VLLM agens for automation possibilities. Using templates for each agent which is then filled in by another agents inputs seems really useful.

Recent Activity

Organizations

None yet

Smorty100's activity

reacted to Reality123b's post with πŸ˜” 8 days ago
view post
Post
2197
https://huggingface.co/posts/Reality123b/533143502736808
Since many of you upvoted that post, I'm open-sourcing this on 19th February 2025.

I don't know, but, this may be the "smartest AI on earth". im not totally sure.
also, i need some kind of help with the UI coz i suck at that.
reacted to lewtun's post with ❀️ 14 days ago
view post
Post
4539
Introducing OpenR1-Math-220k!

open-r1/OpenR1-Math-220k

The community has been busy distilling DeepSeek-R1 from inference providers, but we decided to have a go at doing it ourselves from scratch πŸ’ͺ

What’s new compared to existing reasoning datasets?

β™Ύ Based on AI-MO/NuminaMath-1.5: we focus on math reasoning traces and generate answers for problems in NuminaMath 1.5, an improved version of the popular NuminaMath-CoT dataset.

🐳 800k R1 reasoning traces: We generate two answers for 400k problems using DeepSeek R1. The filtered dataset contains 220k problems with correct reasoning traces.

πŸ“€ 512 H100s running locally: Instead of relying on an API, we leverage vLLM and SGLang to run generations locally on our science cluster, generating 180k reasoning traces per day.

⏳ Automated filtering: We apply Math Verify to only retain problems with at least one correct answer. We also leverage Llama3.3-70B-Instruct as a judge to retrieve more correct examples (e.g for cases with malformed answers that can’t be verified with a rules-based parser)

πŸ“Š We match the performance of DeepSeek-Distill-Qwen-7B by finetuning Qwen-7B-Math-Instruct on our dataset.

πŸ”Ž Read our blog post for all the nitty gritty details: https://huggingface.co/blog/open-r1/update-2
replied to nroggendorff's post 14 days ago
view reply

this is so real...

just like - adress the things u don't like, don't tell it to us through ur weird games.

it'd be fun if it were treated like in kindergarden where u throw a ball around and say a thing. but it's not somehow... but no, these activities are not as self-reflective as u would hope they'd be ;(

reacted to nroggendorff's post with πŸ‘ 14 days ago
view post
Post
2638
Dearest None-yet Team,

I couldn't help but notice that our productivity has room for improvement. To address this, we will be engaging in a company-wide morale-building activity designed to boost teamwork, enthusiasm, and *most importantly* results.

I know you're all as excited as I am for this fun and absolutely required initiative. Participation is not just encouraged, it's mandatory. Think of it as a team-bonding experience you never signed up for but will absolutely tolerate.

More details to follow, but for now, mark your calendars and prepare for an engaging experience that will definitely make us all better, stronger, and more synchronized, or at least give us something to talk about later.

Looking forward to seeing you all there!

Best,
Me
Β·
New activity in huggingchat/chat-ui 15 days ago

[MODELS] Discussion

604
#372 opened about 1 year ago by
victor
reacted to schuler's post with πŸ‘πŸ€― 15 days ago
view post
Post
7219
πŸ“’ New Research Alert: Making Language Models Smaller & Smarter!

Thrilled to share the latest technical report demonstrating how to reduce language model parameters by 77% while maintaining performance.

The secret? Grouped pointwise convolutions. Yes. We brought a method from computer vision to the transformers arena.

πŸ”‘ Key Findings:
β€’ 77% parameter reduction.
β€’ Maintained model capabilities.
β€’ Improved generalization.

Paper: https://www.researchgate.net/publication/388835829_SAVING_77_OF_THE_PARAMETERS_IN_LARGE_LANGUAGE_MODELS_TECHNICAL_REPORT
Code: https://github.com/joaopauloschuler/less-parameters-llm
  • 2 replies
Β·
replied to prithivMLmods's post 26 days ago
reacted to etemiz's post with πŸ˜ŽπŸ‘€ about 1 month ago
view post
Post
1120
Updated the Hoopoe model which is taking faith related and religious texts in.

etemiz/Hoopoe-8B-Llama-3.1

Faith score went from 8% to 54%. Expect more updates and increase in the score. I also did the instruct fine tuning before adding faith to the model. So some of the improvements may be there because I started with llama 3.1 base and not the instruct.

Here are some comparisons with original Llama 3.1:
replied to MonsterMMORPG's post about 1 month ago
view reply

woah, some of these demo images look actually gud... i kinda lost hope for image diffusers there for a minute, but this is impressive. the one with the leaves cought me offguard.

And on such smol GPUs now too? that is super cool!