Konrad Szafer

KonradSzafer

AI & ML interests

None yet

Recent Activity

liked a model 5 days ago
google/gemma-2-2b-it
liked a Space 5 days ago
JeffreyXiang/TRELLIS
View all activity

Organizations

Blog-explorers's profile picture hf-qa-bot's profile picture Auton Lab's profile picture Hugging Face Discord Community's profile picture

KonradSzafer's activity

reacted to gabrielmbmb's post with πŸ”₯ 4 months ago
view post
Post
1822
Yesterday Β  @mattshumer released mattshumer/Reflection-Llama-3.1-70B, an impressive model that achieved incredible results in benchmarks like MMLU. The model was fine-tuned using Reflection-Tuning and the dataset used wasn't released, but I created a small recipe with distilabel that allows generating a dataset with a similar output format:

1. We use MagPie 🐦 in combination with https://huggingface.co/meta-llama/Meta-Llama-3.1-70B-Instruct to generate reasoning instructions.
2. We generate a response again using https://huggingface.co/meta-llama/Meta-Llama-3.1-70B-Instruct, but we steer the LLM to generate an specific output format using a custom system prompt. In the system prompt, we instruct the LLM that it will have first to think πŸ’­ and have reflections that will help resolving ambiguities. After that, we instruct the LLM to generate an output based on the previous thinking

In this dataset gabrielmbmb/distilabel-reflection-tuning you can found 5 rows that I generated with this recipe. You can also found the code of the pipeline in the file called reflection.py.