chansung park's picture

chansung park PRO

chansung

AI & ML interests

None yet

Recent Activity

View all activity

Articles

Organizations

Notebooks-explorers's profile picture various keras sd deployment 's profile picture LLMs's profile picture Hugging Face Fellows's profile picture Gradio-Themes-Party's profile picture Alpaca LoRA's profile picture Webhooks Explorers (BETA)'s profile picture Deploy HF TF ViTs's profile picture Blog-explorers's profile picture Personal Coding Assistant's profile picture ZeroGPU Explorers's profile picture Social Post Explorers's profile picture Top Contributors: Dataset Downloads's profile picture llama-duo's profile picture klcsp's profile picture ExpanLLM's profile picture

chansung's activity

reacted to their post with šŸ‘ 3 days ago
view post
Post
1514
New look for AI powered paper reviews from the list by Hugging Face Daily Papers ( managed by the @akhaliq )

Bookmark the webpage along, check comprehensive reviews by Google DeepMind Gemini 1.5, and listen to audio podcast made by the same tech used in NotebookLM.

Link: https://deep-diver.github.io/ai-paper-reviewer/

This is not an official service by Hugging Face. It is just a service developed by an individual developer using his own money :)
posted an update 3 days ago
view post
Post
1514
New look for AI powered paper reviews from the list by Hugging Face Daily Papers ( managed by the @akhaliq )

Bookmark the webpage along, check comprehensive reviews by Google DeepMind Gemini 1.5, and listen to audio podcast made by the same tech used in NotebookLM.

Link: https://deep-diver.github.io/ai-paper-reviewer/

This is not an official service by Hugging Face. It is just a service developed by an individual developer using his own money :)
reacted to their post with šŸ‘ 4 days ago
view post
Post
1889
Simple summarization of Evolving Deeper LLM Thinking (Google DeepMind)

The process starts by posing a question.
1) The LLM generates initial responses.
2) These generated responses are evaluated according to specific criteria (program-based checker).
3) The LLM critiques the evaluated results.
4) The LLM refines the responses based on the evaluation, critique, and original responses.

The refined response is then fed back into step 2). If it meets the criteria, the process ends. Otherwise, the algorithm generates more responses based on the refined ones (with some being discarded, some remaining, and some responses potentially being merged).

Through this process, it demonstrated excellent performance in complex scheduling problems (travel planning, meeting scheduling, etc.). It's a viable method for finding highly effective solutions in specific scenarios.

However, there are two major drawbacks:
šŸ¤” An excessive number of API calls are required. (While the cost might not be very high, it leads to significant latency.)
šŸ¤” The evaluator is program-based. (This limits its use as a general method. It could potentially be modified/implemented using LLM as Judge, but that would introduce additional API costs for evaluation.)

https://arxiv.org/abs/2501.09891
posted an update 4 days ago
view post
Post
1889
Simple summarization of Evolving Deeper LLM Thinking (Google DeepMind)

The process starts by posing a question.
1) The LLM generates initial responses.
2) These generated responses are evaluated according to specific criteria (program-based checker).
3) The LLM critiques the evaluated results.
4) The LLM refines the responses based on the evaluation, critique, and original responses.

The refined response is then fed back into step 2). If it meets the criteria, the process ends. Otherwise, the algorithm generates more responses based on the refined ones (with some being discarded, some remaining, and some responses potentially being merged).

Through this process, it demonstrated excellent performance in complex scheduling problems (travel planning, meeting scheduling, etc.). It's a viable method for finding highly effective solutions in specific scenarios.

However, there are two major drawbacks:
šŸ¤” An excessive number of API calls are required. (While the cost might not be very high, it leads to significant latency.)
šŸ¤” The evaluator is program-based. (This limits its use as a general method. It could potentially be modified/implemented using LLM as Judge, but that would introduce additional API costs for evaluation.)

https://arxiv.org/abs/2501.09891
reacted to their post with šŸ‘ 5 days ago
view post
Post
1944
Simple Summarization on DeepSeek-R1 from DeepSeek AI

The RL stage is very important.
ā†³ However, it is difficult to create a truly helpful AI for people solely through RL.
ā†³ So, we applied a learning pipeline consisting of four stages: providing a good starting point, reasoning RL, SFT, and safety RL, and achieved performance comparable to o1.
ā†³ Simply fine-tuning other open models with the data generated by R1-Zero (distillation) resulted in performance comparable to o1-mini.

Of course, this is just a brief overview and may not be of much help. All models are accessible on Hugging Face, and the paper can be read through the GitHub repository.


Model: https://huggingface.co/deepseek-ai
Paper: https://github.com/deepseek-ai/DeepSeek-R1
  • 1 reply
Ā·
posted an update 5 days ago
view post
Post
1944
Simple Summarization on DeepSeek-R1 from DeepSeek AI

The RL stage is very important.
ā†³ However, it is difficult to create a truly helpful AI for people solely through RL.
ā†³ So, we applied a learning pipeline consisting of four stages: providing a good starting point, reasoning RL, SFT, and safety RL, and achieved performance comparable to o1.
ā†³ Simply fine-tuning other open models with the data generated by R1-Zero (distillation) resulted in performance comparable to o1-mini.

Of course, this is just a brief overview and may not be of much help. All models are accessible on Hugging Face, and the paper can be read through the GitHub repository.


Model: https://huggingface.co/deepseek-ai
Paper: https://github.com/deepseek-ai/DeepSeek-R1
  • 1 reply
Ā·
upvoted an article 6 days ago
view article
Article

Introducing multi-backends (TRT-LLM, vLLM) support for Text Generation Inference

ā€¢ 60