Jorge Alonso PRO

oieieio

AI & ML interests

AI/ML

Recent Activity

View all activity

Organizations

Hugging Face Discord Community's profile picture OIEIEIO's profile picture Lambda Go Labs's profile picture

oieieio's activity

reacted to burtenshaw's post with 🤗 1 day ago
view post
Post
5121
Now the Hugging Face agent course is getting real! With frameworks like smolagents, LlamaIndex, and LangChain.

🔗 Follow the org for updates https://huggingface.co/agents-course

This week we are releasing the first framework unit in the course and it’s on smolagents. This is what the unit covers:

- why should you use smolagents vs another library?
- how to build agents that use code
- build multiagents systems
- use vision language models for browser use

The team has been working flat out on this for a few weeks. Led by @sergiopaniego and supported by smolagents author @m-ric .
replied to wassemgtk's post 2 days ago
view reply

GESAL Training Update

Working on some tweaks but here are some earlier training logs.

I’ve captured the latest GESAL training logs for my engine/turbine dataset, showing real-time adaptation in action. Using meta-llama/Llama-3.2-1B on Ubuntu 24.04 (4060Ti GPU, 16GB VRAM; Xeon, 32 cores; 256GB RAM) with InfluxDB 3 Core, the logs detail:

  • Embedding generation for 1,326 synthetic engines
  • Node updates (currently 2 nodes, target: 10-15)
  • Distance calculations (e.g., 0.00985)
  • InfluxDB writes for parameters like flight_hours, exhaust_temp, and wear

Note: The attached image shows logs from my GESAL-Training.ipynb notebook, running on February 26, 2025.
Screenshot from 2025-02-26 12-40-07.png

replied to wassemgtk's post 2 days ago
view reply

@wassemgtk — Thank you for your brilliant work on Graph-Enhanced Singular Adaptive Learning (GESAL) and for releasing the code and white paper on GitHub. Your framework’s innovative approach—leveraging Singular Value Factorization (SVF), graph memory, and reinforcement learning (RL) for real-time LLM adaptation—is truly impressive, efficient, and fast. I’ve adapted GESAL for synthetic engine/turbine data for wear prediction and predictive maintenance, as referenced in @oieieio1’s X update, and it’s shown promising results, though I’ve noticed prompt design has a massive impact on outcomes.

My Implementation

I’m using GESAL with meta-llama/Llama-3.2-1B on an Ubuntu 24.04 system with:

  • Hardware: 4060Ti GPU (16GB VRAM), Xeon processors (36 cores), 256GB RAM
  • Software: InfluxDB 3 Core (version 0.1.0 alpha) for storing and querying ~20 engine parameters, including:
    • flight_hours
    • exhaust_temp
    • vibration
    • rpm
    • thrust
    • pressure
    • wear
    • (and others like fuel_flow, oil_temp, etc.)

My latest analysis of 1,326 synthetic engines shows:

  • Nodes: 2 unique nodes
  • Accuracy: 22.48% on maintenance predictions (keywords: “replace,” “maintenance,” “check”)
  • Earlier Results: 42.0% accuracy and 189 nodes with fewer engines (still experimenting with prompts).

For example, a sample input is:
Engine 1—flight_hours=5000h, exhaust_temp=750°C, vibration=0.5g, rpm=12000, thrust=75kN, pressure=1.5bar, wear=3.2. Predict maintenance.

We’re actively experimenting with prompts—e.g., “Predict maintenance” vs. “Identify maintenance needs for wear and predict ‘replace’, ‘maintenance’, or ‘check’”—and finding that slight changes dramatically affect GESAL’s responses, likely due to temperature=0.7 and top_k=50 in generation.

Goals and Challenges

While GESAL’s scalability is excellent, I’m targeting:

  • Accuracy: 70-80%
  • Nodes: 10-15

The drop from 42.0% to 22.48% accuracy may stem from prompt variations or scaling effects, possibly amplified by current hyperparameters.

Seeking Feedback

I’d appreciate your insights on:

  • Optimizing prompts for consistency and accuracy
  • Tuning hyperparameters (e.g., temperature, top_k, distance_threshold)
  • Scaling GESAL for large industrial datasets (e.g., thousands of engines)
  • Benchmarking GESAL for mechanical systems, as @hassenhamdi mentioned

The attached visualization (Wear vs Exhaust Temperature) shows current performance—I’m eager to collaborate further to refine prompts and boost accuracy!


Thanks again for this groundbreaking tool—I’m excited to see how GESAL can evolve!
Screenshot from 2025-02-26 11-23-19.png

replied to wassemgtk's post 2 days ago
view reply

Trained with GESAL on synthetic egine/turbine data, training for wear and predictive maintenance.

Screenshot from 2025-02-26 10-10-08.png

reacted to wassemgtk's post with ❤️ 2 days ago
view post
Post
1769
# GESAL: Real-Time Adaptation for LLMs


We’re excited to unveil **Graph-Enhanced Singular Adaptive Learning (GESAL)**, a framework that lets LLMs like meta-llama/Llama-3.2-1B adapt in real time using user feedback. Check out the code and white paper on GitHub!

🔗 **Code**: [https://github.com/writer/AI-Adaptive-Learning-GESAL](https://github.com/writer/AI-Adaptive-Learning-GESAL)

---

## Why GESAL?

Static LLMs struggle to adapt without heavy retraining. GESAL solves this with:
- **SVF**: Adapts weights via \( W' = U (\Sigma \cdot z) V^T \), using few parameters.
- **Graph Memory**: Stores adaptations in nodes for scalability.
- **RL**: Updates via \( J(z) = \mathbb{E}[\log \pi_z(y|x) r] \) based on feedback.

---

## How It Works

Ask "How many R’s in ‘strawberry’?" If it says "2" and you say "no," GESAL learns to say "3" next time, avoiding repeats.

---

## Try It

Built with Hugging Face’s transformers:
pip install transformers torch numpy
python Adaptive_Learning_(GESAL).py

Needs a Hugging Face token for Llama-3.2-1B.

---

## Results

GESAL hits 95% accuracy after 5 feedbacks vs. LoRA’s 70%. It’s efficient (~0.5M params) and scalable.
·
reacted to stefan-it's post with 👍 4 days ago
view post
Post
5006
She arrived 😍

[Expect more models soon...]
  • 2 replies
·
replied to fdaudens's post 15 days ago
view reply

If you're building with AI at any scale, definitely worth checking out. Yes! Looks great!

reacted to fdaudens's post with 🔥 15 days ago
view post
Post
2681
⭐️ The AI Energy Score project just launched - this is a game-changer for making informed decisions about AI deployment.

You can now see exactly how much energy your chosen model will consume, with a simple 5-star rating system. Think appliance energy labels, but for AI.

Looking at transcription models on the leaderboard is fascinating: choosing between whisper-tiny or whisper-large-v3 can make a 7x difference. Real-time data on these tradeoffs changes everything.

166 models already evaluated across 10 different tasks, from text generation to image classification. The whole thing is public and you can submit your own models to test.

Why this matters:
- Teams can pick efficient models that still get the job done
- Developers can optimize for energy use from day one
- Organizations can finally predict their AI environmental impact

If you're building with AI at any scale, definitely worth checking out.

👉 leaderboard: https://lnkd.in/esrSxetj
👉 blog post: https://lnkd.in/eFJvzHi8

Huge work led by @sasha with @bgamazay @yjernite @sarahooker @regisss @meg
  • 1 reply
·
reacted to ZennyKenny's post with 🤗 15 days ago
view post
Post
3423
I've completed the first unit of the just-launched Hugging Face Agents Course. I would highly recommend it, even for experienced builders, because it is a great walkthrough of the smolagents library and toolkit.
reacted to davidberenstein1957's post with 🤗 15 days ago
view post
Post
3223
🚀 Find banger tools for your smolagents!

I created the Tools gallery, which makes tools specifically developed by/for smolagents searchable and visible. This will help with:
- inspiration
- best practices
- finding cool tools

Space: davidberenstein1957/smolagents-and-tools
  • 1 reply
·
upvoted an article 17 days ago
reacted to lewtun's post with ❤️🔥 17 days ago
view post
Post
4594
Introducing OpenR1-Math-220k!

open-r1/OpenR1-Math-220k

The community has been busy distilling DeepSeek-R1 from inference providers, but we decided to have a go at doing it ourselves from scratch 💪

What’s new compared to existing reasoning datasets?

♾ Based on AI-MO/NuminaMath-1.5: we focus on math reasoning traces and generate answers for problems in NuminaMath 1.5, an improved version of the popular NuminaMath-CoT dataset.

🐳 800k R1 reasoning traces: We generate two answers for 400k problems using DeepSeek R1. The filtered dataset contains 220k problems with correct reasoning traces.

📀 512 H100s running locally: Instead of relying on an API, we leverage vLLM and SGLang to run generations locally on our science cluster, generating 180k reasoning traces per day.

⏳ Automated filtering: We apply Math Verify to only retain problems with at least one correct answer. We also leverage Llama3.3-70B-Instruct as a judge to retrieve more correct examples (e.g for cases with malformed answers that can’t be verified with a rules-based parser)

📊 We match the performance of DeepSeek-Distill-Qwen-7B by finetuning Qwen-7B-Math-Instruct on our dataset.

🔎 Read our blog post for all the nitty gritty details: https://huggingface.co/blog/open-r1/update-2