Impressive performances, huge congrats @patrickvonplaten @sgvaze @pandora-s @devendrachaplot @sophiamyang and team!
Very nice to have SOTA Multilingual OCR and Chart understanding in an open-weights model
Still not being able to get those impressive marks, trying to reproduce something simple with wikitext.. not much "performance" out of it.
Anyone has made this to work and get positive results?
latest 3.5 version of Claude model is even more impressive.. like SEVERAL problems (AI/ML) basically torch, where GPT4o fails epically.. were solved by Claude in 0-Shot.
But also to be said, GPT4o is very impressive using its sandbox.. kudos to that!
UNA is a modification of the modeling_$model.py
of transformers. I port it to to the different transformer version and models, keeping it clean and performant, So it works with any of these frameworks like #axolotl
we working on it my friend, LASER team is awesome. We are investigating further these two together how they amplify. The improvements on performance are larger than the usual tho we are empirically testing such thing.