AI & ML interests

None defined yet.

Recent Activity

thomwolfย  authored a paper about 2 months ago
Robot Learning: A Tutorial
sanchit-gandhiย  authored a paper 5 months ago
Magistral
sanchit-gandhiย  authored a paper 5 months ago
Voxtral
View all activity

mrfakenameย 
posted an update 1 day ago
view post
Post
307
Excited to share that I've joined the Hugging Face Fellows program! ๐Ÿค—

Looking forward to contributing to & working more closely with the open-source ecosystem - huge thanks to everyone who's supported me on this journey! ๐Ÿš€
mrfakenameย 
posted an update about 1 month ago
view post
Post
5969
Trained a model for emotion-controllable TTS based on MiMo audio on LAION's dataset.

Still very early and does have an issue with hallucinating but results seem pretty good so far, given that it is very early into the training run.

Will probably kick off a new run later with some settings tweaked.

Put up a demo here: https://huggingface.co/spaces/mrfakename/EmoAct-MiMo

(Turn ๐Ÿ”Š on to hear audio samples)
ยท
multimodalartย 
posted an update about 2 months ago
view post
Post
5886
Want to iterate on a Hugging Face Space with an LLM?

Now you can easily convert any HF entire repo (Model, Dataset or Space) to a text file and feed it to a language model!

multimodalart/repo2txt
multimodalartย 
posted an update 6 months ago
view post
Post
17860
Self-Forcing - a real-time video distilled model from Wan 2.1 by @adobe is out, and they open sourced it ๐Ÿ

I've built a live real time demo on Spaces ๐Ÿ“น๐Ÿ’จ

multimodalart/self-forcing
ยท
cbensimonย 
posted an update 6 months ago
view post
Post
4079
๐Ÿš€ ZeroGPU now supports PyTorch native quantization via torchao

While it hasnโ€™t been battle-tested yet, Int8WeightOnlyConfig is already working flawlessly in our tests.

Let us know if you run into any issues โ€” and weโ€™re excited to see what the community will build!

import spaces
from diffusers import FluxPipeline
from torchao.quantization.quant_api import Int8WeightOnlyConfig, quantize_

pipeline = FluxPipeline.from_pretrained(...).to('cuda')
quantize_(pipeline.transformer, Int8WeightOnlyConfig()) # Or any other component(s)

@spaces.GPU
def generate(prompt: str):
    return pipeline(prompt).images[0]
ยท
cbensimonย 
posted an update 7 months ago
view post
Post
6087
๐Ÿš€ ZeroGPU medium size is now available as a power-user feature

Nothing too fancy for nowโ€”ZeroGPU Spaces still default to large (70GB VRAM)โ€”but this paves the way for:
- ๐Ÿ’ฐ size-based quotas / pricing (medium will offer significantly more usage than large)
- ๐Ÿฆฃ the upcoming xlarge size (141GB VRAM)

You can as of now control GPU size via a Space variable. Accepted values:
- auto (future default)
- medium
- large (current default)

The auto mode checks total CUDA tensor size during startup:
- More than 30GB โ†’ large
- Otherwise โ†’ medium
ยท
thomwolfย 
posted an update 8 months ago
view post
Post
7500
If you've followed the progress of robotics in the past 18 months, you've likely noticed how robotics is increasingly becoming the next frontier that AI will unlock.

At Hugging Faceโ€”in robotics and across all AI fieldsโ€”we believe in a future where AI and robots are open-source, transparent, and affordable; community-built and safe; hackable and fun. We've had so much mutual understanding and passion working with the Pollen Robotics team over the past year that we decided to join forces!

You can already find our open-source humanoid robot platform Reachy 2 on the Pollen website and the Pollen community and people here on the hub at pollen-robotics

We're so excited to build and share more open-source robots with the world in the coming months!
  • 1 reply
ยท
mrfakenameย 
posted an update 8 months ago
view post
Post
3613
Papla P1 from Papla Media is now available on the TTS Arena!

Try out Papla's new ultra-realistic TTS model + compare it with other leading models on the TTS Arena: TTS-AGI/TTS-Arena
thomwolfย 
posted an update 8 months ago
view post
Post
3835
The new DeepSite space is really insane for vibe-coders
enzostvs/deepsite

With the wave of vibe-coding-optimized LLMs like the latest open-source DeepSeek model (version V3-0324), you can basically prompt out-of-the-box and create any app and game in one-shot.

It feels so powerful to me, no more complex framework or under-the-hood prompt engineering to have a working text-to-app tool.

AI is eating the world and *open-source* AI is eating AI itself!

PS: and even more meta is that the DeepSite app and DeepSeek model are both fully open-source code => time to start recursively improve?

PPS: you still need some inference hosting unless you're running the 600B param model at home, so check the very nice list of HF Inference Providers for this model: deepseek-ai/DeepSeek-V3-0324
  • 1 reply
ยท
mrfakenameย 
posted an update 9 months ago
mrfakenameย 
posted an update 9 months ago
thomwolfย 
posted an update 9 months ago
view post
Post
3119
We've kept pushing our Open-R1 project, an open initiative to replicate and extend the techniques behind DeepSeek-R1.

And even we were mind-blown by the results we got with this latest model we're releasing: โšก๏ธOlympicCoder ( open-r1/OlympicCoder-7B and open-r1/OlympicCoder-32B)

It's beating Claude 3.7 on (competitive) programming โ€“a domain Anthropic has been historically really strong atโ€“ and it's getting close to o1-mini/R1 on olympiad level coding with just 7B parameters!

And the best part is that we're open-sourcing all about its training dataset, the new IOI benchmark, and more in our Open-R1 progress report #3: https://huggingface.co/blog/open-r1/update-3

Datasets are are releasing:
- open-r1/codeforces
- open-r1/codeforces-cots
- open-r1/ioi
- open-r1/ioi-test-cases
- open-r1/ioi-sample-solutions
- open-r1/ioi-cots
- open-r1/ioi-2024-model-solutions