Julien Chaumond's picture

Julien Chaumond PRO

julien-c

AI & ML interests

<3 ML/AI for everyone, building products to propel communities fwd

Recent Activity

Organizations

Hugging Face's profile picture Safetensors's profile picture Notebooks-explorers's profile picture Nbconvert-internal's profile picture BigScience Workshop's profile picture Spaces-explorers's profile picture Templates's profile picture The LLM Course's profile picture Giskard's profile picture ph-snps's profile picture Text Generation Inference's profile picture Amazon SageMaker Community's profile picture Training Transformers Together's profile picture Hugging Chat's profile picture Atmos Bank's profile picture Godot Engine Demos's profile picture Pyodide Demos's profile picture Huggingface.js's profile picture Webhooks Explorers (BETA)'s profile picture Workshop June 13 Classroom's profile picture HF Canonical Model Maintainers's profile picture TRL's profile picture Scanned Tokens's profile picture HF Legal's profile picture Language Tools's profile picture Stable Diffusion concepts library's profile picture Teven-projects's profile picture Banana-projects's profile picture Exbert-project's profile picture Blog-explorers's profile picture EU org's profile picture Hacktoberfest 2023's profile picture huggingPartyParis's profile picture Enterprise Explorers's profile picture ZeroGPU Explorers's profile picture OpenAI community's profile picture XLNet community's profile picture ALBERT community's profile picture Transformer-XL community's profile picture Facebook AI community's profile picture DistilBERT community's profile picture BERT community's profile picture T5 community's profile picture choosealicense.com mirror's profile picture Social Post Explorers's profile picture Dev Mode Explorers's profile picture Test's profile picture private beta for deeplinks's profile picture Paris AI Running Club's profile picture kmhf's profile picture Hugging Face Party @ PyTorch Conference's profile picture Nerdy Face's profile picture Hugging Face Science's profile picture open/ acc's profile picture DDUF's profile picture Self-serve FTW's profile picture Inference Explorers's profile picture hf-inference's profile picture Hugging Face MCP Course's profile picture Changelog's profile picture Tiny Agents's profile picture

julien-c's activity

reacted to jsulz's post with πŸ”₯ 18 days ago
view post
Post
2131
Heyo @RichardErkhov the xet-team at Hugging face was wondering if you wanted to join the fun and jump over to Xet storage. πŸ€—

We've been onboarding folks https://huggingface.co/blog/xet-on-the-hub know the backend can scale (Llama 4 and Qwen 3 are on Xet), is great for working with quants (see xet-team/quantization-dedup ), and we're pushing on inviting impactful orgs and users on the Hub. You fit the bill.

We'd love to onboard you, get some feedback, and create some excitement πŸŽ‰

The steps are pretty straightforward - join the waitlist at hf.co/join/xet and we'll take care of the rest.

The system is fully backward compatible, so you shouldn't notice a thing. BUT to get the best experience when uploading/downloading, make sure you have hf_xet installed alongside the latest huggingface_hub

What do you think?
  • 4 replies
Β·
reacted to reach-vb's post with πŸ‘ 19 days ago
view post
Post
3690
hey hey @mradermacher - VB from Hugging Face here, we'd love to onboard you over to our optimised xet backend! πŸ’₯

as you know we're in the process of upgrading our storage backend to xet (which helps us scale and offer blazingly fast upload/ download speeds too): https://huggingface.co/blog/xet-on-the-hub and now that we are certain that the backend can scale with even big models like Llama 4/ Qwen 3 - we;re moving to the next phase of inviting impactful orgs and users on the hub over as you are a big part of the open source ML community - we would love to onboard you next and create some excitement about it in the community too!

in terms of actual steps - it should be as simple as one of the org admins to join hf.co/join/xet - we'll take care of the rest.

p.s. you'd need to have a the latest hf_xet version of huggingface_hub lib but everything else should be the same: https://huggingface.co/docs/hub/storage-backends#using-xet-storage

p.p.s. this is fully backwards compatible so everything will work as it should! πŸ€—
Β·
replied to cbensimon's post 22 days ago
reacted to cbensimon's post with πŸ”₯ 22 days ago
view post
Post
5745
πŸš€ ZeroGPU medium size is now available as a power-user feature

Nothing too fancy for nowβ€”ZeroGPU Spaces still default to large (70GB VRAM)β€”but this paves the way for:
- πŸ’° size-based quotas / pricing (medium will offer significantly more usage than large)
- 🦣 the upcoming xlarge size (141GB VRAM)

You can as of now control GPU size via a Space variable. Accepted values:
- auto (future default)
- medium
- large (current default)

The auto mode checks total CUDA tensor size during startup:
- More than 30GB β†’ large
- Otherwise β†’ medium
Β·
replied to their post about 1 month ago
reacted to their post with πŸ‘πŸš€πŸ”₯ about 1 month ago
view post
Post
3894
Important notice 🚨

For Inference Providers who have built support for our Billing API (currently: Fal, Novita, HF-Inference – with more coming soon), we've started enabling Pay as you go (=PAYG)

What this means is that you can use those Inference Providers beyond the free included credits, and they're charged to your HF account.

You can see it on this view: any provider that does not have a "Billing disabled" badge, is PAYG-compatible.
Β·
reacted to danielhanchen's post with β€οΈπŸ€—πŸ”₯ about 1 month ago
view post
Post
5826
πŸ¦₯ Introducing Unsloth Dynamic v2.0 GGUFs!
Our v2.0 quants set new benchmarks on 5-shot MMLU and KL Divergence, meaning you can now run & fine-tune quantized LLMs while preserving as much accuracy as possible.

Llama 4: unsloth/Llama-4-Scout-17B-16E-Instruct-GGUF
DeepSeek-R1: unsloth/DeepSeek-R1-GGUF-UD
Gemma 3: unsloth/gemma-3-27b-it-GGUF

We made selective layer quantization much smarter. Instead of modifying only a subset of layers, we now dynamically quantize all layers so every layer has a different bit. Now, our dynamic method can be applied to all LLM architectures, not just MoE's.

Blog with Details: https://docs.unsloth.ai/basics/dynamic-v2.0

All our future GGUF uploads will leverage Dynamic 2.0 and our hand curated 300K–1.5M token calibration dataset to improve conversational chat performance.

For accurate benchmarking, we built an evaluation framework to match the reported 5-shot MMLU scores of Llama 4 and Gemma 3. This allowed apples-to-apples comparisons between full-precision vs. Dynamic v2.0, QAT and standard iMatrix quants.

Dynamic v2.0 aims to minimize the performance gap between full-precision models and their quantized counterparts.
reacted to their post with πŸ˜ŽπŸ€—πŸ”₯ about 1 month ago
view post
Post
4765
BOOOOM: Today I'm dropping TINY AGENTS

the 50 lines of code Agent in Javascript πŸ”₯

I spent the last few weeks working on this, so I hope you will like it.

I've been diving into MCP (Model Context Protocol) to understand what the hype was all about.

It is fairly simple, but still quite powerful: MCP is a standard API to expose sets of Tools that can be hooked to LLMs.

But while doing that, came my second realization:

Once you have a MCP Client, an Agent is literally just a while loop on top of it. 🀯

➑️ read it exclusively on the official HF blog: https://huggingface.co/blog/tiny-agents
  • 1 reply
Β·
posted an update about 1 month ago
view post
Post
4765
BOOOOM: Today I'm dropping TINY AGENTS

the 50 lines of code Agent in Javascript πŸ”₯

I spent the last few weeks working on this, so I hope you will like it.

I've been diving into MCP (Model Context Protocol) to understand what the hype was all about.

It is fairly simple, but still quite powerful: MCP is a standard API to expose sets of Tools that can be hooked to LLMs.

But while doing that, came my second realization:

Once you have a MCP Client, an Agent is literally just a while loop on top of it. 🀯

➑️ read it exclusively on the official HF blog: https://huggingface.co/blog/tiny-agents
  • 1 reply
Β·
replied to gtvracer's post about 2 months ago
view reply

which provider do you use?

We'll ship a provider="auto" in the coming days BTW, cc @sbrandeis @Wauplin @celinah

In the meantime, the model is served by those providers, you can use one of them, for instance, add provider="novita" to your code:

image.png

replied to OFT's post 2 months ago
replied to jsulz's post 3 months ago
reacted to jsulz's post with β€οΈπŸš€ 3 months ago
view post
Post
1462
It's finally here ❀️

Build faster than ever with lightning fast upload and download speeds starting today on the Hub ⚑

Xet storage is rolling out access across the Hub - join the waitlist here https://huggingface.co/join/xet

You can apply for yourself, or your entire organization. Head over to your account settings for more information or join anywhere you see the Xet logo on a repository you know.

Have questions? Join the conversation below πŸ‘‡ or open a discussion on the Xet team page xet-team/README
Β·