MLX, Llama.cpp, and Candle are performing about equally on an M3 Max now.
π―οΈπ₯[Candle](github.com/huggingface/...) is now much faster on macOS thanks to a contribution by @EricLBuehler, which brings major speed improvements to the Metal backend.ππ
Try it out by running some of our examples with the `--features metal` flag.
#Candle #RustLang #macOS #Metal #HuggingFace
21.07.2025 22:22 β π 2 π 1 π¬ 0 π 0
Building Tensors from Scratch in Rust (Part 2): View Operations
A Blog post by Kyle Birnbaum on Hugging Face
I just published part 2 of my article series about creating tensors from scratch in Rust. This one is about view operations.
#tensors #machine-learning #ml #ai
Take a look here:
huggingface.co/blog/KeighBe...
18.06.2025 23:18 β π 2 π 1 π¬ 1 π 0
Building Tensors From Scratch in Rust: Part 1, Core Structure and Indexing
A Blog post by Kyle Birnbaum on Hugging Face
I'm writing an article series about creating tensors from scratch in Rust. #tensors #machine-learning #ml #ai
huggingface.co/blog/KeighBe...
12.06.2025 23:56 β π 5 π 3 π¬ 0 π 0
The mixture of experts model is also an option:
```
cargo run --example qwen --features metal --release -- --prompt "Write a poem about butterflies. <think></think>." --model "3-moe-a3b"
```
30.05.2025 20:00 β π 0 π 0 π¬ 0 π 0
GitHub - huggingface/candle: Minimalist ML framework for Rust
Minimalist ML framework for Rust. Contribute to huggingface/candle development by creating an account on GitHub.
Qwen 3 is now supported in Candle!
Run the 3-4B model locally with:
```
cargo run --example qwen --release -- --model 3-4b --prompt 'The capital of France is '
```
On macOS, enable Metal for faster inference:
```
--features metal
```
Clone the repo and test it out. github.com/huggingface/...
30.05.2025 20:00 β π 0 π 0 π¬ 1 π 0
microsoft/rifts Β· Datasets at Hugging Face
Weβre on a journey to advance and democratize artificial intelligence through open source and open science.
RIFTS Dataset: Solving Critical LLM Conversation Failures
- LLMs 3x less likely to clarify than humans
- 16x less likely to provide follow-up requests
- Early failures predict later breakdowns
- Includes preliminary intervention strategies
huggingface.co/datasets/mic...
21.03.2025 09:57 β π 11 π 3 π¬ 1 π 0
Welcome Gemma 3: Google's all new multimodal, multilingual, long context open LLM
Weβre on a journey to advance and democratize artificial intelligence through open source and open science.
Google just released Gemma 3, an open, on-device LLM with vision capabilities and support for over 140 different languages. Models range from 1B-27B parameters.
Zero-day support for multiple frameworks including transformers, MLX, llama.cpp, and more! πΌ π
Read more here:
huggingface.co/blog/gemma3
12.03.2025 18:46 β π 3 π 0 π¬ 1 π 0
Made some significant updates to the @hf.co semantic datasets search app. If you love falling into a wiki black hole, you might like this...
huggingface.co/spaces/libra...
13.02.2025 17:14 β π 9 π 5 π¬ 0 π 0
YouTube video by Sasha Rush π€
How DeepSeek Changes the LLM Story
What to know about DeepSeek
youtu.be/0eMzc-WnBfQ?...
In which we attempt to figure out MoE, o1, scaling, tech reporting, modern semiconductors, microeconomics, and international geopolitics.
04.02.2025 15:41 β π 95 π 13 π¬ 1 π 5
Itβs a green light for the Frugal AI Challenge! π
For the next month, we invite all members of the AI community to participate in one of our 3 AI for Climate tasks, with the goal of developing a highly accurate model while consuming as little energy as possible β‘
06.01.2025 17:36 β π 23 π 11 π¬ 2 π 1
GitHub - huggingface/coreml-examples: Swift Core ML Examples
Swift Core ML Examples. Contribute to huggingface/coreml-examples development by creating an account on GitHub.
Weβve got great examples of PyTorch to CoreML conversions in the Huggingface coreml-examples repo. Currently, thereβs one tutorial, but more are coming soon. After converting, you can choose what compute units you want the model to run on!
12.12.2024 19:02 β π 0 π 0 π¬ 0 π 0
Christmas came early! π
π» Today marks the newest release of the HuggingChat π€ update with some really exciting capabilities! First up, automatic context injection!
1) Open a file in a supported app, summon HFChat, and it pre-populates the context window. No more copy-pasting. /cc @hf.co
09.12.2024 19:11 β π 11 π 2 π¬ 1 π 1
Or, My laptop has a 72 Wh battery (~208,512 J assuming only 80% is usable). Running Llama3.2-1B would drain the battery after processing:
- CPU: 674,249 tokens (~518,653 words, ~7 novels)
- GPU: 2,799,550 tokens (~2,153,500 words, ~30 novels)
- ANE: 11,273,184 tokens (~8,671,679 words, ~123 novels)
05.12.2024 20:08 β π 2 π 0 π¬ 0 π 0
To put it in perspective: Llama3.2-1B uses ~280 GFLOPS per 20 tokens. My laptop (~2kg) running the model would be the energy equivalent of:
- CPU (6 J): dropping it from 1 foot (31 cm)
- GPU (1.4 J): dropping it from 3 inches (7 cm)
- ANE (0.3 J): dropping it by just half an inch (1.5 cm)!
05.12.2024 20:08 β π 2 π 0 π¬ 1 π 0
Chart Title: Model Hardware vs Energy per GigaFLOP.
Vertical Axis: mJ/GFLOP(Log)
Horizontal Axis: Hardware Type(CPU, CPU + GPU, CPU + ANE)
CPU: min 6.9 1st quartile 11.7 median 13.4 3rd quartile 35.6 max 53.1
CPU + GPU: 4.6 4.6 4.7 6.2 9.6
CPU + ANE: 0.9 1.0 1.1 1.4 1.8
Preliminary data shows the Apple Neural Engine uses ~94% less energy than the CPU and ~75% less than the GPU π€―
On the On-Device team at Hugging Face, we've been profiling energy usage for CoreML models. Hereβs some data I collected:
05.12.2024 20:08 β π 4 π 1 π¬ 2 π 0
Sr. Software Engineer @ Apple AIML
nsarrazin.com
Technology enthusiast, musician, philosophy and history addict. Senior Backend Engineer, Rustacean.
Building tools for AI datasets. π½
Looking in AI datasets. π
Sharing clean open AI datasets. π»
at https://bsky.app/profile/hf.co
Assistant Professor @JohnsHopkinsAMS, Works in Mathematical Optimization,
Mostly here to share pretty maths/3D prints, sometimes sharing my research
Professor, Programmer in NYC.
Cornell, Hugging Face π€
Bringing the power of machine learning to the web. Currently working on Transformers.js (@huggingface π€)
Sr Mgr & Research Scientist @ServiceNowRSRCH, Montreal
π€ ML at Hugging Face
π² Academic Staff at Stanford University (AIMI Center)
𦴠Radiology AI is my stuff
VP and Distinguished Scientist at Microsoft Research NYC. AI evaluation and measurement, responsible AI, computational social science, machine learning. She/her.
One photo a day since January 2018: https://www.instagram.com/logisticaggression/
Building a co-pilot for hardware designers at Vinci4d. Formerly SVP of Engineering Iron Ox. WiML president, COO Mayfield Robotics Roboticist/Machine Learning Researcher
Assistant Professor in CS: researching ML/AI in sociotechnical systems & teaching Data Science and Dev tools with an emphasis responsible computing
New Englander, NSBE lifetime member
profile pic: me in a purplish sweater with math vaguely on the w
I lead Cohere For AI. Formerly Research
Google Brain. ML Efficiency, LLMs,
@trustworthy_ml.
Climate & AI Lead @HuggingFace, TED speaker, WiML board member, TIME AI 100 (She/her/Dr/π¦)
Director of Responsible AI @GSK | Interested in Responsible AI, AI for healthcare, fairness and causality | Prev. Google DeepMind, Google Health, UCL, Stanford, ULiege | WiML board/volunteer. She/her.
Machine Learning researcher. Former stats faculty. Works for Google Research, and on better days, herself.
Researcher in machine learning and computer vision for science. Senior Group Leader at HHMI Janelia Research Campus. Supporter of DEIB in science and tech. CV: https://bit.ly/BransonCV
Personal Account
Founder: The Distributed AI Research Institute @dairinstitute.bsky.social.
Author: The View from Somewhere, a memoir & manifesto arguing for a technological future that serves our communities (to be published by One Signal / Atria
PhD student in Machine Learning @ MPI-IS TΓΌbingen, TΓΌbingen AI Center, IMPRS-IS
machine learning for health at microsoft research, based in cambridge UK π» she/her