Scaling AI-Based Data Processing with Hugging Face + Dask

The Hugging Face platform has many datasets and pre-trained models that make using and training state-of-the-art machine learning models increasingly accessible. However, it can be hard to scale AI tasks because AI datasets are often large (100s GBs to TBs) and using Hugging Face transformers for model inference can sometimes be computationally expensive. Dask, a Python library for distributed computing, can handle out-of-core computing (processing data that doesn’t fit in memory) by breaking datasets into manageable chunks. This makes it […]

Read more

Welcome, Gradio 5

We’ve been hard at work over the past few months, and we are excited to now announce the stable release of Gradio 5.  With Gradio 5, developers can build production-ready machine learning web applications that are performant, scalable, beautifully designed, accessible, and follow best web security practices, all in a few lines of Python. To give Gradio 5 a spin, simply type    

Read more

Fixing Gradient Accumulation

Our friends at Unsloth shared an issue regarding gradient accumulation yesterday that is affecting the transformers Trainer. The initial report comes from @bnjmn_marie (kudos to him!). Gradient accumulation is supposed to be mathematically equivalent to full batch training; however, losses did not match between training runs where the setting was toggled on and off. Where does it stem from? Inside the modeling code of each model, transformers offers a “default” loss function that’s the most typically    

Read more

Llama 3.2 in Keras

This is going to be the shortest blog post ever. Question: Llama 3.2 landed two weeks ago on Hugging Face / Transformers. When will it be available in Keras? Answer: It has been working from day 1 😀. There is nothing to wait for. Yes, Keras Llama3 can be loaded from any standard (i.e. safetensors) Hugging    

Read more

Deploying Speech-to-Speech on Hugging Face

Speech-to-Speech (S2S) is an exciting new project from Hugging Face that combines several advanced models to create a seamless, almost magical experience: you speak, and the system responds with a synthesized voice. The project implements a cascaded pipeline leveraging models available through the Transformers library on the Hugging Face hub. The pipeline consists of the following components: Voice Activity Detection (VAD) Speech to Text (STT) Language Model (LM) Text to Speech (TTS) What’s more, S2S has multi-language support! It currently […]

Read more

🧨 Diffusers welcomes Stable Diffusion 3.5 Large

Stable Diffusion 3.5 is the improved variant of its predecessor, Stable Diffusion 3. As of today, the models are available on the Hugging Face Hub and can be used with 🧨 Diffusers. The release comes with two checkpoints: A large (8B) model A large (8B) timestep-distilled model enabling few-step inference In this post, we will focus on how to use Stable Diffusion 3.5 (SD3.5) with Diffusers, covering both inference and training. Table Of Contents

Read more
1 47 48 49 50 51 1,024