Introducing smolagents, a simple library to build agents

Today we are launching smolagents, a very simple library that unlocks agentic capabilities for language models. Here’s a glimpse: from smolagents import CodeAgent, DuckDuckGoSearchTool, HfApiModel agent = CodeAgent(tools=[DuckDuckGoSearchTool()], model=HfApiModel()) agent.run(“How many seconds would it take for a leopard at full speed to run through Pont des Arts?”) Table of Contents

Read more

CO₂ Emissions and Models Performance: Insights from the Open LLM Leaderboard

Since June 2024, we have evaluated more than 3,000 models on the Open LLM Leaderboard, a worldwide ranking of open language models performance. Even though we’re trying to run evaluations without wasting resources (we use the spare cycles of our cluster, in other words the GPUs which are active but waiting between jobs), this still represents quite a big amount of energy spent for model inference! In the last year, people have become more and more aware that using large […]

Read more

AI Agents Are Here. What Now?

Introduction The sudden, rapid advancement of LLM capabilities – such as writing fluent sentences and achieving increasingly high scores on benchmarks – has led AI developers and businesses alike to look towards what comes next: What game-changing technology is just on the horizon? One technology very recently taking off is “AI agents”, systems that can take actions in the digital world aligned with a deployer’s goals. Most of today’s AI agents    

Read more

Train 400x faster Static Embedding Models with Sentence Transformers

This blog post introduces a method to train static embedding models that run 100x to 400x faster on CPU than state-of-the-art embedding models, while retaining most of the quality. This unlocks a lot of exciting use cases, including on-device and in-browser execution, edge computing, low power and embedded applications. We apply this recipe to train two extremely efficient embedding models: sentence-transformers/static-retrieval-mrl-en-v1    

Read more

Timm ❤️ Transformers: Use any timm model with transformers

Get lightning-fast inference, quick quantization, torch.compile boosts, and effortless fine-tuning for any timm model—all within the friendly 🤗 transformers ecosystem. Enter TimmWrapper—a simple, yet powerful tool that unlocks this potential. In this post, we’ll cover: How the timm integration works and why it’s a game-changer. How to integrate timm models with 🤗 transformers. Practical examples: pipelines, quantization, fine-tuning, and more. To follow along with this blog post, install the latest version of transformers and timm by running: pip install -Uq […]

Read more
1 48 49 50 51 52 1,021