Llama can now see and run on your device – welcome Llama 3.2

Llama 3.2 is out! Today, we welcome the next iteration of the Llama collection to Hugging Face. This time, we’re excited to collaborate with Meta on the release of multimodal and small models. Ten open-weight models (5 multimodal models and 5 text-only ones) are available on the Hub. Llama 3.2 Vision comes in two sizes: 11B for efficient deployment and development on consumer-size GPU, and 90B for large-scale applications. Both versions come in base and instruction-tuned variants. In addition to […]

Read more

🇨🇿 BenCzechMark – Can your LLM Understand Czech?

The 🇨🇿 BenCzechMark is the first and most comprehensive evaluation suite for assessing the abilities of Large Language Models (LLMs) in the Czech language. It aims to test how well LLMs can: Reason and perform complex tasks in Czech. Generate and verify grammatically and semantically correct Czech. Extract information and store knowledge by answering questions about Czech culture and Czech-related facts. Do what language models were originally trained for—estimate the probability of Czech texts. To achieve this, we’ve sourced 50 tasks […]

Read more

A Short Summary of Chinese AI Global Expansion

In the early 15th century, Zheng He (also known as Chong Ho), a Chinese mariner and explorer during the early Ming Dynasty, led seven major naval expeditions, known as the “Voyages to the Western Oceans”. His journey traced a path that went through Southeast Asia, the Middle East and then reached out to Africa. It was a bold move by China to    

Read more

Introducing the Open FinLLM Leaderboard

Finding the best LLM models for finance use cases The growing complexity of financial language models (LLMs) necessitates evaluations that go beyond general NLP benchmarks. While traditional leaderboards focus on broader NLP tasks like translation or summarization, they often fall short in addressing the specific needs of the finance industry. Financial tasks, such as predicting stock movements, assessing credit risks, and extracting information from financial reports, present unique challenges that require models with specialized skills. This is why we decided […]

Read more

Scaling AI-Based Data Processing with Hugging Face + Dask

The Hugging Face platform has many datasets and pre-trained models that make using and training state-of-the-art machine learning models increasingly accessible. However, it can be hard to scale AI tasks because AI datasets are often large (100s GBs to TBs) and using Hugging Face transformers for model inference can sometimes be computationally expensive. Dask, a Python library for distributed computing, can handle out-of-core computing (processing data that doesn’t fit in memory) by breaking datasets into manageable chunks. This makes it […]

Read more

Welcome, Gradio 5

We’ve been hard at work over the past few months, and we are excited to now announce the stable release of Gradio 5.  With Gradio 5, developers can build production-ready machine learning web applications that are performant, scalable, beautifully designed, accessible, and follow best web security practices, all in a few lines of Python. To give Gradio 5 a spin, simply type    

Read more
1 43 44 45 46 47 1,020