Hyperparameter Search with Transformers and Ray Tune

A guest blog post by Richard Liaw from the Anyscale team With cutting edge research implementations, thousands of trained models easily accessible, the Hugging Face transformers library has become critical to the success and growth of natural language processing today. For any machine learning model to achieve good performance, users often need to implement some form of parameter tuning. Yet, nearly everyone (1, 2) either ends up disregarding hyperparameter tuning or opting to do a simplistic grid search with a […]

Read more

Transformers.js v4: Now Available on NPM!

We’re excited to announce that Transformers.js v4 is now available on NPM! After a year of development (we started in March 2025 🤯), we’re finally ready for you to use it. npm i @huggingface/transformers Performance & Runtime Improvements The biggest change is undoubtedly the adoption of a new WebGPU Runtime, completely rewritten in C++. We’ve worked closely with the ONNX Runtime team to thoroughly test this runtime across our ~200 supported model architectures, as well as many new v4-exclusive architectures. […]

Read more

TRL v1.0: Post-Training Library Built to Move with the Field

We’re releasing TRL v1.0, and it marks a real shift in what TRL is. What started as a research codebase has become a dependable library people build on, with clearer expectations around stability. This isn’t just a version bump. It reflects the reality that TRL now powers production systems, and embraces that responsibility. TRL now implements more than 75 post-training methods. But coverage isn’t the goal by itself. What matters is making these methods easy to try, compare, and actually […]

Read more

Granite 4.0 3B Vision: Compact Multimodal Intelligence for Enterprise Documents

Today we’re excited to announce Granite 4.0 3B Vision, a compact vision-language model (VLM) designed for enterprise document understanding. It’s purpose-built for reliable information extraction from complex documents, forms, and structured visuals. Granite 4.0 3B Vision excels on the following capabilities: Table Extraction: Accurately parsing complex table structures (e.g., multi-row, multi-column, etc.) from document images Chart Understanding: Converting charts and figures into structured machine-readable formats, summaries, or executable code Semantic Key-Value Pair (KVP) Extraction: Identifying and grounding semantically meaningful key-value […]

Read more

Falcon Perception

TL;DR — Falcon Perception is a 0.6B-parameter early-fusion Transformer for open-vocabulary grounding and segmentation from natural language prompts. The model processes image patches + text in one sequence using a hybrid attention mask, and produces variable numbers of instances with a small, structured token interface and lightweight output heads. On SA-Co, Falcon Perception reaches 68.0 Macro-F1 (vs. 62.3 for SAM 3) with the main remaining gap being presence calibration (MCC 0.64 vs. 0.82). We also introduce PBench, a diagnostic benchmark […]

Read more

Holo3: Breaking the Computer Use Frontier

We are proud to unveil Holo3, the latest evolution of our vision for the Autonomous Enterprise. With a score of 78.85% on the OSWorld-Verified benchmark, Holo3 establishes a new state of the art for the industry on the leading desktop computer use benchmark. Holo3 is more than a benchmark leader; it is engineered for production. Built using our agentic flywheel, it has been trained to execute real-world workflows within synthetic enterprise environments. This not only ensures that Holo3 excels in […]

Read more

Welcome Gemma 4: Frontier multimodal intelligence on device

The Gemma 4 family of multimodal models by Google DeepMind is out on Hugging Face, with support for your favorite agents, inference engines, and fine-tuning libraries 🤗 These models are the real deal: truly open with Apache 2 licenses, high quality with pareto frontier arena scores, multimodal including audio, and sizes you can use everywhere including on-device. Gemma 4 builds on advances from previous families and makes them click together. In our tests with pre-release checkpoints we have been impressed […]

Read more

Quiz: Class Concepts: Object-Oriented Programming in Python

Interactive Quiz â‹… 8 QuestionsBy Martin Breuss Share In this quiz, you’ll test your understanding of Class Concepts: Object-Oriented Programming in Python. By working through this quiz, you’ll revisit how to define classes, use instance and class attributes, write different types of methods, and apply the descriptor protocol through properties. You can also deepen your knowledge with the tutorial Python Classes: The Power of Object-Oriented Programming. The quiz contains 8 questions and there is no time limit. You’ll get 1 […]

Read more

Quiz: How to Add Features to a Python Project With Codex CLI

Interactive Quiz â‹… 7 QuestionsBy Joseph Peart Share In this quiz, you’ll test your understanding of How to Add Features to a Python Project With Codex CLI. By working through this quiz, you’ll revisit how to install, configure, and use Codex CLI to implement and refine features in a Python project using natural language prompts. The quiz contains 7 questions and there is no time limit. You’ll get 1 point for each correct answer. At the end of the quiz, […]

Read more

Quiz: Python’s Counter: The Pythonic Way to Count Objects

Interactive Quiz â‹… 11 QuestionsBy Martin Breuss Share In this quiz, you’ll test your understanding of Python’s Counter: The Pythonic Way to Count Objects. By working through this quiz, you’ll revisit how to create Counter objects, update counts, find most common elements, and use counters as multisets with arithmetic operations. This quiz covers practical Counter tasks such as constructing counters from different data types, accessing counts, and working with multiset operations. If you want a deeper walkthrough, review the tutorial […]

Read more
1 2 3 4 1,071