Author: Deep Learner
OpenRAIL: Towards open and responsible AI licensing frameworks
Open & Responsible AI licenses (“OpenRAIL”) are AI-specific licenses enabling open access, use and distribution of AI artifacts while requiring a responsible use of the latter. OpenRAIL licenses could be for open and responsible ML what current open software licenses are to code and Creative Commons to general content: a widespread community licensing tool. Advances in machine learning and other
Read moreHow to train a Language Model with Megatron-LM
Training large language models in Pytorch requires more than a simple training loop. It is usually distributed across multiple devices, with many optimization techniques for a stable and efficient training. Hugging Face 🤗 Accelerate library was created to support distributed training across GPUs and TPUs with very easy integration into the training loops. 🤗 Transformers also support distributed
Read moreTrain your first Decision Transformer
In a previous post, we announced the launch of Decision Transformers in the transformers library. This new technique of using a Transformer as a Decision-making model is getting increasingly popular. So today, you’ll learn to train your first
Read moreEthics and Society Newsletter #1
Hello, world! Originating as an open-source company, Hugging Face was founded on some key ethical values in tech: collaboration, responsibility, and transparency. To code in an open environment means having your code – and the choices within – viewable to the world, associated with your account and available for others to critique and add to. As the research community began using
Read moreSetFit: Efficient Few-Shot Learning Without Prompts
SetFit is significantly more sample efficient and robust to noise than standard fine-tuning. Few-shot learning with pretrained language models has emerged as a promising solution to every data scientist’s nightmare: dealing with data that has few to no labels 😱. Together with our research partners at Intel Labs and the UKP Lab, Hugging Face is excited to introduce SetFit: an efficient framework for few-shot fine-tuning of Sentence Transformers. SetFit achieves high accuracy with little labeled data – for example, with […]
Read moreHow 🤗 Accelerate runs very large models thanks to PyTorch
Meta AI and BigScience recently open-sourced very large language models which won’t fit into memory (RAM or GPU) of most consumer hardware. At Hugging Face, part of our mission is to make even those large models accessible, so we developed tools to allow you to run those models even if you don’t own a supercomputer. All the examples picked in this blog
Read more