Pre-Training BERT with Hugging Face Transformers and Habana Gaudi

Philipp Schmid's avatar

In this Tutorial, you will learn how to pre-train BERT-base from scratch using a Habana Gaudi-based DL1 instance on AWS to take advantage of the cost-performance benefits of Gaudi. We will use the Hugging Face Transformers, Optimum Habana and Datasets libraries to pre-train a BERT-base model using masked-language modeling, one of the two original BERT pre-training

 

 

 

To finish reading, please visit source site