Since their introduction in 2017, transformers have quickly become the
dominant architecture for achieving state-of-the-art results on a
variety of natural language processing tasks. If you're a data scientist
or coder, this practical book -now revised in full color- shows you how
to train and scale these large models using Hugging Face Transformers, a
Python-based deep learning library.
Transformers have been used to write realistic news stories, improve
Google Search queries, and even create chatbots that tell corny jokes.
In this guide, authors Lewis Tunstall, Leandro von Werra, and Thomas
Wolf, among the creators of Hugging Face Transformers, use a hands-on
approach to teach you how transformers work and how to integrate them in
your applications. You'll quickly learn a variety of tasks they can help
you solve.
- Build, debug, and optimize transformer models for core NLP tasks, such
as text classification, named entity recognition, and question
answering
- Learn how transformers can be used for cross-lingual transfer learning
- Apply transformers in real-world scenarios where labeled data is
scarce
- Make transformer models efficient for deployment using techniques such
as distillation, pruning, and quantization
- Train transformers from scratch and learn how to scale to multiple
GPUs and distributed environments