Mastering Transformers - Second Edition : The Journey from BERT to Large Language Models and Stable Diffusion

Mastering Transformers - Second Edition : The Journey from BERT to Large Language Models and Stable Diffusion

PAPERBACK

03 Jun, 2024

Explore transformer-based language models from BERT to GPT, delving into NLP and computer vision tasks, while tackling challenges effectively Key Features-...

No Reviews

International Edition

Ships within 15-17 Business Days

New

₹ 4091
BRAND NEW - Item in perfectly NEW condition.

Used

-
GOOD CONDITION - Used book in GOOD - READABLE condition. The books may contain markings, highlightings and wear due to previous usage. The book is in overall good condition. Great Deal !!!

* This item is NOT Returnable *

ISBN-10:

1837633789

ISBN-13:

9781837633784

Publisher

Packt Publishing

Dimensions

9.25 X 7.50 X 0.93 inches

Language

English

Description

Explore transformer-based language models from BERT to GPT, delving into NLP and computer vision tasks, while tackling challenges effectively

Key Features

- Understand the complexity of deep learning architecture and transformers architecture

- Create solutions to industrial natural language processing (NLP) and computer vision (CV) problems

- Explore challenges in the preparation process, such as problem and language-specific dataset transformation

- Purchase of the print or Kindle book includes a free PDF eBook

Book Description

Transformer-based language models such as BERT, T5, GPT, DALL-E, and ChatGPT have dominated NLP studies and become a new paradigm. Thanks to their accurate and fast fine-tuning capabilities, transformer-based language models have been able to outperform traditional machine learning-based approaches for many challenging natural language understanding (NLU) problems.

Aside from NLP, a fast-growing area in multimodal learning and generative AI has recently been established, showing promising results. Mastering Transformers will help you understand and implement multimodal solutions, including text-to-image. Computer vision solutions that are based on transformers are also explained in the book. You'll get started by understanding various transformer models before learning how to train different autoregressive language models such as GPT and XLNet. The book will also get you up to speed with boosting model performance, as well as tracking model training using the TensorBoard toolkit. In the later chapters, you'll focus on using vision transformers to solve computer vision problems. Finally, you'll discover how to harness the power of transformers to model time series data and for predicting.

By the end of this transformers book, you'll have an understanding of transformer models and how to use them to solve challenges in NLP and CV.

What you will learn

- Focus on solving simple-to-complex NLP problems with Python

- Discover how to solve classification/regression problems with traditional NLP approaches

- Train a language model and explore how to fine-tune models to the downstream tasks

- Understand how to use transformers for generative AI and computer vision tasks

- Build transformer-based NLP apps with the Python transformers library

- Focus on language generation such as machine translation and conversational AI in any language

- Speed up transformer model inference to reduce latency

Who this book is for

This book is for deep learning researchers, hands-on practitioners, and ML/NLP researchers. Educators, as well as students who have a good command of programming subjects, knowledge in the field of machine learning and artificial intelligence, and who want to develop apps in the field of NLP as well as multimodal tasks will also benefit from this book's hands-on approach. Knowledge of Python (or any programming language) and machine learning literature, as well as a basic understanding of computer science, are required.

Table of Contents

- From Bag-of-Words to the Transformer

- A Hands-On Introduction to the Subject

- Autoencoding Language Models

- Autoregressive Language Models

- Fine-Tuning Language Model for Text Classification

- Fine-Tuning Language Models for Token Classification

- Text Representation

- Boosting Your Model Performance

- Parameter Efficient Fine-Tuning

- Zero-Shot and Few-Shot Learning in NLP

- Explainable AI (XAI) for NLP

- Working with Efficient Transformers

- Cross-Lingual Language Modeling

- Serving Transformer Models

(N.B. Please use the Read Sample option to see further chapters)

Product Details

ISBN-10

:1837633789

ISBN-13

:9781837633784

Publisher

:Packt Publishing

Publication date

: 03 Jun, 2024

Category

: Computer & Internet

Format

:PAPERBACK

Language

:English

Reading Level

: All

Dimension

: 9.25 X 7.50 X 0.93 inches

Weight

:789 g

Loading, please wait...

Copyright © 2024. Boganto.com. All Rights Reserved