-
内容大纲
在过去的几年,AI获得了令人惊讶的新语言能力。在深度学习快速发展的推动下,语言AI系统比以往任何时候都能更好地编写和理解文本。这一趋势正在催生新功能、新产品,甚至新的行业。通过本书的可视化教育方式,读者将学习到现在使用这些功能所需的实用工具和概念。
你将了解如何将预训练的大语言模型用于文案撰写和摘要生成等应用场景,创建超越关键字匹配的语义搜索系统,以及使用现有库和预训练的模型进行文本分类、搜索和聚类。 -
作者介绍
-
目录
Preface
Part I. Understanding Language Models
1. An Introduction to Large Language Models
What Is Language AI?
A Recent History of Language AI
Representing Language as a Bag-of-Words
Better Representations with Dense Vector Embeddings
Types of Embeddings
Encoding and Decoding Context with Attention
Attention Is All You Need
Representation Models: Encoder-Only Models
Generative Models: Decoder-Only Models
The Year of Generative AI
The Moving Definition of a "Large Language Model"
The Training Paradigm of Large Language Models
Large Language Model Applications: What Makes Them So Useful?
Responsible LLM Development and Usage
Limited Resources Are All You Need
Interfacing with Large Language Models
Proprietary, Private Models
Open Models
Open Source Frameworks
Generating Your First Text
Summary
2. Tokens and Embeddings
LLM Tokenization
How Tokenizers Prepare the Inputs to the Language Model
Downloading and Running an LLM
How Does the Tokenizer Break Down Text?
Word Versus Subword Versus Character Versus Byte Tokens
Comparing Trained LLM Tokenizers
Tokenizer Properties
Token Embeddings
A Language Model Holds Embeddings for the Vocabulary of Its Tokenizer
Creating Contextualized Word Embeddings with Language Models
Text Embeddings (for Sentences and Whole Documents)
Word Embeddings Beyond LLMs
Using pretrained Word Embeddings
The Word2vec Algorithm and Contrastive Training
Embeddings for Recommendation Systems
Recommending Songs by Embeddings
Training a Song Embedding Model
Summary
3. Looking Inside Large Language Models
An Overview of Transformer Models
The Inputs and Outputs of a Trained Transformer LLM
The Components of the Forward Pass
Choosing a Single Token from the Probability Distribution (Sampling/Decoding)
Parallel Token Processing and Context Size
Speeding Up Generation by Caching Keys and Values
Inside the Transformer Block
Recent Improvements to the Transformer Architecture
More Efficient Attention
The Transformer Block
Positional Embeddings (ROPE)
Other Architectural Experiments and Improvements
Summary
Part II. Using Pretrained Language Models
4. Text Classification
The Sentiment of Movie Reviews
Text Classification with Representation Models
Model Selection
Using a Task-Specific Model
Classification Tasks That Leverage Embeddings
Supervised Classification
What If We Do Not Have Labeled Data?
Text Classification with Generative Models
Using the Text-to-Text Transfer Transformer
ChatGPT for Classification
Summary
5. Text Clustering and Topic Modeling
ArXiv's Articles: Computation and Language
A Common Pipeline for Text Clustering
Embedding Documents
Reducing the Dimensionality of Embeddings
Cluster the Reduced Embeddings
Inspecting the Clusters
From Text Clustering to Topic Modeling
BERTopic: A Modular Topic Modeling Framework
Adding a Special Lego Block
The Text Generation Lego Block
Summary
6. Prompt Engineering
Using Text Generation Models
Choosing a Text Generation Model
Loading a Text Generation Model
Controlling Model Output
Intro to Prompt Engineering
The Basic Ingredients of a Prompt
Instruction-Based Prompting
Advanced Prompt Engineering
The Potential Complexity of a Prompt
In-Context Learning: Providing Examples
Chain Prompting: Breaking up the Problem
Reasoning with Generative Models
Chain-of-Thought: Think Before Answering
Self-Consistency: Sampling Outputs
Tree-of-Thought: Exploring Intermediate Steps
Output Verification
Providing Examples
Grammar: Constrained Sampling
Summary
7. Advanced Text Generation Techniques and Tools
Model I/O: Loading Quantized Models with LangChain
Chains: Extending the Capabilities of LLMs
A Single Link in the Chain: Prompt Template
A Chain with Multiple Prompts
Memory: Helping LLMs to Remember Conversations
Conversation Buffer
Windowed Conversation Buffer
Conversation Summary
Agents: Creating a System of LLMs
The Driving Power Behind Agents: Step-by-step Reasoning
ReAct in LangChain
Summary
8. Semantic Search and Retrieval-Augmented Generation
Overview of Semantic Search and RAG
Semantic Search with Language Models
Dense Retrieval
Reranking
Retrieval Evaluation Metrics
Retrieval-Augmented Generation (RAG)
From Search to RAG
Example: Grounded Generation with an LLM API
Example: RAG with Local Models
Advanced RAG Techniques
RAG Evaluation
Summary
9. Multimoflal Large Language Models
Transformers for Vision
Multimodal Embedding Models
CLIP: Connecting Text and Images
How Can CLIP Generate Multimodal Embeddings?
OpenCLIP
Making Text Generation Models Multimodal
BLIP-2: Bridging the Modality Gap
Preprocessing Multimodal Inputs
Use Case 1: Image Captioning
Use Case 2: Multimodal Chat-Based Prompting
Summary
Part III. Training and Fine-Tuning Language Models
10. Creating Text Embedding Models
Embedding Models
What Is Contrastive Learning?
SBERT
Creating an Embedding Model
Generating Contrastive Examples
Train Model
In-Depth Evaluation
Loss Functions
Fine-Tuning an Embedding Model
Supervised
Augmented SBERT
Unsupervised Learning
Transformer-Based Sequential Denoising Auto-Encoder
Using TSDAE for Domain Adaptation
Summary
11. Fine-Tuning Representation Models for Classification
Supervised Classification
Fine-Tuning a Pretrained BERT Model
Freezing Layers
Few-Shot Classification
SetFit: Efficient Fine-Tuning with Few Training Examples
Fine-Tuning for Few-Shot Classification
Continued Pretraining with Masked Language Modeling
Named-Entity Recognition
Preparing Data for Named-Entity Recognition
Fine-Tuning for Named-Entity Recognition
Summary
12. Fine-Tuning Generation Models
The Three LLM Training Steps: Pretraining, Supervised Fine-Tuning, and
Preference Tuning
Supervised Fine-Tuning (SFT)
Full Fine-Tuning
Parameter-Efficient Fine-Tuning (PEFT)
Instruction Tuning with QLoRA
Templating Instruction Data
Model Quantization
LoRA Configuration
Training Configuration
Training
Merge Weights
Evaluating Generative Models
Word-Level Metrics
Benchmarks
Leaderboards
Automated Evaluation
Human Evaluation
Preference-Tuning / Alignment / RLHF
Automating Preference Evaluation Using Reward Models
The Inputs and Outputs of a Reward Model
Training a Reward Model
Training No Reward Model
Preference Tuning with DPO
Templating Alignment Data
Model Quantization
Training Configuration
Training
Summary
Afterword
Index
同类热销排行榜
- C语言与程序设计教程(高等学校计算机类十二五规划教材)16
- 电机与拖动基础(教育部高等学校自动化专业教学指导分委员会规划工程应用型自动化专业系列教材)13.48
- 传感器与检测技术(第2版高职高专电子信息类系列教材)13.6
- ASP.NET项目开发实战(高职高专计算机项目任务驱动模式教材)15.2
- Access数据库实用教程(第2版十二五职业教育国家规划教材)14.72
- 信号与系统(第3版下普通高等教育九五国家级重点教材)15.08
- 电气控制与PLC(普通高等教育十二五电气信息类规划教材)17.2
- 数字电子技术基础(第2版)17.36
- VB程序设计及应用(第3版十二五职业教育国家规划教材)14.32
- Java Web从入门到精通(附光盘)/软件开发视频大讲堂27.92
推荐书目
-
孩子你慢慢来/人生三书 华人世界率性犀利的一枝笔,龙应台独家授权《孩子你慢慢来》20周年经典新版。她的《...
-
时间简史(插图版) 相对论、黑洞、弯曲空间……这些词给我们的感觉是艰深、晦涩、难以理解而且与我们的...
-
本质(精) 改革开放40年,恰如一部四部曲的年代大戏。技术突变、产品迭代、产业升级、资本对接...