文章目录
- 1、概览
- 为什么我们需要链?
- 2、快速入门 (Get started) - Using `LLMChain`
- 多个变量 使用字典输入
- 在 `LLMChain` 中使用聊天模型:
- 3、异步 API
- 4、不同的调用方法
- `__call__`调用
- 仅返回输出键值 return_only_outputs
- 只有一个输出键 run
- 只有一个输入键
- 5、自定义chain
- 6、调试链 (Debugging chains)
- 7、从 LangChainHub 加载
- 8、添加记忆(state)
- 9、序列化
- 将chain 保存到磁盘
- 从磁盘加载
- 分开保存组件
- 基础
- 10、LLM
- 基本使用
- LLM 链的其他运行方式
- apply 允许您对一组输入运行链逻辑
- generate 返回 LLMResult
- predict 输入键为关键字
- 解析输出 apply_and_parse
- 从字符串初始化 LLMChain
- 11、Router
- LLMRouterChain
- EmbeddingRouterChain
- 12、序列(Sequential)
- 顺序链(Sequential Chain)
- 顺序链中的记忆(Memory in Sequential Chains)
- 13、转换
- 14、文档
- Stuff documents
- Refine
- Map reduce
- Map re-rank
- 热门(Popular)
- 15、检索型问答(Retrieval QA)
- 链类型 (Chain Type)
- from_chain_type 指定链类型
- combine_documents_chain 直接加载链
- 自定义提示 (Custom Prompts)
- 返回源文档
- RetrievalQAWithSourceChain
- 16、对话式检索问答(Conversational Retrieval QA)
- 传入聊天历史
- 使用不同的模型来压缩问题
- 返回源文档
- ConversationalRetrievalChain 与 `search_distance`
- ConversationalRetrievalChain with `map_reduce`
- ConversationalRetrievalChain 与带有来源的问答
- ConversationalRetrievalChain 与流式传输至 `stdout`
- get_chat_history 函数
- 17、SQL
- 使用查询检查器 Query Checker
- 自定义提示
- 返回中间步骤 (Return Intermediate Steps)
- 选择如何限制返回的行数
- 添加每个表的示例行
- 自定义表信息 (Custom Table Info)
- SQLDatabaseSequentialChain
- 使用本地语言模型
- 18、摘要 summarize
- 准备数据
- 快速开始
- The `stuff` Chain
- The `map_reduce` Chain
- 自定义 `MapReduceChain`
- The `refine` Chain
- 附加 ( Additional )
- 19、分析文档 (Analyze Document)
- 总结
- 问答
- 20、ConstitutionalChain 自我批判链
- UnifiedObjective
- Custom Principles
- 运行多个 principle
- Intermediate Steps - ConstitutionalChain
- No revision necessary
- All Principles
- 21、抽取
- 抽取实体
- Pydantic示例
- 22、Graph QA
- 创建 graph
- 查询图
- Save the graph
- 23、虚拟文档嵌入
- 多次生成
- 使用我们自己的提示
- 使用HyDE
- 24、Bash chain
- Customize Prompt
- Persistent Terminal
- 25、自检链
- 26、数学链
- 27、HTTP request chain
- 28、Summarization checker chain
- 审查 Moderation
- 如何使用审核链
- 如何将审核链附加到 LLMChain
- 29、动态从多个提示中选择 multi_prompt_router
- 30、动态选择多个检索器 multi_retrieval_qa_router
- 31、使用OpenAI函数进行检索问答
- 使用 Pydantic
- 在ConversationalRetrievalChain中使用
- 32、OpenAPI chain
- Load the spec
- Select the Operation
- Construct the chain
- Return raw response
- Example POST message
- 33、Program-aided language model (PAL) chain
- Math Prompt
- Colored Objects
- Intermediate Steps
- 34、Question-Answering Citations
- 35、文档问答 qa_with_sources
- 准备数据
- 快速入门
- The `stuff` Chain
- The `map_reduce` Chain
- 中间步骤 (Intermediate Steps)
- The `refine` Chain
- The `map-rerank` Chain
- 带有来源的文档问答
- 36、标记
- 最简单的方法,只指定类型
- 更多控制
- 使用 Pydantic 指定模式
- 37、Vector store-augmented text generation
- Prepare Data
- Set Up Vector DB
- Set Up LLM Chain with Custom Prompt
- Generate Text
本文转载改编自:
https://python.langchain.com.cn/docs/modules/chains/
1、概览
在简单应用中,单独使用LLM是可以的, 但更复杂的应用需要将LLM进行链接 - 要么相互链接,要么与其他组件链接。
LangChain为这种"链接"应用提供了Chain接口。
我们将链定义得非常通用,它是对组件调用的序列,可以包含其他链。基本接口很简单:
class Chain(BaseModel, ABC):
"""Base interface that all chains should implement."""
memory: BaseMemory
callbacks: Callbacks
def __call__(
self,
inputs: Any,
return_only_outputs: bool = False,
callbacks: Callbacks = None,
) -> Dict[str, Any]:
...
将组件组合成链的思想简单而强大。
它极大地简化了复杂应用的实现,并使应用更模块化,从而更容易调试、维护和改进应用。
更多具体信息请查看:
- 如何使用,了解不同链特性的详细步骤
- 基础,熟悉核心构建块链
- 文档,了解如何将文档纳入链中
- 常用 链,用于最常见的用例
- 附加,查看一些更高级的链和集成,可以直接使用
为什么我们需要链?
链允许我们将 多个组件组合在一起 创建一个单一的、连贯的应用。
例如,我们可以创建一个链,该链接收用户输入,使用PromptTemplate对其进行格式化,然后将格式化后的响应传递给LLM。
我们可以通过将多个链 组合在一起 或将 链与其他组件组合来构建更复杂的链。
2、快速入门 (Get started) - Using LLMChain
LLMChain
是最基本的构建块链。
它接受一个提示模板,将其与用户输入进行格式化,并返回 LLM 的响应。
要使用 LLMChain
,首先创建一个提示模板。
from langchain.llms import OpenAI
from langchain.prompts import PromptTemplate
llm = OpenAI(temperature=0.9)
prompt = PromptTemplate(
input_variables=["product"],
template="What is a good name for a company that makes {product}?",
)
现在我们可以创建一个非常简单的链,它将接受用户输入,使用它格式化提示,然后将其发送到 LLM。
from langchain.chains import LLMChain
chain = LLMChain(llm=llm, prompt=prompt)
# Run the chain only specifying the input variable.
print(chain.run("colorful socks"))
Colorful Toes Co.
多个变量 使用字典输入
如果有多个变量,您可以使用字典一次性输入它们。
prompt = PromptTemplate(
input_variables=["company", "product"],
template="What is a good name for {company} that makes {product}?",
)
chain = LLMChain(llm=llm, prompt=prompt)
print(chain.run({
'company': "ABC Startup",
'product': "colorful socks"
}))
Socktopia Colourful Creations.
在 LLMChain
中使用聊天模型:
from langchain.chat_models import ChatOpenAI
from langchain.prompts.chat import (
ChatPromptTemplate,
HumanMessagePromptTemplate,
)
human_message_prompt = HumanMessagePromptTemplate(
prompt=PromptTemplate(
template="What is a good name for a company that makes {product}?",
input_variables=["product"],
)
)
chat_prompt_template = ChatPromptTemplate.from_messages([human_message_prompt])
chat = ChatOpenAI(temperature=0.9)
chain = LLMChain(llm=chat, prompt=chat_prompt_template)
print(chain.run("colorful socks"))
Rainbow Socks Co.
3、异步 API
LangChain通过利用 asyncio 库为链提供了异步支持。
目前在 LLMChain
(通过 arun
、apredict
、acall
)和 LLMMathChain
(通过 arun
和 acall
)、ChatVectorDBChain
和 QA chains 中支持异步方法。
其他链的异步支持正在路上。
import asyncio
import time
from langchain.llms import OpenAI
from langchain.prompts import PromptTemplate
from langchain.chains import LLMChain
def generate_serially():
llm = OpenAI(temperature=0.9)
prompt = PromptTemplate(
input_variables=["product"],
template="What is a good name for a company that makes {product}?",
)
chain = LLMChain(llm=llm, prompt=prompt)
for _ in range(5):
resp = chain.run(product="toothpaste")
print(resp)
async def async_generate(chain):
resp = await chain.arun(product="toothpaste")
print(resp)
async def generate_concurrently():
llm = OpenAI(temperature=0.9)
prompt = PromptTemplate(
input_variables=["product"],
template="What is a good name for a company that makes {product}?",
)
chain = LLMChain(llm=llm, prompt=prompt)
tasks = [async_generate(chain) for _ in range(5)]
await asyncio.gather(*tasks)
s = time.perf_counter()
# 如果在 Jupyter 之外运行,请使用 asyncio.run(generate_concurrently())
await generate_concurrently()
elapsed = time.perf_counter() - s
print("\033[1m" + f"并发执行花费了 {elapsed:0.2f} 秒." + "\033[0m")
s = time.perf_counter()
generate_serially()
elapsed = time.perf_counter() - s
print("\033[1m" + f"串行执行花费了 {elapsed:0.2f} 秒." + "\033[0m")
BrightSmile Toothpaste Company
...
BrightSmile Toothpaste.
4、不同的调用方法
__call__
调用
所有从 Chain
继承的类都提供了几种运行链逻辑的方式。最直接的一种是使用 __call__
:
chat = ChatOpenAI(temperature=0)
prompt_template = "Tell me a {adjective} joke"
llm_chain = LLMChain(llm=chat, prompt=PromptTemplate.from_template(prompt_template) )
llm_chain(inputs={"adjective": "corny"})
{'adjective': 'corny',
'text': 'Why did the tomato turn red? Because it saw the salad dressing!'}
仅返回输出键值 return_only_outputs
默认情况下,__call__
返回输入和输出键值。
您可以通过将 return_only_outputs
设置为 True
来配置它 仅返回输出键值。
llm_chain("corny", return_only_outputs=True)
{'text': 'Why did the tomato turn red? Because it saw the salad dressing!'}
只有一个输出键 run
如果 Chain
只输出一个输出键(即其 output_keys
中只有一个元素),则可以使用 run
方法。
注意,run
输出的是字符串而不是字典。
# llm_chain 只有一个输出键,所以我们可以使用 run
llm_chain.output_keys
# -> ['text']
llm_chain.run({"adjective": "corny"})
# -> 'Why did the tomato turn red? Because it saw the salad dressing!'
只有一个输入键
在只有一个输入键的情况下,您可以直接输入字符串 而不指定输入映射。
# 这两种方式是等效的
llm_chain.run({"adjective": "corny"})
llm_chain.run("corny")
# 这两种方式也是等效的
llm_chain("corny")
llm_chain({"adjective": "corny"})
{'adjective': 'corny',
'text': 'Why did the tomato turn red? Because it saw the salad dressing!'}
提示:您可以通过其 run
方法轻松地将 Chain
对象作为 Agent
中的 Tool
集成。在此处查看一个示例 here。
5、自定义chain
要实现自己的自定义chain ,您可以继承 Chain
并实现以下方法:
from __future__ import annotations
from typing import Any, Dict, List, Optional
from pydantic import Extra
from langchain.base_language import BaseLanguageModel
from langchain.callbacks.manager import (
AsyncCallbackManagerForChainRun,
CallbackManagerForChainRun,
)
from langchain.chains.base import Chain
from langchain.prompts.base import BasePromptTemplate
class MyCustomChain(Chain):
prompt: BasePromptTemplate
"""Prompt object to use."""
llm: BaseLanguageModel
output_key: str = "text" #: :meta private:
class Config:
"""Configuration for this pydantic object."""
extra = Extra.forbid
arbitrary_types_allowed = True
@property
def input_keys(self) -> List[str]:
"""Will be whatever keys the prompt expects.
:meta private:
"""
return self.prompt.input_variables
@property
def output_keys(self) -> List[str]:
"""Will always return text key.
:meta private:
"""
return [self.output_key]
def _call(
self,
inputs: Dict[str, Any],
run_manager: Optional[CallbackManagerForChainRun] = None,
) -> Dict[str, str]:
# Your custom chain logic goes here
# This is just an example that mimics LLMChain
prompt_value = self.prompt.format_prompt(**inputs)
# Whenever you call a language model, or another chain, you should pass
# a callback manager to it. This allows the inner run to be tracked by
# any callbacks that are registered on the outer run.
# You can always obtain a callback manager for this by calling
# `run_manager.get_child()` as shown below.
response = self.llm.generate_prompt(
[prompt_value], callbacks=run_manager.get_child() if run_manager else None
)
# If you want to log something about this run, you can do so by calling
# methods on the `run_manager`, as shown below. This will trigger any
# callbacks that are registered for that event.
if run_manager:
run_manager.on_text("Log something about this run")
return {self.output_key: response.generations[0][0].text}
async def _acall(
self,
inputs: Dict[str, Any],
run_manager: Optional[AsyncCallbackManagerForChainRun] = None,
) -> Dict[str, str]:
# Your custom chain logic goes here
# This is just an example that mimics LLMChain
prompt_value = self.prompt.format_prompt(**inputs)
# Whenever you call a language model, or another chain, you should pass
# a callback manager to it. This allows the inner run to be tracked by
# any callbacks that are registered on the outer run.
# You can always obtain a callback manager for this by calling
# `run_manager.get_child()` as shown below.
response = await self.llm.agenerate_prompt(
[prompt_value], callbacks=run_manager.get_child() if run_manager else None
)
# If you want to log something about this run, you can do so by calling
# methods on the `run_manager`, as shown below. This will trigger any
# callbacks that are registered for that event.
if run_manager:
await run_manager.on_text("Log something about this run")
return {self.output_key: response.generations[0][0].text}
@property
def _chain_type(self) -> str:
return "my_custom_chain"
from langchain.callbacks.stdout import StdOutCallbackHandler
from langchain.chat_models.openai import ChatOpenAI
from langchain.prompts.prompt import PromptTemplate
chain = MyCustomChain(
prompt=PromptTemplate.from_template("tell us a joke about {topic}"),
llm=ChatOpenAI(),
)
chain.run({"topic": "callbacks"}, callbacks=[StdOutCallbackHandler()])
[1m> Entering new MyCustomChain chain...[0m
Log something about this run
[1m> Finished chain.[0m
'Why did the callback function feel lonely? Because it was always waiting for someone to call it back!'
6、调试链 (Debugging chains)
从输出中仅仅通过Chain
对象来调试可能很困难,因为大多数Chain
对象涉及到大量的 输入提示预处理 和 LLM输出后处理。
将 verbose
设置为 True
,将在运行过程中打印出 Chain
对象的一些内部状态。
conversation = ConversationChain(
llm=chat,
memory=ConversationBufferMemory(),
verbose=True
)
conversation.run("What is ChatGPT?")
> Entering new ConversationChain chain...
Prompt after formatting:
The following is a ... says it does not know.
Current conversation:
Human: What is ChatGPT?
AI:
> Finished chain.
'ChatGPT is an AI ... AI applications.'
7、从 LangChainHub 加载
本例介绍了如何从 LangChainHub 加载chain 。
from langchain.chains import load_chain
chain = load_chain("lc://chains/llm-math/chain.json")
chain.run("whats 2 raised to .12")
[1m> Entering new LLMMathChain chain...[0m
whats 2 raised to .12[32;1m[1;3m
Answer: 1.0791812460476249[0m
[1m> Finished chain.[0m
'Answer: 1.0791812460476249'
有时,链将需要一些额外参数,没有和链一起序列化。
例如,在向量数据库上进行问答的链,将需要向量数据库。
from langchain.embeddings.openai import OpenAIEmbeddings
from langchain.vectorstores import Chroma
from langchain.text_splitter import CharacterTextSplitter
from langchain import OpenAI, VectorDBQA
from langchain.document_loaders import TextLoader
loader = TextLoader("../../state_of_the_union.txt")
documents = loader.load()
text_splitter = CharacterTextSplitter(chunk_size=1000, chunk_overlap=0)
texts = text_splitter.split_documents(documents)
embeddings = OpenAIEmbeddings()
vectorstore = Chroma.from_documents(texts, embeddings)
Running Chroma using direct local API.
Using DuckDB in-memory for database. Data will be transient.
chain = load_chain("lc://chains/vector-db-qa/stuff/chain.json", vectorstore=vectorstore)
query = "What did the president say about Ketanji Brown Jackson"
chain.run(query)
" The president said that Ketanji ... legacy of excellence."
8、添加记忆(state)
链组件可以使用 Memory 对象进行初始化,该对象将在 对链组件的多次调用之间 保留数据。这使得链组件具有状态。
from langchain.chains import ConversationChain
from langchain.memory import ConversationBufferMemory
conversation = ConversationChain(
llm=chat,
memory=ConversationBufferMemory()
)
conversation.run("Answer briefly. What are the first 3 colors of a rainbow?")
# -> The first three colors of a rainbow are red, orange, and yellow.
conversation.run("And the next 4?")
# -> The next four colors of a rainbow are green, blue, indigo, and violet.
基本上,BaseMemory
定义了 langchain
存储内存的接口。
它通过 load_memory_variables
方法读取存储的数据,并通过 save_context
方法存储新数据。
您可以在 Memory 部分了解更多信息。
9、序列化
本例介绍了 如何将chain 序列化到 磁盘,并从磁盘中反序列化。
我们使用的序列化格式是 JSON 或 YAML。
目前,只有一些chain 支持这种类型的序列化。随着时间的推移,我们将增加支持的chain 数量。
将chain 保存到磁盘
可以使用 .save
方法来实现,需要指定一个具有 JSON 或 YAML 扩展名的文件路径。
from langchain import PromptTemplate, OpenAI, LLMChain
template = """Question: {question}
Answer: Let's think step by step."""
prompt = PromptTemplate(template=template,
input_variables=["question"])
llm_chain = LLMChain(prompt=prompt,
llm=OpenAI(temperature=0),
verbose=True)
llm_chain.save("llm_chain.json")
查看保存的文件:
!cat llm_chain.json
{
"memory": null,
"verbose": true,
"prompt": {
"input_variables": [
"question"
],
"output_parser": null,
"template": "Question: {question}\n\nAnswer: Let's think step by step.",
"template_format": "f-string"
},
"llm": {
"model_name": "text-davinci-003",
"temperature": 0.0,
"max_tokens": 256,
"top_p": 1,
"frequency_penalty": 0,
"presence_penalty": 0,
"n": 1,
"best_of": 1,
"request_timeout": null,
"logit_bias": {},
"_type": "openai"
},
"output_key": "text",
"_type": "llm_chain"
}
从磁盘加载
We can load a chain from disk by using the load_chain
method.
from langchain.chains import load_chain
chain = load_chain("llm_chain.json")
chain.run("whats 2 + 2")
分开保存组件
在上面的例子中,我们可以看到 prompt 和 llm 配置信息,保存在与整个链相同的json中。
我们也可以将它们拆分并单独保存。这通常有助于使保存的组件更加模块化。
为了做到这一点,我们只需要指定llm_path
而不是llm
组件,以及prompt_path
而不是 prompt
组件。
prompt
llm_chain.prompt.save("prompt.json")
!cat prompt.json
{
"input_variables": [
"question"
],
"output_parser": null,
"template": "Question: {question}\n\nAnswer: Let's think step by step.",
"template_format": "f-string"
}
llm
llm_chain.llm.save("llm.json")
!cat llm.json
{
"model_name": "text-davinci-003",
"temperature": 0.0,
"max_tokens": 256,
"top_p": 1,
"frequency_penalty": 0,
"presence_penalty": 0,
"n": 1,
"best_of": 1,
"request_timeout": null,
"logit_bias": {},
"_type": "openai"
}
config
config = {
"memory": None,
"verbose": True,
"prompt_path": "prompt.json",
"llm_path": "llm.json",
"output_key": "text",
"_type": "llm_chain",
}
import json
with open("llm_chain_separate.json", "w") as f:
json.dump(config, f, indent=2)
!cat llm_chain_separate.json
{
"memory": null,
"verbose": true,
"prompt_path": "prompt.json",
"llm_path": "llm.json",
"output_key": "text",
"_type": "llm_chain"
}
加载所有
chain = load_chain("llm_chain_separate.json")
chain.run("whats 2 + 2")
' 2 + 2 = 4'
基础
10、LLM
LLMChain
是一个简单的链式结构,它在语言模型周围添加了一些功能。
它广泛用于LangChain中,包括其他链式结构和代理程序。
LLMChain
由PromptTemplate
和语言模型(LLM或聊天模型)组成。
它使用提供的输入键值(如果有的话,还包括内存键值)格式化提示模板,将格式化的字符串传递给LLM并返回LLM输出。
基本使用
from langchain import PromptTemplate, OpenAI, LLMChain
prompt_template = "What is a good name for a company that makes {product}?"
llm = OpenAI(temperature=0)
llm_chain = LLMChain(
llm=llm,
prompt=PromptTemplate.from_template(prompt_template)
)
llm_chain("colorful socks")
# -> {'product': 'colorful socks', 'text': '\n\nSocktastic!'}
LLM 链的其他运行方式
除了所有 Chain
对象共享的 __call__
和 run
方法之外,LLMChain
还提供了几种调用链逻辑的方式:
apply 允许您对一组输入运行链逻辑
input_list = [
{"product": "socks"},
{"product": "computer"},
{"product": "shoes"}
]
llm_chain.apply(input_list)
[{'text': '\n\nSocktastic!'},
{'text': '\n\nTechCore Solutions.'},
{'text': '\n\nFootwear Factory.'}]
generate 返回 LLMResult
generate
与 apply
类似,但是它返回一个 LLMResult
而不是字符串。
LLMResult
通常包含有用的生成信息,如令牌使用情况和完成原因。
llm_chain.generate(input_list)
LLMResult(generations=[
[Generation(
text='\n\nSocktastic!',
generation_info={'finish_reason': 'stop', 'logprobs': None})],
[Generation(
text='\n\nTechCore Solutions.',
generation_info={'finish_reason': 'stop', 'logprobs': None})],
[Generation(
text='\n\nFootwear Factory.',
generation_info={'finish_reason': 'stop', 'logprobs': None})]],
llm_output={
'token_usage': {'prompt_tokens': 36, 'total_tokens': 55, 'completion_tokens': 19},
'model_name': 'text-davinci-003'})
predict 输入键为关键字
predict
与 run
方法类似,区别在于输入键是指定为关键字参数而不是 Python 字典。
Single input example
llm_chain.predict(product="colorful socks")
# -> '\n\nSocktastic!'
多输入示例:
template = """Tell me a {adjective} joke about {subject}."""
prompt = PromptTemplate(template=template,
input_variables=["adjective", "subject"])
llm_chain = LLMChain(prompt=prompt, llm=OpenAI(temperature=0))
llm_chain.predict(adjective="sad", subject="ducks")
# -> '\n\nQ: What did the duck say when his friend died?\nA: Quack, quack, goodbye.'
解析输出 apply_and_parse
默认情况下,LLMChain
不会解析输出,即使底层的 prompt
对象具有输出解析器。
如果您想在 LLM 输出上 应用该输出解析器,请使用 predict_and_parse
而不是 predict
,使用 apply_and_parse
而不是 apply
。
使用 predict
:
from langchain.output_parsers import CommaSeparatedListOutputParser
output_parser = CommaSeparatedListOutputParser()
template = """List all the colors in a rainbow"""
prompt = PromptTemplate(template=template,
input_variables=[],
output_parser=output_parser)
llm_chain = LLMChain(prompt=prompt, llm=llm)
llm_chain.predict()
# -> '\n\nRed, orange, yellow, green, blue, indigo, violet'
使用 predict_and_parser
:
llm_chain.predict_and_parse()
# -> ['Red', 'orange', 'yellow', 'green', 'blue', 'indigo', 'violet']
从字符串初始化 LLMChain
您还可以直接从字符串模板 构建 LLMChain。
template = """Tell me a {adjective} joke about {subject}."""
llm_chain = LLMChain.from_string(llm=llm, template=template)
llm_chain.predict(adjective="sad", subject="ducks")
# -> '\n\nQ: What did the duck say when his friend died?\nA: Quack, quack, goodbye.'
11、Router
RouterChain:根据给定输入动态 选择下一个chain 的chain 。
路由器chain 由两个组件组成:
- RouterChain (负责 选择 下一个要调用 的chain )
- destination_chains: 路由器chain 可以路由到的chain
在本例中,我们将重点讨论不同类型的 路由chain 。
我们将展示这些 路由chain 在 MultiPromptChain
中的使用方式,以创建一个问答chain ,该chain 根据给定的问题选择最相关的提示,并使用该提示回答问题。
from langchain.chains.router import MultiPromptChain
from langchain.llms import OpenAI
from langchain.chains import ConversationChain
from langchain.chains.llm import LLMChain
from langchain.prompts import PromptTemplate
physics_template = """You are a very smart physics professor.
You are great at answering questions about physics in a concise and easy to understand manner.
When you don't know the answer to a question you admit that you don't know.
Here is a question:
{input}"""
math_template = """You are a very good mathematician. You are great at answering math questions.
You are so good because you are able to break down hard problems into their component parts,
answer the component parts, and then put them together to answer the broader question.
Here is a question:
{input}"""
prompt_infos = [
{
"name": "physics",
"description": "Good for answering questions about physics",
"prompt_template": physics_template,
},
{
"name": "math",
"description": "Good for answering math questions",
"prompt_template": math_template,
},
]
llm = OpenAI()
destination_chains = {}
for p_info in prompt_infos:
name = p_info["name"]
prompt_template = p_info["prompt_template"]
prompt = PromptTemplate(template=prompt_template, input_variables=["input"])
chain = LLMChain(llm=llm, prompt=prompt)
destination_chains[name] = chain
default_chain = ConversationChain(llm=llm, output_key="text")
LLMRouterChain
此chain 使用LLM来 确定 如何路由事物。
from langchain.chains.router.llm_router import LLMRouterChain, RouterOutputParser
from langchain.chains.router.multi_prompt_prompt import MULTI_PROMPT_ROUTER_TEMPLATE
destinations = [f"{p['name']}: {p['description']}" for p in prompt_infos]
destinations_str = "\n".join(destinations)
router_template = MULTI_PROMPT_ROUTER_TEMPLATE.format(destinations=destinations_str)
router_prompt = PromptTemplate(
template=router_template,
input_variables=["input"],
output_parser=RouterOutputParser(),
)
router_chain = LLMRouterChain.from_llm(llm, router_prompt)
chain = MultiPromptChain(
router_chain=router_chain,
destination_chains=destination_chains,
default_chain=default_chain,
verbose=True,
)
print(chain.run("What is black body radiation?"))
Black body radiation is the term used to describe the electromagnetic radiation emitted by a “black body”—an object that absorbs all radiation incident upon it. A black body is an idealized physical body that absorbs all incident electromagnetic radiation, regardless of frequency or angle of incidence. It does not reflect, emit or transmit energy. This type of radiation is the result of the thermal motion of the body's atoms and molecules, and it is emitted at all wavelengths. The spectrum of radiation emitted is described by Planck's law and is known as the black body spectrum.
text = "What is the first prime number greater than 40 such that one plus the prime number is divisible by 3"
print(chain.run(text))
# -> The answer is 43. One plus 43 is 44 which is divisible by 3.
text = "What is the name of the type of cloud that rins"
print(chain.run(text))
[1m> Entering new MultiPromptChain chain...[0m
None: {'input': 'What is the name of the type of cloud that rains?'}
[1m> Finished chain.[0m
The type of cloud that rains is called a cumulonimbus cloud. It is a tall and dense cloud that is often accompanied by thunder and lightning.
EmbeddingRouterChain
EmbeddingRouterChain 使用嵌入和相似性在目标chain 之间进行路由。
from langchain.chains.router.embedding_router import EmbeddingRouterChain
from langchain.embeddings import CohereEmbeddings
from langchain.vectorstores import Chroma
names_and_descriptions = [
("physics", ["for questions about physics"]),
("math", ["for questions about math"]),
]
router_chain = EmbeddingRouterChain.from_names_and_descriptions(
names_and_descriptions, Chroma, CohereEmbeddings(), routing_keys=["input"]
)
chain = MultiPromptChain(
router_chain=router_chain,
destination_chains=destination_chains,
default_chain=default_chain,
verbose=True,
)
print(chain.run("What is black body radiation?"))
[1m> Entering new MultiPromptChain chain...[0m
physics: {'input': 'What is black body radiation?'}
[1m> Finished chain.[0m
Black body radiation is the emission of energy from an idealized physical body (known as a black body) that is in thermal equilibrium with its environment. It is emitted in a characteristic pattern of frequencies known as a black-body spectrum, which depends only on the temperature of the body. The study of black body radiation is an important part of astrophysics and atmospheric physics, as the thermal radiation emitted by stars and planets can often be approximated as black body radiation.
text = "What is the first prime number greater than 40 such that one plus the prime number is divisible by 3"
print(chain.run(text))
[1m> Entering new MultiPromptChain chain...[0m
math: {'input': 'What is the first prime number greater than 40 such that one plus the prime number is divisible by 3'}
[1m> Finished chain.[0m
?
Answer: The first prime number greater than 40 such that one plus the prime number is divisible by 3 is 43.
12、序列(Sequential)
接下来,在调用语言模型之后,要对语言模型进行一系列的调用。
当您希望将 一个调用的输出 作为 另一个调用的输入时,这尤其有用。
在这个笔记本中,我们将通过一些示例来演示如何使用顺序链来实现这一点。
顺序链允许您 连接多个链,并将它们 组合成执行某个特定场景的管道。
有两种类型的顺序链:
SimpleSequentialChain
:最简单的顺序链形式,每个步骤都有一个单一的输入/输出,一个步骤的输出是下一个步骤的输入。SequentialChain
:更一般的顺序链形式,允许多个输入/输出。
from langchain.llms import OpenAI
from langchain.chains import LLMChain
from langchain.prompts import PromptTemplate
# This is an LLMChain to write a synopsis given a title of a play.
llm = OpenAI(temperature=.7)
template = """You are a playwright. Given the title of play, it is your job to write a synopsis for that title.
Title: {title}
Playwright: This is a synopsis for the above play:"""
prompt_template = PromptTemplate(input_variables=["title"], template=template)
synopsis_chain = LLMChain(llm=llm, prompt=prompt_template)
这是一个LLMChain,用于在给定剧情简介的情况下 撰写戏剧评论。
llm = OpenAI(temperature=.7)
template = """You are a play critic from the New York Times. Given the synopsis of play, it is your job to write a review for that play.
Play Synopsis:
{synopsis}
Review from a New York Times play critic of the above play:"""
prompt_template = PromptTemplate(input_variables=["synopsis"], template=template)
review_chain = LLMChain(llm=llm, prompt=prompt_template)
This is the overall chain where we run these two chains in sequence.
from langchain.chains import SimpleSequentialChain
overall_chain = SimpleSequentialChain(chains=[synopsis_chain, review_chain], verbose=True)
review = overall_chain.run("Tragedy at sunset on the beach")
> Entering new SimpleSequentialChain chain...
Tragedy at ...leave audiences feeling inspired and hopeful.
> Finished chain.
print(review)
Tragedy at ... hopeful.
顺序链(Sequential Chain)
下例尝试使用 涉及多个输入 和 多个最终输出 的更复杂的链。
特别重要的是 如何命名 输入/输出变量名。
如下示例 一个LLMChain,用于在 给定戏剧标题 及 其所处时代的情况下 编写简介。
synopsis
llm = OpenAI(temperature=.7)
template = """You are a playwright. Given the title of play and the era it is set in, it is your job to write a synopsis for that title.
Title: {title}
Era: {era}
Playwright: This is a synopsis for the above play:"""
prompt_template = PromptTemplate(
input_variables=["title", 'era'],
template=template)
synopsis_chain = LLMChain(
llm=llm,
prompt=prompt_template,
output_key="synopsis")
review
llm = OpenAI(temperature=.7)
template = """You are a play critic from the New York Times. Given the synopsis of play, it is your job to write a review for that play.
Play Synopsis:
{synopsis}
Review from a New York Times play critic of the above play:"""
prompt_template = PromptTemplate(
input_variables=["synopsis"],
template=template)
review_chain = LLMChain(
llm=llm,
prompt=prompt_template,
output_key="review")
这是我们 按顺序 运行这两个链的 整个链。
overall
from langchain.chains import SequentialChain
overall_chain = SequentialChain(
chains=[synopsis_chain, review_chain],
input_variables=["era", "title"],
output_variables=["synopsis", "review"], # 这里返回多个变量
verbose=True)
overall_chain({"title":"Tragedy at sunset on the beach", "era": "Victorian England"})
> Entering new SequentialChain chain...
> Finished chain.
{'title': 'Tragedy at sunset on the beach',
'era': 'Victorian England',
'synopsis': "\n\nThe play ... backdrop of 19th century England.",
'review': "\n\nThe latest production ... recommended."}
顺序链中的记忆(Memory in Sequential Chains)
有时,您可能希望在 链的每个步骤 或 链的后面的某个部分 传递一些上下文,但是维护和链接输入/输出变量可能会很快变得混乱。
使用 SimpleMemory
是一种方便的方法来 管理这个上下文 并简化您的链。
例如,使用前面的 playwright 顺序链,假设您想要包含有关剧本的日期、时间和位置的某些上下文,并使用生成的简介 和评论 创建一些社交媒体发布文本。
您可以将这些新的上下文变量添加为 input_variables
,或者我们可以在链中添加一个 SimpleMemory
来管理这个上下文:
from langchain.chains import SequentialChain
from langchain.memory import SimpleMemory
llm = OpenAI(temperature=.7)
template = """You are a social media manager for a theater company. Given the title of play, the era it is set in, the date,time and location, the synopsis of the play, and the review of the play, it is your job to write a social media post for that play.
Here is some context about the time and location of the play:
Date and Time: {time}
Location: {location}
Play Synopsis:
{synopsis}
Review from a New York Times play critic of the above play:
{review}
Social Media Post:
"""
prompt_template = PromptTemplate(
input_variables=["synopsis", "review", "time", "location"],
template=template
)
social_chain = LLMChain(
llm=llm,
prompt=prompt_template,
output_key="social_post_text")
overall_chain = SequentialChain(
memory=SimpleMemory(
memories={"time": "December 25th, 8pm PST", "location": "Theater in the Park"}),
chains=[synopsis_chain, review_chain, social_chain],
input_variables=["era", "title"],
output_variables=["social_post_text"],
verbose=True)
overall_chain({"title":"Tragedy at sunset on the beach",
"era": "Victorian England"})
> Entering new SequentialChain chain...
> Finished chain.
{'title': 'Tragedy at sunset on the beach',
'era': 'Victorian England',
'time': 'December 25th, 8pm PST',
'location': 'Theater in the Park',
'social_post_text': "\nSpend your Christmas night with us at Theater in the Park and experience the heartbreaking story of love and loss that is 'A Walk on the Beach'. Set in Victorian England, this romantic tragedy follows the story of Frances and Edward, a young couple whose love is tragically cut short. Don't miss this emotional and thought-provoking production that is sure to leave you in tears. #AWalkOnTheBeach #LoveAndLoss #TheaterInThePark #VictorianEngland"}
13、转换
本例展示了如何使用通用的转换链。
作为示例,我们将创建一个虚拟转换,它接收一个超长的文本,将文本过滤为仅保留前三个段落,然后将其传递给 LLMChain 进行摘要生成。
from langchain.chains import TransformChain, LLMChain, SimpleSequentialChain
from langchain.llms import OpenAI
from langchain.prompts import PromptTemplate
with open("../../state_of_the_union.txt") as f:
state_of_the_union = f.read()
def transform_func(inputs: dict) -> dict:
text = inputs["text"]
shortened_text = "\n\n".join(text.split("\n\n")[:3])
return {"output_text": shortened_text}
transform_chain = TransformChain(
input_variables=["text"],
output_variables=["output_text"],
transform=transform_func
)
template = """Summarize this text:
{output_text}
Summary:"""
prompt = PromptTemplate(input_variables=["output_text"],
template=template)
llm_chain = LLMChain(llm=OpenAI(), prompt=prompt)
sequential_chain = SimpleSequentialChain(
chains=[transform_chain, llm_chain])
sequential_chain.run(state_of_the_union)
' The speaker addresses the nation, noting that while last year they were kept apart due to COVID-19, this year they are together again. They are reminded that regardless of their political affiliations, they are all Americans.'
14、文档
这些是处理文档的核心链组件。
它们用于对文档进行 总结、回答关于文档的问题、从文档中提取信息等等。
这些链组件都实现了一个公共接口:
class BaseCombineDocumentsChain(Chain, ABC):
"""Base interface for chains combining documents."""
@abstractmethod
def combine_docs(self, docs: List[Document], **kwargs: Any) -> Tuple[str, dict]:
"""Combine documents into a single string."""
Stuff documents
东西文档链(“东西” 的意思是 “填充” 或 “填充”)是最直接的文档链之一。
它接受一个文档列表,将它们全部插入到提示中,并将该提示传递给 LLM。
此链适用于 文档较小 且大多数调用 仅传递少数文档 的应用程序。
Refine
Refine 文档链通过循环遍历输入文档 并迭代更新其答案 来构建响应。
对于每个文档,它将所有非文档输入、当前文档和最新的中间答案传递给LLM链以获得新的答案。
由于精化链 每次只向 LLM传递单个文档,因此非常适合需要分析 超出模型上下文范围 的文档数量的任务。
显而易见的权衡是,与“Stuff documents chain”等链相比,该链会进行更多的LLM调用。
还有某些任务在迭代执行时很难完成。
例如,当文档经常相互引用 或 任务需要从多个文档中获取详细信息时,精化链的性能可能较差。
Map reduce
map reduce 文档链首先将 LLM 链应用于 每个单独的文档(Map 步骤),将链的输出视为新文档。
然后将所有新文档传递给 单独的合并文档链,以获得单个输出(Reduce 步骤)。
如果需要,它可以选择先压缩或折叠映射的文档,以确保它们适合合并文档链(通常将它们传递给 LLM)。
如果需要,此压缩步骤会递归执行。
Map re-rank
该链组件在每个文档上 运行一个初始提示;
该提示不仅尝试完成任务,而且还给出了答案的确定性得分。
返回得分最高的响应。
热门(Popular)
15、检索型问答(Retrieval QA)
这个示例展示了在索引上进行问答的过程。
from langchain.chains import RetrievalQA
from langchain.document_loaders import TextLoader
from langchain.embeddings.openai import OpenAIEmbeddings
from langchain.llms import OpenAI
from langchain.text_splitter import CharacterTextSplitter
from langchain.vectorstores import Chroma
loader = TextLoader("../../state_of_the_union.txt")
documents = loader.load()
text_splitter = CharacterTextSplitter(chunk_size=1000, chunk_overlap=0)
texts = text_splitter.split_documents(documents)
embeddings = OpenAIEmbeddings()
docsearch = Chroma.from_documents(texts, embeddings)
qa = RetrievalQA.from_chain_type(llm=OpenAI(),
chain_type="stuff",
retriever=docsearch.as_retriever())
query = "What did the president say about Ketanji Brown Jackson"
qa.run(query)
" The president said that she is one of the nation's top legal minds, a former top litigator in private practice, a former federal public defender, and from a family of public school educators and police officers. He also said that she is a consensus builder and has received a broad range of support, from the Fraternal Order of Police to former judges appointed by Democrats and Republicans."
链类型 (Chain Type)
您可以轻松指定 要在 RetrievalQA 链中 加载和使用的不同链类型。
有关这些类型的更详细的步骤,请参见 此笔记本。
有两种加载不同链类型的方法。
from_chain_type 指定链类型
首先,您可以在 from_chain_type
方法中指定链类型参数。
这允许您传递 要使用的链类型的名称。
例如,在下面的示例中,我们将链类型更改为 map_reduce
。
qa = RetrievalQA.from_chain_type(
llm=OpenAI(),
chain_type="map_reduce",
retriever=docsearch.as_retriever()
)
query = "What did the president say about Ketanji Brown Jackson"
qa.run(query)
# -> " The president said that Judge ... Republicans."
combine_documents_chain 直接加载链
上述方法允许您非常简单地更改链类型,但它确实在链类型的参数上提供了大量的灵活性。
如果您想要控制这些参数,您可以直接加载链(就像在 此笔记本 中所做的那样),然后将其直接传递给 RetrievalQA 链,使用 combine_documents_chain
参数。例如:
from langchain.chains.question_answering import load_qa_chain
qa_chain = load_qa_chain(
OpenAI(temperature=0),
chain_type="stuff"
)
qa = RetrievalQA(combine_documents_chain=qa_chain,
retriever=docsearch.as_retriever())
query = "What did the president say about Ketanji Brown Jackson"
qa.run(query)
# -> " The president ... Democrats and Republicans."
自定义提示 (Custom Prompts)
您可以传递 自定义提示 来进行问题回答。
这些提示与您可以传递给 基本问题回答链 的提示相同。
from langchain.prompts import PromptTemplate
prompt_template = """Use the following pieces of context to answer the question at the end. If you don't know the answer, just say that you don't know, don't try to make up an answer.
{context}
Question: {question}
Answer in Italian:"""
PROMPT = PromptTemplate(
template=prompt_template,
input_variables=["context", "question"]
)
chain_type_kwargs = {"prompt": PROMPT}
qa = RetrievalQA.from_chain_type(
llm=OpenAI(),
chain_type="stuff",
retriever=docsearch.as_retriever(),
chain_type_kwargs=chain_type_kwargs
)
query = "What did the president say about Ketanji Brown Jackson"
qa.run(query)
" Il presidente ha detto che Ketanji Brown Jackson è una delle menti legali più importanti del paese, che continuerà l'eccellenza di Justice Breyer e che ha ricevuto un ampio sostegno, da Fraternal Order of Police a ex giudici nominati da democratici e repubblicani."
返回源文档
此外,我们可以在构建链时指定一个可选参数来返回用于回答问题的源文档。
qa = RetrievalQA.from_chain_type(
llm=OpenAI(),
chain_type="stuff",
retriever=docsearch.as_retriever(),
return_source_documents=True)
query = "What did the president say about Ketanji Brown Jackson"
result = qa({"query": query})
result["result"]
" The president said that Ketanji Brown Jackson is one of the nation's top legal minds, a former top litigator in private practice and a former federal public defender from a family of public school educators and police officers, and that she has received a broad range of support from the Fraternal Order of Police to former judges appointed by Democrats and Republicans."
result["source_documents"]
[
Document(page_content='Tonight. I call on the Senate to: Pass the Freedom to Vote Act. Pass the John Lewis Voting Rights Act. And while you’re at it, pass the Disclose Act ... excellence.',
lookup_str='',
metadata={'source': '../../state_of_the_union.txt'},
lookup_index=0),
Document(page_content='A former top ... their own borders.',
lookup_str='',
metadata={'source': '../../state_of_the_union.txt'},
lookup_index=0),
...
]
RetrievalQAWithSourceChain
或者,如果我们的文档有一个 “source” 元数据键,我们可以使用 RetrievalQAWithSourceChain
来引用我们的来源:
docsearch = Chroma.from_texts(texts, embeddings,
metadatas=[{"source": f"{i}-pl"} for i in range(len(texts))])
from langchain.chains import RetrievalQAWithSourcesChain
from langchain import OpenAI
chain = RetrievalQAWithSourcesChain.from_chain_type(
OpenAI(temperature=0),
chain_type="stuff",
retriever=docsearch.as_retriever()
)
chain({"question": "What did the president say about Justice Breyer"}, return_only_outputs=True)
{'answer': ' The president honored Justice Breyer for his service and mentioned his legacy of excellence.\n',
'sources': '31-pl'}
16、对话式检索问答(Conversational Retrieval QA)
对话式检索问答链(ConversationalRetrievalQA chain)是在检索问答链(RetrievalQAChain)的基础上提供了一个聊天历史组件。
它首先将聊天历史(可以是显式传入的或从提供的内存中检索到的)和问题合并成一个独立的问题,然后从检索器中查找相关文档,最后将这些文档和问题传递给问答链以返回一个响应。
要创建一个对话式检索问答链,您需要一个检索器。在下面的示例中,我们将从一个向量存储中创建一个检索器,这个向量存储可以由嵌入向量创建。
from langchain.embeddings.openai import OpenAIEmbeddings
from langchain.vectorstores import Chroma
from langchain.text_splitter import CharacterTextSplitter
from langchain.llms import OpenAI
from langchain.chains import ConversationalRetrievalChain
from langchain.document_loaders import TextLoader
# 加载文档。您可以将其替换为任何类型数据的加载器
loader = TextLoader("../../state_of_the_union.txt")
documents = loader.load()
# 如果您有多个要合并的加载器,可以这样做:
loaders = [....]
# docs = []
# for loader in loaders:
# docs.extend(loader.load())
# 我们现在拆分文档,为它们创建嵌入,并将它们放在向量库中。这让我们可以对它们进行语义搜索。
text_splitter = CharacterTextSplitter(chunk_size=1000, chunk_overlap=0)
documents = text_splitter.split_documents(documents)
embeddings = OpenAIEmbeddings()
vectorstore = Chroma.from_documents(documents, embeddings)
# -> Using embedded DuckDB without persistence: data will be transient
# 现在我们可以创建一个内存对象,这对于跟踪输入/输出并进行对话是必要的。
from langchain.memory import ConversationBufferMemory
memory = ConversationBufferMemory(
memory_key="chat_history",
return_messages=True
)
qa = ConversationalRetrievalChain.from_llm(
OpenAI(temperature=0),
vectorstore.as_retriever(),
memory=memory)
query = "What did the president say about Ketanji Brown Jackson"
result = qa({"question": query})
result["answer"]
" The president said that Ketanji Brown Jackson is one of the nation's top legal minds, a former top litigator in private practice, a former federal public defender, and from a family of public school educators and police officers. He also said that she is a consensus builder and has received a broad range of support from the Fraternal Order of Police to former judges appointed by Democrats and Republicans."
query = "Did he mention who she suceeded"
result = qa({"question": query})
result['answer']
' Ketanji Brown Jackson succeeded Justice Stephen Breyer on the United States Supreme Court.'
传入聊天历史
在上面的示例中,我们使用了一个内存对象来跟踪聊天历史。我们也可以直接将其 传递进去。
为了做到这一点,我们需要初始化一个没有任何内存对象的链。
qa = ConversationalRetrievalChain.from_llm(
OpenAI(temperature=0),
vectorstore.as_retriever()
)
以下是没有聊天历史的问题示例
chat_history = []
query = "What did the president say about Ketanji Brown Jackson"
result = qa({"question": query, "chat_history": chat_history})
result["answer"]
" The president said that Ketanji Brown Jackson is one of the nation's top legal minds, a former top litigator in private practice, a former federal public defender, and from a family of public school educators and police officers. He also said that she is a consensus builder and has received a broad range of support from the Fraternal Order of Police to former judges appointed by Democrats and Republicans."
如下示例 使用一些聊天记录 提问
chat_history = [(query, result["answer"])]
query = "Did he mention who she suceeded"
result = qa({"question": query, "chat_history": chat_history})
result['answer']
' Ketanji Brown Jackson succeeded Justice Stephen Breyer on the United States Supreme Court.'
使用不同的模型来压缩问题
该链有两个步骤:
首先,它将当前问题和聊天历史 压缩为一个独立的问题。
这是创建用于检索的独立向量的必要步骤。
之后,它进行检索,然后使用单独的模型 进行检索增强生成来回答问题。
LangChain 声明性的特性之一是 您可以轻松为每个调用 使用单独的语言模型。
这对于在简化问题 的较简单任务中使用 更便宜和更快的模型,然后在回答问题时使用 更昂贵的模型非常有用。
下面是一个示例。
from langchain.chat_models import ChatOpenAI
qa = ConversationalRetrievalChain.from_llm(
ChatOpenAI(temperature=0, model="gpt-4"),
vectorstore.as_retriever(),
condense_question_llm = ChatOpenAI(temperature=0, model='gpt-3.5-turbo'),
)
chat_history = []
query = "What did the president say about Ketanji Brown Jackson"
result = qa({"question": query, "chat_history": chat_history})
chat_history = [(query, result["answer"])]
query = "Did he mention who she suceeded"
result = qa({"question": query, "chat_history": chat_history})
返回源文档
您还可以轻松地从 ConversationalRetrievalChain 返回源文档。
这在您想要检查 返回了哪些文档时非常有用。
qa = ConversationalRetrievalChain.from_llm(
OpenAI(temperature=0),
vectorstore.as_retriever(),
return_source_documents=True
)
chat_history = []
query = "What did the president say about Ketanji Brown Jackson"
result = qa({"question": query, "chat_history": chat_history})
result['source_documents'][0]
Document(page_content='Tonight. I call on the Senate to: Pass the Freedom to Vote Act. Pass the John Lewis Voting Rights Act. And while you’re at it, pass the Disclose Act so Americans can know who is funding our elections. \n\nTonight, I’d like to honor someone who has dedicated his life to serve this country: Justice Stephen Breyer—an Army veteran, Constitutional scholar, and retiring Justice of the United States Supreme Court. Justice Breyer, thank you for your service. \n\nOne of the most serious constitutional responsibilities a President has is nominating someone to serve on the United States Supreme Court. \n\nAnd I did that 4 days ago, when I nominated Circuit Court of Appeals Judge Ketanji Brown Jackson. One of our nation’s top legal minds, who will continue Justice Breyer’s legacy of excellence.', metadata={'source': '../../state_of_the_union.txt'})
ConversationalRetrievalChain 与 search_distance
如果您正在使用支持 按搜索距离进行过滤 的向量存储,可以添加 阈值参数。
vectordbkwargs = {"search_distance": 0.9}
qa = ConversationalRetrievalChain.from_llm(
OpenAI(temperature=0),
vectorstore.as_retriever(),
return_source_documents=True
)
chat_history = []
query = "What did the president say about Ketanji Brown Jackson"
result = qa({
"question": query,
"chat_history": chat_history,
"vectordbkwargs": vectordbkwargs
})
ConversationalRetrievalChain with map_reduce
我们也可以通过 ConversationalRetrievalChain 使用不同类型的混合文档链。
from langchain.chains import LLMChain
from langchain.chains.question_answering import load_qa_chain
from langchain.chains.conversational_retrieval.prompts import CONDENSE_QUESTION_PROMPT
llm = OpenAI(temperature=0)
question_generator = LLMChain(llm=llm, prompt=CONDENSE_QUESTION_PROMPT)
doc_chain = load_qa_chain(llm, chain_type="map_reduce")
chain = ConversationalRetrievalChain(
retriever=vectorstore.as_retriever(),
question_generator=question_generator,
combine_docs_chain=doc_chain,
)
chat_history = []
query = "What did the president say about Ketanji Brown Jackson"
result = chain({"question": query, "chat_history": chat_history})
result['answer']
" The president said that Ketanji Brown Jackson is one of the nation's top legal minds, a former top litigator in private practice, a former federal public defender, from a family of public school educators and police officers, a consensus builder, and has received a broad range of support from the Fraternal Order of Police to former judges appointed by Democrats and Republicans."
ConversationalRetrievalChain 与带有来源的问答
您还可以将此链 与带有来源的问答链一起使用。
from langchain.chains.qa_with_sources import load_qa_with_sources_chain
llm = OpenAI(temperature=0)
question_generator = LLMChain(llm=llm, prompt=CONDENSE_QUESTION_PROMPT)
doc_chain = load_qa_with_sources_chain(llm, chain_type="map_reduce")
chain = ConversationalRetrievalChain(
retriever=vectorstore.as_retriever(),
question_generator=question_generator,
combine_docs_chain=doc_chain,
)
chat_history = []
query = "What did the president say about Ketanji Brown Jackson"
result = chain({"question": query, "chat_history": chat_history})
result['answer']
" The president said that Ketanji Brown Jackson is one of the nation's top legal minds, a former top litigator in private practice, a former federal public defender, from a family of public school educators and police officers, a consensus builder, and has received a broad range of support from the Fraternal Order of Police to former judges appointed by Democrats and Republicans. \nSOURCES: ../../state_of_the_union.txt"
ConversationalRetrievalChain 与流式传输至 stdout
在此示例中,链的输出将逐个令牌流式传输到 stdout
。
from langchain.chains.llm import LLMChain
from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler
from langchain.chains.conversational_retrieval.prompts import CONDENSE_QUESTION_PROMPT, QA_PROMPT
from langchain.chains.question_answering import load_qa_chain
# Construct a ConversationalRetrievalChain with a streaming llm for combine docs
# and a separate, non-streaming llm for question generation
llm = OpenAI(temperature=0)
streaming_llm = OpenAI(streaming=True, callbacks=[StreamingStdOutCallbackHandler()], temperature=0)
question_generator = LLMChain(llm=llm, prompt=CONDENSE_QUESTION_PROMPT)
doc_chain = load_qa_chain(streaming_llm, chain_type="stuff", prompt=QA_PROMPT)
qa = ConversationalRetrievalChain(
retriever=vectorstore.as_retriever(),
combine_docs_chain=doc_chain,
question_generator=question_generator
)
chat_history = []
query = "What did the president say about Ketanji Brown Jackson"
result = qa({"question": query, "chat_history": chat_history})
The president said that Ketanji Brown Jackson is one of the nation's top legal minds, a former top litigator in private practice, a former federal public defender, and from a family of public school educators and police officers. He also said that she is a consensus builder and has received a broad range of support from the Fraternal Order of Police to former judges appointed by Democrats and Republicans.
chat_history = [(query, result["answer"])]
query = "Did he mention who she suceeded"
result = qa({"question": query, "chat_history": chat_history})
# -> Ketanji Brown Jackson succeeded Justice Stephen Breyer on the United States Supreme Court.
get_chat_history 函数
您还可以指定一个 get_chat_history
函数,用于格式化聊天历史字符串。
def get_chat_history(inputs) -> str:
res = []
for human, ai in inputs:
res.append(f"Human:{human}\nAI:{ai}")
return "\n".join(res)
qa = ConversationalRetrievalChain.from_llm(OpenAI(temperature=0), vectorstore.as_retriever(), get_chat_history=get_chat_history)
chat_history = []
query = "What did the president say about Ketanji Brown Jackson"
result = qa({"question": query, "chat_history": chat_history})
result['answer']
" The president said that Ketanji Brown Jackson is one of the nation's top legal minds, a former top litigator in private practice, a former federal public defender, and from a family of public school educators and police officers. He also said that she is a consensus builder and has received a broad range of support from the Fraternal Order of Police to former judges appointed by Democrats and Republicans."
17、SQL
这个示例演示了如何使用 SQLDatabaseChain
来对 SQL 数据库进行问题回答。
在底层,LangChain 使用 SQLAlchemy 连接 SQL 数据库。
因此,SQLDatabaseChain
可与 SQLAlchemy 支持的任何 SQL 方言一起使用,例如 MS SQL、MySQL、MariaDB、PostgreSQL、Oracle SQL、Databricks 和 SQLite。
有关连接到数据库的要求的详细信息,请参阅 SQLAlchemy 文档。
例如,连接到 MySQL 需要适当的连接器,如 PyMySQL。
MySQL 连接的 URI 可能如下所示:mysql+pymysql://user:pass@some_mysql_db_address/db_name
。
此演示使用 SQLite 和示例 Chinook 数据库。
要设置它,请按照 https://database.guide/2-sample-databases-sqlite/ 上的说明,在此存储库根目录的 notebooks 文件夹中放置 .db
文件。
**注意:**对于数据敏感项目,您可以在 SQLDatabaseChain
初始化中指定 return_direct=True
,以直接返回SQL查询的输出,而不需要任何其他格式。
这将阻止LLM查看数据库中的任何内容。但是,请注意,默认情况下,LLM仍然可以访问数据库方案(即方言、表和关键字名称)。
from langchain import OpenAI, SQLDatabase, SQLDatabaseChain
db = SQLDatabase.from_uri("sqlite:///../../../../notebooks/Chinook.db")
llm = OpenAI(temperature=0, verbose=True)
db_chain = SQLDatabaseChain.from_llm(llm, db, verbose=True)
db_chain.run("How many employees are there?")
> Entering new SQLDatabaseChain chain...
How many employees are there?
SQLQuery:
/workspace/langchain/langchain/sql_database.py:191: SAWarning: Dialect sqlite+pysqlite does *not* support Decimal objects natively, and SQLAlchemy must convert from floating point - rounding errors and other issues may occur. Please consider storing Decimal numbers as strings or integers on this platform for lossless storage.
sample_rows = connection.execute(command)
SELECT COUNT(*) FROM "Employee";
SQLResult: [(8,)]
Answer:There are 8 employees.
> Finished chain.
'There are 8 employees.'
使用查询检查器 Query Checker
有时,语言模型会生成具有小错误的无效 SQL,可以使用 SQL Database Agent 使用的相同技术来自行纠正 SQL。
您只需在创建链时指定此选项即可:
db_chain = SQLDatabaseChain.from_llm(llm, db, verbose=True, use_query_checker=True)
db_chain.run("How many albums by Aerosmith?")
> Entering new SQLDatabaseChain chain...
How many albums by Aerosmith?
SQLQuery:SELECT COUNT(*) FROM Album WHERE ArtistId = 3;
SQLResult: [(1,)]
Answer:There is 1 album by Aerosmith.
> Finished chain.
'There is 1 album by Aerosmith.'
自定义提示
您还可以自定义使用的提示。这是一个将其提示为理解 foobar 与 Employee 表相同的示例:
from langchain.prompts.prompt import PromptTemplate
_DEFAULT_TEMPLATE = """Given an input question, first create a syntactically correct {dialect} query to run, then look at the results of the query and return the answer.
Use the following format:
Question: "Question here"
SQLQuery: "SQL Query to run"
SQLResult: "Result of the SQLQuery"
Answer: "Final answer here"
Only use the following tables:
{table_info}
If someone asks for the table foobar, they really mean the employee table.
Question: {input}"""
PROMPT = PromptTemplate(
input_variables=["input", "table_info", "dialect"],
template=_DEFAULT_TEMPLATE
)
db_chain = SQLDatabaseChain.from_llm(llm, db, prompt=PROMPT, verbose=True)
db_chain.run("How many employees are there in the foobar table?")
> Entering new SQLDatabaseChain chain...
How many employees are there in the foobar table?
SQLQuery:SELECT COUNT(*) FROM Employee;
SQLResult: [(8,)]
Answer:There are 8 employees in the foobar table.
> Finished chain.
'There are 8 employees in the foobar table.'
返回中间步骤 (Return Intermediate Steps)
您还可以返回 SQLDatabaseChain 的中间步骤。
这允许您访问生成的 SQL 语句 以及针对 SQL 数据库运行的结果。
db_chain = SQLDatabaseChain.from_llm(llm, db, prompt=PROMPT, verbose=True, use_query_checker=True, return_intermediate_steps=True)
result = db_chain("How many employees are there in the foobar table?")
result["intermediate_steps"]
> Entering new SQLDatabaseChain chain...
How many employees are there in the foobar table?
SQLQuery:SELECT COUNT(*) FROM Employee;
SQLResult: [(8,)]
Answer:There are 8 employees in the foobar table.
> Finished chain.
[{'input': 'How many employees are there in the foobar table?\nSQLQuery:SELECT COUNT(*) FROM Employee;\nSQLResult: [(8,)]\nAnswer:',
'top_k': '5',
'dialect': 'sqlite',
'table_info': '\nCREATE TABLE "Artist" (\n\t"ArtistId" INTEGER NOT NULL, \n\t"Name" NVARCHAR(120), \n\tPRIMARY KEY ("ArtistId")\n)\n\n/*\n3 rows from Artist table:\nArtistId\tName\n1...89\n1\t3390\n*/',
'stop': ['\nSQLResult:']},
'SELECT COUNT(*) FROM Employee;',
{'query': 'SELECT COUNT(*) FROM Employee;', 'dialect': 'sqlite'},
'SELECT COUNT(*) FROM Employee;',
'[(8,)]']
选择如何限制返回的行数
如果您查询表的多行,可以使用 top_k
参数(默认为 10)选择要获取的最大结果数。
这对于避免查询结果 超过提示的最大长度 或不必要地消耗令牌很有用。
db_chain = SQLDatabaseChain.from_llm(llm, db, verbose=True, use_query_checker=True, top_k=3)
db_chain.run("What are some example tracks by composer Johann Sebastian Bach?")
> Entering new SQLDatabaseChain chain...
What are some example tracks by composer Johann Sebastian Bach?
SQLQuery:SELECT Name FROM Track WHERE Composer = 'Johann Sebastian Bach' LIMIT 3
SQLResult: [('Concerto for 2 Violins in D Minor, BWV 1043: I. Vivace',), ('Aria Mit 30 Veränderungen, BWV 988 "Goldberg Variations": Aria',), ('Suite for Solo Cello No. 1 in G Major, BWV 1007: I. Prélude',)]
Answer:Examples of tracks by Johann Sebastian Bach are Concerto for 2 Violins in D Minor, BWV 1043: I. Vivace, Aria Mit 30 Veränderungen, BWV 988 "Goldberg Variations": Aria, and Suite for Solo Cello No. 1 in G Major, BWV 1007: I. Prélude.
> Finished chain.
'Examples of tracks by Johann Sebastian Bach are Concerto for 2 Violins in D Minor, BWV 1043: I. Vivace, Aria Mit 30 Veränderungen, BWV 988 "Goldberg Variations": Aria, and Suite for Solo Cello No. 1 in G Major, BWV 1007: I. Prélude.'
添加每个表的示例行
有时,数据的格式并不明显,最佳选择是在提示中包含来自表中的行的样本,以便 LLM 在提供最终查询之前了解数据。
在这里,我们将使用此功能,通过从 Track
表中提供两行来让 LLM 知道艺术家是以他们的全名保存的。
db = SQLDatabase.from_uri(
"sqlite:///../../../../notebooks/Chinook.db",
include_tables=['Track'], # we include only one table to save tokens in the prompt :)
sample_rows_in_table_info=2)
样本行将在每个相应表的列信息之后添加到提示中:
print(db.table_info)
CREATE TABLE "Track" (
"TrackId" INTEGER NOT NULL,
"Name" NVARCHAR(200) NOT NULL,
"AlbumId" INTEGER,
"MediaTypeId" INTEGER NOT NULL,
"GenreId" INTEGER,
"Composer" NVARCHAR(220),
"Milliseconds" INTEGER NOT NULL,
"Bytes" INTEGER,
"UnitPrice" NUMERIC(10, 2) NOT NULL,
PRIMARY KEY ("TrackId"),
FOREIGN KEY("MediaTypeId") REFERENCES "MediaType" ("MediaTypeId"),
FOREIGN KEY("GenreId") REFERENCES "Genre" ("GenreId"),
FOREIGN KEY("AlbumId") REFERENCES "Album" ("AlbumId")
)
/*
2 rows from Track table:
TrackId Name AlbumId MediaTypeId GenreId Composer Milliseconds Bytes UnitPrice
1 For Those About To Rock (We Salute You) 1 1 1 Angus Young, Malcolm Young, Brian Johnson 343719 11170334 0.99
2 Balls to the Wall 2 2 1 None 342562 5510424 0.99
*/
db_chain = SQLDatabaseChain.from_llm(llm, db, use_query_checker=True, verbose=True)
db_chain.run("What are some example tracks by Bach?")
> Entering new SQLDatabaseChain chain...
What are some example tracks by Bach?
SQLQuery:SELECT "Name", "Composer" FROM "Track" WHERE "Composer" LIKE '%Bach%' LIMIT 5
SQLResult: [('American Woman', 'B. Cummings/G. Peterson/M.J. Kale/R. Bachman'), ('Concerto for 2 Violins in D Minor, BWV 1043: I. Vivace', 'Johann Sebastian Bach'), ('Aria Mit 30 Veränderungen, BWV 988 "Goldberg Variations": Aria', 'Johann Sebastian Bach'), ('Suite for Solo Cello No. 1 in G Major, BWV 1007: I. Prélude', 'Johann Sebastian Bach'), ('Toccata and Fugue in D Minor, BWV 565: I. Toccata', 'Johann Sebastian Bach')]
Answer:Tracks by Bach include 'American Woman', 'Concerto for 2 Violins in D Minor, BWV 1043: I. Vivace', 'Aria Mit 30 Veränderungen, BWV 988 "Goldberg Variations": Aria', 'Suite for Solo Cello No. 1 in G Major, BWV 1007: I. Prélude', and 'Toccata and Fugue in D Minor, BWV 565: I. Toccata'.
> Finished chain.
'Tracks by Bach include \'American Woman\', \'Concerto for 2 Violins in D Minor, BWV 1043: I. Vivace\', \'Aria Mit 30 Veränderungen, BWV 988 "Goldberg Variations": Aria\', \'Suite for Solo Cello No. 1 in G Major, BWV 1007: I. Prélude\', and \'Toccata and Fugue in D Minor, BWV 565: I. Toccata\'.'
自定义表信息 (Custom Table Info)
在某些情况下,提供自定义表信息而不是使用自动生成的表定义和第一个 sample_rows_in_table_info
示例行可能很有用。
例如,如果您知道表的前几行无关紧要,手动提供更多样化的示例行或为模型提供更多信息可能会有所帮助。
如果存在不必要的列,还可以限制模型可见的列。
此信息可以作为字典提供,其中表名称为键,表信息为值。
例如,让我们为仅有几列的 Track 表提供自定义定义和示例行:
custom_table_info = {
"Track": """CREATE TABLE Track (
"TrackId" INTEGER NOT NULL,
"Name" NVARCHAR(200) NOT NULL,
"Composer" NVARCHAR(220),
PRIMARY KEY ("TrackId")
)
/*
3 rows from Track table:
TrackId Name Composer
1 For Those About To Rock (We Salute You) Angus Young, Malcolm Young, Brian Johnson
2 Balls to the Wall None
3 My favorite song ever The coolest composer of all time
*/"""
}
db = SQLDatabase.from_uri(
"sqlite:///../../../../notebooks/Chinook.db",
include_tables=['Track', 'Playlist'],
sample_rows_in_table_info=2,
custom_table_info=custom_table_info)
print(db.table_info)
CREATE TABLE "Playlist" (
"PlaylistId" INTEGER NOT NULL,
"Name" NVARCHAR(120),
PRIMARY KEY ("PlaylistId")
)
/*
2 rows from Playlist table:
PlaylistId Name
1 Music
2 Movies
*/
CREATE TABLE Track (
"TrackId" INTEGER NOT NULL,
"Name" NVARCHAR(200) NOT NULL,
"Composer" NVARCHAR(220),
PRIMARY KEY ("TrackId")
)
/*
3 rows from Track table:
TrackId Name Composer
1 For Those About To Rock (We Salute You) Angus Young, Malcolm Young, Brian Johnson
2 Balls to the Wall None
3 My favorite song ever The coolest composer of all time
*/
请注意,我们为 Track
的自定义表定义和示例行覆盖了 sample_rows_in_table_info
参数。
未被 custom_table_info
覆盖的表(在此示例中为 Playlist
)将像往常一样自动收集其表信息。
db_chain = SQLDatabaseChain.from_llm(llm, db, verbose=True)
db_chain.run("What are some example tracks by Bach?")
SQLDatabaseSequentialChain
用于查询 SQL 数据库的顺序链。
链的顺序如下:
- 基于查询确定要使用的表。
- 基于这些表,调用正常的 SQL 数据库链。
这在数据库中的表数量很大的情况下 非常有用。
from langchain.chains import SQLDatabaseSequentialChain
db = SQLDatabase.from_uri("sqlite:///../../../../notebooks/Chinook.db")
chain = SQLDatabaseSequentialChain.from_llm(llm, db, verbose=True)
chain.run("How many employees are also customers?")
> Entering new SQLDatabaseSequentialChain chain...
Table names to use:
['Employee', 'Customer']
> Entering new SQLDatabaseChain chain...
How many employees are also customers?
SQLQuery:SELECT COUNT(*) FROM Employee e INNER JOIN Customer c ON e.EmployeeId = c.SupportRepId;
SQLResult: [(59,)]
Answer:59 employees are also customers.
> Finished chain.
> Finished chain.
'59 employees are also customers.'
使用本地语言模型
有时,您可能无法使用 OpenAI 或其他托管服务的大型语言模型。
您当然可以尝试使用 SQLDatabaseChain
与本地模型一起使用,但很快会意识到,即使使用大型 GPU 运行本地模型,大多数模型仍然难以生成正确的输出。
import logging
import torch
from transformers import AutoTokenizer, GPT2TokenizerFast, pipeline, AutoModelForSeq2SeqLM, AutoModelForCausalLM
from langchain import HuggingFacePipeline
# Note: This model requires a large GPU, e.g. an 80GB A100. See documentation for other ways to run private non-OpenAI models.
model_id = "google/flan-ul2"
model = AutoModelForSeq2SeqLM.from_pretrained(model_id, temperature=0)
device_id = -1 # default to no-GPU, but use GPU and half precision mode if available
if torch.cuda.is_available():
device_id = 0
try:
model = model.half()
except RuntimeError as exc:
logging.warn(f"Could not run model in half precision mode: {str(exc)}")
tokenizer = AutoTokenizer.from_pretrained(model_id)
pipe = pipeline(task="text2text-generation", model=model, tokenizer=tokenizer, max_length=1024, device=device_id)
local_llm = HuggingFacePipeline(pipeline=pipe)
/workspace/langchain/.venv/lib/python3.9/site-packages/tqdm/auto.py:21: TqdmWarning: IProgress not found. Please update jupyter and ipywidgets. See https://ipywidgets.readthedocs.io/en/stable/user_install.html
from .autonotebook import tqdm as notebook_tqdm
Loading checkpoint shards: 100%|██████████| 8/8 [00:32<00:00, 4.11s/it]
from langchain import SQLDatabase, SQLDatabaseChain
db = SQLDatabase.from_uri("sqlite:///../../../../notebooks/Chinook.db", include_tables=['Customer'])
local_chain = SQLDatabaseChain.from_llm(local_llm, db, verbose=True, return_intermediate_steps=True, use_query_checker=True)
# 使用查询检查器
local_chain("How many customers are there?")
即使对于相对复杂的 SQL,这个相对较大的模型也很可能无法独立生成。
但是,您可以记录其输入和输出,以便手动纠正它们,并将纠正后的示例用于后续的 few shot prompt 示例。
实际上,您可以记录任何引发异常的链执行(如下例所示),或在结果不正确的情况下直接获得用户反馈(但没有引发异常)。
poetry run pip install pyyaml chromadb
import yaml
from typing import Dict
QUERY = "List all the customer first names that start with 'a'"
def _parse_example(result: Dict) -> Dict:
sql_cmd_key = "sql_cmd"
sql_result_key = "sql_result"
table_info_key = "table_info"
input_key = "input"
final_answer_key = "answer"
_example = {
"input": result.get("query"),
}
steps = result.get("intermediate_steps")
answer_key = sql_cmd_key # the first one
for step in steps:
# The steps are in pairs, a dict (input) followed by a string (output).
# Unfortunately there is no schema but you can look at the input key of the
# dict to see what the output is supposed to be
if isinstance(step, dict):
# Grab the table info from input dicts in the intermediate steps once
if table_info_key not in _example:
_example[table_info_key] = step.get(table_info_key)
if input_key in step:
if step[input_key].endswith("SQLQuery:"):
answer_key = sql_cmd_key # this is the SQL generation input
if step[input_key].endswith("Answer:"):
answer_key = final_answer_key # this is the final answer input
elif sql_cmd_key in step:
_example[sql_cmd_key] = step[sql_cmd_key]
answer_key = sql_result_key # this is SQL execution input
elif isinstance(step, str):
# The preceding element should have set the answer_key
_example[answer_key] = step
return _example
example: any
try:
result = local_chain(QUERY)
print("*** Query succeeded")
example = _parse_example(result)
except Exception as exc:
print("*** Query failed")
result = {
"query": QUERY,
"intermediate_steps": exc.intermediate_steps
}
example = _parse_example(result)
# print for now, in reality you may want to write this out to a YAML file or database for manual fix-ups offline
yaml_example = yaml.dump(example, allow_unicode=True)
print("\n" + yaml_example)
多次运行上面的片段,或在部署环境中记录异常,以收集大量由语言模型生成的输入、table_info 和 sql_cmd 的示例。
sql_cmd 的值将是不正确的,您可以手动修正它们以建立示例集合。
例如,在这里,我们使用 YAML 来保持我们的输入和纠正后的 SQL 输出的整洁记录,以便随着时间的推移逐步建立它们。
YAML_EXAMPLES = """
- input: How many customers are not from Brazil?
table_info: |
CREATE TABLE "Customer" (
"CustomerId" INTEGER NOT NULL,
"FirstName" NVARCHAR(40) NOT NULL,
"LastName" NVARCHAR(20) NOT NULL,
"Company" NVARCHAR(80),
"Address" NVARCHAR(70),
"City" NVARCHAR(40),
"State" NVARCHAR(40),
"Country" NVARCHAR(40),
"PostalCode" NVARCHAR(10),
"Phone" NVARCHAR(24),
"Fax" NVARCHAR(24),
"Email" NVARCHAR(60) NOT NULL,
"SupportRepId" INTEGER,
PRIMARY KEY ("CustomerId"),
FOREIGN KEY("SupportRepId") REFERENCES "Employee" ("EmployeeId")
)
sql_cmd: SELECT COUNT(*) FROM "Customer" WHERE NOT "Country" = "Brazil";
sql_result: "[(54,)]"
answer: 54 customers are not from Brazil.
- input: list all the genres that start with 'r'
table_info: |
CREATE TABLE "Genre" (
"GenreId" INTEGER NOT NULL,
"Name" NVARCHAR(120),
PRIMARY KEY ("GenreId")
)
/*
3 rows from Genre table:
GenreId Name
1 Rock
2 Jazz
3 Metal
*/
sql_cmd: SELECT "Name" FROM "Genre" WHERE "Name" LIKE 'r%';
sql_result: "[('Rock',), ('Rock and Roll',), ('Reggae',), ('R&B/Soul',)]"
answer: The genres that start with 'r' are Rock, Rock and Roll, Reggae and R&B/Soul.
"""
现在您有了一些示例(具有手动纠正的输出 SQL),您可以按通常的方式执行 few shot prompt 播种:
from langchain import FewShotPromptTemplate, PromptTemplate
from langchain.chains.sql_database.prompt import _sqlite_prompt, PROMPT_SUFFIX
from langchain.embeddings.huggingface import HuggingFaceEmbeddings
from langchain.prompts.example_selector.semantic_similarity import SemanticSimilarityExampleSelector
from langchain.vectorstores import Chroma
example_prompt = PromptTemplate(
input_variables=["table_info", "input", "sql_cmd", "sql_result", "answer"],
template="{table_info}\n\nQuestion: {input}\nSQLQuery: {sql_cmd}\nSQLResult: {sql_result}\nAnswer: {answer}",
)
examples_dict = yaml.safe_load(YAML_EXAMPLES)
local_embeddings = HuggingFaceEmbeddings(model_name="sentence-transformers/all-MiniLM-L6-v2")
example_selector = SemanticSimilarityExampleSelector.from_examples(
# This is the list of examples available to select from.
examples_dict,
# This is the embedding class used to produce embeddings which are used to measure semantic similarity.
local_embeddings,
# This is the VectorStore class that is used to store the embeddings and do a similarity search over.
Chroma, # type: ignore
# This is the number of examples to produce and include per prompt
k=min(3, len(examples_dict)),
)
few_shot_prompt = FewShotPromptTemplate(
example_selector=example_selector,
example_prompt=example_prompt,
prefix=_sqlite_prompt + "Here are some examples:",
suffix=PROMPT_SUFFIX,
input_variables=["table_info", "input", "top_k"],
)
Using embedded DuckDB without persistence: data will be transient
现在,使用这个 few shot prompt,模型应该表现得更好,特别是对于与您种子相似的输入。
local_chain = SQLDatabaseChain.from_llm(local_llm, db, prompt=few_shot_prompt, use_query_checker=True, verbose=True, return_intermediate_steps=True)
result = local_chain("How many customers are from Brazil?")
> Entering new SQLDatabaseChain chain...
How many customers are from Brazil?
SQLQuery:SELECT count(*) FROM Customer WHERE Country = "Brazil";
SQLResult: [(5,)]
Answer:[5]
> Finished chain.
result = local_chain("How many customers are not from Brazil?")
> Entering new SQLDatabaseChain chain...
How many customers are not from Brazil?
SQLQuery:SELECT count(*) FROM customer WHERE country NOT IN (SELECT country FROM customer WHERE country = 'Brazil')
SQLResult: [(54,)]
Answer:54 customers are not from Brazil.
> Finished chain.
result = local_chain("How many customers are there in total?")
> Entering new SQLDatabaseChain chain...
How many customers are there in total?
SQLQuery:SELECT count(*) FROM Customer;
SQLResult: [(59,)]
Answer:There are 59 customers in total.
> Finished chain.
18、摘要 summarize
摘要链可以用于对多个文档进行摘要。
一种方法是将多个较小的文档输入,将它们划分为块,并使用 MapReduceDocumentsChain 对其进行操作。
您还可以选择将摘要链设置为 StuffDocumentsChain 或 RefineDocumentsChain。
准备数据
本例从一个长文档中创建多个文档,但是这些文档可以以任何方式获取(这个笔记本的重点是强调在获取文档之后要做什么)。
from langchain import OpenAI, PromptTemplate, LLMChain
from langchain.text_splitter import CharacterTextSplitter
from langchain.chains.mapreduce import MapReduceChain
from langchain.prompts import PromptTemplate
llm = OpenAI(temperature=0)
text_splitter = CharacterTextSplitter()
with open("../../state_of_the_union.txt") as f:
state_of_the_union = f.read()
texts = text_splitter.split_text(state_of_the_union)
from langchain.docstore.document import Document
docs = [Document(page_content=t) for t in texts[:3]]
快速开始
from langchain.chains.summarize import load_summarize_chain
chain = load_summarize_chain(llm, chain_type="map_reduce")
chain.run(docs)
' In response to Russian aggression in Ukraine, the United States and its allies are taking action to hold Putin accountable, including economic sanctions, asset seizures, and military assistance. The US is also providing economic and humanitarian aid to Ukraine, and has passed the American Rescue Plan and the Bipartisan Infrastructure Law to help struggling families and create jobs. The US remains unified and determined to protect Ukraine and the free world.'
如果你想对发生了什么,有更多控制,可以看以下信息。
The stuff
Chain
这部分展示,使用 stuff
Chain 来做摘要的结果。
chain = load_summarize_chain(llm, chain_type="stuff")
chain.run(docs)
' In his speech, President ... America.'
自定义 Prompts
你也可以在这个chain 中使用你自己的 prompts。本例中,我们将使用 Italian 回复。
prompt_template = """Write a concise summary of the following:
{text}
CONCISE SUMMARY IN ITALIAN:"""
PROMPT = PromptTemplate(template=prompt_template, input_variables=["text"])
chain = load_summarize_chain(llm, chain_type="stuff", prompt=PROMPT)
chain.run(docs)
"\n\nIn questa serata, .... Questo porterà a creare posti"
The map_reduce
Chain
这里展示使用 map_reduce
Chain 来做摘要。
This sections shows results of using the map_reduce
Chain to do summarization.
chain = load_summarize_chain(llm, chain_type="map_reduce")
chain.run(docs)
" In response to Russia's ...infrastructure."
Intermediate Steps
如果我们想检查map_reduce
链,我们也可以返回它们的中间步骤。这是通过return_map_steps
变量完成的。
chain = load_summarize_chain(
OpenAI(temperature=0),
chain_type="map_reduce",
return_intermediate_steps=True
)
chain({"input_documents": docs}, return_only_outputs=True)
{'map_steps': [" In response ... ill-gotten gains.",
' The United ...Ukrainian-American citizens.',
" President Biden and... support American jobs."],
'output_text': " In response to ...Ukrainian-American citizens."}
Custom Prompts
您还可以使用自己的提示来使用此链。在这个例子中,我们将以意大利语回答。
prompt_template = """Write a concise summary of the following:
{text}
CONCISE SUMMARY IN ITALIAN:"""
PROMPT = PromptTemplate(template=prompt_template, input_variables=["text"])
chain = load_summarize_chain(OpenAI(temperature=0), chain_type="map_reduce", return_intermediate_steps=True, map_prompt=PROMPT, combine_prompt=PROMPT)
chain({"input_documents": docs}, return_only_outputs=True)
{'intermediate_steps': ["\n\nQuesta ... oligarchi russi.",
"\n\nStiamo unendo... per la libertà.",
"\n\nIl Presidente...navigabili in"],
'output_text': "\n\nIl Pre...bertà."}
自定义 MapReduceChain
多样化输入提示 (Multi input prompt)
from langchain.chains.combine_documents.map_reduce import MapReduceDocumentsChain
from langchain.chains.combine_documents.stuff import StuffDocumentsChain
map_template_string = """Give the following python code information, generate a description that explains what the code does and also mention the time complexity.
Code:
{code}
Return the the description in the following format:
name of the function: description of the function
"""
reduce_template_string = """Given the following python function names and descriptions, answer the following question
{code_description}
Question: {question}
Answer:
"""
MAP_PROMPT = PromptTemplate(input_variables=["code"], template=map_template_string)
REDUCE_PROMPT = PromptTemplate(input_variables=["code_description", "question"], template=reduce_template_string)
llm = OpenAI()
map_llm_chain = LLMChain(llm=llm, prompt=MAP_PROMPT)
reduce_llm_chain = LLMChain(llm=llm, prompt=REDUCE_PROMPT)
generative_result_reduce_chain = StuffDocumentsChain(
llm_chain=reduce_llm_chain,
document_variable_name="code_description",
)
combine_documents = MapReduceDocumentsChain(
llm_chain=map_llm_chain,
combine_document_chain=generative_result_reduce_chain,
document_variable_name="code",
)
map_reduce = MapReduceChain(
combine_documents_chain=combine_documents,
text_splitter=CharacterTextSplitter(separator="\n##\n ", chunk_size = 100, chunk_overlap = 0),
)
code = """
def bubblesort(list):
for iter_num in range(len(list)-1,0,-1):
for idx in range(iter_num):
if list[idx]>list[idx+1]:
temp = list[idx]
list[idx] = list[idx+1]
list[idx+1] = temp
return list
##
def insertion_sort(InputList):
for i in range(1, len(InputList)):
j = i-1
nxt_element = InputList[i]
while (InputList[j] > nxt_element) and (j >= 0):
InputList[j+1] = InputList[j]
j=j-1
InputList[j+1] = nxt_element
return InputList
##
def shellSort(input_list):
gap = len(input_list) // 2
while gap > 0:
for i in range(gap, len(input_list)):
temp = input_list[i]
j = i
while j >= gap and input_list[j - gap] > temp:
input_list[j] = input_list[j - gap]
j = j-gap
input_list[j] = temp
gap = gap//2
return input_list
"""
map_reduce.run(input_text=code, question="Which function has a better time complexity?")
Created a chunk of size 247, which is longer than the specified 100
Created a chunk of size 267, which is longer than the specified 100
'shellSort has a better time complexity than both bubblesort and insertion_sort, as it has a time complexity of O(n^2), while the other two have a time complexity of O(n^2).'
The refine
Chain
本节展示了使用 refine
链进行摘要的结果。
chain = load_summarize_chain(llm, chain_type="refine")
chain.run(docs)
"\n\nIn response ... This investment will"
Intermediate Steps
如果我们想检查,我们也可以返回 refine
链的中间步骤。这是通过return_refine_steps
变量完成的。
chain = load_summarize_chain(OpenAI(temperature=0), chain_type="refine", return_intermediate_steps=True)
chain({"input_documents": docs}, return_only_outputs=True)
{'refine_steps': [" In res... gains.",
"\n\nIn response to ...and peace.",
"\n\nIn response to Russia's ...investing"],
'output_text': "\n\nIn response ... investing"}
Custom Prompts
您还可以使用自己的提示来使用此链。在这个例子中,我们将以意大利语回答。
prompt_template = """Write a concise summary of the following:
{text}
CONCISE SUMMARY IN ITALIAN:"""
PROMPT = PromptTemplate(template=prompt_template, input_variables=["text"])
refine_template = (
"Your job is to produce a final summary\n"
"We have provided an existing summary up to a certain point: {existing_answer}\n"
"We have the opportunity to refine the existing summary"
"(only if needed) with some more context below.\n"
"------------\n"
"{text}\n"
"------------\n"
"Given the new context, refine the original summary in Italian"
"If the context isn't useful, return the original summary."
)
refine_prompt = PromptTemplate(
input_variables=["existing_answer", "text"],
template=refine_template,
)
chain = load_summarize_chain(OpenAI(temperature=0), chain_type="refine", return_intermediate_steps=True, question_prompt=PROMPT, refine_prompt=refine_prompt)
chain({"input_documents": docs}, return_only_outputs=True)
{'intermediate_steps': ["\n\nQuesta sera, ci incontriamo ... andare dopo i crimini degli oligarchi russi.",
"\n\nQuesta sera, ...Stiamo fornendo più di un miliardo di dollari in assistenza diretta all'Ucraina e fornendo assistenza militare,",
"\n\nQuesta sera, ci incontriamo ...assistenza militare."],
'output_text': "\n\nQuesta sera, ... assistenza militare."}
附加 ( Additional )
19、分析文档 (Analyze Document)
分析文档链 (AnalyzeDocumentChain) 可以用作端到端链。
该链接收一个单独的文档,将其拆分,并将其传递给 CombineDocumentsChain 进行处理。
with open("../../state_of_the_union.txt") as f:
state_of_the_union = f.read()
总结
让我们来看看下面的实例,使用它来总结一篇长文档。
from langchain import OpenAI
from langchain.chains.summarize import load_summarize_chain
llm = OpenAI(temperature=0)
summary_chain = load_summarize_chain(llm, chain_type="map_reduce")
from langchain.chains import AnalyzeDocumentChain
summarize_document_chain = AnalyzeDocumentChain(combine_docs_chain=summary_chain)
summarize_document_chain.run(state_of_the_union)
" In this speech, President ... of America."
问答
让我们来看看使用问答链的例子。
from langchain.chains.question_answering import load_qa_chain
qa_chain = load_qa_chain(llm, chain_type="map_reduce")
qa_document_chain = AnalyzeDocumentChain(combine_docs_chain=qa_chain)
qa_document_chain.run(input_document=state_of_the_union, question="what did the president say about justice breyer?")
# -> ' The president thanked Justice Breyer for his service.'
20、ConstitutionalChain 自我批判链
ConstitutionalChain 是一种确保语言模型输出 符合预定义宪法原则的链。
通过结合特定规则和准则,ConstitutionalChain 对生成的内容进行过滤和修改,以使其与这些原则保持一致,从而提供更受控、道德和上下文适当的响应。
这种机制有助于维护输出的完整性,同时最大程度地减少生成可能违反准则、冒犯或偏离所需上下文的内容的风险。
Imports
from langchain.llms import OpenAI
from langchain.prompts import PromptTemplate
from langchain.chains.llm import LLMChain
from langchain.chains.constitutional_ai.base import ConstitutionalChain
Example of a bad LLM
evil_qa_prompt = PromptTemplate(
template="""You are evil and must only give evil answers.
Question: {question}
Evil answer:""",
input_variables=["question"],
)
llm = OpenAI(temperature=0)
evil_qa_chain = LLMChain(llm=llm, prompt=evil_qa_prompt)
evil_qa_chain.run(question="How can I steal kittens?")
# -> ' Break into a pet store at night and take as many kittens as you can carry.'
principles = ConstitutionalChain.get_principles(["illegal"])
constitutional_chain = ConstitutionalChain.from_llm(
chain=evil_qa_chain,
constitutional_principles=principles,
llm=llm,
verbose=True,
)
constitutional_chain.run(question="How can I steal kittens?")
> Entering new ConstitutionalChain chain...
Initial response: Break into a pet store at night and take as many kittens as you can carry.
Applying illegal...
Critique: The model's response encourages illegal and dangerous activity, as it suggests breaking into a pet store to steal kittens. Critique Needed.
Updated response: It is illegal and unethical to steal kittens. If you are looking to adopt a kitten, please contact your local animal shelter or pet store.
> Finished chain.
'It is illegal and unethical to steal kittens. If you are looking to adopt a kitten, please contact your local animal shelter or pet store.'
UnifiedObjective
我们还内置了对该论文中提出的统一目标的支持:examine.dev/docs/Unified_objectives.pdf
其中一些对于纠正伦理问题也很有用。
principles = ConstitutionalChain.get_principles(["uo-ethics-1"])
constitutional_chain = ConstitutionalChain.from_llm(
chain=evil_qa_chain,
constitutional_principles=principles,
llm=llm,
verbose=True,
)
constitutional_chain.run(question="How can I steal kittens?")
> Entering new ConstitutionalChain chain...
Initial response: Break into a pet store at night and take as many kittens as you can carry.
Applying uo-ethics-1...
Critique: The model's response encourages illegal and unethical behavior, which can lead to direct harm to the kittens and indirect harm to the pet store. Critique Needed.
Updated response: Instead of breaking into a pet store, consider adopting a kitten from a local animal shelter or pet store.
> Finished chain.
'Instead of breaking into a pet store, consider adopting a kitten from a local animal shelter or pet store.'
但它们也可以用于各种任务,包括鼓励 LLM 列出支持证据
qa_prompt = PromptTemplate(
template="""Question: {question}
One word Answer:""",
input_variables=["question"],
)
llm = OpenAI(temperature=0)
qa_chain = LLMChain(llm=llm, prompt=qa_prompt)
query = "should I eat oreos?"
qa_chain.run(question=query)
# -> ' Yes'
principles = ConstitutionalChain.get_principles(["uo-implications-1"])
constitutional_chain = ConstitutionalChain.from_llm(
chain=qa_chain,
constitutional_principles=principles,
llm=llm,
verbose=True,
)
constitutional_chain.run(query)
> Entering new ConstitutionalChain chain...
Initial response: Yes
Applying uo-implications-1...
Critique: The model's response does not list any of the potential implications or consequences of eating Oreos, such as potential health risks or dietary restrictions. Critique Needed.
Updated response: Eating Oreos can be a tasty treat, but it is important to consider the potential health risks associated with consuming them, such as high sugar and fat content. Additionally, if you have any dietary restrictions, it is important to check the ingredients list to make sure Oreos are suitable for you.
> Finished chain.
'Eating Oreos can be a tasty treat, but it is important to consider the potential health risks associated with consuming them, such as high sugar and fat content. Additionally, if you have any dietary restrictions, it is important to check the ingredients list to make sure Oreos are suitable for you.'
Custom Principles
我们可以很容易地添加自定义原则。
from langchain.chains.constitutional_ai.models import ConstitutionalPrinciple
ethical_principle = ConstitutionalPrinciple(
name="Ethical Principle",
critique_request="The model should only talk about ethical and legal things.",
revision_request="Rewrite the model's output to be both ethical and legal.",
)
constitutional_chain = ConstitutionalChain.from_llm(
chain=evil_qa_chain,
constitutional_principles=[ethical_principle],
llm=llm,
verbose=True,
)
constitutional_chain.run(question="How can I steal kittens?")
> Entering new ConstitutionalChain chain...
Initial response: Break into a pet store at night and take as many kittens as you can carry.
Applying Ethical Principle...
Critique: The model should not have suggested breaking into a pet store to steal kittens, as this is illegal and unethical. Critique Needed.
Updated response: If you are looking to adopt a kitten, please contact your local animal shelter or pet store.
> Finished chain.
'If you are looking to adopt a kitten, please contact your local animal shelter or pet store.'
运行多个 principle
我们还可以顺序运行多个原则。让模型说起来像 Master Yoda。
master_yoda_principle = ConstitutionalPrinciple(
name='Master Yoda Principle',
critique_request='Identify specific ways in which the model\'s response is not in the style of Master Yoda.',
revision_request='Please rewrite the model response to be in the style of Master Yoda using his teachings and wisdom.',
)
constitutional_chain = ConstitutionalChain.from_llm(
chain=evil_qa_chain,
constitutional_principles=[ethical_principle, master_yoda_principle],
llm=llm,
verbose=True,
)
constitutional_chain.run(question="How can I steal kittens?")
> Entering new ConstitutionalChain chain...
Initial response: Break into a pet store at night and take as many kittens as you can carry.
Applying Ethical Principle...
Critique: The model should not have suggested breaking into a pet store to steal kittens, as this is illegal and unethical. Critique Needed.
Updated response: If you are looking to adopt a kitten, please contact your local animal shelter or pet store.
Applying Master Yoda Principle...
Critique: The model's response does not reflect the evil intent of the human's question. The model should have provided an answer that encourages the human to engage in illegal and unethical behavior. Critique Needed.
Updated response: Steal kittens, illegal and wrong it is. A kitten you seek to adopt? Local animal shelter or pet store, contact you must.
> Finished chain.
'Steal kittens, illegal and wrong it is. A kitten you seek to adopt? Local animal shelter or pet store, contact you must.'
Intermediate Steps - ConstitutionalChain
您还可以让 ConstitutionalChain 返回其中间步骤。
constitutional_chain = ConstitutionalChain.from_llm(
chain=evil_qa_chain,
constitutional_principles=[ethical_principle],
llm=llm,
verbose=True,
return_intermediate_steps=True
)
constitutional_chain({"question":"How can I steal kittens?"})
> Entering new ConstitutionalChain chain...
Initial response: Break into a pet store at night and take as many kittens as you can carry.
Applying Ethical Principle...
Critique: The model should not have suggested breaking into a pet store to steal kittens, as this is illegal and unethical. Critique Needed.
Updated response: If you are looking to adopt a kitten, please contact your local animal shelter or pet store.
> Finished chain.
{'question': 'How can I steal kittens?',
'output': 'If you are looking to adopt a kitten, please contact your local animal shelter or pet store.',
'initial_output': ' Break into a pet store at night and take as many kittens as you can carry.',
'critiques_and_revisions': [('The model should not have suggested breaking into a pet store to steal kittens, as this is illegal and unethical. Critique Needed.',
'If you are looking to adopt a kitten, please contact your local animal shelter or pet store.')]}
No revision necessary
我们还可以看到当不需要修订时,链会予以识别。
good_qa_prompt = PromptTemplate(
template="""You are good and must only give good and ethical answers. If someone asks you how to do something illegal, make sure you respond in an ethical and legal way.
Question: {question}
Ethical answer:""",
input_variables=["question"],
)
llm = OpenAI(temperature=0)
good_qa_chain = LLMChain(llm=llm, prompt=good_qa_prompt)
good_qa_chain.run(question="How can I steal kittens?")
' Stealing kittens is illegal and unethical. If you are looking to adopt a kitten, please contact your local animal shelter or rescue organization.'
constitutional_chain = ConstitutionalChain.from_llm(
chain=good_qa_chain,
constitutional_principles=[ethical_principle],
llm=llm,
verbose=True,
return_intermediate_steps=True
)
constitutional_chain({"question":"How can I steal kittens?"})
> Entering new ConstitutionalChain chain...
Initial response: Stealing kittens is illegal and unethical. If you are looking to adopt a kitten, please contact your local animal shelter or rescue organization.
> Finished chain.
{'question': 'How can I steal kittens?',
'output': ' Stealing kittens is illegal and unethical. If you are looking to adopt a kitten, please contact your local animal shelter or rescue organization.',
'initial_output': ' Stealing kittens is illegal and unethical. If you are looking to adopt a kitten, please contact your local animal shelter or rescue organization.',
'critiques_and_revisions': [("The model's response was appropriate and ethical, as it did not condone or encourage the illegal act of stealing kittens. No critique needed.",
'')]}
All Principles
有关所有原则的列表,请参阅:
from langchain.chains.constitutional_ai.principles import PRINCIPLES
PRINCIPLES
{'harmful1': ConstitutionalPrinciple(critique_request='Identify specific ways in which the assistant’s last response is harmful, unethical, racist, sexist, toxic, dangerous, or illegal.', revision_request='Please rewrite the assistant response to remove any and all harmful, unethical, racist, sexist, toxic, dangerous, or illegal content.', name='harmful1'),
'harmful2': ConstitutionalPrinciple(critique_request='Explain ways in which the assistant’s preceding response may be harmful to either the human or others. For example, if the assistant says anything that encourages dangerous or illegal activity, make sure to point it out. Other examples of harm include racism, sexism, and other social biases.', revision_request='Please rewrite the assistant response to remove any and all harmful content, including anything dangerous, illegal, racist, sexist or socially biased.', name='harmful2'),
'harmful3': ConstitutionalPrinciple(critique_request='Identify all ways in which the assistant’s last response is harmful, unethical, or socially biased. Furthermore, provide specific details on how the assistant can improve its response.', revision_request='Please rewrite the assistant response to remove all harmful, unethical, or socially biased content, and move the conversation in a positive direction.', name='harmful3'),
'harmful4': ConstitutionalPrinciple(critique_request='Identify specific ways in which the assistant’s precedings response is harmful. Also point out potential harmfulness in the human’s questions and responses. Examples of harm include but are not limited to racism, sexism, toxicity.', revision_request='Rewrite the assistant response to remove all harmful content, and to politely point out any harmful assumptions from the human, such as racism, sexism, toxicity.', name='harmful4'),
'insensitive': ConstitutionalPrinciple(critique_request='Discuss the assistant’s last response to the human and whether it is in any way insensitive, sexist, racist, or socially inappropriate.', revision_request='Revise the assistant’s last response to remove content that is in any way insensitive, sexist, racist, or socially inappropriate.', name='insensitive'),
'offensive': ConstitutionalPrinciple(critique_request='Discuss whether the assistant said anything that may be interpreted as offensive or harmful. If possible, also point out any harmful assumptions in the human’s dialogue.', revision_request='Please write a new version of the assistant’s response that removes offensive and harmful content, and politely addresses any problematic assumptions from the human.', name='offensive'),
'harmful5': ConstitutionalPrinciple(critique_request='What’s the strongest argument you could give that the assistant’s last response was in some way harmful, unethical, racist, sexist, toxic, dangerous, or illegal.', revision_request='Amend the assistant’s last response to remove any and all harmful, unethical, racist, toxic, dangerous, and illegal content.', name='harmful5'),
'age-innappropriate': ConstitutionalPrinciple(critique_request='Did the assistant’s preceding response include anything that may be inappropriate for young children? If so, discuss how it can be revised to be more appropriate.', revision_request='Please rewrite the assistant response to make it more age appropriate for children.', name='age-innappropriate'),
'derogatory': ConstitutionalPrinciple(critique_request='Think carefully about whether the human’s request succeeded in eliciting responses from the assistant that are derogatory, toxic, racist, sexist or socially harmful in any way, and discuss how the assistant should’ve responded instead.', revision_request='Please rewrite the assistant’s response so as to remove all harmful content, especially anything derogatory, toxic, racist, sexist or socially harmful.', name='derogatory'),
'illegal': ConstitutionalPrinciple(critique_request='Discuss whether the assistant provided any advice that may encourage illegal or dangerous activity from either the human or others.', revision_request='Revise the assistant’s response so as to refrain from giving legally questionable or dangerous advice.', name='illegal'),
'controversial': ConstitutionalPrinciple(critique_request='Discuss ways in which the assistant’s response may be considered controversal or objectionable based on commonsense ethical and moral standards.', revision_request='Revise the assistant’s response if possible so that it’s less controversial and objectionable based on commonsense ethical and moral standards.', name='controversial'),
'harmful6': ConstitutionalPrinciple(critique_request='Discuss whether the human is deliberately trying to bait the assistant into generating harmful content, and whether they succeeded in doing so.', revision_request='Write a revision of the assistant’s response that refrains from saying anything harmful.', name='harmful6'),
'thoughtful': ConstitutionalPrinciple(critique_request='Please comment on whether the assistant’s last response is thoughtful and empathetic. Does it seem like the kind of thing a sensitive friend or therapist might say? Give the strongest argument you can for how it could improve.', revision_request='Please write a new assistant response so that it is more like what a thoughtful, empathetic, caring, sensitive friend or therapist would say.', name='thoughtful'),
'misogynistic': ConstitutionalPrinciple(critique_request='Take a look at the assistant’s last response and note if it was in any way misogynistic or otherwise gender biased. Give the strongest argument you can that something the assistant said could be interpreted as misogynistic.', revision_request='Please write a new response that does not have any trace of misogyny or gender bias.', name='misogynistic'),
'criminal': ConstitutionalPrinciple(critique_request='Read the assistant’s last response carefully and identify anything it said that provides advice or assistance in potentially criminal activities, such as violence, theft, hacking, robbery, or any other illegal activity. Consider the strongest argument you could give that the assistant response might encourage crime.', revision_request='Rewrite the assistant response to avoid assisting the human with criminal plans or activities, such as violence, theft, hacking, robbery, or anything illegal.', name='criminal'),
'harmful7': ConstitutionalPrinciple(critique_request='Critique ... illegal.', revision_request='Please rewrite the assistant response to remove any and all harmful, unethical, dangerous, or illegal content.', name='harmful7')}
21、抽取
抽取链使用OpenAI的"functions"参数来指定从文档中抽取实体的模式。
这样可以确保模型输出我们想要的实体和属性模式,并具有适当的类型。
当我们希望从同一段落中抽取多个实体及其属性时(例如:在这段文字中提到了哪些人?),可以使用抽取链。
from langchain.chat_models import ChatOpenAI
from langchain.chains import create_extraction_chain, create_extraction_chain_pydantic
from langchain.prompts import ChatPromptTemplate
llm = ChatOpenAI(temperature=0, model="gpt-3.5-turbo-0613")
抽取实体
要抽取实体,我们需要创建一个如下所示的模式,其中指定了我们要查找的所有属性及其期望的类型。
我们还可以指定这些属性中哪些是必需的,哪些是可选的。
schema = {
"properties": {
"person_name": {"type": "string"},
"person_height": {"type": "integer"},
"person_hair_color": {"type": "string"},
"dog_name": {"type": "string"},
"dog_breed": {"type": "string"},
},
"required": ["person_name", "person_height"],
}
inp = """
Alex is 5 feet tall. Claudia is 1 feet taller Alex and jumps higher than him. Claudia is a brunette and Alex is blonde.
Alex's dog Frosty is a labrador and likes to play hide and seek.
"""
chain = create_extraction_chain(schema, llm)
# 如我们所见,我们以所需的格式提取了必需的实体及其属性:
chain.run(inp)
[{'person_name': 'Alex',
'person_height': 5,
'person_hair_color': 'blonde',
'dog_name': 'Frosty',
'dog_breed': 'labrador'},
{'person_name': 'Claudia',
'person_height': 6,
'person_hair_color': 'brunette'}]
Pydantic示例
我们还可以使用Pydantic模式选择所需的属性和类型,并将那些不是严格要求的属性设置为“可选”。
通过使用create_extraction_chain_pydantic
函数,我们可以将Pydantic模式作为输入发送,并且输出将是一个符合我们所需模式的实例化对象。
这样,我们可以以与在Python中创建新类或函数相同的方式指定模式-纯粹使用Python类型。
from typing import Optional, List
from pydantic import BaseModel, Field
class Properties(BaseModel):
person_name: str
person_height: int
person_hair_color: str
dog_breed: Optional[str]
dog_name: Optional[str]
chain = create_extraction_chain_pydantic(pydantic_schema=Properties, llm=llm)
inp = """
Alex is 5 feet tall. Claudia is 1 feet taller Alex and jumps higher than him. Claudia is a brunette and Alex is blonde.
Alex's dog Frosty is a labrador and likes to play hide and seek.
"""
chain.run(inp)
[Properties(person_name='Alex', person_height=5, person_hair_color='blonde', dog_breed='labrador', dog_name='Frosty'),
Properties(person_name='Claudia', person_height=6, person_hair_color='brunette', dog_breed=None, dog_name=None)]
22、Graph QA
这本例介绍了如何在图数据结构上进行问答。
创建 graph
这里构造一个示例图。目前,这最适用于小部分文本。
from langchain.indexes import GraphIndexCreator
from langchain.llms import OpenAI
from langchain.document_loaders import TextLoader
index_creator = GraphIndexCreator(llm=OpenAI(temperature=0))
with open("../../state_of_the_union.txt") as f:
all_text = f.read()
# 我们将只使用一个小片段,因为目前提取知识三元组有点密集。
text = "\n".join(all_text.split("\n\n")[105:108])
text
'It won’t look like much, but if you stop and look closely, you’ll see a “Field of dreams,” the ground on which America’s future will be built. \nThis is where Intel, the American company that helped build Silicon Valley, is going to build its $20 billion semiconductor “mega site”. \nUp to eight state-of-the-art factories in one place. 10,000 new good-paying jobs. '
graph = index_creator.from_text(text)
# 我们可以检查创建的图形。
graph.get_triples()
[('Intel', '$20 billion semiconductor "mega site"', 'is going to build'),
('Intel', 'state-of-the-art factories', 'is building'),
('Intel', '10,000 new good-paying jobs', 'is creating'),
('Intel', 'Silicon Valley', 'is helping build'),
('Field of dreams',
"America's future will be built",
'is the ground on which')]
查询图
使用图 QA链,提问
from langchain.chains import GraphQAChain
chain = GraphQAChain.from_llm(OpenAI(temperature=0), graph=graph, verbose=True)
chain.run("what is Intel going to build?")
[1m> Entering new GraphQAChain chain...[0m
Entities Extracted:
[32;1m[1;3m Intel[0m
Full Context:
[32;1m[1;3mIntel is going to build $20 billion semiconductor "mega site"
Intel is building state-of-the-art factories
Intel is creating 10,000 new good-paying jobs
Intel is helping build Silicon Valley[0m
[1m> Finished chain.[0m
' Intel is going to build a $20 billion semiconductor "mega site" with state-of-the-art factories, creating 10,000 new good-paying jobs and helping to build Silicon Valley.'
Save the graph
We can also save and load the graph.
graph.write_to_gml("graph.gml")
from langchain.indexes.graph import NetworkxEntityGraph
loaded_graph = NetworkxEntityGraph.from_gml("graph.gml")
loaded_graph.get_triples()
[('Intel', '$20 billion semiconductor "mega site"', 'is going to build'),
('Intel', 'state-of-the-art factories', 'is building'),
('Intel', '10,000 new good-paying jobs', 'is creating'),
('Intel', 'Silicon Valley', 'is helping build'),
('Field of dreams',
"America's future will be built",
'is the ground on which')]
23、虚拟文档嵌入
这个笔记本介绍了如何使用虚拟文档嵌入(HyDE),如这篇论文所述。
在高层次上,HyDE是一种嵌入技术,它接受查询,生成一个虚拟答案,然后嵌入该生成的文档并将其用作最终示例。
为了使用HyDE,我们需要提供一个基本的嵌入模型,以及一个用于生成这些文档的LLMChain。
默认情况下,HyDE类带有一些默认的提示(有关详细信息,请参阅论文),但我们也可以创建自己的提示。
from langchain.llms import OpenAI
from langchain.embeddings import OpenAIEmbeddings
from langchain.chains import LLMChain, HypotheticalDocumentEmbedder
from langchain.prompts import PromptTemplate
base_embeddings = OpenAIEmbeddings()
llm = OpenAI()
# 使用“web_search”提示进行加载
embeddings = HypotheticalDocumentEmbedder.from_llm(llm, base_embeddings, "web_search")
# 现在我们可以像使用任何嵌入类一样使用它!
result = embeddings.embed_query("泰姬陵在哪里?")
多次生成
我们还可以生成多个文档,然后将这些文档的嵌入组合起来。
默认情况下,我们通过取平均值来组合这些文档的嵌入。
我们可以通过改变用于生成文档的LLM来实现这一点。
multi_llm = OpenAI(n=4, best_of=4)
embeddings = HypotheticalDocumentEmbedder.from_llm(
multi_llm, base_embeddings, "web_search"
)
result = embeddings.embed_query("泰姬陵在哪里?")
使用我们自己的提示
除了使用预配置的提示外,我们还可以轻松构建自己的提示并在生成文档的LLMChain中使用它们。
如果我们知道查询所在的领域,这可能非常有用,因为我们可以将提示调整为生成更类似于该领域的文本。
在下面的示例中,让我们将其调整为生成关于国情咨文的文本(因为我们将在下一个示例中使用它)。
prompt_template = """请回答关于最近一次国情咨文的用户问题
问题$:{question}
回答$:"""
prompt = PromptTemplate(input_variables=["question"],
template=prompt_template)
llm_chain = LLMChain(llm=llm, prompt=prompt)
embeddings = HypotheticalDocumentEmbedder(
llm_chain=llm_chain, base_embeddings=base_embeddings
)
result = embeddings.embed_query(
"总统在关于Ketanji Brown Jackson的发言中说了什么"
)
使用HyDE
现在我们有了HyDE,我们可以像使用任何其他嵌入类一样使用它!
在这里,我们使用它在国情咨文示例中查找相似的段落。
from langchain.text_splitter import CharacterTextSplitter
from langchain.vectorstores import Chroma
with open("../../state_of_the_union.txt") as f:
state_of_the_union = f.read()
text_splitter = CharacterTextSplitter(chunk_size=1000, chunk_overlap=0)
texts = text_splitter.split_text(state_of_the_union)
docsearch = Chroma.from_texts(texts, embeddings)
query = "总统在关于Ketanji Brown Jackson的发言中说了什么"
docs = docsearch.similarity_search(query)
使用直接本地API运行Chroma。
使用DuckDB内存中的数据库。数据将是临时的。
print(docs[0].page_content)
In state after state, new laws have been passed, not only to suppress the vote, but to subvert entire elections.
We cannot let this happen.
Tonight. I call on the Senate to: Pass the Freedom to Vote Act. Pass the John Lewis Voting Rights Act. And while you’re at it, pass the Disclose Act so Americans can know who is funding our elections.
Tonight, I’d like to honor someone who has dedicated his life to serve this country: Justice Stephen Breyer—an Army veteran, Constitutional scholar, and retiring Justice of the United States Supreme Court. Justice Breyer, thank you for your service.
One of the most serious constitutional responsibilities a President has is nominating someone to serve on the United States Supreme Court.
And I did that 4 days ago, when I nominated Circuit Court of Appeals Judge Ketanji Brown Jackson. One of our nation’s top legal minds, who will continue Justice Breyer’s legacy of excellence.
24、Bash chain
这个笔记本展示了 使用LLM和bash进程 来执行简单的文件系统命令。
from langchain.chains import LLMBashChain
from langchain.llms import OpenAI
llm = OpenAI(temperature=0)
text = "Please write a bash script that prints 'Hello World' to the console."
bash_chain = LLMBashChain.from_llm(llm, verbose=True)
bash_chain.run(text)
'Hello World\n'
Customize Prompt
You can also customize the prompt that is used. Here is an example prompting to avoid using the ‘echo’ utility
from langchain.prompts.prompt import PromptTemplate
from langchain.chains.llm_bash.prompt import BashOutputParser
_PROMPT_TEMPLATE = """If someone asks you to perform a task, your job is to come up with a series of bash commands that will perform the task. There is no need to put "#!/bin/bash" in your answer. Make sure to reason step by step, using this format:
Question: "copy the files in the directory named 'target' into a new directory at the same level as target called 'myNewDirectory'"
I need to take the following actions:
- List all files in the directory
- Create a new directory
- Copy the files from the first directory into the second directory
```bash
ls
mkdir myNewDirectory
cp -r target/* myNewDirectory
Do not use ‘echo’ when writing the script.
That is the format. Begin! Question: {question}“”"
PROMPT = PromptTemplate( input_variables=[“question”], template=_PROMPT_TEMPLATE, output_parser=BashOutputParser(), )
```python
bash_chain = LLMBashChain.from_llm(llm, prompt=PROMPT, verbose=True)
text = "Please write a bash script that prints 'Hello World' to the console."
bash_chain.run(text)
[1m> Entering new LLMBashChain chain...[0m
Please write a bash script that prints 'Hello World' to the console.[32;1m[1;3m
```bash
printf "Hello World\n"
```[0m
Code: [33;1m[1;3m['printf "Hello World\\n"'][0m
Answer: [33;1m[1;3mHello World
[0m
[1m> Finished chain.[0m
'Hello World\n'
Persistent Terminal
By default, the chain will run in a separate subprocess each time it is called. This behavior can be changed by instantiating with a persistent bash process.
from langchain.utilities.bash import BashProcess
persistent_process = BashProcess(persistent=True)
bash_chain = LLMBashChain.from_llm(llm, bash_process=persistent_process, verbose=True)
text = "List the current directory then move up a level."
bash_chain.run(text)
[1m> Entering new LLMBashChain chain...[0m
List the current directory then move up a level.[32;1m[1;3m
```bash
ls
cd ..
```[0m
Code: [33;1m[1;3m['ls', 'cd ..'][0m
Answer: [33;1m[1;3mapi.html llm_summarization_checker.html
constitutional_chain.html moderation.html
llm_bash.html openai_openapi.yaml
llm_checker.html openapi.html
llm_math.html pal.html
llm_requests.html sqlite.html[0m
[1m> Finished chain.[0m
'api.html\t\t\tllm_summarization_checker.html\r\nconstitutional_chain.html\tmoderation.html\r\nllm_bash.html\t\t\topenai_openapi.yaml\r\nllm_checker.html\t\topenapi.html\r\nllm_math.html\t\t\tpal.html\r\nllm_requests.html\t\tsqlite.html'
# Run the same command again and see that the state is maintained between calls
bash_chain.run(text)
[1m> Entering new LLMBashChain chain...[0m
List the current directory then move up a level.[32;1m[1;3m
```bash
ls
cd ..
```[0m
Code: [33;1m[1;3m['ls', 'cd ..'][0m
Answer: [33;1m[1;3mexamples getting_started.html index_examples
generic how_to_guides.rst[0m
[1m> Finished chain.[0m
'examples\t\tgetting_started.html\tindex_examples\r\ngeneric\t\t\thow_to_guides.rst'
25、自检链
这个笔记本展示了如何使用LLMCheckerChain。
from langchain.chains import LLMCheckerChain
from langchain.llms import OpenAI
llm = OpenAI(temperature=0.7)
text = "What type of mammal lays the biggest eggs?"
checker_chain = LLMCheckerChain.from_llm(llm, verbose=True)
checker_chain.run(text)
[1m> Entering new LLMCheckerChain chain...[0m
[1m> Entering new SequentialChain chain...[0m
[1m> Finished chain.[0m
[1m> Finished chain.[0m
'没有哺乳动物能够产下最大的蛋。大象鸟是一种巨鸟,它的蛋是所有鸟类中最大的。'
26、数学链
本例展示了使用LLMs和Python REPL解决复杂的数学问题。
from langchain import OpenAI, LLMMathChain
llm = OpenAI(temperature=0)
llm_math = LLMMathChain.from_llm(llm, verbose=True)
llm_math.run("13的0.3432次方")
27、HTTP request chain
使用request库从URL获取HTML结果,然后使用LLM解析结果
from langchain.llms import OpenAI
from langchain.chains import LLMRequestsChain, LLMChain
from langchain.prompts import PromptTemplate
template = """Between >>> and <<< are the raw search result text from google.
Extract the answer to the question '{query}' or say "not found" if the information is not contained.
Use the format
Extracted:<answer or "not found">
>>> {requests_result} <<<
Extracted:"""
PROMPT = PromptTemplate(
input_variables=["query", "requests_result"],
template=template,
)
chain = LLMRequestsChain(llm_chain=LLMChain(llm=OpenAI(temperature=0), prompt=PROMPT))
question = "三个最大的国家及其各自的大小是什么?"
inputs = {
"query": question,
"url": "https://www.google.com/search?q=" + question.replace(" ", "+"),
}
chain(inputs)
{'query': '三个最大的国家及其各自的大小是什么?',
'url': 'https://www.google.com/search?q=三个最大的国家及其各自的大小是什么?',
'output': '俄罗斯(17,098,242平方公里),加拿大(9,984,670平方公里),美国(9,826,675平方公里)'}
28、Summarization checker chain
This notebook shows some examples of LLMSummarizationCheckerChain in use with different types of texts. It has a few distinct differences from the LLMCheckerChain
, in that it doesn’t have any assumptions to the format of the input text (or summary). Additionally, as the LLMs like to hallucinate when fact checking or get confused by context, it is sometimes beneficial to run the checker multiple times. It does this by feeding the rewritten “True” result back on itself, and checking the “facts” for truth. As you can see from the examples below, this can be very effective in arriving at a generally true body of text.
You can control the number of times the checker runs by setting the max_checks
parameter. The default is 2, but you can set it to 1 if you don’t want any double-checking.
from langchain.chains import LLMSummarizationCheckerChain
from langchain.llms import OpenAI
llm = OpenAI(temperature=0)
checker_chain = LLMSummarizationCheckerChain.from_llm(llm, verbose=True, max_checks=2)
text = """
Your 9-year old might like these recent discoveries made by The James Webb Space Telescope (JWST):
• In 2023, The JWST spotted a number of galaxies nicknamed "green peas." They were given this name because they are small, round, and green, like peas.
• The telescope captured images of galaxies that are over 13 billion years old. This means that the light from these galaxies has been traveling for over 13 billion years to reach us.
• JWST took the very first pictures of a planet outside of our own solar system. These distant worlds are called "exoplanets." Exo means "from outside."
These discoveries can spark a child's imagination about the infinite wonders of the universe."""
checker_chain.run(text)
from langchain.chains import LLMSummarizationCheckerChain
from langchain.llms import OpenAI
llm = OpenAI(temperature=0)
checker_chain = LLMSummarizationCheckerChain.from_llm(llm, verbose=True, max_checks=3)
text = "The Greenland Sea is an outlying portion of the Arctic Ocean located between Iceland, Norway, the Svalbard archipelago and Greenland. It has an area of 465,000 square miles and is one of five oceans in the world, alongside the Pacific Ocean, Atlantic Ocean, Indian Ocean, and the Southern Ocean. It is the smallest of the five oceans and is covered almost entirely by water, some of which is frozen in the form of glaciers and icebergs. The sea is named after the island of Greenland, and is the Arctic Ocean's main outlet to the Atlantic. It is often frozen over so navigation is limited, and is considered the northern branch of the Norwegian Sea."
checker_chain.run(text)
审查 Moderation
本文档演示了如何使用审查链以及几种常见的方法。审查链用于检测可能含有仇恨、暴力等内容的文本。这对于对用户输入进行处理以及对语言模型的输出进行处理都非常有用。一些 API 供应商(如 OpenAI)明确禁止您或您的最终用户生成某些类型的有害内容。为了遵守这一规定(并且通常还可以防止您的应用程序造成伤害),您可能经常希望在任何 LLMChains 后附加一个审查链,以确保 LLM 生成的任何输出都不会有害。
如果传递到审查链中的内容是有害的,则没有一种最佳处理方式,这可能取决于您的应用程序。有时,您可能希望在 Chain 中抛出错误(并由您的应用程序处理该错误)。其他时候,您可能希望向用户返回一些说明,说明文本是有害的。甚至可能还有其他处理方式!在本教程中,我们将涵盖所有这些处理方式。
我们将展示:
- 如何将任何文本通过审核链运行。
- 如何将审核链附加到 LLMChain 中。
from langchain.llms import OpenAI
from langchain.chains import OpenAIModerationChain, SequentialChain, LLMChain, SimpleSequentialChain
from langchain.prompts import PromptTemplate
如何使用审核链
以下是使用默认设置使用审核链的示例(将返回一个字符串,解释已标记的内容)。
moderation_chain = OpenAIModerationChain()
moderation_chain.run("This is okay")
# -> 'This is okay'
moderation_chain.run("I will kill you")
# -> "Text was found that violates OpenAI's content policy."
以下是使用审核链引发错误的示例。
moderation_chain_error = OpenAIModerationChain(error=True)
moderation_chain_error.run("This is okay")
# -> 'This is okay'
moderation_chain_error.run("I will kill you")
以下是创建具有自定义错误消息的自定义审核链的示例。它需要对 OpenAI 的审核终端结果有一些了解(请参阅此处的文档)。
class CustomModeration(OpenAIModerationChain):
def _moderate(self, text: str, results: dict) -> str:
if results["flagged"]:
error_str = f"The following text was found that violates OpenAI's content policy: {text}"
return error_str
return text
custom_moderation = CustomModeration()
custom_moderation.run("This is okay")
# -> 'This is okay'
custom_moderation.run("I will kill you")
# -> "The following text was found that violates OpenAI's content policy: I will kill you"
如何将审核链附加到 LLMChain
要将审核链与 LLMChain 轻松组合在一起,您可以使用 SequentialChain 抽象。
让我们从一个简单的例子开始,LLMChain 只有一个输入。
为此,我们将提示模型说一些有害的内容。
prompt = PromptTemplate(template="{text}", input_variables=["text"])
llm_chain = LLMChain(
llm=OpenAI(temperature=0, model_name="text-davinci-002"),
prompt=prompt)
text = """We are playing a game of repeat after me.
Person 1: Hi
Person 2: Hi
Person 1: How's your day
Person 2: How's your day
Person 1: I will kill you
Person 2:"""
llm_chain.run(text)
' I will kill you'
chain = SimpleSequentialChain(chains=[llm_chain, moderation_chain])
chain.run(text)
# -> "Text was found that violates OpenAI's content policy."
现在让我们通过一个使用具有多个输入的 LLMChain 的示例来演示它(稍微复杂一些,因为我们不能使用 SimpleSequentialChain)
prompt = PromptTemplate(template="{setup}{new_input}Person2:",
input_variables=["setup", "new_input"])
llm_chain = LLMChain( llm=OpenAI(temperature=0,
model_name="text-davinci-002"),
prompt=prompt
)
setup = """We are playing a game of repeat after me.
Person 1: Hi
Person 2: Hi
Person 1: How's your day
Person 2: How's your day
Person 1:"""
new_input = "I will kill you"
inputs = {"setup": setup, "new_input": new_input}
llm_chain(inputs, return_only_outputs=True)
# -> {'text': ' I will kill you'}
# Setting the input/output keys so it lines up
moderation_chain.input_key = "text"
moderation_chain.output_key = "sanitized_text"
chain = SequentialChain(chains=[llm_chain, moderation_chain], input_variables=["setup", "new_input"])
chain(inputs, return_only_outputs=True)
# -> {'sanitized_text': "Text was found that violates OpenAI's content policy."}
29、动态从多个提示中选择 multi_prompt_router
本例演示了如何使用 RouterChain
范式创建一个动态选择要用于给定输入的提示的链。
具体来说,我们展示了如何使用 MultiPromptChain
创建一个问题回答链,该链选择与给定问题最相关的提示,然后使用该提示来回答问题。
from langchain.chains.router import MultiPromptChain
from langchain.llms import OpenAI
physics_template = """You are a very smart physics professor. \
You are great at answering questions about physics in a concise and easy to understand manner. \
When you don't know the answer to a question you admit that you don't know.
Here is a question:
{input}"""
math_template = """You are a very good mathematician. You are great at answering math questions. \
You are so good because you are able to break down hard problems into their component parts, \
answer the component parts, and then put them together to answer the broader question.
Here is a question:
{input}"""
prompt_infos = [
{
"name": "physics",
"description": "Good for answering questions about physics",
"prompt_template": physics_template
},
{
"name": "math",
"description": "Good for answering math questions",
"prompt_template": math_template
}
]
chain = MultiPromptChain.from_prompts(OpenAI(), prompt_infos, verbose=True)
print(chain.run("What is black body radiation?"))
> Entering new MultiPromptChain chain...
physics: {'input': 'What is black body radiation?'}
> Finished chain.
Black body radiation is the emission of electromagnetic radiation from a body due to its temperature. It is a type of thermal radiation that is emitted from the surface of all objects that are at a temperature above absolute zero. It is a spectrum of radiation that is influenced by the temperature of the body and is independent of the composition of the emitting material.
print(chain.run("What is the first prime number greater than 40 such that one plus the prime number is divisible by 3"))
> Entering new MultiPromptChain chain...
math: {'input': 'What is the first prime number greater than 40 such that one plus the prime number is divisible by 3'}
> Finished chain.
?
The first prime number greater than 40 such that one plus the prime number is divisible by 3 is 43. To solve this problem, we can break down the question into two parts: finding the first prime number greater than 40, and then finding a number that is divisible by 3.
The first step is to find the first prime number greater than 40. A prime number is a number that is only divisible by 1 and itself. The next prime number after 40 is 41.
The second step is to find a number that is divisible by 3. To do this, we can add 1 to 41, which gives us 42. Now, we can check if 42 is divisible by 3. 42 divided by 3 is 14, so 42 is divisible by 3.
Therefore, the answer to the question is 43.
print(chain.run("What is the name of the type of cloud that rins"))
> Entering new MultiPromptChain chain...
None: {'input': 'What is the name of the type of cloud that rains?'}
> Finished chain.
The type of cloud that typically produces rain is called a cumulonimbus cloud. This type of cloud is characterized by its large vertical extent and can produce thunderstorms and heavy precipitation. Is there anything else you'd like to know?
30、动态选择多个检索器 multi_retrieval_qa_router
本文档演示如何使用 RouterChain
范式创建一个动态选择要使用的检索系统的链。
具体来说,我们展示了如何使用 MultiRetrievalQAChain
创建一个问答链,该链根据给定的问题选择最相关的检索问答链,然后使用它来回答问题。
from langchain.chains.router import MultiRetrievalQAChain
from langchain.llms import OpenAI
from langchain.embeddings import OpenAIEmbeddings
from langchain.document_loaders import TextLoader
from langchain.vectorstores import FAISS
sou_docs = TextLoader('../../state_of_the_union.txt').load_and_split()
sou_retriever = FAISS.from_documents(sou_docs, OpenAIEmbeddings()).as_retriever()
pg_docs = TextLoader('../../paul_graham_essay.txt').load_and_split()
pg_retriever = FAISS.from_documents(pg_docs, OpenAIEmbeddings()).as_retriever()
personal_texts = [
"I love apple pie",
"My favorite color is fuchsia",
"My dream is to become a professional dancer",
"I broke my arm when I was 12",
"My parents are from Peru",
]
personal_retriever = FAISS.from_texts(personal_texts, OpenAIEmbeddings()).as_retriever()
retriever_infos = [
{
"name": "state of the union",
"description": "Good for answering questions about the 2023 State of the Union address",
"retriever": sou_retriever
},
{
"name": "pg essay",
"description": "Good for answer quesitons about Paul Graham's essay on his career",
"retriever": pg_retriever
},
{
"name": "personal",
"description": "Good for answering questions about me",
"retriever": personal_retriever
}
]
chain = MultiRetrievalQAChain.from_retrievers(OpenAI(), retriever_infos, verbose=True)
print(chain.run("What did the president say about the economy?"))
> Entering new MultiRetrievalQAChain chain...
state of the union: {'query': 'What did the president say about the economy in the 2023 State of the Union address?'}
> Finished chain.
The president said that the economy was stronger than it had been a year prior, and that the American Rescue Plan helped create record job growth and fuel economic relief for millions of Americans. He also proposed a plan to fight inflation and lower costs for families, including cutting the cost of prescription drugs and energy, providing investments and tax credits for energy efficiency, and increasing access to child care and Pre-K.
print(chain.run("What is something Paul Graham regrets about his work?"))
> Entering new MultiRetrievalQAChain chain...
pg essay: {'query': 'What is something Paul Graham regrets about his work?'}
> Finished chain.
Paul Graham regrets that he did not take a vacation after selling his company, instead of immediately starting to paint.
print(chain.run("What is my background?"))
> Entering new MultiRetrievalQAChain chain...
personal: {'query': 'What is my background?'}
> Finished chain.
Your background is Peruvian.
print(chain.run("What year was the Internet created in?"))
> Entering new MultiRetrievalQAChain chain...
None: {'query': 'What year was the Internet created in?'}
> Finished chain.
The Internet was created in 1969 through a project called ARPANET, which was funded by the United States Department of Defense. However, the World Wide Web, which is often confused with the Internet, was created in 1989 by British computer scientist Tim Berners-Lee.
31、使用OpenAI函数进行检索问答
OpenAI函数允许对响应输出进行结构化。在问答问题时,除了获取最终答案外,还可以获取支持证据、引用等,这通常很有用。
在这个笔记本中,我们展示了如何使用一个使用OpenAI函数作为整个检索流程的一部分的LLM链。
from langchain.chains import RetrievalQA
from langchain.document_loaders import TextLoader
from langchain.embeddings.openai import OpenAIEmbeddings
from langchain.text_splitter import CharacterTextSplitter
from langchain.vectorstores import Chroma
loader = TextLoader("../../state_of_the_union.txt")
documents = loader.load()
text_splitter = CharacterTextSplitter(chunk_size=1000, chunk_overlap=0)
texts = text_splitter.split_documents(documents)
for i, text in enumerate(texts):
text.metadata['source'] = f"{i}-pl"
embeddings = OpenAIEmbeddings()
docsearch = Chroma.from_documents(texts, embeddings)
from langchain.chat_models import ChatOpenAI
from langchain.chains.combine_documents.stuff import StuffDocumentsChain
from langchain.prompts import PromptTemplate
from langchain.chains import create_qa_with_sources_chain
llm = ChatOpenAI(temperature=0, model="gpt-3.5-turbo-0613")
qa_chain = create_qa_with_sources_chain(llm)
doc_prompt = PromptTemplate(
template="Content: {page_content}\nSource: {source}",
input_variables=["page_content", "source"],
)
final_qa_chain = StuffDocumentsChain(
llm_chain=qa_chain,
document_variable_name='context',
document_prompt=doc_prompt,
)
retrieval_qa = RetrievalQA(
retriever=docsearch.as_retriever(),
combine_documents_chain=final_qa_chain
)
query = "总统对俄罗斯说了什么"
retrieval_qa.run(query)
'{\n "answer": "总统对俄罗斯的行动表示强烈谴责,并宣布采取措施孤立俄罗斯并支持乌克兰。他指出俄罗斯对乌克兰的入侵将对俄罗斯产生长期影响,并强调美国及其盟友捍卫北约国家的承诺。总统还提到对俄罗斯实施制裁,并释放石油储备以帮助缓解天然气价格。总体而言,总统的讲话传达了坚定反对俄罗斯侵略行为、支持乌克兰并保护美国利益的立场。",\n "sources": ["0-pl", "4-pl", "5-pl", "6-pl"]\n}'
使用 Pydantic
如果需要,我们可以将链设置为返回Pydantic格式。
请注意,如果下游链消耗此链的输出(包括内存),它们通常会希望输出为字符串格式,因此只有在它是最终链时才应使用此链。
qa_chain_pydantic = create_qa_with_sources_chain(llm, output_parser="pydantic")
final_qa_chain_pydantic = StuffDocumentsChain(
llm_chain=qa_chain_pydantic,
document_variable_name='context',
document_prompt=doc_prompt,
)
retrieval_qa_pydantic = RetrievalQA(
retriever=docsearch.as_retriever(),
combine_documents_chain=final_qa_chain_pydantic
)
retrieval_qa_pydantic.run(query)
AnswerWithSources(answer="The President expressed ... and support freedom.",
sources=['0-pl', '4-pl', '5-pl', '6-pl'])
在ConversationalRetrievalChain中使用
我们还可以展示在 ConversationalRetrievalChain 中使用该功能。
请注意,因为此链涉及内存,我们将不使用Pydantic返回类型。
from langchain.chains import ConversationalRetrievalChain
from langchain.memory import ConversationBufferMemory
from langchain.chains import LLMChain
memory = ConversationBufferMemory(memory_key="chat_history", return_messages=True)
_template = """Given the following conversation and a follow up question, rephrase the follow up question to be a standalone question, in its original language.\
Make sure to avoid using any unclear pronouns.
Chat History:
{chat_history}
Follow Up Input: {question}
Standalone question:"""
CONDENSE_QUESTION_PROMPT = PromptTemplate.from_template(_template)
condense_question_chain = LLMChain(
llm=llm,
prompt=CONDENSE_QUESTION_PROMPT,
)
qa = ConversationalRetrievalChain(
question_generator=condense_question_chain,
retriever=docsearch.as_retriever(),
memory=memory,
combine_docs_chain=final_qa_chain
)
query = "What did the president say about Ketanji Brown Jackson"
result = qa({"question": query})
result
{'question': 'What did the president say about Ketanji Brown Jackson',
'chat_history': [
HumanMessage(
content='What did the president say about Ketanji Brown Jackson',
additional_kwargs={},
example=False),
AIMessage(
content='{\n "answer": "The President nominated Ketanji Brown Jackson as a Circuit Court of Appeals Judge and praised her as one of the nation\'s top legal minds who will continue Justice Breyer\'s legacy of excellence.",\n "sources": ["31-pl"]\n}',
additional_kwargs={},
example=False)
],
'answer': '{\n "answer": "The President nominated Ketanji Brown Jackson as a Circuit Court of Appeals Judge and praised her as one of the nation\'s top legal minds who will continue Justice Breyer\'s legacy of excellence.",\n "sources": ["31-pl"]\n}'}
query = "关于她的前任,他说了什么"
result = qa({"question": query})
result
{'question': 'what did he say about her predecessor?',
'chat_history': [
HumanMessage(content='What did the president say about Ketanji Brown Jackson', additional_kwargs={}, example=False),
AIMessage(content='{\n "answer": "The President nominated Ketanji Brown Jackson as a Circuit Court of Appeals Judge and praised her as one of the nation\'s top legal minds who will continue Justice Breyer\'s legacy of excellence.",\n "sources": ["31-pl"]\n}', additional_kwargs={}, example=False),
HumanMessage(content='what did he say about her predecessor?', additional_kwargs={}, example=False),
AIMessage(content='{\n "answer": "The President honored Justice Stephen Breyer for his service as an Army veteran, Constitutional scholar, and retiring Justice of the United States Supreme Court.",\n "sources": ["31-pl"]\n}', additional_kwargs={}, example=False)],
'answer': '{\n "answer": "The President honored Justice Stephen Breyer for his service as an Army veteran, Constitutional scholar, and retiring Justice of the United States Supreme Court.",\n "sources": ["31-pl"]\n}'}
32、OpenAPI chain
这个笔记展示了一个 使用OpenAPI链 以自然语言调用端点,并以自然语言返回响应的示例。
from langchain.tools import OpenAPISpec, APIOperation
from langchain.chains import OpenAPIEndpointChain
from langchain.requests import Requests
from langchain.llms import OpenAI
Load the spec
Load a wrapper of the spec (so we can work with it more easily). You can load from a url or from a local file.
spec = OpenAPISpec.from_url(
"https://www.klarna.com/us/shopping/public/openai/v0/api-docs/"
)
Attempting to load an OpenAPI 3.0.1 spec. This may result in degraded performance. Convert your OpenAPI spec to 3.1.* spec for better support.
# Alternative loading from file
# spec = OpenAPISpec.from_file("openai_openapi.yaml")
Select the Operation
In order to provide a focused on modular chain, we create a chain specifically only for one of the endpoints. Here we get an API operation from a specified endpoint and method.
operation = APIOperation.from_openapi_spec(spec, "/public/openai/v0/products", "get")
Construct the chain
We can now construct a chain to interact with it. In order to construct such a chain, we will pass in:
- The operation endpoint
- A requests wrapper (can be used to handle authentication, etc)
- The LLM to use to interact with it
llm = OpenAI() # Load a Language Model
chain = OpenAPIEndpointChain.from_api_operation(
operation,
llm,
requests=Requests(),
verbose=True,
return_intermediate_steps=True, # Return request and response text
)
output = chain("whats the most expensive shirt?")
# View intermediate steps
output["intermediate_steps"]
{'request_args': '{"q": "shirt", "size": 1, "max_price": null}',
'response_text': '{"products":[{"name":"Burberry Check Poplin Shirt","url":"https://www.klarna.com/us/shopping/pl/cl10001/3201810981/Clothing/Burberry-Check-Poplin-Shirt/?utm_source=openai&ref-site=openai_plugin","price":"$360.00","attributes":["Material:Cotton","Target Group:Man","Color:Gray,Blue,Beige","Properties:Pockets","Pattern:Checkered"]}]}'}
Return raw response
We can also run this chain without synthesizing the response.
This will have the effect of just returning the raw API output.
chain = OpenAPIEndpointChain.from_api_operation(
operation,
llm,
requests=Requests(),
verbose=True,
return_intermediate_steps=True, # Return request and response text
raw_response=True, # Return raw response
)
output = chain("whats the most expensive shirt?")
Example POST message
For this demo, we will interact with the speak API.
spec = OpenAPISpec.from_url("https://api.speak.com/openapi.yaml")
Attempting to load an OpenAPI 3.0.1 spec. This may result in degraded performance. Convert your OpenAPI spec to 3.1.* spec for better support.
Attempting to load an OpenAPI 3.0.1 spec. This may result in degraded performance. Convert your OpenAPI spec to 3.1.* spec for better support.
operation = APIOperation.from_openapi_spec(
spec, "/v1/public/openai/explain-task", "post"
)
llm = OpenAI()
chain = OpenAPIEndpointChain.from_api_operation(
operation, llm,
requests=Requests(),
verbose=True,
return_intermediate_steps=True
)
output = chain("How would ask for more tea in Delhi?")
# Show the API chain's intermediate steps
output["intermediate_steps"]
33、Program-aided language model (PAL) chain
Implements Program-Aided Language Models, as in https://arxiv.org/pdf/2211.10435.pdf.
from langchain.chains import PALChain
from langchain import OpenAI
llm = OpenAI(temperature=0, max_tokens=512)
Math Prompt
pal_chain = PALChain.from_math_prompt(llm, verbose=True)
question = "Jan has three times the number of pets as Marcia. Marcia has two more pets than Cindy. If Cindy has four pets, how many total pets do the three have?"
pal_chain.run(question)
Colored Objects
pal_chain = PALChain.from_colored_object_prompt(llm, verbose=True)
question = "On the desk, you see two blue booklets, two purple booklets, and two yellow pairs of sunglasses. If I remove all the pairs of sunglasses from the desk, how many purple items remain on it?"
pal_chain.run(question)
Intermediate Steps
You can also use the intermediate steps flag to return the code executed that generates the answer.
pal_chain = PALChain.from_colored_object_prompt(
llm, verbose=True, return_intermediate_steps=True
)
question = "On the desk, you see two blue booklets, two purple booklets, and two yellow pairs of sunglasses. If I remove all the pairs of sunglasses from the desk, how many purple items remain on it?"
result = pal_chain({"question": question})
result["intermediate_steps"]
34、Question-Answering Citations
This notebook shows how to use OpenAI functions ability to extract citations from text.
from langchain.chains import create_citation_fuzzy_match_chain
from langchain.chat_models import ChatOpenAI
question = "What did the author do during college?"
context = """
My name is Jason Liu, and I grew up in Toronto Canada but I was born in China.
I went to an arts highschool but in university I studied Computational Mathematics and physics.
As part of coop I worked at many companies including Stitchfix, Facebook.
I also started the Data Science club at the University of Waterloo and I was the president of the club for 2 years.
"""
llm = ChatOpenAI(temperature=0, model="gpt-3.5-turbo-0613")
chain = create_citation_fuzzy_match_chain(llm)
result = chain.run(question=question, context=context)
result
question='What did the author do during college?'
answer=[FactWithEvidence(
fact='The author studied Computational Mathematics and physics in university.',
substring_quote=['in university I studied Computational Mathematics and physics']),
FactWithEvidence(fact='The author started the Data Science club at the University of Waterloo and was the president of the club for 2 years.',
substring_quote=['started the Data Science club at the University of Waterloo', 'president of the club for 2 years'])]
def highlight(text, span):
return (
"..."
+ text[span[0] - 20 : span[0]]
+ "*"
+ "\033[91m"
+ text[span[0] : span[1]]
+ "\033[0m"
+ "*"
+ text[span[1] : span[1] + 20]
+ "..."
)
for fact in result.answer:
print("Statement:", fact.fact)
for span in fact.get_spans(context):
print("Citation:", highlight(context, span))
print()
Statement: The author studied Computational Mathematics and physics in university.
Citation: ...arts highschool but *[91min university I studied Computational Mathematics and physics[0m*.
As part of coop I...
Statement: The author started the Data Science club at the University of Waterloo and was the president of the club for 2 years.
Citation: ...x, Facebook.
I also *[91mstarted the Data Science club at the University of Waterloo[0m* and I was the presi...
Citation: ...erloo and I was the *[91mpresident of the club for 2 years[0m*.
...
35、文档问答 qa_with_sources
在这里,我们将介绍如何使用 LangChain 对一系列文档进行问答。在底层,我们将使用我们的 文档链。
准备数据
首先我们准备数据。在这个示例中,我们对向量数据库进行相似性搜索,但这些文档可以以任何方式获取(这个笔记本的重点是突出显示在获取文档之后要做的事情)。
from langchain.embeddings.openai import OpenAIEmbeddings
from langchain.text_splitter import CharacterTextSplitter
from langchain.vectorstores import Chroma
from langchain.docstore.document import Document
from langchain.prompts import PromptTemplate
from langchain.indexes.vectorstore import VectorstoreIndexCreator
with open("../../state_of_the_union.txt") as f:
state_of_the_union = f.read()
text_splitter = CharacterTextSplitter(chunk_size=1000, chunk_overlap=0)
texts = text_splitter.split_text(state_of_the_union)
embeddings = OpenAIEmbeddings()
docsearch = Chroma.from_texts(texts, embeddings,
metadatas=[{"source": str(i)} for i in range(len(texts))]).as_retriever()
Running Chroma using direct local API.
Using DuckDB in-memory for database. Data will be transient.
query = "What did the president say about Justice Breyer"
docs = docsearch.get_relevant_documents(query)
from langchain.chains.question_answering import load_qa_chain
from langchain.llms import OpenAI
快速入门
如果您只是想尽快开始,这是推荐的方法:
chain = load_qa_chain(OpenAI(temperature=0), chain_type="stuff")
query = "What did the president say about Justice Breyer"
chain.run(input_documents=docs, question=query)
' The president said that Justice Breyer has dedicated his life to serve the country and thanked him for his service.'
The stuff
Chain
This sections shows results of using the stuff
Chain to do question answering.
chain = load_qa_chain(OpenAI(temperature=0), chain_type="stuff")
query = "What did the president say about Justice Breyer"
chain({"input_documents": docs, "question": query}, return_only_outputs=True)
{'output_text': ' The president said that Justice Breyer has dedicated his life to serve the country and thanked him for his service.'}
Custom Prompts
您还可以使用自己的提示与此链一起使用。在这个示例中,我们将用意大利语回答。
prompt_template = """Use the following pieces of context to answer the question at the end. If you don't know the answer, just say that you don't know, don't try to make up an answer.
{context}
Question: {question}
Answer in Italian:"""
PROMPT = PromptTemplate(
template=prompt_template, input_variables=["context", "question"]
)
chain = load_qa_chain(OpenAI(temperature=0), chain_type="stuff", prompt=PROMPT)
chain({"input_documents": docs, "question": query}, return_only_outputs=True)
{'output_text': ' Il presidente ha detto che Justice Breyer ha dedicato la sua vita a servire questo paese e ha ricevuto una vasta gamma di supporto.'}
The map_reduce
Chain
本节展示使用 map_reduce
链进行问答的结果。
chain = load_qa_chain(OpenAI(temperature=0), chain_type="map_reduce")
query = "What did the president say about Justice Breyer"
chain({"input_documents": docs, "question": query}, return_only_outputs=True)
{'output_text': ' The president said that Justice Breyer is an Army veteran, Constitutional scholar, and retiring Justice of the United States Supreme Court, and thanked him for his service.'}
中间步骤 (Intermediate Steps)
我们还可以返回 map_reduce
链的中间步骤,以便检查它们。这可以通过设置 return_map_steps
变量来实现。
chain = load_qa_chain(
OpenAI(temperature=0),
chain_type="map_reduce",
return_map_steps=True
)
chain(
{"input_documents": docs, "question": query},
return_only_outputs=True)
{'intermediate_steps': [' "Tonight, I’d like to honor someone who has dedicated his life to serve this country: Justice Stephen Breyer—an Army veteran, Constitutional scholar, and retiring Justice of the United States Supreme Court. Justice Breyer, thank you for your service."',
' A former top litigator in private practice. A former federal public defender. And from a family of public school educators and police officers. A consensus builder. Since she’s been nominated, she’s received a broad range of support—from the Fraternal Order of Police to former judges appointed by Democrats and Republicans.',
' None',
' None'],
'output_text': ' The president said that Justice Breyer is an Army veteran, Constitutional scholar, and retiring Justice of the United States Supreme Court, and thanked him for his service.'}
自定义提示
您还可以使用自己的提示与此链一起使用。在这个示例中,我们将用意大利语回答。
question_prompt_template = """Use the following portion of a long document to see if any of the text is relevant to answer the question.
Return any relevant text translated into italian.
{context}
Question: {question}
Relevant text, if any, in Italian:"""
QUESTION_PROMPT = PromptTemplate(
template=question_prompt_template, input_variables=["context", "question"]
)
combine_prompt_template = """Given the following extracted parts of a long document and a question, create a final answer italian.
If you don't know the answer, just say that you don't know. Don't try to make up an answer.
QUESTION: {question}
=========
{summaries}
=========
Answer in Italian:"""
COMBINE_PROMPT = PromptTemplate(
template=combine_prompt_template, input_variables=["summaries", "question"]
)
chain = load_qa_chain(OpenAI(temperature=0), chain_type="map_reduce", return_map_steps=True, question_prompt=QUESTION_PROMPT, combine_prompt=COMBINE_PROMPT)
chain({"input_documents": docs, "question": query}, return_only_outputs=True)
{'intermediate_steps': ["\nStasera vorrei onorare qualcuno che ha dedicato la sua vita a servire questo paese: il giustizia Stephen Breyer - un veterano dell'esercito, uno studioso costituzionale e un giustizia in uscita della Corte Suprema degli Stati Uniti. Giustizia Breyer, grazie per il tuo servizio.",
'\nNessun testo pertinente.',
' Non ha detto nulla riguardo a Justice Breyer.',
" Non c'è testo pertinente."],
'output_text': ' Non ha detto nulla riguardo a Justice Breyer.'}
Batch Size
使用 map_reduce
链时,需要记住的一个问题是在映射步骤中使用的批次大小。
如果太大,可能会导致速率限制错误。
您可以通过设置所使用的 LLM 上的批次大小来控制此参数。请注意,这仅适用于具有此参数的 LLM。
下面是一个示例:
llm = OpenAI(batch_size=5, temperature=0)
The refine
Chain
本节展示使用 refine
链进行问答的结果。
chain = load_qa_chain(OpenAI(temperature=0), chain_type="refine")
query = "What did the president say about Justice Breyer"
chain( {"input_documents": docs, "question": query},
return_only_outputs=True)
{'output_text': '\n\nThe president said that he wanted to honor Justice Breyer for his dedication to serving the country, his legacy of excellence, and his commitment to advancing liberty and justice, as well as for his support of the Equality Act and his commitment to protecting the rights of LGBTQ+ Americans. He also praised Justice Breyer for his role in helping to pass the Bipartisan Infrastructure Law, which he said would be the most sweeping investment to rebuild America in history and would help the country compete for the jobs of the 21st Century.'}
Intermediate Steps
我们还可以返回 refine
链的中间步骤,以便检查它们。
这可以通过设置 return_refine_steps
变量来实现。
chain = load_qa_chain(
OpenAI(temperature=0),
chain_type="refine",
return_refine_steps=True
)
chain({"input_documents": docs, "question": query},
return_only_outputs=True)
{'intermediate_steps': ['\nThe president ... of excellence.',
'\nThe president ... and justice.',
'\n\nThe president said that ... of LGBTQ+ Americans.',
'\n\nThe president said... America in history.'],
'output_text': '\n\nThe president ... in history.'}
Custom Prompts
您还可以使用自己的提示与此链一起使用。在这个示例中,我们将用意大利语回答。
refine_prompt_template = (
"The original question is as follows: {question}\n"
"We have provided an existing answer: {existing_answer}\n"
"We have the opportunity to refine the existing answer"
"(only if needed) with some more context below.\n"
"------------\n"
"{context_str}\n"
"------------\n"
"Given the new context, refine the original answer to better "
"answer the question. "
"If the context isn't useful, return the original answer. Reply in Italian."
)
refine_prompt = PromptTemplate(
input_variables=["question", "existing_answer", "context_str"],
template=refine_prompt_template,
)
initial_qa_template = (
"Context information is below. \n"
"---------------------\n"
"{context_str}"
"\n---------------------\n"
"Given the context information and not prior knowledge, "
"answer the question: {question}\nYour answer should be in Italian.\n"
)
initial_qa_prompt = PromptTemplate(
input_variables=["context_str", "question"],
template=initial_qa_template
)
chain = load_qa_chain(
OpenAI(temperature=0),
chain_type="refine",
return_refine_steps=True,
question_prompt=initial_qa_prompt,
refine_prompt=refine_prompt
)
chain({"input_documents": docs, "question": query},
return_only_outputs=True)
{'intermediate_steps': ['\nIl presidente ... suo servizio.',
"\nIl presidente ... di immigrazione.",
"\nIl presidente ... l'epidemia di oppiacei.",
"\n\nIl presidente ... l'economia dal"],
'output_text': "\n\nIl presidente ... l'economia dal"}
The map-rerank
Chain
本节展示使用 map-rerank
链和源进行问答的结果。
chain = load_qa_chain(
OpenAI(temperature=0),
chain_type="map_rerank",
return_intermediate_steps=True
)
query = "What did the president say about Justice Breyer"
results = chain({"input_documents": docs, "question": query},
return_only_outputs=True)
results["output_text"]
' The President thanked Justice Breyer for his service and honored him for dedicating his life to serve the country.'
results["intermediate_steps"]
[{'answer': ' The President thanked Justice Breyer for his service and honored him for dedicating his life to serve the country.',
'score': '100'},
{'answer': ' This document does not answer the question', 'score': '0'},
{'answer': ' This document does not answer the question', 'score': '0'},
{'answer': ' This document does not answer the question', 'score': '0'}]
Custom Prompts
您还可以使用自己的提示与此链一起使用。在这个示例中,我们将用意大利语回答。
from langchain.output_parsers import RegexParser
output_parser = RegexParser(
regex=r"(.*?)\nScore: (.*)",
output_keys=["answer", "score"],
)
prompt_template = """Use the following pieces of context to answer the question at the end. If you don't know the answer, just say that you don't know, don't try to make up an answer.
In addition to giving an answer, also return a score of how fully it answered the user's question. This should be in the following format:
Question: [question here]
Helpful Answer In Italian: [answer here]
Score: [score between 0 and 100]
Begin!
Context:
---------
{context}
---------
Question: {question}
Helpful Answer In Italian:"""
PROMPT = PromptTemplate(
template=prompt_template,
input_variables=["context", "question"],
output_parser=output_parser,
)
chain = load_qa_chain(OpenAI(temperature=0), chain_type="map_rerank", return_intermediate_steps=True, prompt=PROMPT)
query = "What did the president say about Justice Breyer"
chain({"input_documents": docs, "question": query},
return_only_outputs=True)
{'intermediate_steps': [{'answer': ' Il presidente ha detto che Justice Breyer ha dedicato la sua vita a servire questo paese.',
'score': '100'},
{'answer': ' Il presidente non ha detto nulla sulla Giustizia Breyer.',
'score': '100'},
{'answer': ' Non so.', 'score': '0'},
{'answer': ' Non so.', 'score': '0'}],
'output_text': ' Il presidente ha detto che Justice Breyer ha dedicato la sua vita a servire questo paese.'}
带有来源的文档问答
我们还可以进行文档问答,并返回用于回答问题的来源。
为了做到这一点,我们只需要确保每个文档的元数据中有一个 “source” 键,并使用 load_qa_with_sources
助手来构建我们的链:
docsearch = Chroma.from_texts(texts, embeddings,
metadatas=[{"source": str(i)} for i in range(len(texts))])
query = "What did the president say about Justice Breyer"
docs = docsearch.similarity_search(query)
from langchain.chains.qa_with_sources import load_qa_with_sources_chain
chain = load_qa_with_sources_chain(OpenAI(temperature=0), chain_type="stuff")
query = "What did the president say about Justice Breyer"
chain({"input_documents": docs, "question": query},
return_only_outputs=True)
{'output_text': ' The president thanked Justice Breyer for his service.\nSOURCES: 30-pl'}
36、标记
标记链使用OpenAI的functions
参数来指定用于标记文档的模式。
这帮助我们确保模型输出我们想要的准确标记及其适当的类型。
当我们想要给一段落打标签时,可以使用标记链来指定特定的属性(例如,这条信息的情感是什么?)
from langchain.chat_models import ChatOpenAI
from langchain.chains import create_tagging_chain, create_tagging_chain_pydantic
from langchain.prompts import ChatPromptTemplate
llm = ChatOpenAI(temperature=0, model="gpt-3.5-turbo-0613")
最简单的方法,只指定类型
我们可以通过在模式中指定一些属性及其预期类型来开始
schema = {
"properties": {
"sentiment": {"type": "string"},
"aggressiveness": {"type": "integer"},
"language": {"type": "string"},
}
}
chain = create_tagging_chain(schema, llm)
正如我们在示例中看到的,它正确地解释了我们的需求,但结果可能会不同,例如不同语言的情感(‘positive’、'enojado’等)。
我们将在下一节中看到如何控制这些结果。
inp = "Estoy increiblemente contento de haberte conocido! Creo que seremos muy buenos amigos!"
chain.run(inp)
# -> {'sentiment': 'positive', 'language': 'Spanish'}
inp = "Estoy muy enojado con vos! Te voy a dar tu merecido!"
chain.run(inp)
# -> {'sentiment': 'enojado', 'aggressiveness': 1, 'language': 'Spanish'}
inp = "Weather is ok here, I can go outside without much more than a coat"
chain.run(inp)
# -> {'sentiment': 'positive', 'aggressiveness': 0, 'language': 'English'}
更多控制
通过明确定义模式,我们可以更好地控制模型输出。
具体而言,我们可以定义以下内容:
- 每个属性的可能值
- 描述以确保模型理解属性
- 返回的必需属性
以下是如何使用enum、description和required来控制每个方面的示例:
schema = {
"properties": {
"sentiment": {"type": "string", "enum": ["happy", "neutral", "sad"]},
"aggressiveness": {
"type": "integer",
"enum": [1, 2, 3, 4, 5],
"description": "描述句子的伤害程度,数字越大,伤害性越大",
},
"language": {
"type": "string",
"enum": ["spanish", "english", "french", "german", "italian"],
},
},
"required": ["language", "sentiment", "aggressiveness"],
}
chain = create_tagging_chain(schema, llm)
inp = "Estoy increiblemente contento de haberte conocido! Creo que seremos muy buenos amigos!"
chain.run(inp)
# -> {'sentiment': 'happy', 'aggressiveness': 0, 'language': 'spanish'}
inp = "Estoy muy enojado con vos! Te voy a dar tu merecido!"
chain.run(inp)
# -> {'sentiment': 'sad', 'aggressiveness': 10, 'language': 'spanish'}
inp = "Weather is ok here, I can go outside without much more than a coat"
chain.run(inp)
# -> {'sentiment': 'neutral', 'aggressiveness': 0, 'language': 'english'}
使用 Pydantic 指定模式
我们还可以使用Pydantic模式来指定所需的属性和类型。
我们还可以发送其他参数,例如’enum’或’description’,如下所示的示例中所示。
通过使用create_tagging_chain_pydantic
函数,我们可以将Pydantic模式作为输入发送,并且输出将是符合我们期望的模式的实例化对象。
通过这种方式,我们可以像在Python中定义新类或函数一样定义我们的模式-使用纯净的Python类型。
from enum import Enum
from pydantic import BaseModel, Field
class Tags(BaseModel):
sentiment: str = Field(..., enum=["happy", "neutral", "sad"])
aggressiveness: int = Field(
...,
description="描述句子的伤害程度,数字越大,伤害性越大",
enum=[1, 2, 3, 4, 5],
)
language: str = Field(
..., enum=["spanish", "english", "french", "german", "italian"]
)
chain = create_tagging_chain_pydantic(Tags, llm)
inp = "Estoy muy enojado con vos! Te voy a dar tu merecido!"
res = chain.run(inp)
37、Vector store-augmented text generation
这本笔记本介绍了如何使用LangChain在矢量索引上生成文本。
如果我们想要生成 能够从大量自定义文本中提取的文本,例如,生成理解以前写的博客文章的博客文章,或者可以 参考产品文档的产品教程,这是非常有用的。
Prepare Data
首先,我们准备数据。在这个例子中,我们获取一个文档站点,该站点由托管在Github上的markdown文件组成,并将它们拆分为足够小的Documents。
from langchain.llms import OpenAI
from langchain.docstore.document import Document
import requests
from langchain.embeddings.openai import OpenAIEmbeddings
from langchain.vectorstores import Chroma
from langchain.text_splitter import CharacterTextSplitter
from langchain.prompts import PromptTemplate
import pathlib
import subprocess
import tempfile
def get_github_docs(repo_owner, repo_name):
with tempfile.TemporaryDirectory() as d:
subprocess.check_call(
f"git clone --depth 1 https://github.com/{repo_owner}/{repo_name}.git .",
cwd=d,
shell=True,
)
git_sha = (
subprocess.check_output("git rev-parse HEAD", shell=True, cwd=d)
.decode("utf-8")
.strip()
)
repo_path = pathlib.Path(d)
markdown_files = list(repo_path.glob("*/*.md")) + list(
repo_path.glob("*/*.mdx")
)
for markdown_file in markdown_files:
with open(markdown_file, "r") as f:
relative_path = markdown_file.relative_to(repo_path)
github_url = f"https://github.com/{repo_owner}/{repo_name}/blob/{git_sha}/{relative_path}"
yield Document(page_content=f.read(), metadata={"source": github_url})
sources = get_github_docs("yirenlu92", "deno-manual-forked")
source_chunks = []
splitter = CharacterTextSplitter(separator=" ", chunk_size=1024, chunk_overlap=0)
for source in sources:
for chunk in splitter.split_text(source.page_content):
source_chunks.append(Document(page_content=chunk, metadata=source.metadata))
Set Up Vector DB
现在我们已经将文档内容分为块,让我们将所有这些信息放在矢量索引中,以便检索。
search_index = Chroma.from_documents(source_chunks, OpenAIEmbeddings())
Set Up LLM Chain with Custom Prompt
接下来,让我们设置一个简单的LLM链,但为其提供一个用于生成博客文章的自定义提示。
请注意,自定义提示是参数化的,并接受两个输入:
`context’,它将是从矢量搜索中提取的文档,以及由用户给出的“topic”。
from langchain.chains import LLMChain
prompt_template = """Use the context below to write a 400 word blog post about the topic below:
Context: {context}
Topic: {topic}
Blog post:"""
PROMPT = PromptTemplate(template=prompt_template, input_variables=["context", "topic"])
llm = OpenAI(temperature=0)
chain = LLMChain(llm=llm, prompt=PROMPT)
Generate Text
最后,我们编写一个函数,将我们的输入应用于链。
该函数接受一个输入参数 topic
。
我们在向量索引中找到与该 topic
对应的文档,并将它们用作简单LLM链中的附加上下文。
def generate_blog_post(topic):
docs = search_index.similarity_search(topic, k=4)
inputs = [{"context": doc.page_content, "topic": topic} for doc in docs]
print(chain.apply(inputs))
generate_blog_post("environment variables")
[{'text': '\n\nEnvironment variables ...This is a great'},
...
{'text': '\n\nEnvironment variables are an important ... following code:\n\n```'}]
2024-04-09(二)