
from langchain.text_splitter import RecursiveCharacterTextSplitter
#加载要切割的文档
with open("test.txt") as f:
zuizhonghuanxiang = f.read()
#初始化切割器
text_splitter = RecursiveCharacterTextSplitter(
chunk_size=50,#切分的文本块大小,一般通过长度函数计算
chunk_overlap=20,#切分的文本块重叠大小,一般通过长度函数计算
length_function=len,#长度函数,也可以传递tokenize函数
add_start_index=True,#是否添加起始索引
)
text = text_splitter.create_documents([zuizhonghuanxiang])
text[0]
text[1]from langchain.text_splitter import CharacterTextSplitter
#加载要切分的文档
with open("test.txt") as f:
zuizhonghuanxiang = f.read()
#初始化切分器
text_splitter = CharacterTextSplitter(
separator="。",#切割的标志字符,默认是\n\n
chunk_size=50,#切分的文本块大小,一般通过长度函数计算
chunk_overlap=20,#切分的文本块重叠大小,一般通过长度函数计算
length_function=len,#长度函数,也可以传递tokenize函数
add_start_index=True,#是否添加起始索引
is_separator_regex=False,#是否是正则表达式
)
text = text_splitter.create_documents([zuizhonghuanxiang])
print(text[0])
from langchain.text_splitter import (
RecursiveCharacterTextSplitter,
Language,
)
#支持解析的编程语言
#[e.value for e in Language]
#要切割的代码文档
PYTHON_CODE = """
def hello_world():
print("Hello, World!")
#调用函数
hello_world()
"""
py_spliter = RecursiveCharacterTextSplitter.from_language(
language=Language.PYTHON,
chunk_size=50,
chunk_overlap=10,
)
python_docs = py_spliter.create_documents([PYTHON_CODE])
python_docsfrom langchain.text_splitter import CharacterTextSplitter
#要切割的文档
with open("test.txt") as f:
zuizhonghuanxiang = f.read()
#初始化切分器
text_splitter = CharacterTextSplitter.from_tiktoken_encoder(
chunk_size=4000,#切分的文本块大小,一般通过长度函数计算
chunk_overlap=30,#切分的文本块重叠大小,一般通过长度函数计算
)
text = text_splitter.create_documents([zuizhonghuanxiang])
print(text[0])
先装包:
! pip install doctran==0.0.14
先加载文档:
with open("letter.txt") as f:
content = f.read()from dotenv import load_dotenv
import os
load_dotenv("openai.env")
OPENAI_API_KEY = os.environ.get("OPEN_API_KEY")
OPENAI_API_BASE = os.environ.get("OPENAI_API_BASE")
OPENAI_MODEL = "gpt-3.5-turbo-16k"
OPENAI_TOKEN_LIMIT = 8000
from doctran import Doctran
doctrans = Doctran(
openai_api_key=OPENAI_API_KEY,
openai_model=OPENAI_MODEL,
openai_token_limit=OPENAI_TOKEN_LIMIT,
)
documents = doctrans.parse(content=content)summary = documents.summarize(token_limit=100).execute()
print(summary.transformed_content)
translation = documents.translate(language="chinese").execute()
print(translation.transformed_content)
删除除了某个主题或关键词之外的内容,仅保留与主题相关的内容
refined = documents.refine(topics=["marketing","Development"]).execute()
print(refined.transformed_content)
关注我,紧跟本系列专栏文章,咱们下篇再续!
作者简介:魔都架构师,多家大厂后端一线研发经验,在分布式系统设计、数据平台架构和AI应用开发等领域都有丰富实践经验。 各大技术社区头部专家博主。具有丰富的引领团队经验,深厚业务架构和解决方案的积累。 负责:
参考: