==========代码: from paperqa import Docs, LangchainVectorStore, LangchainEmbeddingModel, SentenceTransformerEmbeddingModel, SparseEmbeddingModel from langchain_community.chat_models import ChatZhipuAI from langchain_community.embeddings import ZhipuAIEmbeddings, QianfanEmbeddingsEndpoint, SparkLLMTextEmbeddings from langchain_community.vectorstores.faiss import FAISS # get a list of paths my_docs = [ '/code/火星峡谷网研究的现状_问题与展望_史语桐.pdf', # './bert.pdf', ] # my_online_docs = [ # 'https://arxiv.org/pdf/1810.04805', # ] docs = Docs( llm="langchain", client=ChatZhipuAI(), # 使用本地的embeding库 embedding="sentence-transformers" # 使用线上模型 # embedding="langchain", # embedding_client=ZhipuAIEmbeddings(model="embedding-3"), # 有bug # embedding_client=SparkLLMTextEmbeddings() # embedding_client=QianfanEmbeddingsEndpoint() # 可用,但连接有时候不稳定 ) # 本地文件读取 for d in my_docs: docs.add(path=d, chunk_chars=300, disable_check=True) # 网络文件读取 # for u in my_online_docs: # docs.add_url(url=u, chunk_chars=300) # answer = docs.query( # "What are the key features that differentiate BERT from previous language representation models?" # ) answer = docs.query( "这篇文章讲的什么内容?请用中文回答" ) print(answer.formatted_answer) ==========报错: root@pcb8d0abed7b40e68b6eb7e53ce17548-task0-0:/# cd / ; /usr/bin/env /opt/conda/bin/python /root/.local/share/code-server/extensions/ms-python.debugpy-2024.8.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher 55121 -- /code/chat-paper-glm4.py /opt/conda/lib/python3.10/site-packages/huggingface_hub/file_download.py:1132: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`. warnings.warn( Ignoring wrong pointing object 12 0 (offset 0) Ignoring wrong pointing object 59 0 (offset 0) Ignoring wrong pointing object 75 0 (offset 0) Ignoring wrong pointing object 77 0 (offset 0) Ignoring wrong pointing object 107 0 (offset 0) Traceback (most recent call last): File "/code/chat-paper-glm4.py", line 31, in docs.add(path=d, chunk_chars=300, disable_check=True) File "/code/paper-qa/paperqa/docs.py", line 353, in add return loop.run_until_complete( File "/opt/conda/lib/python3.10/asyncio/base_events.py", line 649, in run_until_complete return future.result() File "/code/paper-qa/paperqa/docs.py", line 388, in aadd chain_result = await cite_chain({"text": texts[0].text}, None) File "/code/paper-qa/paperqa/llms.py", line 305, in execute output = await self.achat(client, messages) File "/code/paper-qa/paperqa/llms.py", line 731, in achat return (await client.ainvoke(lc_messages, **self.config)).content File "/opt/conda/lib/python3.10/site-packages/langchain_core/language_models/chat_models.py", line 307, in ainvoke llm_result = await self.agenerate_prompt( File "/opt/conda/lib/python3.10/site-packages/langchain_core/language_models/chat_models.py", line 796, in agenerate_prompt return await self.agenerate( File "/opt/conda/lib/python3.10/site-packages/langchain_core/language_models/chat_models.py", line 756, in agenerate raise exceptions[0] File "/opt/conda/lib/python3.10/site-packages/langchain_core/language_models/chat_models.py", line 924, in _agenerate_with_cache result = await self._agenerate( File "/opt/conda/lib/python3.10/site-packages/langchain_community/chat_models/zhipuai.py", line 653, in _agenerate response.raise_for_status() File "/opt/conda/lib/python3.10/site-packages/httpx/_models.py", line 761, in raise_for_status raise HTTPStatusError(message, request=request, response=self) httpx.HTTPStatusError: Client error '429 Too Many Requests' for url 'https://open.bigmodel.cn/api/paas/v4/chat/completions' For more information check: https://developer.mozilla.org/en-US/docs/Web/HTTP/Status/429 root@pcb8d0abed7b40e68b6eb7e53ce17548-task0-0:/#