We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
问题描述 / Problem Description 调用agent查询本地知识库报错: chatchat终端报错信息:
Entering new AgentExecutor chain... INFO: 127.0.0.1:38928 - "POST /v1/chat/completions HTTP/1.1" 200 OK 2024-08-09 09:18:30,390 httpx 22987 INFO HTTP Request: POST http://127.0.0.1:7861/v1/chat/completions "HTTP/1.1 200 OK" 2024-08-09 09:18:30,398 httpx 22987 INFO HTTP Request: POST http://127.0.0.1:9997/v1/chat/completions "HTTP/1.1 200 OK" 2024-08-09 09:18:31.734 | ERROR | chatchat.server.api_server.openai_routes:generator:105 - openai request error: An error occurred during streaming 2024-08-09 09:18:31.758 | ERROR | chatchat.server.utils:wrap_done:46 - APIError: Caught exception: An error occurred during streaming
chatchat 可视化web页面报错信息: 问:大模型在推理过程中有哪些优化显存的策略 答:这个问题涉及到大模型的优化策略,我需要调用一下大模型优化的知识库来回答这个问题。 An error occurred during streaming
xinference报错信息: 2024-08-09 09:18:31,730 xinference.api.restful_api 11477 ERROR Chat completion stream got an error: [address=0.0.0.0:45209, pid=12401] unhashable type: 'slice' Traceback (most recent call last): File "/home/envs/xinference/lib/python3.10/site-packages/xinference/model/llm/utils.py", line 587, in _to_chat_completion_chunks for i, chunk in enumerate(chunks): File "/home/envs/xinference/lib/python3.10/site-packages/xinference/model/llm/pytorch/chatglm.py", line 497, in _stream_generator chunk_text = chunk_text[last_chunk_text_length:] TypeError: [address=0.0.0.0:45209, pid=12401] unhashable type: 'slice'
复现问题的步骤 / Steps to Reproduce
环境信息 / Environment Information
The text was updated successfully, but these errors were encountered:
我也是,用的chatglm3 + bge large 也是报错An error occurred during streaming 换成qwen虽然不报错 但不知道为什么卡的不行
Sorry, something went wrong.
有解决办法嘛
有解决办法嘛 不用transformers换成vllm启动试试
这个问题已经被标记为 stale ,因为它已经超过 30 天没有任何活动。
stale
这个问题已经被自动关闭,因为它被标为 stale 后超过 14 天没有任何活动。
No branches or pull requests
问题描述 / Problem Description
调用agent查询本地知识库报错:
chatchat终端报错信息:
chatchat 可视化web页面报错信息:
问:大模型在推理过程中有哪些优化显存的策略
答:这个问题涉及到大模型的优化策略,我需要调用一下大模型优化的知识库来回答这个问题。
An error occurred during streaming
xinference报错信息:
2024-08-09 09:18:31,730 xinference.api.restful_api 11477 ERROR Chat completion stream got an error: [address=0.0.0.0:45209, pid=12401] unhashable type: 'slice'
Traceback (most recent call last):
File "/home/envs/xinference/lib/python3.10/site-packages/xinference/model/llm/utils.py", line 587, in _to_chat_completion_chunks
for i, chunk in enumerate(chunks):
File "/home/envs/xinference/lib/python3.10/site-packages/xinference/model/llm/pytorch/chatglm.py", line 497, in _stream_generator
chunk_text = chunk_text[last_chunk_text_length:]
TypeError: [address=0.0.0.0:45209, pid=12401] unhashable type: 'slice'
复现问题的步骤 / Steps to Reproduce
环境信息 / Environment Information
The text was updated successfully, but these errors were encountered: