You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
here is a screenshot to demonstrate this issue(the input question is: where is China's capital city?):
same issue here with following models on Xiaomi 14 (Xiaomi 14 is available since Oct 2023, Xiaomi 14 contains a very very very powerful mobile SoC ------ Qualcomm SM8650-AB Snapdragon 8 Gen 3 (4 nm) ------ and was used for personal device-side AI PoC development activity).
zhouwg
changed the title
[llama.cpp] AI answer does not stop automatically when inference is launched
[llama.cpp] AI answer does not stop automatically when inference is launched on Android phone
Apr 15, 2024
CPU-only reference on Xiaomi 14.
here is a screenshot to demonstrate this issue(the input question is: where is China's capital city?):
same issue here with following models on Xiaomi 14 (Xiaomi 14 is available since Oct 2023, Xiaomi 14 contains a very very very powerful mobile SoC ------ Qualcomm SM8650-AB Snapdragon 8 Gen 3 (4 nm) ------ and was used for personal device-side AI PoC development activity).
original similar issue report could be found at upstream llama.cpp:
Baichuan2-7B-Chat model converted to ggml-model-q4_0.gguf, AI answer does not stop automatically when inference is made ggerganov/llama.cpp#5034
Infinite loop of "context shift" ggerganov/llama.cpp#3969
The text was updated successfully, but these errors were encountered: