This repository has been archived by the owner on Jun 24, 2024. It is now read-only.
Runs out of ggml context's memory pool on larger models #115
Labels
issue:bug
Something isn't working
Running LLaMA 30B and 65B models:
ggml_new_tensor_impl: not enough space in the context's memory pool (needed 1073742848, available 1073741824)
followed by Segmentation Fault.
To reproduce, run following: (sorry for silly prompt (mostly taken from llama.cpp examples), it's almost what my current project "poop-gpt" uses.)
The text was updated successfully, but these errors were encountered: