- Notifications
You must be signed in to change notification settings - Fork2.4k
Closed as not planned
Description
The problem is: I only can see the output calculate by cpu, if I switch using device_type mps, it can run ok, but I cannot see the output, when I pprint(memory.load_memory_variables({}))
there are many characters like this:
'AI: \x1c' '\x1c' '\x1c' '\x1c' '\x1c' '\x1c' '\x1c' '\x1c' '\x1c' '\x1c' '\x1c' '\x1c' '\x1c' '\x1c' '\x1c' '\x1c' '\x1c' '\x1c' '\x1c' '\x1c' '\x1c' '\x1c' '\x1c' '\x1c' '\x1c' '\x1c' '\x1c'
My environment is this:
CPU and GPU: Apple M1 16G
Python version: Python 3.11.4
llama-cpp-python == 0.1.78
I have tried to only run llama without langchain, the mps is run ok, and can output string, mps fast than cpu.
So I don't know the problem is langchain or somewhat, and I don't know how to fix this, I really don't wanna use cpu, it's slow and make my Mac very hot.
Metadata
Metadata
Assignees
Labels
No labels