You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
At the moment I've seen two different scenarios:
A) app crashes (ex. Llama-3.2-1B-Instruct-Q4_0_4_4)
B) initLlama fails with the following error (ex. Llama-3.2-3B-Instruct-Q6_K)
C) crash: app killed by system because it is using too much memory; even with Extended Virtual Addressing capability enabled (ex. Llama-3.2-3B-Instruct-Q4_0)
I've investigated scenario A), apparently ggml-metal throws the following error MUL MAT-MAT not implemented through function LM_GGML_ABORT, that as far as I understand means that there is no matrix multiplication available for matrices involved in the kind of quantized species used.
This scenario seems not to be managed by llama.rn, in my opinion a good solution would be a more verbose error message + not crashing. I will try to fix this in the weekend.
As per scenario B) and C) I'm still trying to understand what's going on, posting it here in case anyone else is having these same problems.
The text was updated successfully, but these errors were encountered:
Loading models with n_gpu_layers: 0 works fine but otherwise doesn't work.
I'm using an iPhone 13 with iOS 18.1.
Minimum reproducible code is almost same as the example, but for the record:
At the moment I've seen two different scenarios:
A) app crashes (ex. Llama-3.2-1B-Instruct-Q4_0_4_4)
B) initLlama fails with the following error (ex. Llama-3.2-3B-Instruct-Q6_K)
C) crash: app killed by system because it is using too much memory; even with Extended Virtual Addressing capability enabled (ex. Llama-3.2-3B-Instruct-Q4_0)
I've investigated scenario A), apparently ggml-metal throws the following error
MUL MAT-MAT not implemented
through functionLM_GGML_ABORT
, that as far as I understand means that there is no matrix multiplication available for matrices involved in the kind of quantized species used.This scenario seems not to be managed by llama.rn, in my opinion a good solution would be a more verbose error message + not crashing. I will try to fix this in the weekend.
As per scenario B) and C) I'm still trying to understand what's going on, posting it here in case anyone else is having these same problems.
The text was updated successfully, but these errors were encountered: