You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
$ ./build/bin/llama-cli --version
register_backend: registered backend Metal (1 devices)
register_device: registered device Metal (Apple M3)
register_backend: registered backend BLAS (1 devices)
register_device: registered device BLAS (Accelerate)
register_backend: registered backend CPU (1 devices)
register_device: registered device CPU (Apple M3)
version: 4830 (d78b4df0)
built with Apple clang version 16.0.0 (clang-1600.0.26.6) for arm64-apple-darwin23.6.0
Operating systems
Mac
Which llama.cpp modules do you know to be affected?
Other (Please specify in the next section)
Command line
Problem description & steps to reproduce
When running the LlamaSwiftUI example in a simulator and trying to load a model, the following error occurs:
ggml_metal_init: loaded kernel_ssm_conv_f32 0x0 | th_max = 0 | th_width = 0ggml_metal_init: error: load pipeline error: Error Domain=CompilerError Code=2 "only 14 constant buffers binding are supported in the simulator but 16 were used" UserInfo={NSLocalizedDescription=only 14 constant buffers binding are supported in the simulator but 16 were used}ggml_backend_metal_device_init: error: failed to allocate contextllama_init_from_model: failed to initialize Metal backendCould not load context!Error: The operation couldn’t be completed. (llama_swiftui.LlamaError error 0.)
This seems to be coming from the kernel_ssm_conv_f32 kernel function:
This kernel function has 18 parameters and 16 are constant buffers. But the limit for
the simulator seems to be 14. I tried simply adding structs for the kernel functions that it complained about and this allowed the example to run (load a model and perform inference).
There is already an open issue and ongoing work to create these structs. I'm just opening this issue for awareness/tracking.
First Bad Commit
No response
Relevant log output
The text was updated successfully, but these errors were encountered:
Name and Version
$ ./build/bin/llama-cli --version
register_backend: registered backend Metal (1 devices)
register_device: registered device Metal (Apple M3)
register_backend: registered backend BLAS (1 devices)
register_device: registered device BLAS (Accelerate)
register_backend: registered backend CPU (1 devices)
register_device: registered device CPU (Apple M3)
version: 4830 (d78b4df0)
built with Apple clang version 16.0.0 (clang-1600.0.26.6) for arm64-apple-darwin23.6.0
Operating systems
Mac
Which llama.cpp modules do you know to be affected?
Other (Please specify in the next section)
Command line
Problem description & steps to reproduce
When running the LlamaSwiftUI example in a simulator and trying to load a model, the following error occurs:
This seems to be coming from the
kernel_ssm_conv_f32
kernel function:This kernel function has 18 parameters and 16 are constant buffers. But the limit for
the simulator seems to be 14. I tried simply adding structs for the kernel functions that it complained about and this allowed the example to run (load a model and perform inference).
There is already an open issue and ongoing work to create these structs. I'm just opening this issue for awareness/tracking.
First Bad Commit
No response
Relevant log output
The text was updated successfully, but these errors were encountered: