ports/misc/llama-cpp/pkg-message
Yuri Victorovich dd116a4403 misc/llama-cpp: Add the RC service llama-server
... that allows to communicate with AI models in a browser.
Submitted by:	Daniel Morante <daniel@morante.net> (initial version of the service file)
2025-04-08 19:41:51 -07:00

28 lines
651 B
Plaintext

[
{ type: install
message: <<EOM
You installed LLaMA-cpp: Facebook's LLaMA model runner.
In order to experience LLaMA-cpp please download some
AI model in the GGUF format, for example from huggingface.com,
run the script below, and open localhost:9011 in your browser
to communicate with this AI model.
$ llama-server -m $MODEL \
--host 0.0.0.0 \
--port 9011 \
-ngl 15
or
you can add the following lines to /etc/rc.conf,
start the llama-server service,
and navigate to http://localhost:8080:
> llama_server_enable=YES
> llama_server_model=/path/to/models/llama-2-7b-chat.Q4_K_M.gguf
> llama_server_args="--device Vulkan0 -ngl 27"
EOM
}
]