1.0 KiB
1.0 KiB
llm-hosting
This is an extended article to not have to write everything in the main readme. This chapter takes care of hosting llm models on the server.
deploy
kubectl apply -f llm/llama_cpp_hosting.yaml
development
links
Two examples of model files that are currently tried out: