Llama.cpp server support

#5
by vigneshR - opened

Hi

Thanks for the excellent work. I am trying to see if it is possible to run the gguf version as a Llama CPP http server? Could you give me an example of how I can run it please?

Thanks

I have the same problem. Otherwise, the efficiency is too low.

OpenBMB org

Hi
we are glad that our work can provide some help to the open source community.
However, we have limited manpower and may continue to support it for some time in the future, but we are not yet able to give any clear time point.

I can run it in normal (text) mode with a recent build of llama.cpp main and server programs.
Image loading not working yet.

Sign up or log in to comment