Most llama.cpp openai servers are pretty close to vanilla llama.cpp, albeit without the batching support.