I want to use embedding model services provided by Ollama or vLLM, but the current configuration file doesn't allow me to specify them directly. I tried modifying server.py, and it seems to work, but I feel like my code isn't very clean or standard. Does the official team have any plans to support this?