Loading request...
Currently, chat and embedding models are hardcoded. Request to dynamically query and list all available models (e.g., via '/models' endpoint for OpenAI-compatible APIs or by allowing users to select any GGUF model for Ollama embeddings), which would improve flexibility, especially for aggregated API services like One-API.
1,各个模型貌似是写死的,比如MOONSHOT_MODELS中就只有 "moonshot-v1-8k", "moonshot-v1-32k", "moonshot-v1-128k";我记得各家兼容openai的都有/models ,可以query,直接查询得到所有的models,这样有个好处,比如使用的是one-api这种聚合类的api,就一次把所有模型都接进来了。不用设置那么多。 2,embedding模型貌似也是只能选写死的几个,Ollama可以载入所有的gguf模型,这面也包括embedding模型,后期应该让用户可以自己选。