.. |
__init__.py
|
feat: server multi models support (#799)
|
2023-08-12 00:57:00 +08:00 |
anthropic_provider.py
|
feat: advanced prompt backend (#1301)
|
2023-10-12 10:13:10 -05:00 |
azure_openai_provider.py
|
feat: advanced prompt backend (#1301)
|
2023-10-12 10:13:10 -05:00 |
baichuan_provider.py
|
feat: advanced prompt backend (#1301)
|
2023-10-12 10:13:10 -05:00 |
base.py
|
feat: advanced prompt backend (#1301)
|
2023-10-12 10:13:10 -05:00 |
chatglm_provider.py
|
feat: advanced prompt backend (#1301)
|
2023-10-12 10:13:10 -05:00 |
hosted.py
|
feat: add zhipuai (#1188)
|
2023-09-18 17:32:31 +08:00 |
huggingface_hub_provider.py
|
feat: advanced prompt backend (#1301)
|
2023-10-12 10:13:10 -05:00 |
localai_provider.py
|
feat: advanced prompt backend (#1301)
|
2023-10-12 10:13:10 -05:00 |
minimax_provider.py
|
feat: advanced prompt backend (#1301)
|
2023-10-12 10:13:10 -05:00 |
openai_provider.py
|
fix: max tokens of OpenAI gpt-3.5-turbo-instruct to 4097 (#1338)
|
2023-10-13 02:07:07 -05:00 |
openllm_provider.py
|
feat: advanced prompt backend (#1301)
|
2023-10-12 10:13:10 -05:00 |
replicate_provider.py
|
feat: advanced prompt backend (#1301)
|
2023-10-12 10:13:10 -05:00 |
spark_provider.py
|
feat: advanced prompt backend (#1301)
|
2023-10-12 10:13:10 -05:00 |
tongyi_provider.py
|
feat: advanced prompt backend (#1301)
|
2023-10-12 10:13:10 -05:00 |
wenxin_provider.py
|
feat: advanced prompt backend (#1301)
|
2023-10-12 10:13:10 -05:00 |
xinference_provider.py
|
feat: use xinference client instead of xinference (#1339)
|
2023-10-13 02:46:09 -05:00 |
zhipuai_provider.py
|
feat: advanced prompt backend (#1301)
|
2023-10-12 10:13:10 -05:00 |