Commit Graph

30 Commits

Author SHA1 Message Date
takatost
4dfbcd0b4e
feat: support chatglm_turbo model #1443 (#1460) 2023-11-06 04:33:05 -06:00
takatost
7c9b585a47
feat: support weixin ernie-bot-4 and chat mode (#1375) 2023-10-18 02:35:24 -05:00
takatost
3efaa713da
feat: use xinference client instead of xinference (#1339) 2023-10-13 02:46:09 -05:00
takatost
f4be2b8bcd
fix: raise error in minimax stream generate (#1336) 2023-10-12 23:48:28 -05:00
takatost
2851a9f04e
feat: optimize minimax llm call (#1312) 2023-10-11 07:17:41 -05:00
takatost
875dfbbf0e
fix: openllm completion start with prompt, remove it (#1303) 2023-10-10 04:44:19 -05:00
takatost
4ab4bcc074
feat: support openllm embedding (#1293) 2023-10-09 23:09:35 -05:00
takatost
1d4f019de4
feat: add baichuan llm support (#1294)
Co-authored-by: zxhlyh <jasonapring2015@outlook.com>
2023-10-09 23:09:26 -05:00
Garfield Dai
e409895c02
Feat/huggingface embedding support (#1211)
Co-authored-by: StyleZhang <jasonapring2015@outlook.com>
2023-09-22 13:59:02 +08:00
takatost
ae3f1ac0a9
feat: support gpt-3.5-turbo-instruct model (#1195) 2023-09-19 02:05:04 +08:00
takatost
827c97f0d3
feat: add zhipuai (#1188) 2023-09-18 17:32:31 +08:00
takatost
c4d8bdc3db
fix: hf hosted inference check (#1128) 2023-09-09 00:29:48 +08:00
takatost
d75e8aeafa
feat: disable anthropic retry (#1067) 2023-08-31 16:44:46 +08:00
takatost
2eba98a465
feat: optimize anthropic connection pool (#1066) 2023-08-31 16:18:59 +08:00
takatost
417c19577a
feat: add LocalAI local embedding model support (#1021)
Co-authored-by: StyleZhang <jasonapring2015@outlook.com>
2023-08-29 22:22:02 +08:00
takatost
0796791de5
feat: hf inference endpoint stream support (#1028) 2023-08-26 19:48:34 +08:00
Uranus
2d9616c29c
fix: xinference last token being ignored (#1013) 2023-08-25 18:15:05 +08:00
takatost
9ae91a2ec3
feat: optimize xinference request max token key and stop reason (#998) 2023-08-24 18:11:15 +08:00
takatost
bd3a9b2f8d
fix: xinference-chat-stream-response (#991) 2023-08-24 14:39:34 +08:00
takatost
18d3877151
feat: optimize xinference stream (#989) 2023-08-24 13:58:34 +08:00
takatost
a76fde3d23
feat: optimize hf inference endpoint (#975) 2023-08-23 19:47:50 +08:00
takatost
78d3aa5fcd
fix: embedding init err (#956) 2023-08-22 17:43:59 +08:00
takatost
4f3053a8cc
fix: xinference chat completion error (#952) 2023-08-22 15:58:04 +08:00
takatost
866ee5da91
fix: openllm generate cutoff (#945) 2023-08-22 13:43:36 +08:00
takatost
e0a48c4972
fix: xinference chat support (#939) 2023-08-21 20:44:29 +08:00
takatost
6c832ee328
fix: remove openllm pypi package because of this package too large (#931) 2023-08-21 02:12:28 +08:00
takatost
f42e7d1a61
feat: add spark v2 support (#885) 2023-08-17 15:08:57 +08:00
takatost
c4d759dfba
fix: wenxin error not raise when stream mode (#884) 2023-08-17 13:40:00 +08:00
takatost
cc52cdc2a9
Feat/add free provider apply (#829) 2023-08-14 12:44:35 +08:00
takatost
5fa2161b05
feat: server multi models support (#799) 2023-08-12 00:57:00 +08:00