takatost
|
d75e8aeafa
|
feat: disable anthropic retry (#1067)
|
2023-08-31 16:44:46 +08:00 |
|
takatost
|
2eba98a465
|
feat: optimize anthropic connection pool (#1066)
|
2023-08-31 16:18:59 +08:00 |
|
takatost
|
417c19577a
|
feat: add LocalAI local embedding model support (#1021)
Co-authored-by: StyleZhang <jasonapring2015@outlook.com>
|
2023-08-29 22:22:02 +08:00 |
|
takatost
|
0796791de5
|
feat: hf inference endpoint stream support (#1028)
|
2023-08-26 19:48:34 +08:00 |
|
Uranus
|
2d9616c29c
|
fix: xinference last token being ignored (#1013)
|
2023-08-25 18:15:05 +08:00 |
|
takatost
|
9ae91a2ec3
|
feat: optimize xinference request max token key and stop reason (#998)
|
2023-08-24 18:11:15 +08:00 |
|
takatost
|
bd3a9b2f8d
|
fix: xinference-chat-stream-response (#991)
|
2023-08-24 14:39:34 +08:00 |
|
takatost
|
18d3877151
|
feat: optimize xinference stream (#989)
|
2023-08-24 13:58:34 +08:00 |
|
takatost
|
a76fde3d23
|
feat: optimize hf inference endpoint (#975)
|
2023-08-23 19:47:50 +08:00 |
|
takatost
|
78d3aa5fcd
|
fix: embedding init err (#956)
|
2023-08-22 17:43:59 +08:00 |
|
takatost
|
4f3053a8cc
|
fix: xinference chat completion error (#952)
|
2023-08-22 15:58:04 +08:00 |
|
takatost
|
866ee5da91
|
fix: openllm generate cutoff (#945)
|
2023-08-22 13:43:36 +08:00 |
|
takatost
|
e0a48c4972
|
fix: xinference chat support (#939)
|
2023-08-21 20:44:29 +08:00 |
|
takatost
|
6c832ee328
|
fix: remove openllm pypi package because of this package too large (#931)
|
2023-08-21 02:12:28 +08:00 |
|
takatost
|
0cc0b6e052
|
fix: error raise status code not exist (#888)
|
2023-08-17 15:33:35 +08:00 |
|
takatost
|
f42e7d1a61
|
feat: add spark v2 support (#885)
|
2023-08-17 15:08:57 +08:00 |
|
takatost
|
c4d759dfba
|
fix: wenxin error not raise when stream mode (#884)
|
2023-08-17 13:40:00 +08:00 |
|
takatost
|
cc52cdc2a9
|
Feat/add free provider apply (#829)
|
2023-08-14 12:44:35 +08:00 |
|
takatost
|
5fa2161b05
|
feat: server multi models support (#799)
|
2023-08-12 00:57:00 +08:00 |
|