Uranus
|
2d9616c29c
|
fix: xinference last token being ignored (#1013)
|
2023-08-25 18:15:05 +08:00 |
|
takatost
|
9ae91a2ec3
|
feat: optimize xinference request max token key and stop reason (#998)
|
2023-08-24 18:11:15 +08:00 |
|
takatost
|
bd3a9b2f8d
|
fix: xinference-chat-stream-response (#991)
|
2023-08-24 14:39:34 +08:00 |
|
takatost
|
18d3877151
|
feat: optimize xinference stream (#989)
|
2023-08-24 13:58:34 +08:00 |
|
takatost
|
a76fde3d23
|
feat: optimize hf inference endpoint (#975)
|
2023-08-23 19:47:50 +08:00 |
|
takatost
|
78d3aa5fcd
|
fix: embedding init err (#956)
|
2023-08-22 17:43:59 +08:00 |
|
takatost
|
4f3053a8cc
|
fix: xinference chat completion error (#952)
|
2023-08-22 15:58:04 +08:00 |
|
takatost
|
866ee5da91
|
fix: openllm generate cutoff (#945)
|
2023-08-22 13:43:36 +08:00 |
|
takatost
|
e0a48c4972
|
fix: xinference chat support (#939)
|
2023-08-21 20:44:29 +08:00 |
|
takatost
|
6c832ee328
|
fix: remove openllm pypi package because of this package too large (#931)
|
2023-08-21 02:12:28 +08:00 |
|
takatost
|
0cc0b6e052
|
fix: error raise status code not exist (#888)
|
2023-08-17 15:33:35 +08:00 |
|
takatost
|
f42e7d1a61
|
feat: add spark v2 support (#885)
|
2023-08-17 15:08:57 +08:00 |
|
takatost
|
c4d759dfba
|
fix: wenxin error not raise when stream mode (#884)
|
2023-08-17 13:40:00 +08:00 |
|
takatost
|
cc52cdc2a9
|
Feat/add free provider apply (#829)
|
2023-08-14 12:44:35 +08:00 |
|
takatost
|
5fa2161b05
|
feat: server multi models support (#799)
|
2023-08-12 00:57:00 +08:00 |
|