mirror of
https://github.com/langgenius/dify.git
synced 2024-11-16 03:32:23 +08:00
parent
a9ee18300e
commit
4dfbcd0b4e
|
@ -26,6 +26,11 @@ class ZhipuAIProvider(BaseModelProvider):
|
|||
def _get_fixed_model_list(self, model_type: ModelType) -> list[dict]:
|
||||
if model_type == ModelType.TEXT_GENERATION:
|
||||
return [
|
||||
{
|
||||
'id': 'chatglm_turbo',
|
||||
'name': 'chatglm_turbo',
|
||||
'mode': ModelMode.CHAT.value,
|
||||
},
|
||||
{
|
||||
'id': 'chatglm_pro',
|
||||
'name': 'chatglm_pro',
|
||||
|
|
|
@ -11,6 +11,12 @@
|
|||
},
|
||||
"model_flexibility": "fixed",
|
||||
"price_config": {
|
||||
"chatglm_turbo": {
|
||||
"prompt": "0.005",
|
||||
"completion": "0.005",
|
||||
"unit": "0.001",
|
||||
"currency": "RMB"
|
||||
},
|
||||
"chatglm_pro": {
|
||||
"prompt": "0.01",
|
||||
"completion": "0.01",
|
||||
|
|
|
@ -96,7 +96,7 @@ class ZhipuAIChatLLM(BaseChatModel):
|
|||
return True
|
||||
|
||||
client: Any = None #: :meta private:
|
||||
model: str = "chatglm_lite"
|
||||
model: str = "chatglm_turbo"
|
||||
"""Model name to use."""
|
||||
temperature: float = 0.95
|
||||
"""A non-negative float that tunes the degree of randomness in generation."""
|
||||
|
|
Loading…
Reference in New Issue
Block a user