diff --git a/api/core/model_runtime/model_providers/openrouter/llm/_position.yaml b/api/core/model_runtime/model_providers/openrouter/llm/_position.yaml index fd4ed1109d..7e00dd3f4b 100644 --- a/api/core/model_runtime/model_providers/openrouter/llm/_position.yaml +++ b/api/core/model_runtime/model_providers/openrouter/llm/_position.yaml @@ -12,6 +12,9 @@ - google/gemini-pro - cohere/command-r-plus - cohere/command-r +- meta-llama/llama-3.1-405b-instruct +- meta-llama/llama-3.1-70b-instruct +- meta-llama/llama-3.1-8b-instruct - meta-llama/llama-3-70b-instruct - meta-llama/llama-3-8b-instruct - mistralai/mixtral-8x22b-instruct diff --git a/api/core/model_runtime/model_providers/openrouter/llm/llama-3.1-405b-instruct.yaml b/api/core/model_runtime/model_providers/openrouter/llm/llama-3.1-405b-instruct.yaml new file mode 100644 index 0000000000..7d68e708b7 --- /dev/null +++ b/api/core/model_runtime/model_providers/openrouter/llm/llama-3.1-405b-instruct.yaml @@ -0,0 +1,23 @@ +model: meta-llama/llama-3.1-405b-instruct +label: + en_US: llama-3.1-405b-instruct +model_type: llm +model_properties: + mode: chat + context_size: 128000 +parameter_rules: + - name: temperature + use_template: temperature + - name: top_p + use_template: top_p + - name: max_tokens + use_template: max_tokens + required: true + default: 512 + min: 1 + max: 128000 +pricing: + input: "3" + output: "3" + unit: "0.000001" + currency: USD diff --git a/api/core/model_runtime/model_providers/openrouter/llm/llama-3.1-70b-instruct.yaml b/api/core/model_runtime/model_providers/openrouter/llm/llama-3.1-70b-instruct.yaml new file mode 100644 index 0000000000..78e3b45435 --- /dev/null +++ b/api/core/model_runtime/model_providers/openrouter/llm/llama-3.1-70b-instruct.yaml @@ -0,0 +1,23 @@ +model: meta-llama/llama-3.1-70b-instruct +label: + en_US: llama-3.1-70b-instruct +model_type: llm +model_properties: + mode: chat + context_size: 128000 +parameter_rules: + - name: temperature + use_template: temperature + - name: top_p + use_template: top_p + - name: max_tokens + use_template: max_tokens + required: true + default: 512 + min: 1 + max: 128000 +pricing: + input: "0.9" + output: "0.9" + unit: "0.000001" + currency: USD diff --git a/api/core/model_runtime/model_providers/openrouter/llm/llama-3.1-8b-instruct.yaml b/api/core/model_runtime/model_providers/openrouter/llm/llama-3.1-8b-instruct.yaml new file mode 100644 index 0000000000..6e69b7deb7 --- /dev/null +++ b/api/core/model_runtime/model_providers/openrouter/llm/llama-3.1-8b-instruct.yaml @@ -0,0 +1,23 @@ +model: meta-llama/llama-3.1-8b-instruct +label: + en_US: llama-3.1-8b-instruct +model_type: llm +model_properties: + mode: chat + context_size: 128000 +parameter_rules: + - name: temperature + use_template: temperature + - name: top_p + use_template: top_p + - name: max_tokens + use_template: max_tokens + required: true + default: 512 + min: 1 + max: 128000 +pricing: + input: "0.2" + output: "0.2" + unit: "0.000001" + currency: USD