Provider | Provider ID |
---|---|
OpenAI | openai |
Azure OpenAI | azure-openai |
AWS Bedrock | bedrock |
Anthropic | anthropic |
google | |
watsonx.ai | watsonx |
Mistral | mistral-ai |
OpenRouter | openrouter |
Ollama | ollama |
api_key
value to authenticate with the LLM service. Although you can use this value in the JSON string configuration, the safest way to store secret values, such as API keys, is to use a connection:
api_key
(Required)custom_host
url_to_fetch
forward_headers
request_timeout
transform_to_form_data
api_key
(Required)azure_resource_name
(Required)azure_deployment_id
(Required)azure_api_version
(Required)azure_model_name'
(Required)custom_host
url_to_fetch
forward_headers
request_timeout
transform_to_form_data
api_key
(Required)aws_secret_access_key
(Required)aws_access_key_id
(Required)custom_host
url_to_fetch
forward_headers
request_timeout
transform_to_form_data
aws_session_token
aws_region
aws_auth_type
aws_role_arn
aws_external_id
aws_s3_bucket
aws_s3_object_key
aws_bedrock_model
aws_server_side_encryption
aws_server_side_encryption_kms_key_id
api_key
(Required)anthropic_beta
anthropic_version
custom_host
url_to_fetch
forward_headers
request_timeout
transform_to_form_data
api_key
(Required)custom_host
url_to_fetch
forward_headers
request_timeout
transform_to_form_data
api_key
(Required)watsonx_space_id
(Required)watsonx_project_id
(Required)watsonx_deployment_id
(Required)watsonx_cpd_url
(Required in on-premises environments)watsonx_cpd_username
(Required in on-premises environments)watsonx_cpd_password
(Required in on-premises environments)watsonx_version
custom_host
url_to_fetch
forward_headers
request_timeout
transform_to_form_data
api_key
(Required)mistral_fim_completion
custom_host
url_to_fetch
forward_headers
request_timeout
transform_to_form_data
api_key
(Required)custom_host
url_to_fetch
forward_headers
request_timeout
transform_to_form_data
api_key
custom_host
url_to_fetch
forward_headers
request_timeout
transform_to_form_data
orchestrate models add
command to add a custom LLM to your active environment.
--name
(-n
): The name of the model you want to add. This name must follow the pattern <provider>
/<model_name>
. The provider must be exactly as outlined in the Supported providers section. And the model_name
must be exactly the same as the name that appears on the provider’s API documentaion.--description
(-d
): An optional description to appear alongside the model in the list view.--display-name
: An Optional display name for the model in the UI.--provider-config
: A JSON string of configuration options. These can also be provided via the connection referenced in --app-id
, especially secret values. You can use the --provider-config
alongside an --app-id
to provide non-required values.--type
- The type of model that is being created. These are the supported types:
chat
: Model that supports chat capabilities.chat_vision
: Model that supports chat and image capabilities.completion
: Model used for completion engines.embedding
: Embedding model used for transforming data.--app-id
(-a
): The app ID of a key_value
connection containing provider configuration details. These will be merged with the values provided in --provider-config
..env
file when you start the watsonx Orchestrate Developer Edition.
For that, your .env
file must contain either:
WATSONX_APIKEY
and WATSONX_SPACE_ID
environment variables.WO_INSTANCE
and WO_API_KEY
environment variables..env
file with these credentials, see Installing the watsonx Orchestrate Developer Edition.
To register the watsonx model through this method, you must create an api_key
credential with the value “gateway
”. You also don’t need to specify a space_id
when you add the model. See the following example:
orchestrate models import
command
--file
(-f
): File path of the spec file containing the model configuration--app-id
(-a
): The app id of a key_value
connection containing provider configuration details. These will be merged with the values provided in the provider_config
section of the spec.orchestrate models list
command to see all available LLMs in your active environment.
--raw
(-r
) argument.orchestrate models remove
command and use the --name
(-n
) argument to specify the LLM you want to remove.
--name
(-n
): The name of the policy you want to add.--description
(-d
): An optional description to appear a long side the policy in the list view.--display-name
: An optional display name for the policy in the UI--strategy
(-s
): The policy mode you want to use.
loadbalance
: These models operate together by distributing the load of requests between them, following the distribution of weight values. By default, both weight values are attributed as 1
, so the loads are evenly balanced between the models. If you want to customize the weight values, see Importing model policies.fallback
: If one of the models is unavailable, the agent will try to use the other one as a fallback alternative.single
: Uses a only one model, but allows for --retry-on-code
and --retry-attempts
.--strategy-on-code
: A list of HTTP error codes which triggers the strategy. Used for fallback
strategy.--retry-on-code
: A list of HTTP error codes for which the model should retry the request.--retry-attempts
: How many attempts it should make before stopping.my_spec.yaml
file follows this structure:
--file
(-f
): File path of the spec file containing the model policy configuration.add
or import
commands with the name of the model policy that you want to update to update the model policy.
--name
(-n
): The name of the model policy that you want to remove.