Skip to main content

TeleportInferenceModel

Report an Issue

This guide is a comprehensive reference to the fields in the TeleportInferenceModel resource, which you can apply after installing the Teleport Kubernetes operator.

resources.teleport.dev/v1

apiVersion: resources.teleport.dev/v1

FieldTypeDescription
apiVersionstringAPIVersion defines the versioned schema of this representation of an object. Servers should convert recognized schemas to the latest internal value, and may reject unrecognized values. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#resources
kindstringKind is a string value representing the REST resource this object represents. Servers may infer this from the endpoint the client submits requests to. Cannot be updated. In CamelCase. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds
metadataobject
specobjectInferenceModel resource definition v1 from Teleport

spec

FieldTypeDescription
bedrockobjectBedrock indicates that this model uses Amazon Bedrock as the inference provider and specifies Bedrock-specific parameters.
max_session_length_bytesintegerMaxSessionLengthBytes is the maximum session length that can be sent to inference provider. Currently, it's determined by the size of model's context window; future versions of Teleport will allow summarizing larger sessions by splitting them. Inference providers will reject requests that are larger than given model's context window. Since context windows are usually sized in tokens, this value is an approximation. Assuming 2 bytes per input token should be safe. Currently, Teleport will outright reject sessions larger than this limit; future versions will split sessions in chunks, treating this size as a maximum. If unset or set to 0, defaults to 1MB.
openaiobjectOpenai indicates that this model uses OpenAI as the inference provider and specifies OpenAI-specific parameters.

spec.bedrock

FieldTypeDescription
bedrock_model_idstringBedrockModelId specifies a model ID or an inference profile as understood by the Bedrock API.
integrationstringIntegration is the AWS OIDC Integration name. If unset, Teleport will use AWS credentials available on the Auth Service machine; otherwise, it will use the specified OIDC integration for assuming appropriate role.
regionstringRegion is the AWS region which will be used for inference.
temperaturenumberTemperature controls the randomness of the model's output.

spec.openai

FieldTypeDescription
api_key_secret_refstringApiKeySecretRef is a reference to an InferenceSecret that contains the OpenAI API key.
base_urlstringBaseUrl is the OpenAI API base URL. Optional, defaults to the public OpenAI API URL. May be used to point to a custom OpenAI-compatible API, such as LiteLLM. In such case, the api_key_secret_ref must point to a secret that contains the API key for that custom API.
openai_model_idstringOpenaiModelId specifies the model ID, as understood by the OpenAI API.
temperaturenumberTemperature controls the randomness of the model's output.