Semantic Conventions for OpenAI operations

Status: Experimental

The Semantic Conventions for OpenAI extend and override the semantic conventions for Gen AI Spans and Gen AI Metrics.

gen_ai.system MUST be set to "openai".

OpenAI Spans

These attributes track input data and metadata for a request to an OpenAI model. The attributes include general Generative AI attributes and ones specific the OpenAI.

AttributeTypeDescriptionExamplesRequirement LevelStability
gen_ai.operation.namestringThe name of the operation being performed. [1]chat; text_completion; embeddingsRequiredExperimental
gen_ai.request.modelstringThe name of the GenAI model a request is being made to. [2]gpt-4RequiredExperimental
error.typestringDescribes a class of error the operation ended with. [3]timeout; java.net.UnknownHostException; server_certificate_invalid; 500Conditionally Required if the operation ended in an errorStable
gen_ai.openai.request.response_formatstringThe response format that is requested.jsonConditionally Required if the request includes a response_formatExperimental
gen_ai.openai.request.seedintRequests with same seed value more likely to return same result.100Conditionally Required if the request includes a seedExperimental
gen_ai.openai.request.service_tierstringThe service tier requested. May be a specific tier, default, or auto.auto; defaultConditionally Required [4]Experimental
gen_ai.openai.response.service_tierstringThe service tier used for the response.scale; defaultConditionally Required [5]Experimental
server.portintGenAI server port. [6]80; 8080; 443Conditionally Required If server.address is set.Stable
gen_ai.openai.response.system_fingerprintstringA fingerprint to track any eventual change in the Generative AI environment.fp_44709d6fcbRecommendedExperimental
gen_ai.request.encoding_formatsstring[]The encoding formats requested in an embeddings operation, if specified. [7]["base64"]; ["float", "binary"]RecommendedExperimental
gen_ai.request.frequency_penaltydoubleThe frequency penalty setting for the GenAI request.0.1RecommendedExperimental
gen_ai.request.max_tokensintThe maximum number of tokens the model generates for a request.100RecommendedExperimental
gen_ai.request.presence_penaltydoubleThe presence penalty setting for the GenAI request.0.1RecommendedExperimental
gen_ai.request.stop_sequencesstring[]List of sequences that the model will use to stop generating further tokens.["forest", "lived"]RecommendedExperimental
gen_ai.request.temperaturedoubleThe temperature setting for the GenAI request.0.0RecommendedExperimental
gen_ai.request.top_pdoubleThe top_p sampling setting for the GenAI request.1.0RecommendedExperimental
gen_ai.response.finish_reasonsstring[]Array of reasons the model stopped generating tokens, corresponding to each generation received.["stop"]; ["stop", "length"]RecommendedExperimental
gen_ai.response.idstringThe unique identifier for the completion.chatcmpl-123RecommendedExperimental
gen_ai.response.modelstringThe name of the model that generated the response. [8]gpt-4-0613RecommendedExperimental
gen_ai.usage.input_tokensintThe number of tokens used in the prompt sent to OpenAI.100RecommendedExperimental
gen_ai.usage.output_tokensintThe number of tokens used in the completions from OpenAI.180RecommendedExperimental
server.addressstringGenAI server address. [9]example.com; 10.1.2.80; /tmp/my.sockRecommendedStable

[1] gen_ai.operation.name: If one of the predefined values applies, but specific system uses a different name it’s RECOMMENDED to document it in the semantic conventions for specific GenAI system and use system-specific name in the instrumentation. If a different name is not documented, instrumentation libraries SHOULD use applicable predefined value.

[2] gen_ai.request.model: The name of the GenAI model a request is being made to. If the model is supplied by a vendor, then the value must be the exact name of the model requested. If the model is a fine-tuned custom model, the value should have a more specific name than the base model that’s been fine-tuned.

[3] error.type: The error.type SHOULD match the error code returned by the Generative AI provider or the client library, the canonical name of exception that occurred, or another low-cardinality error identifier. Instrumentations SHOULD document the list of errors they report.

[4] gen_ai.openai.request.service_tier: if the request includes a service_tier and the value is not ‘auto’

[5] gen_ai.openai.response.service_tier: if the response was received and includes a service_tier

[6] server.port: When observed from the client side, and when communicating through an intermediary, server.port SHOULD represent the server port behind any intermediaries, for example proxies, if it’s available.

[7] gen_ai.request.encoding_formats: In some GenAI systems the encoding formats are called embedding types. Also, some GenAI systems only accept a single format per request.

[8] gen_ai.response.model: If available. The name of the GenAI model that provided the response. If the model is supplied by a vendor, then the value must be the exact name of the model actually used. If the model is a fine-tuned custom model, the value should have a more specific name than the base model that’s been fine-tuned.

[9] server.address: When observed from the client side, and when communicating through an intermediary, server.address SHOULD represent the server address behind any intermediaries, for example proxies, if it’s available.


error.type has the following list of well-known values. If one of them applies, then the respective value MUST be used; otherwise, a custom value MAY be used.

ValueDescriptionStability
_OTHERA fallback error value to be used when the instrumentation doesn’t define a custom value.Stable

gen_ai.openai.request.response_format has the following list of well-known values. If one of them applies, then the respective value MUST be used; otherwise, a custom value MAY be used.

ValueDescriptionStability
json_objectJSON object response formatExperimental
json_schemaJSON schema response formatExperimental
textText response formatExperimental

gen_ai.openai.request.service_tier has the following list of well-known values. If one of them applies, then the respective value MUST be used; otherwise, a custom value MAY be used.

ValueDescriptionStability
autoThe system will utilize scale tier credits until they are exhausted.Experimental
defaultThe system will utilize the default scale tier.Experimental

gen_ai.operation.name has the following list of well-known values. If one of them applies, then the respective value MUST be used; otherwise, a custom value MAY be used.

ValueDescriptionStability
chatChat completion operation such as OpenAI Chat APIExperimental
embeddingsEmbeddings operation such as OpenAI Create embeddings APIExperimental
text_completionText completions operation such as OpenAI Completions API (Legacy)Experimental

OpenAI Metrics

OpenAI metrics follow Generative AI metrics with the noted additional attributes. Individual systems may include additional system-specific attributes. It is recommended to check system-specific documentation, if available.

Metric: gen_ai.client.token.usage

Reports the usage of tokens following the common gen_ai.client.token.usage definition.

Additional attributes:

AttributeTypeDescriptionExamplesRequirement LevelStability
gen_ai.openai.response.service_tierstringThe service tier used for the response.scale; defaultRecommendedExperimental
gen_ai.openai.response.system_fingerprintstringA fingerprint to track any eventual change in the Generative AI environment.fp_44709d6fcbRecommendedExperimental

Metric: gen_ai.client.operation.duration

Measures the to complete an operation following the common gen_ai.client.operation.duration definition.

Additional attributes:

AttributeTypeDescriptionExamplesRequirement LevelStability
gen_ai.openai.response.service_tierstringThe service tier used for the response.scale; defaultRecommendedExperimental
gen_ai.openai.response.system_fingerprintstringA fingerprint to track any eventual change in the Generative AI environment.fp_44709d6fcbRecommendedExperimental