LlamaLlmInferenceResponse¶
-
class
oci.generative_ai_inference.models.LlamaLlmInferenceResponse(**kwargs)¶ Bases:
oci.generative_ai_inference.models.llm_inference_response.LlmInferenceResponseThe generated text result to return.
Attributes
RUNTIME_TYPE_COHEREstr(object=’’) -> str RUNTIME_TYPE_LLAMAstr(object=’’) -> str choices[Required] Gets the choices of this LlamaLlmInferenceResponse. created[Required] Gets the created of this LlamaLlmInferenceResponse. runtime_type[Required] Gets the runtime_type of this LlmInferenceResponse. Methods
__init__(**kwargs)Initializes a new LlamaLlmInferenceResponse object with values from keyword arguments. get_subtype(object_dictionary)Given the hash representation of a subtype of this class, use the info in the hash to return the class of the subtype. -
RUNTIME_TYPE_COHERE= 'COHERE'¶
-
RUNTIME_TYPE_LLAMA= 'LLAMA'¶
-
__init__(**kwargs)¶ Initializes a new LlamaLlmInferenceResponse object with values from keyword arguments. The default value of the
runtime_typeattribute of this class isLLAMAand it should not be changed. The following keyword arguments are supported (corresponding to the getters/setters of this class):Parameters: - runtime_type (str) – The value to assign to the runtime_type property of this LlamaLlmInferenceResponse. Allowed values for this property are: “COHERE”, “LLAMA”
- created (datetime) – The value to assign to the created property of this LlamaLlmInferenceResponse.
- choices (list[oci.generative_ai_inference.models.Choice]) – The value to assign to the choices property of this LlamaLlmInferenceResponse.
-
choices¶ [Required] Gets the choices of this LlamaLlmInferenceResponse. A list of generated texts. Can be more than one if n is greater than 1.
Returns: The choices of this LlamaLlmInferenceResponse. Return type: list[oci.generative_ai_inference.models.Choice]
-
created¶ [Required] Gets the created of this LlamaLlmInferenceResponse. The Unix timestamp (in seconds) of when the generation was created.
Returns: The created of this LlamaLlmInferenceResponse. Return type: datetime
-
static
get_subtype(object_dictionary)¶ Given the hash representation of a subtype of this class, use the info in the hash to return the class of the subtype.
-
runtime_type¶ [Required] Gets the runtime_type of this LlmInferenceResponse. The runtime of the provided model.
Allowed values for this property are: “COHERE”, “LLAMA”, ‘UNKNOWN_ENUM_VALUE’. Any unrecognized values returned by a service will be mapped to ‘UNKNOWN_ENUM_VALUE’.
Returns: The runtime_type of this LlmInferenceResponse. Return type: str
-