Index of /usr/lib/mysqlsh/lib/python3.9/site-packages/oci/generative_ai_inference/models
..
__init__.py
base_chat_request.py
base_chat_response.py
chat_choice.py
chat_content.py
chat_details.py
chat_result.py
choice.py
citation.py
cohere_chat_request.py
cohere_chat_response.py
cohere_llm_inference_request.py
cohere_llm_inference_response.py
cohere_message.py
dedicated_serving_mode.py
embed_text_details.py
embed_text_result.py
generate_text_details.py
generate_text_result.py
generated_text.py
generic_chat_request.py
generic_chat_response.py
llama_llm_inference_request.py
llama_llm_inference_response.py
llm_inference_request.py
llm_inference_response.py
logprobs.py
message.py
on_demand_serving_mode.py
search_query.py
serving_mode.py
summarize_text_details.py
summarize_text_result.py
text_content.py
token_likelihood.py
__init__.py
base_chat_request.py
base_chat_response.py
chat_choice.py
chat_content.py
chat_details.py
chat_result.py
choice.py
citation.py
cohere_chat_request.py
cohere_chat_response.py
cohere_llm_inference_request.py
cohere_llm_inference_response.py
cohere_message.py
dedicated_serving_mode.py
embed_text_details.py
embed_text_result.py
generate_text_details.py
generate_text_result.py
generated_text.py
generic_chat_request.py
generic_chat_response.py
llama_llm_inference_request.py
llama_llm_inference_response.py
llm_inference_request.py
llm_inference_response.py
logprobs.py
message.py
on_demand_serving_mode.py
search_query.py
serving_mode.py
summarize_text_details.py
summarize_text_result.py
text_content.py
token_likelihood.py