The name of the model to use (e.g., "llama3").
The input text to generate from.
Additional generation options (e.g., temperature, top_k).
Whether to stream the response (ignored in this implementation).
A JSONValue object containing the generated text and metadata.
Generates text based on a prompt using the specified model.