Streaming chat — calls onChunk for every assistant token.
Each chunk contains "message": {"role": "assistant", "content": "<token>"}. The final chunk has "done": true and carries usage metadata.
Model name.
Conversation history.
Callback invoked per chunk; must be @safe.
Optional tool definitions.
Structured output schema or JSONValue("json").
How long to keep the model loaded.
Typed generation options.
See Implementation
Streaming chat — calls onChunk for every assistant token.
Each chunk contains "message": {"role": "assistant", "content": "<token>"}. The final chunk has "done": true and carries usage metadata.