llm.rb v4.11.0 released
Briefly

"llm.rb v4.11.0 introduces streaming tool execution, enabling tools to start while the model is still responding, effectively overlapping latency with output."
"The release adds MCP support over both stdio and HTTP, incorporating connection pooling to enhance performance and efficiency."
"A complete concurrency model is included, featuring threads, fibers, and async tasks, which allows for more robust and responsive applications."
"The local model registry for cost tracking, along with JSON Schema unions and production fixes, makes llm.rb v4.11.0 suitable for real systems."
llm.rb v4.11.0 introduces significant enhancements including streaming tool execution, allowing tools to start while the model responds. It supports MCP over stdio and HTTP with connection pooling, and integrates OpenAI's Responses API. The release features a complete concurrency model with threads, fibers, and async tasks. Additionally, it includes a local model registry for cost tracking, JSON Schema unions, and various production fixes, ensuring readiness for real systems where control and performance are critical.
Read at Rubyflow
Unable to calculate read time
[
|
]