Skip to main content
← Back to index page

0.1.14

Introducing vllora_llm crate, a standalone Rust library that provides a unified interface for interacting with multiple LLM providers through the vLLora AI Gateway. The crate enables seamless chat completions across OpenAI-compatible, Anthropic, Gemini, and Bedrock providers, with built-in streaming support and telemetry integration.

use vllora_llm::client::VlloraLLMClient;
use vllora_llm::types::gateway::{ChatCompletionRequest, ChatCompletionMessage};

#[tokio::main]
async fn main() -> Result<(), Box<dyn std::error::Error>> {
let request = ChatCompletionRequest {
model: "gpt-4.1-mini".to_string(),
messages: vec![
ChatCompletionMessage::new_text("user".to_string(), "Say hello!".to_string()),
],
..Default::default()
};

let client = VlloraLLMClient::new();
let response = client.completions().create(request).await?;
Ok(())
}

Other improvements in this release

  • Enhanced breakpoint management with GlobalBreakpointStateEvent integration
  • Improved error handling for multiple provider scenarios
  • Asynchronous improvements to intercept functionality

Features

Bug Fixes

  • fix: update readme path and handle errors in model execution for multiple providers (@karolisg) (d20de5a)
  • fix: make set_intercept_all asynchronous and ensure continuation of requests when disabled (@karolisg) (23b5cc5)