0.1.14
Introducing vllora_llm crate, a standalone Rust library that provides a unified interface for interacting with multiple LLM providers through the vLLora AI Gateway. The crate enables seamless chat completions across OpenAI-compatible, Anthropic, Gemini, and Bedrock providers, with built-in streaming support and telemetry integration.
use vllora_llm::client::VlloraLLMClient;
use vllora_llm::types::gateway::{ChatCompletionRequest, ChatCompletionMessage};
#[tokio::main]
async fn main() -> Result<(), Box<dyn std::error::Error>> {
let request = ChatCompletionRequest {
model: "gpt-4.1-mini".to_string(),
messages: vec![
ChatCompletionMessage::new_text("user".to_string(), "Say hello!".to_string()),
],
..Default::default()
};
let client = VlloraLLMClient::new();
let response = client.completions().create(request).await?;
Ok(())
}
Other improvements in this release
- Enhanced breakpoint management with GlobalBreakpointStateEvent integration
- Improved error handling for multiple provider scenarios
- Asynchronous improvements to intercept functionality
Features
- feat: Update examples (@karolisg) (08fbebb)
- feat: add GlobalBreakpointStateEvent and integrate it into GatewayEvent handling (@karolisg) (48d8d5f)