slug-code/src/provider/mod.rs
Bryan Ramos b8bf9029fe Initial Slug Code Rust implementation
Core features:
- OpenAI-compatible streaming provider (vLLM, Ollama, OpenAI, etc.)
- Agent loop with tool use (bash, read, write, edit, glob, grep)
- Permission system: ask/yolo/sandbox/allowEdits + glob patterns
- SLUG.md hierarchy loaded every turn (CLAUDE.md equivalent)
- Session persistence with --continue/--resume/--fork-session
- Hook system: 5 lifecycle events, command + prompt types
- Compaction: ToolResultTrim/Truncate strategies, /compact command
- Config via TOML, CLI args, env vars

Co-Authored-By: Claude Sonnet 4.6 <noreply@anthropic.com>
2026-03-31 14:23:04 -04:00

169 lines
5.8 KiB
Rust

mod types;
use anyhow::Result;
use futures::Stream;
use std::pin::Pin;
pub use types::*;
use crate::config::Config;
use crate::tools::ToolDefinition;
/// Trait for LLM providers. Anything that speaks OpenAI chat completions works.
pub trait Provider: Send + Sync {
fn stream_chat(
&self,
messages: &[ChatMessage],
tools: &[ToolDefinition],
) -> Pin<Box<dyn Stream<Item = Result<StreamEvent>> + Send + '_>>;
}
/// OpenAI-compatible provider (works with vLLM, Ollama, llama.cpp, OpenAI, etc.)
pub struct OpenAIProvider {
client: reqwest::Client,
endpoint: String,
api_key: Option<String>,
model: String,
max_tokens: u32,
temperature: Option<f32>,
}
impl OpenAIProvider {
pub fn new(config: &Config) -> Self {
Self {
client: reqwest::Client::new(),
endpoint: config.endpoint.trim_end_matches('/').to_string(),
api_key: config.api_key.clone(),
model: config.model.clone(),
max_tokens: config.max_tokens,
temperature: config.temperature,
}
}
fn build_request_body(
&self,
messages: &[ChatMessage],
tools: &[ToolDefinition],
) -> serde_json::Value {
let mut body = serde_json::json!({
"model": self.model,
"messages": messages,
"max_tokens": self.max_tokens,
"stream": true,
});
if let Some(temp) = self.temperature {
body["temperature"] = serde_json::json!(temp);
}
if !tools.is_empty() {
let tool_defs: Vec<serde_json::Value> = tools
.iter()
.map(|t| {
serde_json::json!({
"type": "function",
"function": {
"name": t.name,
"description": t.description,
"parameters": t.parameters,
}
})
})
.collect();
body["tools"] = serde_json::json!(tool_defs);
}
body
}
}
impl Provider for OpenAIProvider {
fn stream_chat(
&self,
messages: &[ChatMessage],
tools: &[ToolDefinition],
) -> Pin<Box<dyn Stream<Item = Result<StreamEvent>> + Send + '_>> {
let body = self.build_request_body(messages, tools);
let url = format!("{}/chat/completions", self.endpoint);
let mut req = self.client.post(&url).json(&body);
if let Some(ref key) = self.api_key {
req = req.bearer_auth(key);
}
Box::pin(async_stream::stream! {
let response = match req.send().await {
Ok(r) => r,
Err(e) => {
yield Err(anyhow::anyhow!(e));
return;
}
};
if !response.status().is_success() {
let status = response.status();
let text = response.text().await.unwrap_or_default();
yield Err(anyhow::anyhow!("API error {status}: {text}"));
return;
}
use futures::StreamExt;
let mut stream = response.bytes_stream();
let mut buffer = String::new();
while let Some(chunk) = stream.next().await {
let chunk = match chunk {
Ok(c) => c,
Err(e) => {
yield Err(anyhow::anyhow!(e));
return;
}
};
buffer.push_str(&String::from_utf8_lossy(&chunk));
// Process complete SSE lines
while let Some(line_end) = buffer.find('\n') {
let line = buffer[..line_end].trim().to_string();
buffer = buffer[line_end + 1..].to_string();
if line.is_empty() || line.starts_with(':') {
continue;
}
if let Some(data) = line.strip_prefix("data: ") {
if data == "[DONE]" {
yield Ok(StreamEvent::Done);
return;
}
match serde_json::from_str::<StreamChunk>(data) {
Ok(chunk) => {
for choice in &chunk.choices {
if let Some(ref content) = choice.delta.content {
yield Ok(StreamEvent::Text(content.clone()));
}
if let Some(ref tool_calls) = choice.delta.tool_calls {
for tc in tool_calls {
yield Ok(StreamEvent::ToolCallDelta(ToolCallDelta {
index: tc.index,
id: tc.id.clone(),
name: tc.function.as_ref().and_then(|f| f.name.clone()),
arguments_delta: tc.function.as_ref().and_then(|f| f.arguments.clone()),
}));
}
}
if choice.finish_reason.is_some() {
yield Ok(StreamEvent::Finish);
}
}
}
Err(e) => {
tracing::warn!("Failed to parse SSE chunk: {e}: {data}");
}
}
}
}
}
})
}
}