Skip to main content

Quickstart

use ferriclink_core::{
messages::AnyMessage,
language_models::{mock_chat_model, GenerationConfig},
runnables::Runnable,
};

#[tokio::main]
async fn main() -> Result<(), Box<dyn std::error::Error>> {
// Create a mock chat model
let chat_model = mock_chat_model("gpt-4o-mini");

// Create a conversation
let messages = vec![
AnyMessage::human("Hello, how are you?"),
];

// Generate a response
let response = chat_model.generate_chat(
messages,
Some(GenerationConfig::new().with_temperature(0.7)),
None,
).await?;

println!("Response: {}", response.text());
Ok(())
}