async fn make_llm_request( messages: Vec<Message>, temperature: f32, max_tokens: Option<i32>, endpoint: &Url, api_key: &str, ) -> Result<LlmCompletionResponse>
Makes a non-streaming request to an LLM