pub async fn make_streaming_llm_request( messages: Vec<Message>, temperature: f32, max_tokens: Option<i32>, app_config: &ApplicationConfiguration, ) -> Result<Response>
Makes a streaming request to an LLM