feat: add client.chat_async() method
This commit is contained in:
@@ -16,9 +16,11 @@ repository = "https://github.com/ivangabriele/mistralai-client-rs"
|
|||||||
|
|
||||||
[dependencies]
|
[dependencies]
|
||||||
minreq = { version = "2.11.0", features = ["https-rustls", "json-using-serde"] }
|
minreq = { version = "2.11.0", features = ["https-rustls", "json-using-serde"] }
|
||||||
|
reqwest = { version = "0.11.24", features = ["json"] }
|
||||||
serde = { version = "1.0.197", features = ["derive"] }
|
serde = { version = "1.0.197", features = ["derive"] }
|
||||||
serde_json = "1.0.114"
|
serde_json = "1.0.114"
|
||||||
thiserror = "1.0.57"
|
thiserror = "1.0.57"
|
||||||
|
tokio = { version = "1.36.0", features = ["full"] }
|
||||||
|
|
||||||
[dev-dependencies]
|
[dev-dependencies]
|
||||||
jrest = "0.2.3"
|
jrest = "0.2.3"
|
||||||
|
|||||||
46
README.md
46
README.md
@@ -16,19 +16,26 @@ Rust client for the Mistral AI API.
|
|||||||
- [As a client argument](#as-a-client-argument)
|
- [As a client argument](#as-a-client-argument)
|
||||||
- [Usage](#usage)
|
- [Usage](#usage)
|
||||||
- [Chat without streaming](#chat-without-streaming)
|
- [Chat without streaming](#chat-without-streaming)
|
||||||
|
- [Chat without streaming (async)](#chat-without-streaming-async)
|
||||||
- [Chat with streaming](#chat-with-streaming)
|
- [Chat with streaming](#chat-with-streaming)
|
||||||
- [Embeddings](#embeddings)
|
- [Embeddings](#embeddings)
|
||||||
|
- [Embeddings (async)](#embeddings-async)
|
||||||
- [List models](#list-models)
|
- [List models](#list-models)
|
||||||
|
- [List models (async)](#list-models-async)
|
||||||
|
|
||||||
---
|
---
|
||||||
|
|
||||||
## Supported APIs
|
## Supported APIs
|
||||||
|
|
||||||
- [x] Chat without streaming
|
- [x] Chat without streaming
|
||||||
|
- [x] Chat without streaming (async)
|
||||||
- [ ] Chat with streaming
|
- [ ] Chat with streaming
|
||||||
- [x] Embedding
|
- [x] Embedding
|
||||||
|
- [ ] Embedding (async)
|
||||||
- [x] List models
|
- [x] List models
|
||||||
|
- [ ] List models (async)
|
||||||
- [ ] Function Calling
|
- [ ] Function Calling
|
||||||
|
- [ ] Function Calling (async)
|
||||||
|
|
||||||
## Installation
|
## Installation
|
||||||
|
|
||||||
@@ -90,6 +97,37 @@ fn main() {
|
|||||||
}
|
}
|
||||||
```
|
```
|
||||||
|
|
||||||
|
### Chat without streaming (async)
|
||||||
|
|
||||||
|
```rs
|
||||||
|
use mistralai_client::v1::{
|
||||||
|
chat_completion::{ChatCompletionMessage, ChatCompletionMessageRole, ChatCompletionRequestOptions},
|
||||||
|
client::Client,
|
||||||
|
constants::Model,
|
||||||
|
};
|
||||||
|
|
||||||
|
#[tokio::main]
|
||||||
|
async fn main() {
|
||||||
|
// This example suppose you have set the `MISTRAL_API_KEY` environment variable.
|
||||||
|
let client = Client::new(None, None, None, None).unwrap();
|
||||||
|
|
||||||
|
let model = Model::OpenMistral7b;
|
||||||
|
let messages = vec![ChatCompletionMessage {
|
||||||
|
role: ChatCompletionMessageRole::user,
|
||||||
|
content: "Just guess the next word: \"Eiffel ...\"?".to_string(),
|
||||||
|
}];
|
||||||
|
let options = ChatCompletionRequestOptions {
|
||||||
|
temperature: Some(0.0),
|
||||||
|
random_seed: Some(42),
|
||||||
|
..Default::default()
|
||||||
|
};
|
||||||
|
|
||||||
|
let result = client.chat(model, messages, Some(options)).await.unwrap();
|
||||||
|
println!("Assistant: {}", result.choices[0].message.content);
|
||||||
|
// => "Assistant: Tower. [...]"
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
### Chat with streaming
|
### Chat with streaming
|
||||||
|
|
||||||
_In progress._
|
_In progress._
|
||||||
@@ -116,6 +154,10 @@ fn main() {
|
|||||||
}
|
}
|
||||||
```
|
```
|
||||||
|
|
||||||
|
### Embeddings (async)
|
||||||
|
|
||||||
|
_In progress._
|
||||||
|
|
||||||
### List models
|
### List models
|
||||||
|
|
||||||
```rs
|
```rs
|
||||||
@@ -130,3 +172,7 @@ fn main() {
|
|||||||
// => "First Model ID: open-mistral-7b"
|
// => "First Model ID: open-mistral-7b"
|
||||||
}
|
}
|
||||||
```
|
```
|
||||||
|
|
||||||
|
### List models (async)
|
||||||
|
|
||||||
|
_In progress._
|
||||||
|
|||||||
@@ -1,5 +1,6 @@
|
|||||||
use crate::v1::error::ApiError;
|
use crate::v1::error::ApiError;
|
||||||
use minreq::Response;
|
use minreq::Response;
|
||||||
|
use reqwest::{Client as ReqwestClient, Error as ReqwestError};
|
||||||
|
|
||||||
use crate::v1::{
|
use crate::v1::{
|
||||||
chat_completion::{
|
chat_completion::{
|
||||||
@@ -57,7 +58,7 @@ impl Client {
|
|||||||
request
|
request
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn get(&self, path: &str) -> Result<Response, ApiError> {
|
pub fn get_sync(&self, path: &str) -> Result<Response, ApiError> {
|
||||||
let url = format!("{}{}", self.endpoint, path);
|
let url = format!("{}{}", self.endpoint, path);
|
||||||
let request = self.build_request(minreq::get(url));
|
let request = self.build_request(minreq::get(url));
|
||||||
|
|
||||||
@@ -78,11 +79,11 @@ impl Client {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
Err(error) => Err(self.new_error(error)),
|
Err(error) => Err(self.new_minreq_error(error)),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn post<T: serde::ser::Serialize + std::fmt::Debug>(
|
pub fn post_sync<T: serde::ser::Serialize + std::fmt::Debug>(
|
||||||
&self,
|
&self,
|
||||||
path: &str,
|
path: &str,
|
||||||
params: &T,
|
params: &T,
|
||||||
@@ -109,7 +110,7 @@ impl Client {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
Err(error) => Err(self.new_error(error)),
|
Err(error) => Err(self.new_minreq_error(error)),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -121,11 +122,46 @@ impl Client {
|
|||||||
) -> Result<ChatCompletionResponse, ApiError> {
|
) -> Result<ChatCompletionResponse, ApiError> {
|
||||||
let request = ChatCompletionRequest::new(model, messages, options);
|
let request = ChatCompletionRequest::new(model, messages, options);
|
||||||
|
|
||||||
let response = self.post("/chat/completions", &request)?;
|
let response = self.post_sync("/chat/completions", &request)?;
|
||||||
let result = response.json::<ChatCompletionResponse>();
|
let result = response.json::<ChatCompletionResponse>();
|
||||||
match result {
|
match result {
|
||||||
Ok(response) => Ok(response),
|
Ok(response) => Ok(response),
|
||||||
Err(error) => Err(self.new_error(error)),
|
Err(error) => Err(self.new_minreq_error(error)),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub async fn chat_async(
|
||||||
|
&self,
|
||||||
|
model: Model,
|
||||||
|
messages: Vec<ChatCompletionMessage>,
|
||||||
|
options: Option<ChatCompletionParams>,
|
||||||
|
) -> Result<ChatCompletionResponse, ApiError> {
|
||||||
|
let request = ChatCompletionRequest::new(model, messages, options);
|
||||||
|
let client = ReqwestClient::new();
|
||||||
|
|
||||||
|
let response = client
|
||||||
|
.post(format!("{}{}", self.endpoint, "/chat/completions"))
|
||||||
|
.json(&request)
|
||||||
|
.bearer_auth(&self.api_key)
|
||||||
|
.header(
|
||||||
|
"User-Agent",
|
||||||
|
format!("mistralai-client-rs/{}", env!("CARGO_PKG_VERSION")),
|
||||||
|
)
|
||||||
|
.send()
|
||||||
|
.await
|
||||||
|
.map_err(|e| self.new_reqwest_error(e))?;
|
||||||
|
|
||||||
|
if response.status().is_success() {
|
||||||
|
response
|
||||||
|
.json::<ChatCompletionResponse>()
|
||||||
|
.await
|
||||||
|
.map_err(|e| self.new_reqwest_error(e))
|
||||||
|
} else {
|
||||||
|
let status = response.status();
|
||||||
|
let text = response.text().await.unwrap_or_default();
|
||||||
|
Err(ApiError {
|
||||||
|
message: format!("{}: {}", status, text),
|
||||||
|
})
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -137,24 +173,30 @@ impl Client {
|
|||||||
) -> Result<EmbeddingResponse, ApiError> {
|
) -> Result<EmbeddingResponse, ApiError> {
|
||||||
let request = EmbeddingRequest::new(model, input, options);
|
let request = EmbeddingRequest::new(model, input, options);
|
||||||
|
|
||||||
let response = self.post("/embeddings", &request)?;
|
let response = self.post_sync("/embeddings", &request)?;
|
||||||
let result = response.json::<EmbeddingResponse>();
|
let result = response.json::<EmbeddingResponse>();
|
||||||
match result {
|
match result {
|
||||||
Ok(response) => Ok(response),
|
Ok(response) => Ok(response),
|
||||||
Err(error) => Err(self.new_error(error)),
|
Err(error) => Err(self.new_minreq_error(error)),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn list_models(&self) -> Result<ModelListResponse, ApiError> {
|
pub fn list_models(&self) -> Result<ModelListResponse, ApiError> {
|
||||||
let response = self.get("/models")?;
|
let response = self.get_sync("/models")?;
|
||||||
let result = response.json::<ModelListResponse>();
|
let result = response.json::<ModelListResponse>();
|
||||||
match result {
|
match result {
|
||||||
Ok(response) => Ok(response),
|
Ok(response) => Ok(response),
|
||||||
Err(error) => Err(self.new_error(error)),
|
Err(error) => Err(self.new_minreq_error(error)),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn new_error(&self, err: minreq::Error) -> ApiError {
|
fn new_minreq_error(&self, err: minreq::Error) -> ApiError {
|
||||||
|
ApiError {
|
||||||
|
message: err.to_string(),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn new_reqwest_error(&self, err: ReqwestError) -> ApiError {
|
||||||
ApiError {
|
ApiError {
|
||||||
message: err.to_string(),
|
message: err.to_string(),
|
||||||
}
|
}
|
||||||
|
|||||||
38
tests/v1_client_chat_async_test.rs
Normal file
38
tests/v1_client_chat_async_test.rs
Normal file
@@ -0,0 +1,38 @@
|
|||||||
|
use jrest::expect;
|
||||||
|
use mistralai_client::v1::{
|
||||||
|
chat_completion::{ChatCompletionMessage, ChatCompletionMessageRole, ChatCompletionParams},
|
||||||
|
client::Client,
|
||||||
|
constants::Model,
|
||||||
|
};
|
||||||
|
|
||||||
|
#[tokio::test]
|
||||||
|
async fn test_client_chat_async() {
|
||||||
|
let client = Client::new(None, None, None, None).unwrap();
|
||||||
|
|
||||||
|
let model = Model::OpenMistral7b;
|
||||||
|
let messages = vec![ChatCompletionMessage {
|
||||||
|
role: ChatCompletionMessageRole::user,
|
||||||
|
content: "Just guess the next word: \"Eiffel ...\"?".to_string(),
|
||||||
|
}];
|
||||||
|
let options = ChatCompletionParams {
|
||||||
|
temperature: Some(0.0),
|
||||||
|
random_seed: Some(42),
|
||||||
|
..Default::default()
|
||||||
|
};
|
||||||
|
|
||||||
|
let response = client
|
||||||
|
.chat_async(model, messages, Some(options))
|
||||||
|
.await
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
expect!(response.model).to_be(Model::OpenMistral7b);
|
||||||
|
expect!(response.object).to_be("chat.completion".to_string());
|
||||||
|
expect!(response.choices.len()).to_be(1);
|
||||||
|
expect!(response.choices[0].index).to_be(0);
|
||||||
|
expect!(response.choices[0].message.role.clone()).to_be(ChatCompletionMessageRole::assistant);
|
||||||
|
expect!(response.choices[0].message.content.clone())
|
||||||
|
.to_be("Tower. The Eiffel Tower is a famous landmark in Paris, France.".to_string());
|
||||||
|
expect!(response.usage.prompt_tokens).to_be_greater_than(0);
|
||||||
|
expect!(response.usage.completion_tokens).to_be_greater_than(0);
|
||||||
|
expect!(response.usage.total_tokens).to_be_greater_than(0);
|
||||||
|
}
|
||||||
Reference in New Issue
Block a user