Compare commits
10 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
5217fcfb94 | ||
|
|
6b1cc5c058 | ||
|
|
4a4219d3ea | ||
|
|
f91e794d71 | ||
|
|
7c96a4a88d | ||
|
|
14437bf609 | ||
|
|
3c228914f7 | ||
|
|
b69f7c617c | ||
|
|
75788b9395 | ||
|
|
a862b92c98 |
2
.github/workflows/test.yml
vendored
2
.github/workflows/test.yml
vendored
@@ -22,7 +22,7 @@ jobs:
|
|||||||
env:
|
env:
|
||||||
MISTRAL_API_KEY: ${{ secrets.MISTRAL_API_KEY }}
|
MISTRAL_API_KEY: ${{ secrets.MISTRAL_API_KEY }}
|
||||||
- name: Upload tests coverage
|
- name: Upload tests coverage
|
||||||
uses: codecov/codecov-action@v3
|
uses: codecov/codecov-action@v4
|
||||||
with:
|
with:
|
||||||
fail_ci_if_error: true
|
fail_ci_if_error: true
|
||||||
token: ${{ secrets.CODECOV_TOKEN }}
|
token: ${{ secrets.CODECOV_TOKEN }}
|
||||||
|
|||||||
22
CHANGELOG.md
22
CHANGELOG.md
@@ -1,3 +1,25 @@
|
|||||||
|
## [0.6.0](https://github.com/ivangabriele/mistralai-client-rs/compare/v0.5.0...v) (2024-03-04)
|
||||||
|
|
||||||
|
|
||||||
|
### ⚠ BREAKING CHANGES
|
||||||
|
|
||||||
|
* You can't set the `stream` option for `client.chat*()`.
|
||||||
|
|
||||||
|
Either use `client.chat_stream()` if you want to use streams
|
||||||
|
or use `client.chat()` / `client.chat_async()` otherwise.
|
||||||
|
|
||||||
|
### Features
|
||||||
|
|
||||||
|
* add client.chat_stream() method ([4a4219d](https://github.com/ivangabriele/mistralai-client-rs/commit/4a4219d3eaa8f0ae953ee6182b36bf464d1c4a21))
|
||||||
|
|
||||||
|
## [0.5.0](https://github.com/ivangabriele/mistralai-client-rs/compare/v0.4.0...v) (2024-03-04)
|
||||||
|
|
||||||
|
|
||||||
|
### Features
|
||||||
|
|
||||||
|
* add client.embeddings_async() method ([3c22891](https://github.com/ivangabriele/mistralai-client-rs/commit/3c228914f78b0edd4a592091265b88d0bc55568b))
|
||||||
|
* add client.list_models_async() method ([b69f7c6](https://github.com/ivangabriele/mistralai-client-rs/commit/b69f7c617c15dd63abb61d004636512916d766bb))
|
||||||
|
|
||||||
## [0.4.0](https://github.com/ivangabriele/mistralai-client-rs/compare/v0.3.0...v) (2024-03-04)
|
## [0.4.0](https://github.com/ivangabriele/mistralai-client-rs/compare/v0.3.0...v) (2024-03-04)
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
@@ -2,7 +2,7 @@
|
|||||||
name = "mistralai-client"
|
name = "mistralai-client"
|
||||||
description = "Mistral AI API client library for Rust (unofficial)."
|
description = "Mistral AI API client library for Rust (unofficial)."
|
||||||
license = "Apache-2.0"
|
license = "Apache-2.0"
|
||||||
version = "0.4.0"
|
version = "0.6.0"
|
||||||
|
|
||||||
edition = "2021"
|
edition = "2021"
|
||||||
rust-version = "1.76.0"
|
rust-version = "1.76.0"
|
||||||
@@ -15,8 +15,8 @@ readme = "README.md"
|
|||||||
repository = "https://github.com/ivangabriele/mistralai-client-rs"
|
repository = "https://github.com/ivangabriele/mistralai-client-rs"
|
||||||
|
|
||||||
[dependencies]
|
[dependencies]
|
||||||
minreq = { version = "2.11.0", features = ["https-rustls", "json-using-serde"] }
|
futures = "0.3.30"
|
||||||
reqwest = { version = "0.11.24", features = ["json"] }
|
reqwest = { version = "0.11.24", features = ["json", "blocking", "stream"] }
|
||||||
serde = { version = "1.0.197", features = ["derive"] }
|
serde = { version = "1.0.197", features = ["derive"] }
|
||||||
serde_json = "1.0.114"
|
serde_json = "1.0.114"
|
||||||
thiserror = "1.0.57"
|
thiserror = "1.0.57"
|
||||||
|
|||||||
96
README.md
96
README.md
@@ -17,7 +17,7 @@ Rust client for the Mistral AI API.
|
|||||||
- [Usage](#usage)
|
- [Usage](#usage)
|
||||||
- [Chat without streaming](#chat-without-streaming)
|
- [Chat without streaming](#chat-without-streaming)
|
||||||
- [Chat without streaming (async)](#chat-without-streaming-async)
|
- [Chat without streaming (async)](#chat-without-streaming-async)
|
||||||
- [Chat with streaming](#chat-with-streaming)
|
- [Chat with streaming (async)](#chat-with-streaming-async)
|
||||||
- [Embeddings](#embeddings)
|
- [Embeddings](#embeddings)
|
||||||
- [Embeddings (async)](#embeddings-async)
|
- [Embeddings (async)](#embeddings-async)
|
||||||
- [List models](#list-models)
|
- [List models](#list-models)
|
||||||
@@ -29,11 +29,11 @@ Rust client for the Mistral AI API.
|
|||||||
|
|
||||||
- [x] Chat without streaming
|
- [x] Chat without streaming
|
||||||
- [x] Chat without streaming (async)
|
- [x] Chat without streaming (async)
|
||||||
- [ ] Chat with streaming
|
- [x] Chat with streaming
|
||||||
- [x] Embedding
|
- [x] Embedding
|
||||||
- [ ] Embedding (async)
|
- [x] Embedding (async)
|
||||||
- [x] List models
|
- [x] List models
|
||||||
- [ ] List models (async)
|
- [x] List models (async)
|
||||||
- [ ] Function Calling
|
- [ ] Function Calling
|
||||||
- [ ] Function Calling (async)
|
- [ ] Function Calling (async)
|
||||||
|
|
||||||
@@ -71,7 +71,7 @@ fn main() {
|
|||||||
|
|
||||||
```rs
|
```rs
|
||||||
use mistralai_client::v1::{
|
use mistralai_client::v1::{
|
||||||
chat_completion::{ChatCompletionMessage, ChatCompletionMessageRole, ChatCompletionRequestOptions},
|
chat_completion::{ChatCompletionParams, ChatMessage, ChatMessageRole},
|
||||||
client::Client,
|
client::Client,
|
||||||
constants::Model,
|
constants::Model,
|
||||||
};
|
};
|
||||||
@@ -81,8 +81,8 @@ fn main() {
|
|||||||
let client = Client::new(None, None, None, None).unwrap();
|
let client = Client::new(None, None, None, None).unwrap();
|
||||||
|
|
||||||
let model = Model::OpenMistral7b;
|
let model = Model::OpenMistral7b;
|
||||||
let messages = vec![ChatCompletionMessage {
|
let messages = vec![ChatMessage {
|
||||||
role: ChatCompletionMessageRole::user,
|
role: ChatMessageRole::user,
|
||||||
content: "Just guess the next word: \"Eiffel ...\"?".to_string(),
|
content: "Just guess the next word: \"Eiffel ...\"?".to_string(),
|
||||||
}];
|
}];
|
||||||
let options = ChatCompletionRequestOptions {
|
let options = ChatCompletionRequestOptions {
|
||||||
@@ -101,7 +101,7 @@ fn main() {
|
|||||||
|
|
||||||
```rs
|
```rs
|
||||||
use mistralai_client::v1::{
|
use mistralai_client::v1::{
|
||||||
chat_completion::{ChatCompletionMessage, ChatCompletionMessageRole, ChatCompletionRequestOptions},
|
chat_completion::{ChatCompletionParams, ChatMessage, ChatMessageRole},
|
||||||
client::Client,
|
client::Client,
|
||||||
constants::Model,
|
constants::Model,
|
||||||
};
|
};
|
||||||
@@ -112,8 +112,8 @@ async fn main() {
|
|||||||
let client = Client::new(None, None, None, None).unwrap();
|
let client = Client::new(None, None, None, None).unwrap();
|
||||||
|
|
||||||
let model = Model::OpenMistral7b;
|
let model = Model::OpenMistral7b;
|
||||||
let messages = vec![ChatCompletionMessage {
|
let messages = vec![ChatMessage {
|
||||||
role: ChatCompletionMessageRole::user,
|
role: ChatMessageRole::user,
|
||||||
content: "Just guess the next word: \"Eiffel ...\"?".to_string(),
|
content: "Just guess the next word: \"Eiffel ...\"?".to_string(),
|
||||||
}];
|
}];
|
||||||
let options = ChatCompletionRequestOptions {
|
let options = ChatCompletionRequestOptions {
|
||||||
@@ -122,15 +122,50 @@ async fn main() {
|
|||||||
..Default::default()
|
..Default::default()
|
||||||
};
|
};
|
||||||
|
|
||||||
let result = client.chat(model, messages, Some(options)).await.unwrap();
|
let result = client.chat_async(model, messages, Some(options)).await.unwrap();
|
||||||
println!("Assistant: {}", result.choices[0].message.content);
|
println!("Assistant: {}", result.choices[0].message.content);
|
||||||
// => "Assistant: Tower. [...]"
|
// => "Assistant: Tower. [...]"
|
||||||
}
|
}
|
||||||
```
|
```
|
||||||
|
|
||||||
### Chat with streaming
|
### Chat with streaming (async)
|
||||||
|
|
||||||
_In progress._
|
```rs
|
||||||
|
use futures::stream::StreamExt;
|
||||||
|
use mistralai_client::v1::{
|
||||||
|
chat_completion::{ChatCompletionParams, ChatMessage, ChatMessageRole},
|
||||||
|
client::Client,
|
||||||
|
constants::Model,
|
||||||
|
};
|
||||||
|
|
||||||
|
[#tokio::main]
|
||||||
|
async fn main() {
|
||||||
|
// This example suppose you have set the `MISTRAL_API_KEY` environment variable.
|
||||||
|
let client = Client::new(None, None, None, None).unwrap();
|
||||||
|
|
||||||
|
let model = Model::OpenMistral7b;
|
||||||
|
let messages = vec![ChatMessage {
|
||||||
|
role: ChatMessageRole::user,
|
||||||
|
content: "Just guess the next word: \"Eiffel ...\"?".to_string(),
|
||||||
|
}];
|
||||||
|
let options = ChatCompletionParams {
|
||||||
|
temperature: Some(0.0),
|
||||||
|
random_seed: Some(42),
|
||||||
|
..Default::default()
|
||||||
|
};
|
||||||
|
|
||||||
|
let stream_result = client.chat_stream(model, messages, Some(options)).await;
|
||||||
|
let mut stream = stream_result.expect("Failed to create stream.");
|
||||||
|
while let Some(chunk_result) = stream.next().await {
|
||||||
|
match chunk_result {
|
||||||
|
Ok(chunk) => {
|
||||||
|
println!("Assistant (message chunk): {}", chunk.choices[0].delta.content);
|
||||||
|
}
|
||||||
|
Err(e) => eprintln!("Error processing chunk: {:?}", e),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
### Embeddings
|
### Embeddings
|
||||||
|
|
||||||
@@ -156,7 +191,26 @@ fn main() {
|
|||||||
|
|
||||||
### Embeddings (async)
|
### Embeddings (async)
|
||||||
|
|
||||||
_In progress._
|
```rs
|
||||||
|
use mistralai_client::v1::{client::Client, constants::EmbedModel};
|
||||||
|
|
||||||
|
#[tokio::main]
|
||||||
|
async fn main() {
|
||||||
|
// This example suppose you have set the `MISTRAL_API_KEY` environment variable.
|
||||||
|
let client: Client = Client::new(None, None, None, None).unwrap();
|
||||||
|
|
||||||
|
let model = EmbedModel::MistralEmbed;
|
||||||
|
let input = vec!["Embed this sentence.", "As well as this one."]
|
||||||
|
.iter()
|
||||||
|
.map(|s| s.to_string())
|
||||||
|
.collect();
|
||||||
|
let options = None;
|
||||||
|
|
||||||
|
let response = client.embeddings_async(model, input, options).await.unwrap();
|
||||||
|
println!("Embeddings: {:?}", response.data);
|
||||||
|
// => "Embeddings: [{...}, {...}]"
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
### List models
|
### List models
|
||||||
|
|
||||||
@@ -175,4 +229,16 @@ fn main() {
|
|||||||
|
|
||||||
### List models (async)
|
### List models (async)
|
||||||
|
|
||||||
_In progress._
|
```rs
|
||||||
|
use mistralai_client::v1::client::Client;
|
||||||
|
|
||||||
|
#[tokio::main]
|
||||||
|
async fn main() {
|
||||||
|
// This example suppose you have set the `MISTRAL_API_KEY` environment variable.
|
||||||
|
let client = Client::new(None, None, None, None).await.unwrap();
|
||||||
|
|
||||||
|
let result = client.list_models_async().unwrap();
|
||||||
|
println!("First Model ID: {:?}", result.data[0].id);
|
||||||
|
// => "First Model ID: open-mistral-7b"
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|||||||
@@ -2,6 +2,25 @@ use serde::{Deserialize, Serialize};
|
|||||||
|
|
||||||
use crate::v1::{common, constants};
|
use crate::v1::{common, constants};
|
||||||
|
|
||||||
|
// -----------------------------------------------------------------------------
|
||||||
|
// Definitions
|
||||||
|
|
||||||
|
#[derive(Clone, Debug, Deserialize, Serialize)]
|
||||||
|
pub struct ChatMessage {
|
||||||
|
pub role: ChatMessageRole,
|
||||||
|
pub content: String,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Clone, Debug, Eq, PartialEq, Deserialize, Serialize)]
|
||||||
|
#[allow(non_camel_case_types)]
|
||||||
|
pub enum ChatMessageRole {
|
||||||
|
assistant,
|
||||||
|
user,
|
||||||
|
}
|
||||||
|
|
||||||
|
// -----------------------------------------------------------------------------
|
||||||
|
// Request
|
||||||
|
|
||||||
#[derive(Debug)]
|
#[derive(Debug)]
|
||||||
pub struct ChatCompletionParams {
|
pub struct ChatCompletionParams {
|
||||||
pub tools: Option<String>,
|
pub tools: Option<String>,
|
||||||
@@ -9,7 +28,6 @@ pub struct ChatCompletionParams {
|
|||||||
pub max_tokens: Option<u32>,
|
pub max_tokens: Option<u32>,
|
||||||
pub top_p: Option<f32>,
|
pub top_p: Option<f32>,
|
||||||
pub random_seed: Option<u32>,
|
pub random_seed: Option<u32>,
|
||||||
pub stream: Option<bool>,
|
|
||||||
pub safe_prompt: Option<bool>,
|
pub safe_prompt: Option<bool>,
|
||||||
}
|
}
|
||||||
impl Default for ChatCompletionParams {
|
impl Default for ChatCompletionParams {
|
||||||
@@ -20,7 +38,6 @@ impl Default for ChatCompletionParams {
|
|||||||
max_tokens: None,
|
max_tokens: None,
|
||||||
top_p: None,
|
top_p: None,
|
||||||
random_seed: None,
|
random_seed: None,
|
||||||
stream: None,
|
|
||||||
safe_prompt: None,
|
safe_prompt: None,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -28,7 +45,7 @@ impl Default for ChatCompletionParams {
|
|||||||
|
|
||||||
#[derive(Debug, Serialize, Deserialize)]
|
#[derive(Debug, Serialize, Deserialize)]
|
||||||
pub struct ChatCompletionRequest {
|
pub struct ChatCompletionRequest {
|
||||||
pub messages: Vec<ChatCompletionMessage>,
|
pub messages: Vec<ChatMessage>,
|
||||||
pub model: constants::Model,
|
pub model: constants::Model,
|
||||||
#[serde(skip_serializing_if = "Option::is_none")]
|
#[serde(skip_serializing_if = "Option::is_none")]
|
||||||
pub tools: Option<String>,
|
pub tools: Option<String>,
|
||||||
@@ -40,8 +57,7 @@ pub struct ChatCompletionRequest {
|
|||||||
pub top_p: Option<f32>,
|
pub top_p: Option<f32>,
|
||||||
#[serde(skip_serializing_if = "Option::is_none")]
|
#[serde(skip_serializing_if = "Option::is_none")]
|
||||||
pub random_seed: Option<u32>,
|
pub random_seed: Option<u32>,
|
||||||
#[serde(skip_serializing_if = "Option::is_none")]
|
pub stream: bool,
|
||||||
pub stream: Option<bool>,
|
|
||||||
#[serde(skip_serializing_if = "Option::is_none")]
|
#[serde(skip_serializing_if = "Option::is_none")]
|
||||||
pub safe_prompt: Option<bool>,
|
pub safe_prompt: Option<bool>,
|
||||||
// TODO Check this prop (seen in official Python client but not in API doc).
|
// TODO Check this prop (seen in official Python client but not in API doc).
|
||||||
@@ -52,7 +68,8 @@ pub struct ChatCompletionRequest {
|
|||||||
impl ChatCompletionRequest {
|
impl ChatCompletionRequest {
|
||||||
pub fn new(
|
pub fn new(
|
||||||
model: constants::Model,
|
model: constants::Model,
|
||||||
messages: Vec<ChatCompletionMessage>,
|
messages: Vec<ChatMessage>,
|
||||||
|
stream: bool,
|
||||||
options: Option<ChatCompletionParams>,
|
options: Option<ChatCompletionParams>,
|
||||||
) -> Self {
|
) -> Self {
|
||||||
let ChatCompletionParams {
|
let ChatCompletionParams {
|
||||||
@@ -61,7 +78,6 @@ impl ChatCompletionRequest {
|
|||||||
max_tokens,
|
max_tokens,
|
||||||
top_p,
|
top_p,
|
||||||
random_seed,
|
random_seed,
|
||||||
stream,
|
|
||||||
safe_prompt,
|
safe_prompt,
|
||||||
} = options.unwrap_or_default();
|
} = options.unwrap_or_default();
|
||||||
|
|
||||||
@@ -79,6 +95,9 @@ impl ChatCompletionRequest {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// -----------------------------------------------------------------------------
|
||||||
|
// Response
|
||||||
|
|
||||||
#[derive(Clone, Debug, Deserialize, Serialize)]
|
#[derive(Clone, Debug, Deserialize, Serialize)]
|
||||||
pub struct ChatCompletionResponse {
|
pub struct ChatCompletionResponse {
|
||||||
pub id: String,
|
pub id: String,
|
||||||
@@ -86,28 +105,45 @@ pub struct ChatCompletionResponse {
|
|||||||
/// Unix timestamp (in seconds).
|
/// Unix timestamp (in seconds).
|
||||||
pub created: u32,
|
pub created: u32,
|
||||||
pub model: constants::Model,
|
pub model: constants::Model,
|
||||||
pub choices: Vec<ChatCompletionChoice>,
|
pub choices: Vec<ChatCompletionResponseChoice>,
|
||||||
pub usage: common::ResponseUsage,
|
pub usage: common::ResponseUsage,
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Clone, Debug, Deserialize, Serialize)]
|
#[derive(Clone, Debug, Deserialize, Serialize)]
|
||||||
pub struct ChatCompletionChoice {
|
pub struct ChatCompletionResponseChoice {
|
||||||
pub index: u32,
|
pub index: u32,
|
||||||
pub message: ChatCompletionMessage,
|
pub message: ChatMessage,
|
||||||
pub finish_reason: String,
|
pub finish_reason: String,
|
||||||
// TODO Check this prop (seen in API responses but undocumented).
|
// TODO Check this prop (seen in API responses but undocumented).
|
||||||
// pub logprobs: ???
|
// pub logprobs: ???
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Clone, Debug, Deserialize, Serialize)]
|
// -----------------------------------------------------------------------------
|
||||||
pub struct ChatCompletionMessage {
|
// Stream
|
||||||
pub role: ChatCompletionMessageRole,
|
|
||||||
pub content: String,
|
#[derive(Debug, Deserialize)]
|
||||||
|
pub struct ChatCompletionStreamChunk {
|
||||||
|
pub id: String,
|
||||||
|
pub object: String,
|
||||||
|
/// Unix timestamp (in seconds).
|
||||||
|
pub created: u32,
|
||||||
|
pub model: constants::Model,
|
||||||
|
pub choices: Vec<ChatCompletionStreamChunkChoice>,
|
||||||
|
// TODO Check this prop (seen in API responses but undocumented).
|
||||||
|
// pub usage: ???,
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Clone, Debug, Eq, PartialEq, Deserialize, Serialize)]
|
#[derive(Debug, Deserialize)]
|
||||||
#[allow(non_camel_case_types)]
|
pub struct ChatCompletionStreamChunkChoice {
|
||||||
pub enum ChatCompletionMessageRole {
|
pub index: u32,
|
||||||
assistant,
|
pub delta: ChatCompletionStreamChunkChoiceDelta,
|
||||||
user,
|
pub finish_reason: Option<String>,
|
||||||
|
// TODO Check this prop (seen in API responses but undocumented).
|
||||||
|
// pub logprobs: ???,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Deserialize)]
|
||||||
|
pub struct ChatCompletionStreamChunkChoiceDelta {
|
||||||
|
pub role: Option<ChatMessageRole>,
|
||||||
|
pub content: String,
|
||||||
}
|
}
|
||||||
|
|||||||
373
src/v1/client.rs
373
src/v1/client.rs
@@ -1,10 +1,13 @@
|
|||||||
|
use futures::stream::StreamExt;
|
||||||
|
use futures::Stream;
|
||||||
|
use reqwest::Error as ReqwestError;
|
||||||
|
use serde_json::from_str;
|
||||||
|
|
||||||
use crate::v1::error::ApiError;
|
use crate::v1::error::ApiError;
|
||||||
use minreq::Response;
|
|
||||||
use reqwest::{Client as ReqwestClient, Error as ReqwestError};
|
|
||||||
|
|
||||||
use crate::v1::{
|
use crate::v1::{
|
||||||
chat_completion::{
|
chat_completion::{
|
||||||
ChatCompletionMessage, ChatCompletionParams, ChatCompletionRequest, ChatCompletionResponse,
|
ChatCompletionParams, ChatCompletionRequest, ChatCompletionResponse, ChatMessage,
|
||||||
},
|
},
|
||||||
constants::{EmbedModel, Model, API_URL_BASE},
|
constants::{EmbedModel, Model, API_URL_BASE},
|
||||||
embedding::{EmbeddingRequest, EmbeddingRequestOptions, EmbeddingResponse},
|
embedding::{EmbeddingRequest, EmbeddingRequestOptions, EmbeddingResponse},
|
||||||
@@ -12,6 +15,8 @@ use crate::v1::{
|
|||||||
model_list::ModelListResponse,
|
model_list::ModelListResponse,
|
||||||
};
|
};
|
||||||
|
|
||||||
|
use super::chat_completion::ChatCompletionStreamChunk;
|
||||||
|
|
||||||
pub struct Client {
|
pub struct Client {
|
||||||
pub api_key: String,
|
pub api_key: String,
|
||||||
pub endpoint: String,
|
pub endpoint: String,
|
||||||
@@ -42,127 +47,80 @@ impl Client {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn build_request(&self, request: minreq::Request) -> minreq::Request {
|
|
||||||
let authorization = format!("Bearer {}", self.api_key);
|
|
||||||
let user_agent = format!(
|
|
||||||
"ivangabriele/mistralai-client-rs/{}",
|
|
||||||
env!("CARGO_PKG_VERSION")
|
|
||||||
);
|
|
||||||
|
|
||||||
let request = request
|
|
||||||
.with_header("Authorization", authorization)
|
|
||||||
.with_header("Accept", "application/json")
|
|
||||||
.with_header("Content-Type", "application/json")
|
|
||||||
.with_header("User-Agent", user_agent);
|
|
||||||
|
|
||||||
request
|
|
||||||
}
|
|
||||||
|
|
||||||
pub fn get_sync(&self, path: &str) -> Result<Response, ApiError> {
|
|
||||||
let url = format!("{}{}", self.endpoint, path);
|
|
||||||
let request = self.build_request(minreq::get(url));
|
|
||||||
|
|
||||||
let result = request.send();
|
|
||||||
match result {
|
|
||||||
Ok(response) => {
|
|
||||||
print!("{:?}", response.as_str().unwrap());
|
|
||||||
|
|
||||||
if (200..=299).contains(&response.status_code) {
|
|
||||||
Ok(response)
|
|
||||||
} else {
|
|
||||||
Err(ApiError {
|
|
||||||
message: format!(
|
|
||||||
"{}: {}",
|
|
||||||
response.status_code,
|
|
||||||
response.as_str().unwrap()
|
|
||||||
),
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
Err(error) => Err(self.new_minreq_error(error)),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
pub fn post_sync<T: serde::ser::Serialize + std::fmt::Debug>(
|
|
||||||
&self,
|
|
||||||
path: &str,
|
|
||||||
params: &T,
|
|
||||||
) -> Result<Response, ApiError> {
|
|
||||||
// print!("{:?}", params);
|
|
||||||
|
|
||||||
let url = format!("{}{}", self.endpoint, path);
|
|
||||||
let request = self.build_request(minreq::post(url));
|
|
||||||
|
|
||||||
let result = request.with_json(params).unwrap().send();
|
|
||||||
match result {
|
|
||||||
Ok(response) => {
|
|
||||||
print!("{:?}", response.as_str().unwrap());
|
|
||||||
|
|
||||||
if (200..=299).contains(&response.status_code) {
|
|
||||||
Ok(response)
|
|
||||||
} else {
|
|
||||||
Err(ApiError {
|
|
||||||
message: format!(
|
|
||||||
"{}: {}",
|
|
||||||
response.status_code,
|
|
||||||
response.as_str().unwrap()
|
|
||||||
),
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
Err(error) => Err(self.new_minreq_error(error)),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
pub fn chat(
|
pub fn chat(
|
||||||
&self,
|
&self,
|
||||||
model: Model,
|
model: Model,
|
||||||
messages: Vec<ChatCompletionMessage>,
|
messages: Vec<ChatMessage>,
|
||||||
options: Option<ChatCompletionParams>,
|
options: Option<ChatCompletionParams>,
|
||||||
) -> Result<ChatCompletionResponse, ApiError> {
|
) -> Result<ChatCompletionResponse, ApiError> {
|
||||||
let request = ChatCompletionRequest::new(model, messages, options);
|
let request = ChatCompletionRequest::new(model, messages, false, options);
|
||||||
|
|
||||||
let response = self.post_sync("/chat/completions", &request)?;
|
let response = self.post_sync("/chat/completions", &request)?;
|
||||||
let result = response.json::<ChatCompletionResponse>();
|
let result = response.json::<ChatCompletionResponse>();
|
||||||
match result {
|
match result {
|
||||||
Ok(response) => Ok(response),
|
Ok(response) => Ok(response),
|
||||||
Err(error) => Err(self.new_minreq_error(error)),
|
Err(error) => Err(self.to_api_error(error)),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub async fn chat_async(
|
pub async fn chat_async(
|
||||||
&self,
|
&self,
|
||||||
model: Model,
|
model: Model,
|
||||||
messages: Vec<ChatCompletionMessage>,
|
messages: Vec<ChatMessage>,
|
||||||
options: Option<ChatCompletionParams>,
|
options: Option<ChatCompletionParams>,
|
||||||
) -> Result<ChatCompletionResponse, ApiError> {
|
) -> Result<ChatCompletionResponse, ApiError> {
|
||||||
let request = ChatCompletionRequest::new(model, messages, options);
|
let request = ChatCompletionRequest::new(model, messages, false, options);
|
||||||
let client = ReqwestClient::new();
|
|
||||||
|
|
||||||
let response = client
|
let response = self.post_async("/chat/completions", &request).await?;
|
||||||
.post(format!("{}{}", self.endpoint, "/chat/completions"))
|
let result = response.json::<ChatCompletionResponse>().await;
|
||||||
.json(&request)
|
match result {
|
||||||
.bearer_auth(&self.api_key)
|
Ok(response) => Ok(response),
|
||||||
.header(
|
Err(error) => Err(self.to_api_error(error)),
|
||||||
"User-Agent",
|
}
|
||||||
format!("mistralai-client-rs/{}", env!("CARGO_PKG_VERSION")),
|
}
|
||||||
)
|
|
||||||
.send()
|
pub async fn chat_stream(
|
||||||
|
&self,
|
||||||
|
model: Model,
|
||||||
|
messages: Vec<ChatMessage>,
|
||||||
|
options: Option<ChatCompletionParams>,
|
||||||
|
) -> Result<impl Stream<Item = Result<ChatCompletionStreamChunk, ApiError>>, ApiError> {
|
||||||
|
let request = ChatCompletionRequest::new(model, messages, true, options);
|
||||||
|
let response = self
|
||||||
|
.post_stream("/chat/completions", &request)
|
||||||
.await
|
.await
|
||||||
.map_err(|e| self.new_reqwest_error(e))?;
|
.map_err(|e| ApiError {
|
||||||
|
message: e.to_string(),
|
||||||
if response.status().is_success() {
|
})?;
|
||||||
response
|
if !response.status().is_success() {
|
||||||
.json::<ChatCompletionResponse>()
|
|
||||||
.await
|
|
||||||
.map_err(|e| self.new_reqwest_error(e))
|
|
||||||
} else {
|
|
||||||
let status = response.status();
|
let status = response.status();
|
||||||
let text = response.text().await.unwrap_or_default();
|
let text = response.text().await.unwrap_or_default();
|
||||||
Err(ApiError {
|
return Err(ApiError {
|
||||||
message: format!("{}: {}", status, text),
|
message: format!("{}: {}", status, text),
|
||||||
})
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
|
let deserialized_stream =
|
||||||
|
response
|
||||||
|
.bytes_stream()
|
||||||
|
.map(|item| -> Result<ChatCompletionStreamChunk, ApiError> {
|
||||||
|
match item {
|
||||||
|
Ok(bytes) => {
|
||||||
|
let text = String::from_utf8(bytes.to_vec()).map_err(|e| ApiError {
|
||||||
|
message: e.to_string(),
|
||||||
|
})?;
|
||||||
|
let text_trimmed = text.trim_start_matches("data: ");
|
||||||
|
from_str(&text_trimmed).map_err(|e| ApiError {
|
||||||
|
message: e.to_string(),
|
||||||
|
})
|
||||||
|
}
|
||||||
|
Err(e) => Err(ApiError {
|
||||||
|
message: e.to_string(),
|
||||||
|
}),
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
Ok(deserialized_stream)
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn embeddings(
|
pub fn embeddings(
|
||||||
@@ -177,7 +135,23 @@ impl Client {
|
|||||||
let result = response.json::<EmbeddingResponse>();
|
let result = response.json::<EmbeddingResponse>();
|
||||||
match result {
|
match result {
|
||||||
Ok(response) => Ok(response),
|
Ok(response) => Ok(response),
|
||||||
Err(error) => Err(self.new_minreq_error(error)),
|
Err(error) => Err(self.to_api_error(error)),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub async fn embeddings_async(
|
||||||
|
&self,
|
||||||
|
model: EmbedModel,
|
||||||
|
input: Vec<String>,
|
||||||
|
options: Option<EmbeddingRequestOptions>,
|
||||||
|
) -> Result<EmbeddingResponse, ApiError> {
|
||||||
|
let request = EmbeddingRequest::new(model, input, options);
|
||||||
|
|
||||||
|
let response = self.post_async("/embeddings", &request).await?;
|
||||||
|
let result = response.json::<EmbeddingResponse>().await;
|
||||||
|
match result {
|
||||||
|
Ok(response) => Ok(response),
|
||||||
|
Err(error) => Err(self.to_api_error(error)),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -186,17 +160,204 @@ impl Client {
|
|||||||
let result = response.json::<ModelListResponse>();
|
let result = response.json::<ModelListResponse>();
|
||||||
match result {
|
match result {
|
||||||
Ok(response) => Ok(response),
|
Ok(response) => Ok(response),
|
||||||
Err(error) => Err(self.new_minreq_error(error)),
|
Err(error) => Err(self.to_api_error(error)),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn new_minreq_error(&self, err: minreq::Error) -> ApiError {
|
pub async fn list_models_async(&self) -> Result<ModelListResponse, ApiError> {
|
||||||
ApiError {
|
let response = self.get_async("/models").await?;
|
||||||
message: err.to_string(),
|
let result = response.json::<ModelListResponse>().await;
|
||||||
|
match result {
|
||||||
|
Ok(response) => Ok(response),
|
||||||
|
Err(error) => Err(self.to_api_error(error)),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn new_reqwest_error(&self, err: ReqwestError) -> ApiError {
|
fn build_request_sync(
|
||||||
|
&self,
|
||||||
|
request: reqwest::blocking::RequestBuilder,
|
||||||
|
) -> reqwest::blocking::RequestBuilder {
|
||||||
|
let user_agent = format!(
|
||||||
|
"ivangabriele/mistralai-client-rs/{}",
|
||||||
|
env!("CARGO_PKG_VERSION")
|
||||||
|
);
|
||||||
|
|
||||||
|
let request_builder = request
|
||||||
|
.bearer_auth(&self.api_key)
|
||||||
|
.header("Accept", "application/json")
|
||||||
|
.header("Content-Type", "application/json")
|
||||||
|
.header("User-Agent", user_agent);
|
||||||
|
|
||||||
|
request_builder
|
||||||
|
}
|
||||||
|
|
||||||
|
fn build_request_async(&self, request: reqwest::RequestBuilder) -> reqwest::RequestBuilder {
|
||||||
|
let user_agent = format!(
|
||||||
|
"ivangabriele/mistralai-client-rs/{}",
|
||||||
|
env!("CARGO_PKG_VERSION")
|
||||||
|
);
|
||||||
|
|
||||||
|
let request_builder = request
|
||||||
|
.bearer_auth(&self.api_key)
|
||||||
|
.header("Accept", "application/json")
|
||||||
|
.header("Content-Type", "application/json")
|
||||||
|
.header("User-Agent", user_agent);
|
||||||
|
|
||||||
|
request_builder
|
||||||
|
}
|
||||||
|
|
||||||
|
fn build_request_stream(&self, request: reqwest::RequestBuilder) -> reqwest::RequestBuilder {
|
||||||
|
let user_agent = format!(
|
||||||
|
"ivangabriele/mistralai-client-rs/{}",
|
||||||
|
env!("CARGO_PKG_VERSION")
|
||||||
|
);
|
||||||
|
|
||||||
|
let request_builder = request
|
||||||
|
.bearer_auth(&self.api_key)
|
||||||
|
.header("Accept", "text/event-stream")
|
||||||
|
.header("Content-Type", "application/json")
|
||||||
|
.header("User-Agent", user_agent);
|
||||||
|
|
||||||
|
request_builder
|
||||||
|
}
|
||||||
|
|
||||||
|
fn get_sync(&self, path: &str) -> Result<reqwest::blocking::Response, ApiError> {
|
||||||
|
let reqwest_client = reqwest::blocking::Client::new();
|
||||||
|
let url = format!("{}{}", self.endpoint, path);
|
||||||
|
let request = self.build_request_sync(reqwest_client.get(url));
|
||||||
|
|
||||||
|
let result = request.send();
|
||||||
|
match result {
|
||||||
|
Ok(response) => {
|
||||||
|
if response.status().is_success() {
|
||||||
|
Ok(response)
|
||||||
|
} else {
|
||||||
|
let status = response.status();
|
||||||
|
let text = response.text().unwrap();
|
||||||
|
Err(ApiError {
|
||||||
|
message: format!("{}: {}", status, text),
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
Err(error) => Err(ApiError {
|
||||||
|
message: error.to_string(),
|
||||||
|
}),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn get_async(&self, path: &str) -> Result<reqwest::Response, ApiError> {
|
||||||
|
let reqwest_client = reqwest::Client::new();
|
||||||
|
let url = format!("{}{}", self.endpoint, path);
|
||||||
|
let request_builder = reqwest_client.get(url);
|
||||||
|
let request = self.build_request_async(request_builder);
|
||||||
|
|
||||||
|
let result = request.send().await;
|
||||||
|
match result {
|
||||||
|
Ok(response) => {
|
||||||
|
if response.status().is_success() {
|
||||||
|
Ok(response)
|
||||||
|
} else {
|
||||||
|
let status = response.status();
|
||||||
|
let text = response.text().await.unwrap_or_default();
|
||||||
|
Err(ApiError {
|
||||||
|
message: format!("{}: {}", status, text),
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
Err(error) => Err(ApiError {
|
||||||
|
message: error.to_string(),
|
||||||
|
}),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn post_sync<T: serde::ser::Serialize + std::fmt::Debug>(
|
||||||
|
&self,
|
||||||
|
path: &str,
|
||||||
|
params: &T,
|
||||||
|
) -> Result<reqwest::blocking::Response, ApiError> {
|
||||||
|
let reqwest_client = reqwest::blocking::Client::new();
|
||||||
|
let url = format!("{}{}", self.endpoint, path);
|
||||||
|
let request_builder = reqwest_client.post(url).json(params);
|
||||||
|
let request = self.build_request_sync(request_builder);
|
||||||
|
|
||||||
|
let result = request.send();
|
||||||
|
match result {
|
||||||
|
Ok(response) => {
|
||||||
|
if response.status().is_success() {
|
||||||
|
Ok(response)
|
||||||
|
} else {
|
||||||
|
let status = response.status();
|
||||||
|
let text = response.text().unwrap_or_default();
|
||||||
|
Err(ApiError {
|
||||||
|
message: format!("{}: {}", status, text),
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
Err(error) => Err(ApiError {
|
||||||
|
message: error.to_string(),
|
||||||
|
}),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn post_async<T: serde::ser::Serialize + std::fmt::Debug>(
|
||||||
|
&self,
|
||||||
|
path: &str,
|
||||||
|
params: &T,
|
||||||
|
) -> Result<reqwest::Response, ApiError> {
|
||||||
|
let reqwest_client = reqwest::Client::new();
|
||||||
|
let url = format!("{}{}", self.endpoint, path);
|
||||||
|
let request_builder = reqwest_client.post(url).json(params);
|
||||||
|
let request = self.build_request_async(request_builder);
|
||||||
|
|
||||||
|
let result = request.send().await;
|
||||||
|
match result {
|
||||||
|
Ok(response) => {
|
||||||
|
if response.status().is_success() {
|
||||||
|
Ok(response)
|
||||||
|
} else {
|
||||||
|
let status = response.status();
|
||||||
|
let text = response.text().await.unwrap_or_default();
|
||||||
|
Err(ApiError {
|
||||||
|
message: format!("{}: {}", status, text),
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
Err(error) => Err(ApiError {
|
||||||
|
message: error.to_string(),
|
||||||
|
}),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn post_stream<T: serde::ser::Serialize + std::fmt::Debug>(
|
||||||
|
&self,
|
||||||
|
path: &str,
|
||||||
|
params: &T,
|
||||||
|
) -> Result<reqwest::Response, ApiError> {
|
||||||
|
let reqwest_client = reqwest::Client::new();
|
||||||
|
let url = format!("{}{}", self.endpoint, path);
|
||||||
|
let request_builder = reqwest_client.post(url).json(params);
|
||||||
|
let request = self.build_request_stream(request_builder);
|
||||||
|
|
||||||
|
let result = request.send().await;
|
||||||
|
match result {
|
||||||
|
Ok(response) => {
|
||||||
|
if response.status().is_success() {
|
||||||
|
Ok(response)
|
||||||
|
} else {
|
||||||
|
let status = response.status();
|
||||||
|
let text = response.text().await.unwrap_or_default();
|
||||||
|
Err(ApiError {
|
||||||
|
message: format!("{}: {}", status, text),
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
Err(error) => Err(ApiError {
|
||||||
|
message: error.to_string(),
|
||||||
|
}),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn to_api_error(&self, err: ReqwestError) -> ApiError {
|
||||||
ApiError {
|
ApiError {
|
||||||
message: err.to_string(),
|
message: err.to_string(),
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -16,4 +16,6 @@ impl Error for ApiError {}
|
|||||||
pub enum ClientError {
|
pub enum ClientError {
|
||||||
#[error("You must either set the `MISTRAL_API_KEY` environment variable or specify it in `Client::new(api_key, ...).")]
|
#[error("You must either set the `MISTRAL_API_KEY` environment variable or specify it in `Client::new(api_key, ...).")]
|
||||||
ApiKeyError,
|
ApiKeyError,
|
||||||
|
#[error("Failed to read the response text.")]
|
||||||
|
ReadResponseTextError,
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
use jrest::expect;
|
use jrest::expect;
|
||||||
use mistralai_client::v1::{
|
use mistralai_client::v1::{
|
||||||
chat_completion::{ChatCompletionMessage, ChatCompletionMessageRole, ChatCompletionParams},
|
chat_completion::{ChatCompletionParams, ChatMessage, ChatMessageRole},
|
||||||
client::Client,
|
client::Client,
|
||||||
constants::Model,
|
constants::Model,
|
||||||
};
|
};
|
||||||
@@ -10,8 +10,8 @@ async fn test_client_chat_async() {
|
|||||||
let client = Client::new(None, None, None, None).unwrap();
|
let client = Client::new(None, None, None, None).unwrap();
|
||||||
|
|
||||||
let model = Model::OpenMistral7b;
|
let model = Model::OpenMistral7b;
|
||||||
let messages = vec![ChatCompletionMessage {
|
let messages = vec![ChatMessage {
|
||||||
role: ChatCompletionMessageRole::user,
|
role: ChatMessageRole::user,
|
||||||
content: "Just guess the next word: \"Eiffel ...\"?".to_string(),
|
content: "Just guess the next word: \"Eiffel ...\"?".to_string(),
|
||||||
}];
|
}];
|
||||||
let options = ChatCompletionParams {
|
let options = ChatCompletionParams {
|
||||||
@@ -29,7 +29,7 @@ async fn test_client_chat_async() {
|
|||||||
expect!(response.object).to_be("chat.completion".to_string());
|
expect!(response.object).to_be("chat.completion".to_string());
|
||||||
expect!(response.choices.len()).to_be(1);
|
expect!(response.choices.len()).to_be(1);
|
||||||
expect!(response.choices[0].index).to_be(0);
|
expect!(response.choices[0].index).to_be(0);
|
||||||
expect!(response.choices[0].message.role.clone()).to_be(ChatCompletionMessageRole::assistant);
|
expect!(response.choices[0].message.role.clone()).to_be(ChatMessageRole::assistant);
|
||||||
expect!(response.choices[0].message.content.clone())
|
expect!(response.choices[0].message.content.clone())
|
||||||
.to_be("Tower. The Eiffel Tower is a famous landmark in Paris, France.".to_string());
|
.to_be("Tower. The Eiffel Tower is a famous landmark in Paris, France.".to_string());
|
||||||
expect!(response.usage.prompt_tokens).to_be_greater_than(0);
|
expect!(response.usage.prompt_tokens).to_be_greater_than(0);
|
||||||
|
|||||||
40
tests/v1_client_chat_stream_test.rs
Normal file
40
tests/v1_client_chat_stream_test.rs
Normal file
@@ -0,0 +1,40 @@
|
|||||||
|
use futures::stream::StreamExt;
|
||||||
|
use jrest::expect;
|
||||||
|
use mistralai_client::v1::{
|
||||||
|
chat_completion::{ChatCompletionParams, ChatMessage, ChatMessageRole},
|
||||||
|
client::Client,
|
||||||
|
constants::Model,
|
||||||
|
};
|
||||||
|
|
||||||
|
#[tokio::test]
|
||||||
|
async fn test_client_chat_stream() {
|
||||||
|
let client = Client::new(None, None, None, None).unwrap();
|
||||||
|
|
||||||
|
let model = Model::OpenMistral7b;
|
||||||
|
let messages = vec![ChatMessage {
|
||||||
|
role: ChatMessageRole::user,
|
||||||
|
content: "Just guess the next word: \"Eiffel ...\"?".to_string(),
|
||||||
|
}];
|
||||||
|
let options = ChatCompletionParams {
|
||||||
|
temperature: Some(0.0),
|
||||||
|
random_seed: Some(42),
|
||||||
|
..Default::default()
|
||||||
|
};
|
||||||
|
|
||||||
|
let stream_result = client.chat_stream(model, messages, Some(options)).await;
|
||||||
|
let mut stream = stream_result.expect("Failed to create stream.");
|
||||||
|
while let Some(chunk_result) = stream.next().await {
|
||||||
|
match chunk_result {
|
||||||
|
Ok(chunk) => {
|
||||||
|
if chunk.choices[0].delta.role == Some(ChatMessageRole::assistant)
|
||||||
|
|| chunk.choices[0].finish_reason == Some("stop".to_string())
|
||||||
|
{
|
||||||
|
expect!(chunk.choices[0].delta.content.len()).to_be(0);
|
||||||
|
} else {
|
||||||
|
expect!(chunk.choices[0].delta.content.len()).to_be_greater_than(0);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
Err(e) => eprintln!("Error processing chunk: {:?}", e),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -1,6 +1,6 @@
|
|||||||
use jrest::expect;
|
use jrest::expect;
|
||||||
use mistralai_client::v1::{
|
use mistralai_client::v1::{
|
||||||
chat_completion::{ChatCompletionMessage, ChatCompletionMessageRole, ChatCompletionParams},
|
chat_completion::{ChatCompletionParams, ChatMessage, ChatMessageRole},
|
||||||
client::Client,
|
client::Client,
|
||||||
constants::Model,
|
constants::Model,
|
||||||
};
|
};
|
||||||
@@ -10,8 +10,8 @@ fn test_client_chat() {
|
|||||||
let client = Client::new(None, None, None, None).unwrap();
|
let client = Client::new(None, None, None, None).unwrap();
|
||||||
|
|
||||||
let model = Model::OpenMistral7b;
|
let model = Model::OpenMistral7b;
|
||||||
let messages = vec![ChatCompletionMessage {
|
let messages = vec![ChatMessage {
|
||||||
role: ChatCompletionMessageRole::user,
|
role: ChatMessageRole::user,
|
||||||
content: "Just guess the next word: \"Eiffel ...\"?".to_string(),
|
content: "Just guess the next word: \"Eiffel ...\"?".to_string(),
|
||||||
}];
|
}];
|
||||||
let options = ChatCompletionParams {
|
let options = ChatCompletionParams {
|
||||||
@@ -26,7 +26,7 @@ fn test_client_chat() {
|
|||||||
expect!(response.object).to_be("chat.completion".to_string());
|
expect!(response.object).to_be("chat.completion".to_string());
|
||||||
expect!(response.choices.len()).to_be(1);
|
expect!(response.choices.len()).to_be(1);
|
||||||
expect!(response.choices[0].index).to_be(0);
|
expect!(response.choices[0].index).to_be(0);
|
||||||
expect!(response.choices[0].message.role.clone()).to_be(ChatCompletionMessageRole::assistant);
|
expect!(response.choices[0].message.role.clone()).to_be(ChatMessageRole::assistant);
|
||||||
expect!(response.choices[0].message.content.clone())
|
expect!(response.choices[0].message.content.clone())
|
||||||
.to_be("Tower. The Eiffel Tower is a famous landmark in Paris, France.".to_string());
|
.to_be("Tower. The Eiffel Tower is a famous landmark in Paris, France.".to_string());
|
||||||
expect!(response.usage.prompt_tokens).to_be_greater_than(0);
|
expect!(response.usage.prompt_tokens).to_be_greater_than(0);
|
||||||
|
|||||||
29
tests/v1_client_embeddings_async_test.rs
Normal file
29
tests/v1_client_embeddings_async_test.rs
Normal file
@@ -0,0 +1,29 @@
|
|||||||
|
use jrest::expect;
|
||||||
|
use mistralai_client::v1::{client::Client, constants::EmbedModel};
|
||||||
|
|
||||||
|
#[tokio::test]
|
||||||
|
async fn test_client_embeddings_async() {
|
||||||
|
let client: Client = Client::new(None, None, None, None).unwrap();
|
||||||
|
|
||||||
|
let model = EmbedModel::MistralEmbed;
|
||||||
|
let input = vec!["Embed this sentence.", "As well as this one."]
|
||||||
|
.iter()
|
||||||
|
.map(|s| s.to_string())
|
||||||
|
.collect();
|
||||||
|
let options = None;
|
||||||
|
|
||||||
|
let response = client
|
||||||
|
.embeddings_async(model, input, options)
|
||||||
|
.await
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
expect!(response.model).to_be(EmbedModel::MistralEmbed);
|
||||||
|
expect!(response.object).to_be("list".to_string());
|
||||||
|
expect!(response.data.len()).to_be(2);
|
||||||
|
expect!(response.data[0].index).to_be(0);
|
||||||
|
expect!(response.data[0].object.clone()).to_be("embedding".to_string());
|
||||||
|
expect!(response.data[0].embedding.len()).to_be_greater_than(0);
|
||||||
|
expect!(response.usage.prompt_tokens).to_be_greater_than(0);
|
||||||
|
expect!(response.usage.completion_tokens).to_be(0);
|
||||||
|
expect!(response.usage.total_tokens).to_be_greater_than(0);
|
||||||
|
}
|
||||||
20
tests/v1_client_list_models_async_test.rs
Normal file
20
tests/v1_client_list_models_async_test.rs
Normal file
@@ -0,0 +1,20 @@
|
|||||||
|
use jrest::expect;
|
||||||
|
use mistralai_client::v1::client::Client;
|
||||||
|
|
||||||
|
#[tokio::test]
|
||||||
|
async fn test_client_list_models_async() {
|
||||||
|
let client = Client::new(None, None, None, None).unwrap();
|
||||||
|
|
||||||
|
let response = client.list_models_async().await.unwrap();
|
||||||
|
|
||||||
|
expect!(response.object).to_be("list".to_string());
|
||||||
|
expect!(response.data.len()).to_be_greater_than(0);
|
||||||
|
|
||||||
|
// let open_mistral_7b_data_item = response
|
||||||
|
// .data
|
||||||
|
// .iter()
|
||||||
|
// .find(|item| item.id == "open-mistral-7b")
|
||||||
|
// .unwrap();
|
||||||
|
|
||||||
|
// expect!(open_mistral_7b_data_item.id).to_be("open-mistral-7b".to_string());
|
||||||
|
}
|
||||||
Reference in New Issue
Block a user