Compare commits
7 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
47c9b9b4fe | ||
|
|
791bef34b3 | ||
|
|
1dd59f6704 | ||
|
|
33876183e4 | ||
|
|
1deab88251 | ||
|
|
b0a3f10c9f | ||
|
|
bbba6b9878 |
@@ -1,3 +1,3 @@
|
||||
# This key is only used for development purposes.
|
||||
# You'll only need one if you want to contribute to this library.
|
||||
MISTRAL_API_KEY=
|
||||
export MISTRAL_API_KEY=
|
||||
|
||||
14
CHANGELOG.md
14
CHANGELOG.md
@@ -1,3 +1,17 @@
|
||||
## [0.4.0](https://github.com/ivangabriele/mistralai-client-rs/compare/v0.3.0...v) (2024-03-04)
|
||||
|
||||
|
||||
### ⚠ BREAKING CHANGES
|
||||
|
||||
* `Client::new()` now returns a `Result`.
|
||||
* `APIError` is renamed to `ApiError`.
|
||||
|
||||
### Features
|
||||
|
||||
* add client.chat_async() method ([1dd59f6](https://github.com/ivangabriele/mistralai-client-rs/commit/1dd59f67048c10458ab0382af8fdfe4ed21c82fa))
|
||||
* add missing api key error ([1deab88](https://github.com/ivangabriele/mistralai-client-rs/commit/1deab88251fc706e0415a5e416ab9aee4b52f6f3))
|
||||
* wrap Client::new() return in a Result ([3387618](https://github.com/ivangabriele/mistralai-client-rs/commit/33876183e41340f426aa1dd1b6d8b5c05c8e15b9))
|
||||
|
||||
## [0.3.0](https://github.com/ivangabriele/mistralai-client-rs/compare/v0.2.0...v) (2024-03-04)
|
||||
|
||||
|
||||
|
||||
@@ -27,8 +27,15 @@ Then run:
|
||||
git clone https://github.com/ivangabriele/mistralai-client-rs.git # or your fork
|
||||
cd ./mistralai-client-rs
|
||||
cargo build
|
||||
cp .env.example .env
|
||||
```
|
||||
|
||||
Then edit the `.env` file to set your `MISTRAL_API_KEY`.
|
||||
|
||||
> [!NOTE]
|
||||
> All tests use either the `open-mistral-7b` or `mistral-embed` models and only consume a few dozen tokens.
|
||||
> So you would have to run them thousands of times to even reach a single dollar of usage.
|
||||
|
||||
### Optional requirements
|
||||
|
||||
- [cargo-watch](https://github.com/watchexec/cargo-watch#install) for `make test-*-watch`.
|
||||
@@ -51,5 +58,4 @@ Help us keep this project open and inclusive. Please read and follow our [Code o
|
||||
|
||||
## Commit Message Format
|
||||
|
||||
This repository follow the [Conventional Commits](https://www.conventionalcommits.org/en/v1.0.0/) specification and
|
||||
specificaly the [Angular Commit Message Guidelines](https://github.com/angular/angular/blob/main/CONTRIBUTING.md#commit).
|
||||
This repository follow the [Conventional Commits](https://www.conventionalcommits.org/en/v1.0.0/) specification.
|
||||
|
||||
@@ -2,7 +2,7 @@
|
||||
name = "mistralai-client"
|
||||
description = "Mistral AI API client library for Rust (unofficial)."
|
||||
license = "Apache-2.0"
|
||||
version = "0.3.0"
|
||||
version = "0.4.0"
|
||||
|
||||
edition = "2021"
|
||||
rust-version = "1.76.0"
|
||||
@@ -16,11 +16,11 @@ repository = "https://github.com/ivangabriele/mistralai-client-rs"
|
||||
|
||||
[dependencies]
|
||||
minreq = { version = "2.11.0", features = ["https-rustls", "json-using-serde"] }
|
||||
reqwest = { version = "0.11.24", features = ["json"] }
|
||||
serde = { version = "1.0.197", features = ["derive"] }
|
||||
serde_json = "1.0.114"
|
||||
thiserror = "1.0.57"
|
||||
tokio = { version = "1.36.0", features = ["full"] }
|
||||
|
||||
[dev-dependencies]
|
||||
dotenv = "0.15.0"
|
||||
jrest = "0.2.3"
|
||||
|
||||
18
Makefile
18
Makefile
@@ -1,7 +1,9 @@
|
||||
SHELL := /bin/bash
|
||||
|
||||
.PHONY: test
|
||||
|
||||
define RELEASE_TEMPLATE
|
||||
conventional-changelog -p conventionalcommits -i CHANGELOG.md -s
|
||||
conventional-changelog -p conventionalcommits -i ./CHANGELOG.md -s
|
||||
git add .
|
||||
git commit -m "docs(changelog): update"
|
||||
git push origin HEAD
|
||||
@@ -9,13 +11,6 @@ define RELEASE_TEMPLATE
|
||||
git push origin HEAD --tags
|
||||
endef
|
||||
|
||||
test:
|
||||
cargo test --no-fail-fast
|
||||
test-cover:
|
||||
cargo tarpaulin --frozen --no-fail-fast --out Xml --skip-clean
|
||||
test-watch:
|
||||
cargo watch -x "test -- --nocapture"
|
||||
|
||||
release-patch:
|
||||
$(call RELEASE_TEMPLATE,patch)
|
||||
|
||||
@@ -24,3 +19,10 @@ release-minor:
|
||||
|
||||
release-major:
|
||||
$(call RELEASE_TEMPLATE,major)
|
||||
|
||||
test:
|
||||
@source ./.env && cargo test --all-targets --no-fail-fast
|
||||
test-cover:
|
||||
cargo tarpaulin --all-targets --frozen --no-fail-fast --out Xml --skip-clean
|
||||
test-watch:
|
||||
cargo watch -x "test -- --all-targets --nocapture"
|
||||
|
||||
56
README.md
56
README.md
@@ -16,19 +16,26 @@ Rust client for the Mistral AI API.
|
||||
- [As a client argument](#as-a-client-argument)
|
||||
- [Usage](#usage)
|
||||
- [Chat without streaming](#chat-without-streaming)
|
||||
- [Chat without streaming (async)](#chat-without-streaming-async)
|
||||
- [Chat with streaming](#chat-with-streaming)
|
||||
- [Embeddings](#embeddings)
|
||||
- [Embeddings (async)](#embeddings-async)
|
||||
- [List models](#list-models)
|
||||
- [List models (async)](#list-models-async)
|
||||
|
||||
---
|
||||
|
||||
## Supported APIs
|
||||
|
||||
- [x] Chat without streaming
|
||||
- [x] Chat without streaming (async)
|
||||
- [ ] Chat with streaming
|
||||
- [x] Embedding
|
||||
- [ ] Embedding (async)
|
||||
- [x] List models
|
||||
- [ ] List models (async)
|
||||
- [ ] Function Calling
|
||||
- [ ] Function Calling (async)
|
||||
|
||||
## Installation
|
||||
|
||||
@@ -54,7 +61,7 @@ use mistralai_client::v1::client::Client;
|
||||
fn main() {
|
||||
let api_key = "your_api_key";
|
||||
|
||||
let client = Client::new(Some(api_key), None, None, None);
|
||||
let client = Client::new(Some(api_key), None, None, None).unwrap();
|
||||
}
|
||||
```
|
||||
|
||||
@@ -71,7 +78,7 @@ use mistralai_client::v1::{
|
||||
|
||||
fn main() {
|
||||
// This example suppose you have set the `MISTRAL_API_KEY` environment variable.
|
||||
let client = Client::new(None, None, None, None);
|
||||
let client = Client::new(None, None, None, None).unwrap();
|
||||
|
||||
let model = Model::OpenMistral7b;
|
||||
let messages = vec![ChatCompletionMessage {
|
||||
@@ -90,6 +97,37 @@ fn main() {
|
||||
}
|
||||
```
|
||||
|
||||
### Chat without streaming (async)
|
||||
|
||||
```rs
|
||||
use mistralai_client::v1::{
|
||||
chat_completion::{ChatCompletionMessage, ChatCompletionMessageRole, ChatCompletionRequestOptions},
|
||||
client::Client,
|
||||
constants::Model,
|
||||
};
|
||||
|
||||
#[tokio::main]
|
||||
async fn main() {
|
||||
// This example suppose you have set the `MISTRAL_API_KEY` environment variable.
|
||||
let client = Client::new(None, None, None, None).unwrap();
|
||||
|
||||
let model = Model::OpenMistral7b;
|
||||
let messages = vec![ChatCompletionMessage {
|
||||
role: ChatCompletionMessageRole::user,
|
||||
content: "Just guess the next word: \"Eiffel ...\"?".to_string(),
|
||||
}];
|
||||
let options = ChatCompletionRequestOptions {
|
||||
temperature: Some(0.0),
|
||||
random_seed: Some(42),
|
||||
..Default::default()
|
||||
};
|
||||
|
||||
let result = client.chat(model, messages, Some(options)).await.unwrap();
|
||||
println!("Assistant: {}", result.choices[0].message.content);
|
||||
// => "Assistant: Tower. [...]"
|
||||
}
|
||||
```
|
||||
|
||||
### Chat with streaming
|
||||
|
||||
_In progress._
|
||||
@@ -101,7 +139,7 @@ use mistralai_client::v1::{client::Client, constants::EmbedModel};
|
||||
|
||||
fn main() {
|
||||
// This example suppose you have set the `MISTRAL_API_KEY` environment variable.
|
||||
let client: Client = Client::new(None, None, None, None);
|
||||
let client: Client = Client::new(None, None, None, None).unwrap();
|
||||
|
||||
let model = EmbedModel::MistralEmbed;
|
||||
let input = vec!["Embed this sentence.", "As well as this one."]
|
||||
@@ -116,6 +154,10 @@ fn main() {
|
||||
}
|
||||
```
|
||||
|
||||
### Embeddings (async)
|
||||
|
||||
_In progress._
|
||||
|
||||
### List models
|
||||
|
||||
```rs
|
||||
@@ -123,10 +165,14 @@ use mistralai_client::v1::client::Client;
|
||||
|
||||
fn main() {
|
||||
// This example suppose you have set the `MISTRAL_API_KEY` environment variable.
|
||||
let client = Client::new(None, None, None, None);
|
||||
let client = Client::new(None, None, None, None).unwrap();
|
||||
|
||||
let result = client.list_models(model, messages, Some(options)).unwrap();
|
||||
let result = client.list_models().unwrap();
|
||||
println!("First Model ID: {:?}", result.data[0].id);
|
||||
// => "First Model ID: open-mistral-7b"
|
||||
}
|
||||
```
|
||||
|
||||
### List models (async)
|
||||
|
||||
_In progress._
|
||||
|
||||
@@ -1,2 +1,4 @@
|
||||
# https://github.com/crate-ci/cargo-release/blob/master/docs/reference.md
|
||||
allow-branch = ["main"]
|
||||
pre-release-commit-message = "ci(release): v{{version}}"
|
||||
pre-release-replacements = [{ file = "CHANGELOG.md", search = "## \\[\\]", replace = "## [{{version}}]" }]
|
||||
|
||||
@@ -1,5 +1,6 @@
|
||||
use crate::v1::error::APIError;
|
||||
use crate::v1::error::ApiError;
|
||||
use minreq::Response;
|
||||
use reqwest::{Client as ReqwestClient, Error as ReqwestError};
|
||||
|
||||
use crate::v1::{
|
||||
chat_completion::{
|
||||
@@ -7,6 +8,7 @@ use crate::v1::{
|
||||
},
|
||||
constants::{EmbedModel, Model, API_URL_BASE},
|
||||
embedding::{EmbeddingRequest, EmbeddingRequestOptions, EmbeddingResponse},
|
||||
error::ClientError,
|
||||
model_list::ModelListResponse,
|
||||
};
|
||||
|
||||
@@ -23,18 +25,21 @@ impl Client {
|
||||
endpoint: Option<String>,
|
||||
max_retries: Option<u32>,
|
||||
timeout: Option<u32>,
|
||||
) -> Self {
|
||||
let api_key = api_key.unwrap_or(std::env::var("MISTRAL_API_KEY").unwrap());
|
||||
) -> Result<Self, ClientError> {
|
||||
let api_key = api_key.unwrap_or(match std::env::var("MISTRAL_API_KEY") {
|
||||
Ok(api_key_from_env) => api_key_from_env,
|
||||
Err(_) => return Err(ClientError::ApiKeyError),
|
||||
});
|
||||
let endpoint = endpoint.unwrap_or(API_URL_BASE.to_string());
|
||||
let max_retries = max_retries.unwrap_or(5);
|
||||
let timeout = timeout.unwrap_or(120);
|
||||
|
||||
Self {
|
||||
Ok(Self {
|
||||
api_key,
|
||||
endpoint,
|
||||
max_retries,
|
||||
timeout,
|
||||
}
|
||||
})
|
||||
}
|
||||
|
||||
pub fn build_request(&self, request: minreq::Request) -> minreq::Request {
|
||||
@@ -53,7 +58,7 @@ impl Client {
|
||||
request
|
||||
}
|
||||
|
||||
pub fn get(&self, path: &str) -> Result<Response, APIError> {
|
||||
pub fn get_sync(&self, path: &str) -> Result<Response, ApiError> {
|
||||
let url = format!("{}{}", self.endpoint, path);
|
||||
let request = self.build_request(minreq::get(url));
|
||||
|
||||
@@ -65,7 +70,7 @@ impl Client {
|
||||
if (200..=299).contains(&response.status_code) {
|
||||
Ok(response)
|
||||
} else {
|
||||
Err(APIError {
|
||||
Err(ApiError {
|
||||
message: format!(
|
||||
"{}: {}",
|
||||
response.status_code,
|
||||
@@ -74,15 +79,15 @@ impl Client {
|
||||
})
|
||||
}
|
||||
}
|
||||
Err(error) => Err(self.new_error(error)),
|
||||
Err(error) => Err(self.new_minreq_error(error)),
|
||||
}
|
||||
}
|
||||
|
||||
pub fn post<T: serde::ser::Serialize + std::fmt::Debug>(
|
||||
pub fn post_sync<T: serde::ser::Serialize + std::fmt::Debug>(
|
||||
&self,
|
||||
path: &str,
|
||||
params: &T,
|
||||
) -> Result<Response, APIError> {
|
||||
) -> Result<Response, ApiError> {
|
||||
// print!("{:?}", params);
|
||||
|
||||
let url = format!("{}{}", self.endpoint, path);
|
||||
@@ -96,7 +101,7 @@ impl Client {
|
||||
if (200..=299).contains(&response.status_code) {
|
||||
Ok(response)
|
||||
} else {
|
||||
Err(APIError {
|
||||
Err(ApiError {
|
||||
message: format!(
|
||||
"{}: {}",
|
||||
response.status_code,
|
||||
@@ -105,7 +110,7 @@ impl Client {
|
||||
})
|
||||
}
|
||||
}
|
||||
Err(error) => Err(self.new_error(error)),
|
||||
Err(error) => Err(self.new_minreq_error(error)),
|
||||
}
|
||||
}
|
||||
|
||||
@@ -114,14 +119,49 @@ impl Client {
|
||||
model: Model,
|
||||
messages: Vec<ChatCompletionMessage>,
|
||||
options: Option<ChatCompletionParams>,
|
||||
) -> Result<ChatCompletionResponse, APIError> {
|
||||
) -> Result<ChatCompletionResponse, ApiError> {
|
||||
let request = ChatCompletionRequest::new(model, messages, options);
|
||||
|
||||
let response = self.post("/chat/completions", &request)?;
|
||||
let response = self.post_sync("/chat/completions", &request)?;
|
||||
let result = response.json::<ChatCompletionResponse>();
|
||||
match result {
|
||||
Ok(response) => Ok(response),
|
||||
Err(error) => Err(self.new_error(error)),
|
||||
Err(error) => Err(self.new_minreq_error(error)),
|
||||
}
|
||||
}
|
||||
|
||||
pub async fn chat_async(
|
||||
&self,
|
||||
model: Model,
|
||||
messages: Vec<ChatCompletionMessage>,
|
||||
options: Option<ChatCompletionParams>,
|
||||
) -> Result<ChatCompletionResponse, ApiError> {
|
||||
let request = ChatCompletionRequest::new(model, messages, options);
|
||||
let client = ReqwestClient::new();
|
||||
|
||||
let response = client
|
||||
.post(format!("{}{}", self.endpoint, "/chat/completions"))
|
||||
.json(&request)
|
||||
.bearer_auth(&self.api_key)
|
||||
.header(
|
||||
"User-Agent",
|
||||
format!("mistralai-client-rs/{}", env!("CARGO_PKG_VERSION")),
|
||||
)
|
||||
.send()
|
||||
.await
|
||||
.map_err(|e| self.new_reqwest_error(e))?;
|
||||
|
||||
if response.status().is_success() {
|
||||
response
|
||||
.json::<ChatCompletionResponse>()
|
||||
.await
|
||||
.map_err(|e| self.new_reqwest_error(e))
|
||||
} else {
|
||||
let status = response.status();
|
||||
let text = response.text().await.unwrap_or_default();
|
||||
Err(ApiError {
|
||||
message: format!("{}: {}", status, text),
|
||||
})
|
||||
}
|
||||
}
|
||||
|
||||
@@ -130,28 +170,34 @@ impl Client {
|
||||
model: EmbedModel,
|
||||
input: Vec<String>,
|
||||
options: Option<EmbeddingRequestOptions>,
|
||||
) -> Result<EmbeddingResponse, APIError> {
|
||||
) -> Result<EmbeddingResponse, ApiError> {
|
||||
let request = EmbeddingRequest::new(model, input, options);
|
||||
|
||||
let response = self.post("/embeddings", &request)?;
|
||||
let response = self.post_sync("/embeddings", &request)?;
|
||||
let result = response.json::<EmbeddingResponse>();
|
||||
match result {
|
||||
Ok(response) => Ok(response),
|
||||
Err(error) => Err(self.new_error(error)),
|
||||
Err(error) => Err(self.new_minreq_error(error)),
|
||||
}
|
||||
}
|
||||
|
||||
pub fn list_models(&self) -> Result<ModelListResponse, APIError> {
|
||||
let response = self.get("/models")?;
|
||||
pub fn list_models(&self) -> Result<ModelListResponse, ApiError> {
|
||||
let response = self.get_sync("/models")?;
|
||||
let result = response.json::<ModelListResponse>();
|
||||
match result {
|
||||
Ok(response) => Ok(response),
|
||||
Err(error) => Err(self.new_error(error)),
|
||||
Err(error) => Err(self.new_minreq_error(error)),
|
||||
}
|
||||
}
|
||||
|
||||
fn new_error(&self, err: minreq::Error) -> APIError {
|
||||
APIError {
|
||||
fn new_minreq_error(&self, err: minreq::Error) -> ApiError {
|
||||
ApiError {
|
||||
message: err.to_string(),
|
||||
}
|
||||
}
|
||||
|
||||
fn new_reqwest_error(&self, err: ReqwestError) -> ApiError {
|
||||
ApiError {
|
||||
message: err.to_string(),
|
||||
}
|
||||
}
|
||||
|
||||
@@ -2,14 +2,18 @@ use std::error::Error;
|
||||
use std::fmt;
|
||||
|
||||
#[derive(Debug)]
|
||||
pub struct APIError {
|
||||
pub struct ApiError {
|
||||
pub message: String,
|
||||
}
|
||||
|
||||
impl fmt::Display for APIError {
|
||||
impl fmt::Display for ApiError {
|
||||
fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
|
||||
write!(f, "APIError: {}", self.message)
|
||||
write!(f, "ApiError: {}", self.message)
|
||||
}
|
||||
}
|
||||
impl Error for ApiError {}
|
||||
|
||||
impl Error for APIError {}
|
||||
#[derive(Debug, PartialEq, thiserror::Error)]
|
||||
pub enum ClientError {
|
||||
#[error("You must either set the `MISTRAL_API_KEY` environment variable or specify it in `Client::new(api_key, ...).")]
|
||||
ApiKeyError,
|
||||
}
|
||||
|
||||
38
tests/v1_client_chat_async_test.rs
Normal file
38
tests/v1_client_chat_async_test.rs
Normal file
@@ -0,0 +1,38 @@
|
||||
use jrest::expect;
|
||||
use mistralai_client::v1::{
|
||||
chat_completion::{ChatCompletionMessage, ChatCompletionMessageRole, ChatCompletionParams},
|
||||
client::Client,
|
||||
constants::Model,
|
||||
};
|
||||
|
||||
#[tokio::test]
|
||||
async fn test_client_chat_async() {
|
||||
let client = Client::new(None, None, None, None).unwrap();
|
||||
|
||||
let model = Model::OpenMistral7b;
|
||||
let messages = vec![ChatCompletionMessage {
|
||||
role: ChatCompletionMessageRole::user,
|
||||
content: "Just guess the next word: \"Eiffel ...\"?".to_string(),
|
||||
}];
|
||||
let options = ChatCompletionParams {
|
||||
temperature: Some(0.0),
|
||||
random_seed: Some(42),
|
||||
..Default::default()
|
||||
};
|
||||
|
||||
let response = client
|
||||
.chat_async(model, messages, Some(options))
|
||||
.await
|
||||
.unwrap();
|
||||
|
||||
expect!(response.model).to_be(Model::OpenMistral7b);
|
||||
expect!(response.object).to_be("chat.completion".to_string());
|
||||
expect!(response.choices.len()).to_be(1);
|
||||
expect!(response.choices[0].index).to_be(0);
|
||||
expect!(response.choices[0].message.role.clone()).to_be(ChatCompletionMessageRole::assistant);
|
||||
expect!(response.choices[0].message.content.clone())
|
||||
.to_be("Tower. The Eiffel Tower is a famous landmark in Paris, France.".to_string());
|
||||
expect!(response.usage.prompt_tokens).to_be_greater_than(0);
|
||||
expect!(response.usage.completion_tokens).to_be_greater_than(0);
|
||||
expect!(response.usage.total_tokens).to_be_greater_than(0);
|
||||
}
|
||||
@@ -6,13 +6,8 @@ use mistralai_client::v1::{
|
||||
};
|
||||
|
||||
#[test]
|
||||
fn test_chat_completion() {
|
||||
extern crate dotenv;
|
||||
|
||||
use dotenv::dotenv;
|
||||
dotenv().ok();
|
||||
|
||||
let client = Client::new(None, None, None, None);
|
||||
fn test_client_chat() {
|
||||
let client = Client::new(None, None, None, None).unwrap();
|
||||
|
||||
let model = Model::OpenMistral7b;
|
||||
let messages = vec![ChatCompletionMessage {
|
||||
@@ -2,13 +2,8 @@ use jrest::expect;
|
||||
use mistralai_client::v1::{client::Client, constants::EmbedModel};
|
||||
|
||||
#[test]
|
||||
fn test_embeddings() {
|
||||
extern crate dotenv;
|
||||
|
||||
use dotenv::dotenv;
|
||||
dotenv().ok();
|
||||
|
||||
let client: Client = Client::new(None, None, None, None);
|
||||
fn test_client_embeddings() {
|
||||
let client: Client = Client::new(None, None, None, None).unwrap();
|
||||
|
||||
let model = EmbedModel::MistralEmbed;
|
||||
let input = vec!["Embed this sentence.", "As well as this one."]
|
||||
@@ -2,13 +2,8 @@ use jrest::expect;
|
||||
use mistralai_client::v1::client::Client;
|
||||
|
||||
#[test]
|
||||
fn test_list_models() {
|
||||
extern crate dotenv;
|
||||
|
||||
use dotenv::dotenv;
|
||||
dotenv().ok();
|
||||
|
||||
let client = Client::new(None, None, None, None);
|
||||
fn test_client_list_models() {
|
||||
let client = Client::new(None, None, None, None).unwrap();
|
||||
|
||||
let response = client.list_models().unwrap();
|
||||
|
||||
@@ -1,12 +1,13 @@
|
||||
use jrest::expect;
|
||||
use mistralai_client::v1::client::Client;
|
||||
use mistralai_client::v1::{client::Client, error::ClientError};
|
||||
|
||||
#[test]
|
||||
fn test_client_new_with_none_params() {
|
||||
let maybe_original_mistral_api_key = std::env::var("MISTRAL_API_KEY").ok();
|
||||
std::env::remove_var("MISTRAL_API_KEY");
|
||||
std::env::set_var("MISTRAL_API_KEY", "test_api_key_from_env");
|
||||
|
||||
let client = Client::new(None, None, None, None);
|
||||
let client = Client::new(None, None, None, None).unwrap();
|
||||
|
||||
expect!(client.api_key).to_be("test_api_key_from_env".to_string());
|
||||
expect!(client.endpoint).to_be("https://api.mistral.ai/v1".to_string());
|
||||
@@ -24,6 +25,7 @@ fn test_client_new_with_none_params() {
|
||||
#[test]
|
||||
fn test_client_new_with_all_params() {
|
||||
let maybe_original_mistral_api_key = std::env::var("MISTRAL_API_KEY").ok();
|
||||
std::env::remove_var("MISTRAL_API_KEY");
|
||||
std::env::set_var("MISTRAL_API_KEY", "test_api_key_from_env");
|
||||
|
||||
let api_key = Some("test_api_key_from_param".to_string());
|
||||
@@ -36,7 +38,8 @@ fn test_client_new_with_all_params() {
|
||||
endpoint.clone(),
|
||||
max_retries.clone(),
|
||||
timeout.clone(),
|
||||
);
|
||||
)
|
||||
.unwrap();
|
||||
|
||||
expect!(client.api_key).to_be(api_key.unwrap());
|
||||
expect!(client.endpoint).to_be(endpoint.unwrap());
|
||||
@@ -50,3 +53,23 @@ fn test_client_new_with_all_params() {
|
||||
None => std::env::remove_var("MISTRAL_API_KEY"),
|
||||
}
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_client_new_with_missing_api_key() {
|
||||
let maybe_original_mistral_api_key = std::env::var("MISTRAL_API_KEY").ok();
|
||||
std::env::remove_var("MISTRAL_API_KEY");
|
||||
|
||||
let call = || Client::new(None, None, None, None);
|
||||
|
||||
match call() {
|
||||
Ok(_) => panic!("Expected `ClientError::ApiKeyError` but got Ok.`"),
|
||||
Err(error) => assert_eq!(error, ClientError::ApiKeyError),
|
||||
}
|
||||
|
||||
match maybe_original_mistral_api_key {
|
||||
Some(original_mistral_api_key) => {
|
||||
std::env::set_var("MISTRAL_API_KEY", original_mistral_api_key)
|
||||
}
|
||||
None => std::env::remove_var("MISTRAL_API_KEY"),
|
||||
}
|
||||
}
|
||||
Reference in New Issue
Block a user