Add synchronous client feature (#8)
This commit is contained in:
parent
8a2f6400b4
commit
9ded1062d1
8
.github/workflows/rust.yml
vendored
8
.github/workflows/rust.yml
vendored
@ -20,6 +20,14 @@ jobs:
|
||||
- uses: actions-rs/cargo@v1
|
||||
with:
|
||||
command: check
|
||||
- uses: actions-rs/cargo@v1
|
||||
with:
|
||||
command: check
|
||||
args: --features=sync --no-default-features
|
||||
- uses: actions-rs/cargo@v1
|
||||
with:
|
||||
command: check
|
||||
args: --features=async --no-default-features
|
||||
|
||||
test:
|
||||
name: Test Suite
|
||||
|
17
Cargo.toml
17
Cargo.toml
@ -11,22 +11,25 @@ keywords = ["openai", "gpt3"]
|
||||
categories = ["api-bindings", "asynchronous"]
|
||||
|
||||
[features]
|
||||
default = ["hyper"]
|
||||
|
||||
hyper = ["surf/hyper-client"]
|
||||
curl = ["surf/curl-client"]
|
||||
h1 = ["surf/h1-client"]
|
||||
default = ["sync", "async"]
|
||||
sync = ["ureq"]
|
||||
async = ["surf/hyper-client"]
|
||||
|
||||
[dependencies]
|
||||
surf = { version = "^2.1.0", default-features = false }
|
||||
thiserror = "^1.0.22"
|
||||
serde = { version = "^1.0.117", features = ["derive"] }
|
||||
derive_builder = "^0.9.0"
|
||||
log = "^0.4.11"
|
||||
|
||||
# Used by sync feature
|
||||
ureq = { version = "^1.5.4", optional=true, features = ["json", "tls"] }
|
||||
serde_json = { version="^1.0"}
|
||||
# Used by async feature
|
||||
surf = { version = "^2.1.0", optional=true, default-features=false}
|
||||
|
||||
[dev-dependencies]
|
||||
mockito = "0.28.0"
|
||||
maplit = "1.0.2"
|
||||
tokio = { version = "^0.2.5", features = ["full"]}
|
||||
serde_json = "^1.0"
|
||||
env_logger = "0.8.2"
|
||||
serde_json = "^1.0"
|
||||
|
@ -1,6 +1,6 @@
|
||||
use openai_api::{
|
||||
api::{CompletionArgs, Engine},
|
||||
OpenAIClient,
|
||||
Client,
|
||||
};
|
||||
|
||||
const START_PROMPT: &str = "
|
||||
@ -11,7 +11,7 @@ AI: I am an AI. How can I help you today?";
|
||||
#[tokio::main]
|
||||
async fn main() {
|
||||
let api_token = std::env::var("OPENAI_SK").unwrap();
|
||||
let client = OpenAIClient::new(&api_token);
|
||||
let client = Client::new(&api_token);
|
||||
let mut context = String::from(START_PROMPT);
|
||||
let mut args = CompletionArgs::builder();
|
||||
args.engine(Engine::Davinci)
|
||||
|
@ -1,14 +1,14 @@
|
||||
///! Example that prints out a story from a prompt. Used in the readme.
|
||||
use openai_api::OpenAIClient;
|
||||
use openai_api::Client;
|
||||
|
||||
#[tokio::main]
|
||||
async fn main() {
|
||||
let api_token = std::env::var("OPENAI_SK").unwrap();
|
||||
let client = OpenAIClient::new(&api_token);
|
||||
let client = Client::new(&api_token);
|
||||
let prompt = String::from("Once upon a time,");
|
||||
println!(
|
||||
"{}{}",
|
||||
prompt,
|
||||
client.complete(prompt.as_str()).await.unwrap()
|
||||
client.complete_prompt(prompt.as_str()).await.unwrap()
|
||||
);
|
||||
}
|
||||
|
510
src/lib.rs
510
src/lib.rs
@ -4,14 +4,14 @@ extern crate derive_builder;
|
||||
|
||||
use thiserror::Error;
|
||||
|
||||
type Result<T> = std::result::Result<T, OpenAIError>;
|
||||
type Result<T> = std::result::Result<T, Error>;
|
||||
|
||||
#[allow(clippy::default_trait_access)]
|
||||
pub mod api {
|
||||
//! Data types corresponding to requests and responses from the API
|
||||
use std::collections::HashMap;
|
||||
use std::{collections::HashMap, fmt::Display};
|
||||
|
||||
use super::OpenAIClient;
|
||||
use super::Client;
|
||||
use serde::{Deserialize, Serialize};
|
||||
|
||||
/// Container type. Used in the api, but not useful for clients of this library
|
||||
@ -115,9 +115,15 @@ pub mod api {
|
||||
/// Request a completion from the api
|
||||
///
|
||||
/// # Errors
|
||||
/// `OpenAIError::APIError` if the api returns an error
|
||||
pub async fn complete(&self, client: &OpenAIClient) -> super::Result<Completion> {
|
||||
client.complete(self.clone()).await
|
||||
/// `Error::APIError` if the api returns an error
|
||||
#[cfg(feature = "async")]
|
||||
pub async fn complete_prompt(self, client: &Client) -> super::Result<Completion> {
|
||||
client.complete_prompt(self).await
|
||||
}
|
||||
|
||||
#[cfg(feature = "sync")]
|
||||
pub fn complete_prompt_sync(self, client: &Client) -> super::Result<Completion> {
|
||||
client.complete_prompt_sync(self)
|
||||
}
|
||||
}
|
||||
|
||||
@ -125,10 +131,16 @@ pub mod api {
|
||||
/// Request a completion from the api
|
||||
///
|
||||
/// # Errors
|
||||
/// `OpenAIError::BadArguments` if the arguments to complete are not valid
|
||||
/// `OpenAIError::APIError` if the api returns an error
|
||||
pub async fn complete(&self, client: &OpenAIClient) -> super::Result<Completion> {
|
||||
client.complete(self.build()?).await
|
||||
/// `Error::BadArguments` if the arguments to complete are not valid
|
||||
/// `Error::APIError` if the api returns an error
|
||||
#[cfg(feature = "async")]
|
||||
pub async fn complete_prompt(&self, client: &Client) -> super::Result<Completion> {
|
||||
client.complete_prompt(self.build()?).await
|
||||
}
|
||||
|
||||
#[cfg(feature = "sync")]
|
||||
pub fn complete_prompt_sync(&self, client: &Client) -> super::Result<Completion> {
|
||||
client.complete_prompt_sync(self.build()?)
|
||||
}
|
||||
}
|
||||
|
||||
@ -199,55 +211,70 @@ pub mod api {
|
||||
pub error_type: String,
|
||||
}
|
||||
|
||||
impl Display for ErrorMessage {
|
||||
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
|
||||
self.message.fmt(f)
|
||||
}
|
||||
}
|
||||
|
||||
/// API-level wrapper used in deserialization
|
||||
#[derive(Deserialize, Debug)]
|
||||
pub(super) struct ErrorWrapper {
|
||||
pub(crate) struct ErrorWrapper {
|
||||
pub error: ErrorMessage,
|
||||
}
|
||||
}
|
||||
|
||||
/// This library's main `Error` type.
|
||||
#[derive(Error, Debug)]
|
||||
pub enum OpenAIError {
|
||||
pub enum Error {
|
||||
/// An error returned by the API itself
|
||||
#[error("API Returned an Error document")]
|
||||
#[error("API returned an Error: {}", .0.message)]
|
||||
APIError(api::ErrorMessage),
|
||||
/// An error the client discovers before talking to the API
|
||||
#[error("Bad arguments")]
|
||||
#[error("Bad arguments: {0}")]
|
||||
BadArguments(String),
|
||||
/// Network / protocol related errors
|
||||
#[error("Error at the protocol level")]
|
||||
ProtocolError(surf::Error),
|
||||
#[cfg(feature = "async")]
|
||||
#[error("Error at the protocol level: {0}")]
|
||||
AsyncProtocolError(surf::Error),
|
||||
#[cfg(feature = "sync")]
|
||||
#[error("Error at the protocol level, sync client")]
|
||||
SyncProtocolError(ureq::Error),
|
||||
}
|
||||
|
||||
impl From<api::ErrorMessage> for OpenAIError {
|
||||
impl From<api::ErrorMessage> for Error {
|
||||
fn from(e: api::ErrorMessage) -> Self {
|
||||
OpenAIError::APIError(e)
|
||||
Error::APIError(e)
|
||||
}
|
||||
}
|
||||
|
||||
impl From<String> for OpenAIError {
|
||||
impl From<String> for Error {
|
||||
fn from(e: String) -> Self {
|
||||
OpenAIError::BadArguments(e)
|
||||
Error::BadArguments(e)
|
||||
}
|
||||
}
|
||||
|
||||
impl From<surf::Error> for OpenAIError {
|
||||
#[cfg(feature = "async")]
|
||||
impl From<surf::Error> for Error {
|
||||
fn from(e: surf::Error) -> Self {
|
||||
OpenAIError::ProtocolError(e)
|
||||
Error::AsyncProtocolError(e)
|
||||
}
|
||||
}
|
||||
|
||||
/// Client object. Must be constructed to talk to the API.
|
||||
pub struct OpenAIClient {
|
||||
client: surf::Client,
|
||||
#[cfg(feature = "sync")]
|
||||
impl From<ureq::Error> for Error {
|
||||
fn from(e: ureq::Error) -> Self {
|
||||
Error::SyncProtocolError(e)
|
||||
}
|
||||
}
|
||||
|
||||
/// Authentication middleware
|
||||
#[cfg(feature = "async")]
|
||||
struct BearerToken {
|
||||
token: String,
|
||||
}
|
||||
|
||||
#[cfg(feature = "async")]
|
||||
impl std::fmt::Debug for BearerToken {
|
||||
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
|
||||
// Get the first few characters to help debug, but not accidentally log key
|
||||
@ -259,6 +286,7 @@ impl std::fmt::Debug for BearerToken {
|
||||
}
|
||||
}
|
||||
|
||||
#[cfg(feature = "async")]
|
||||
impl BearerToken {
|
||||
fn new(token: &str) -> Self {
|
||||
Self {
|
||||
@ -267,6 +295,7 @@ impl BearerToken {
|
||||
}
|
||||
}
|
||||
|
||||
#[cfg(feature = "async")]
|
||||
#[surf::utils::async_trait]
|
||||
impl surf::middleware::Middleware for BearerToken {
|
||||
async fn handle(
|
||||
@ -277,41 +306,100 @@ impl surf::middleware::Middleware for BearerToken {
|
||||
) -> surf::Result<surf::Response> {
|
||||
log::debug!("Request: {:?}", req);
|
||||
req.insert_header("Authorization", format!("Bearer {}", self.token));
|
||||
let response = next.run(req, client).await?;
|
||||
let response: surf::Response = next.run(req, client).await?;
|
||||
log::debug!("Response: {:?}", response);
|
||||
Ok(response)
|
||||
}
|
||||
}
|
||||
|
||||
impl OpenAIClient {
|
||||
/// Creates a new `OpenAIClient` given an api token
|
||||
#[cfg(feature = "async")]
|
||||
fn async_client(token: &str, base_url: &str) -> surf::Client {
|
||||
let mut async_client = surf::client();
|
||||
async_client.set_base_url(surf::Url::parse(base_url).expect("Static string should parse"));
|
||||
async_client.with(BearerToken::new(token))
|
||||
}
|
||||
|
||||
#[cfg(feature = "sync")]
|
||||
fn sync_client(token: &str) -> ureq::Agent {
|
||||
ureq::agent().auth_kind("Bearer", token).build()
|
||||
}
|
||||
|
||||
/// Client object. Must be constructed to talk to the API.
|
||||
pub struct Client {
|
||||
#[cfg(feature = "async")]
|
||||
async_client: surf::Client,
|
||||
#[cfg(feature = "sync")]
|
||||
sync_client: ureq::Agent,
|
||||
#[cfg(feature = "sync")]
|
||||
base_url: String,
|
||||
}
|
||||
|
||||
impl Client {
|
||||
// Creates a new `Client` given an api token
|
||||
#[must_use]
|
||||
pub fn new(token: &str) -> Self {
|
||||
let mut client = surf::client();
|
||||
client.set_base_url(
|
||||
surf::Url::parse("https://api.openai.com/v1/").expect("Static string should parse"),
|
||||
);
|
||||
client = client.with(BearerToken::new(token));
|
||||
Self { client }
|
||||
let base_url = String::from("https://api.openai.com/v1/");
|
||||
Self {
|
||||
#[cfg(feature = "async")]
|
||||
async_client: async_client(token, &base_url),
|
||||
#[cfg(feature = "sync")]
|
||||
sync_client: sync_client(token),
|
||||
#[cfg(feature = "sync")]
|
||||
base_url,
|
||||
}
|
||||
}
|
||||
|
||||
/// Allow setting the api root in the tests
|
||||
// Allow setting the api root in the tests
|
||||
#[cfg(test)]
|
||||
fn set_api_root(&mut self, url: surf::Url) {
|
||||
self.client.set_base_url(url);
|
||||
fn set_api_root(mut self, base_url: &str) -> Self {
|
||||
#[cfg(feature = "async")]
|
||||
{
|
||||
self.async_client.set_base_url(
|
||||
surf::Url::parse(base_url).expect("static URL expected to parse correctly"),
|
||||
);
|
||||
}
|
||||
#[cfg(feature = "sync")]
|
||||
{
|
||||
self.base_url = String::from(base_url);
|
||||
}
|
||||
self
|
||||
}
|
||||
|
||||
/// Private helper for making gets
|
||||
async fn get<T: serde::de::DeserializeOwned>(&self, endpoint: &str) -> Result<T> {
|
||||
let mut response = self.client.get(endpoint).await?;
|
||||
#[cfg(feature = "async")]
|
||||
async fn get<T>(&self, endpoint: &str) -> Result<T>
|
||||
where
|
||||
T: serde::de::DeserializeOwned,
|
||||
{
|
||||
let mut response = self.async_client.get(endpoint).await?;
|
||||
if let surf::StatusCode::Ok = response.status() {
|
||||
Ok(response.body_json::<T>().await?)
|
||||
} else {
|
||||
{
|
||||
let err = response.body_json::<api::ErrorWrapper>().await?.error;
|
||||
Err(OpenAIError::APIError(err))
|
||||
Err(Error::APIError(err))
|
||||
}
|
||||
}
|
||||
|
||||
#[cfg(feature = "sync")]
|
||||
fn get_sync<T>(&self, endpoint: &str) -> Result<T>
|
||||
where
|
||||
T: serde::de::DeserializeOwned,
|
||||
{
|
||||
let response = dbg!(self
|
||||
.sync_client
|
||||
.get(&format!("{}{}", self.base_url, endpoint)))
|
||||
.call();
|
||||
if let 200 = response.status() {
|
||||
Ok(response
|
||||
.into_json_deserialize()
|
||||
.expect("Bug: client couldn't deserialize api response"))
|
||||
} else {
|
||||
let err = response
|
||||
.into_json_deserialize::<api::ErrorWrapper>()
|
||||
.expect("Bug: client couldn't deserialize api error response")
|
||||
.error;
|
||||
Err(Error::APIError(err))
|
||||
}
|
||||
}
|
||||
|
||||
/// Lists the currently available engines.
|
||||
@ -319,35 +407,55 @@ impl OpenAIClient {
|
||||
/// Provides basic information about each one such as the owner and availability.
|
||||
///
|
||||
/// # Errors
|
||||
/// - `OpenAIError::APIError` if the server returns an error
|
||||
/// - `Error::APIError` if the server returns an error
|
||||
#[cfg(feature = "async")]
|
||||
pub async fn engines(&self) -> Result<Vec<api::EngineInfo>> {
|
||||
self.get("engines").await.map(|r: api::Container<_>| r.data)
|
||||
}
|
||||
|
||||
/// Lists the currently available engines.
|
||||
///
|
||||
/// Provides basic information about each one such as the owner and availability.
|
||||
///
|
||||
/// # Errors
|
||||
/// - `Error::APIError` if the server returns an error
|
||||
#[cfg(feature = "sync")]
|
||||
pub fn engines_sync(&self) -> Result<Vec<api::EngineInfo>> {
|
||||
self.get_sync("engines").map(|r: api::Container<_>| r.data)
|
||||
}
|
||||
|
||||
/// Retrieves an engine instance
|
||||
///
|
||||
/// Provides basic information about the engine such as the owner and availability.
|
||||
///
|
||||
/// # Errors
|
||||
/// - `OpenAIError::APIError` if the server returns an error
|
||||
/// - `Error::APIError` if the server returns an error
|
||||
#[cfg(feature = "async")]
|
||||
pub async fn engine(&self, engine: api::Engine) -> Result<api::EngineInfo> {
|
||||
self.get(&format!("engines/{}", engine)).await
|
||||
}
|
||||
|
||||
#[cfg(feature = "sync")]
|
||||
pub fn engine_sync(&self, engine: api::Engine) -> Result<api::EngineInfo> {
|
||||
self.get_sync(&format!("engines/{}", engine))
|
||||
}
|
||||
|
||||
// Private helper to generate post requests. Needs to be a bit more flexible than
|
||||
// get because it should support SSE eventually
|
||||
#[cfg(feature = "async")]
|
||||
async fn post<B, R>(&self, endpoint: &str, body: B) -> Result<R>
|
||||
where
|
||||
B: serde::ser::Serialize,
|
||||
R: serde::de::DeserializeOwned,
|
||||
{
|
||||
let mut response = self
|
||||
.client
|
||||
.async_client
|
||||
.post(endpoint)
|
||||
.body(surf::Body::from_json(&body)?)
|
||||
.await?;
|
||||
match response.status() {
|
||||
surf::StatusCode::Ok => Ok(response.body_json::<R>().await?),
|
||||
_ => Err(OpenAIError::APIError(
|
||||
_ => Err(Error::APIError(
|
||||
response
|
||||
.body_json::<api::ErrorWrapper>()
|
||||
.await
|
||||
@ -356,11 +464,38 @@ impl OpenAIClient {
|
||||
)),
|
||||
}
|
||||
}
|
||||
|
||||
#[cfg(feature = "sync")]
|
||||
fn post_sync<B, R>(&self, endpoint: &str, body: B) -> Result<R>
|
||||
where
|
||||
B: serde::ser::Serialize,
|
||||
R: serde::de::DeserializeOwned,
|
||||
{
|
||||
let response = self
|
||||
.sync_client
|
||||
.post(&format!("{}{}", self.base_url, endpoint))
|
||||
.send_json(
|
||||
serde_json::to_value(body).expect("Bug: client couldn't serialize its own type"),
|
||||
);
|
||||
match response.status() {
|
||||
200 => Ok(response
|
||||
.into_json_deserialize()
|
||||
.expect("Bug: client couldn't deserialize api response")),
|
||||
_ => Err(Error::APIError(
|
||||
response
|
||||
.into_json_deserialize::<api::ErrorWrapper>()
|
||||
.expect("Bug: client couldn't deserialize api error response")
|
||||
.error,
|
||||
)),
|
||||
}
|
||||
}
|
||||
|
||||
/// Get predicted completion of the prompt
|
||||
///
|
||||
/// # Errors
|
||||
/// - `OpenAIError::APIError` if the api returns an error
|
||||
pub async fn complete(
|
||||
/// - `Error::APIError` if the api returns an error
|
||||
#[cfg(feature = "async")]
|
||||
pub async fn complete_prompt(
|
||||
&self,
|
||||
prompt: impl Into<api::CompletionArgs>,
|
||||
) -> Result<api::Completion> {
|
||||
@ -369,20 +504,60 @@ impl OpenAIClient {
|
||||
.post(&format!("engines/{}/completions", args.engine), args)
|
||||
.await?)
|
||||
}
|
||||
|
||||
/// Get predicted completion of the prompt synchronously
|
||||
///
|
||||
/// # Error
|
||||
/// - `Error::APIError` if the api returns an error
|
||||
#[cfg(feature = "sync")]
|
||||
pub fn complete_prompt_sync(
|
||||
&self,
|
||||
prompt: impl Into<api::CompletionArgs>,
|
||||
) -> Result<api::Completion> {
|
||||
let args = prompt.into();
|
||||
self.post_sync(&format!("engines/{}/completions", args.engine), args)
|
||||
}
|
||||
}
|
||||
|
||||
// TODO: add a macro to de-boilerplate the sync and async tests
|
||||
|
||||
#[allow(unused_macros)]
|
||||
macro_rules! async_test {
|
||||
($test_name: ident, $test_body: block) => {
|
||||
#[cfg(feature = "async")]
|
||||
#[tokio::test]
|
||||
async fn $test_name() -> crate::Result<()> {
|
||||
$test_body;
|
||||
Ok(())
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
#[allow(unused_macros)]
|
||||
macro_rules! sync_test {
|
||||
($test_name: ident, $test_body: expr) => {
|
||||
#[cfg(feature = "sync")]
|
||||
#[test]
|
||||
fn $test_name() -> crate::Result<()> {
|
||||
$test_body;
|
||||
Ok(())
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
#[cfg(test)]
|
||||
mod unit {
|
||||
|
||||
use crate::{api, OpenAIClient, OpenAIError};
|
||||
use mockito::Mock;
|
||||
|
||||
fn mocked_client() -> OpenAIClient {
|
||||
use crate::{
|
||||
api::{self, Completion, CompletionArgs, Engine, EngineInfo},
|
||||
Client, Error,
|
||||
};
|
||||
|
||||
fn mocked_client() -> Client {
|
||||
let _ = env_logger::builder().is_test(true).try_init();
|
||||
let mut client = OpenAIClient::new("bogus");
|
||||
client.set_api_root(
|
||||
surf::Url::parse(&mockito::server_url()).expect("mockito url didn't parse"),
|
||||
);
|
||||
client
|
||||
Client::new("bogus").set_api_root(&format!("{}/", mockito::server_url()))
|
||||
}
|
||||
|
||||
#[test]
|
||||
@ -410,10 +585,8 @@ mod unit {
|
||||
Ok(())
|
||||
}
|
||||
|
||||
#[tokio::test]
|
||||
async fn parse_engines() -> crate::Result<()> {
|
||||
use api::{Engine, EngineInfo};
|
||||
let _m = mockito::mock("GET", "/engines")
|
||||
fn mock_engines() -> (Mock, Vec<EngineInfo>) {
|
||||
let mock = mockito::mock("GET", "/engines")
|
||||
.with_status(200)
|
||||
.with_header("content-type", "application/json")
|
||||
.with_body(
|
||||
@ -460,6 +633,7 @@ mod unit {
|
||||
}"#,
|
||||
)
|
||||
.create();
|
||||
|
||||
let expected = vec![
|
||||
EngineInfo {
|
||||
id: Engine::Ada,
|
||||
@ -492,14 +666,23 @@ mod unit {
|
||||
ready: true,
|
||||
},
|
||||
];
|
||||
let response = mocked_client().engines().await?;
|
||||
assert_eq!(response, expected);
|
||||
Ok(())
|
||||
(mock, expected)
|
||||
}
|
||||
|
||||
#[tokio::test]
|
||||
async fn engine_error_response() -> crate::Result<()> {
|
||||
let _m = mockito::mock("GET", "/engines/davinci")
|
||||
async_test!(parse_engines_async, {
|
||||
let (_m, expected) = mock_engines();
|
||||
let response = mocked_client().engines().await?;
|
||||
assert_eq!(response, expected);
|
||||
});
|
||||
|
||||
sync_test!(parse_engines_sync, {
|
||||
let (_m, expected) = mock_engines();
|
||||
let response = mocked_client().engines_sync()?;
|
||||
assert_eq!(response, expected);
|
||||
});
|
||||
|
||||
fn mock_engine() -> (Mock, api::ErrorMessage) {
|
||||
let mock = mockito::mock("GET", "/engines/davinci")
|
||||
.with_status(404)
|
||||
.with_header("content-type", "application/json")
|
||||
.with_body(
|
||||
@ -516,16 +699,26 @@ mod unit {
|
||||
message: "Some kind of error happened".into(),
|
||||
error_type: "some_error_type".into(),
|
||||
};
|
||||
let response = mocked_client().engine(api::Engine::Davinci).await;
|
||||
if let Result::Err(OpenAIError::APIError(msg)) = response {
|
||||
assert_eq!(expected, msg);
|
||||
}
|
||||
Ok(())
|
||||
(mock, expected)
|
||||
}
|
||||
|
||||
#[tokio::test]
|
||||
async fn completion_args() -> crate::Result<()> {
|
||||
let _m = mockito::mock("POST", "/engines/davinci/completions")
|
||||
async_test!(engine_error_response_async, {
|
||||
let (_m, expected) = mock_engine();
|
||||
let response = mocked_client().engine(api::Engine::Davinci).await;
|
||||
if let Result::Err(Error::APIError(msg)) = response {
|
||||
assert_eq!(expected, msg);
|
||||
}
|
||||
});
|
||||
|
||||
sync_test!(engine_error_response_sync, {
|
||||
let (_m, expected) = mock_engine();
|
||||
let response = mocked_client().engine_sync(api::Engine::Davinci);
|
||||
if let Result::Err(Error::APIError(msg)) = response {
|
||||
assert_eq!(expected, msg);
|
||||
}
|
||||
});
|
||||
fn mock_completion() -> crate::Result<(Mock, CompletionArgs, Completion)> {
|
||||
let mock = mockito::mock("POST", "/engines/davinci/completions")
|
||||
.with_status(200)
|
||||
.with_header("content-type", "application/json")
|
||||
.with_body(
|
||||
@ -544,7 +737,17 @@ mod unit {
|
||||
]
|
||||
}"#,
|
||||
)
|
||||
.expect(1)
|
||||
.create();
|
||||
let args = api::CompletionArgs::builder()
|
||||
.engine(api::Engine::Davinci)
|
||||
.prompt("Once upon a time")
|
||||
.max_tokens(5)
|
||||
.temperature(1.0)
|
||||
.top_p(1.0)
|
||||
.n(1)
|
||||
.stop(vec!["\n".into()])
|
||||
.build()?;
|
||||
let expected = api::Completion {
|
||||
id: "cmpl-uqkvlQyYK7bGYrRHQ0eXlWi7".into(),
|
||||
created: 1589478378,
|
||||
@ -556,54 +759,75 @@ mod unit {
|
||||
finish_reason: api::FinishReason::MaxTokensReached,
|
||||
}],
|
||||
};
|
||||
let args = api::CompletionArgs::builder()
|
||||
.engine(api::Engine::Davinci)
|
||||
.prompt("Once upon a time")
|
||||
.max_tokens(5)
|
||||
.temperature(1.0)
|
||||
.top_p(1.0)
|
||||
.n(1)
|
||||
.stop(vec!["\n".into()])
|
||||
.build()?;
|
||||
let response = mocked_client().complete(args).await?;
|
||||
assert_eq!(response.model, expected.model);
|
||||
assert_eq!(response.id, expected.id);
|
||||
assert_eq!(response.created, expected.created);
|
||||
let (resp_choice, expected_choice) = (&response.choices[0], &expected.choices[0]);
|
||||
assert_eq!(resp_choice.text, expected_choice.text);
|
||||
assert_eq!(resp_choice.index, expected_choice.index);
|
||||
assert!(resp_choice.logprobs.is_none());
|
||||
assert_eq!(resp_choice.finish_reason, expected_choice.finish_reason);
|
||||
|
||||
Ok(())
|
||||
Ok((mock, args, expected))
|
||||
}
|
||||
|
||||
// Defines boilerplate here. The Completion can't derive Eq since it contains
|
||||
// floats in various places.
|
||||
fn assert_completion_equal(a: Completion, b: Completion) {
|
||||
assert_eq!(a.model, b.model);
|
||||
assert_eq!(a.id, b.id);
|
||||
assert_eq!(a.created, b.created);
|
||||
let (a_choice, b_choice) = (&a.choices[0], &b.choices[0]);
|
||||
assert_eq!(a_choice.text, b_choice.text);
|
||||
assert_eq!(a_choice.index, b_choice.index);
|
||||
assert!(a_choice.logprobs.is_none());
|
||||
assert_eq!(a_choice.finish_reason, b_choice.finish_reason);
|
||||
}
|
||||
|
||||
async_test!(completion_args_async, {
|
||||
let (m, args, expected) = mock_completion()?;
|
||||
let response = mocked_client().complete_prompt(args).await?;
|
||||
assert_completion_equal(response, expected);
|
||||
m.assert();
|
||||
});
|
||||
|
||||
sync_test!(completion_args_sync, {
|
||||
let (m, args, expected) = mock_completion()?;
|
||||
let response = mocked_client().complete_prompt_sync(args)?;
|
||||
assert_completion_equal(response, expected);
|
||||
m.assert();
|
||||
});
|
||||
}
|
||||
#[cfg(test)]
|
||||
mod integration {
|
||||
|
||||
use api::ErrorMessage;
|
||||
|
||||
use crate::{api, OpenAIClient, OpenAIError};
|
||||
use crate::{
|
||||
api::{self, Completion},
|
||||
Client, Error,
|
||||
};
|
||||
/// Used by tests to get a client to the actual api
|
||||
fn get_client() -> OpenAIClient {
|
||||
fn get_client() -> Client {
|
||||
let _ = env_logger::builder().is_test(true).try_init();
|
||||
let sk = std::env::var("OPENAI_SK").expect(
|
||||
"To run integration tests, you must put set the OPENAI_SK env var to your api token",
|
||||
);
|
||||
OpenAIClient::new(&sk)
|
||||
Client::new(&sk)
|
||||
}
|
||||
|
||||
#[tokio::test]
|
||||
async fn can_get_engines() {
|
||||
async_test!(can_get_engines_async, {
|
||||
let client = get_client();
|
||||
client.engines().await.unwrap();
|
||||
}
|
||||
#[tokio::test]
|
||||
async fn can_get_engine() {
|
||||
client.engines().await?
|
||||
});
|
||||
|
||||
sync_test!(can_get_engines_sync, {
|
||||
let client = get_client();
|
||||
let result = client.engine(api::Engine::Ada).await;
|
||||
let engines = client
|
||||
.engines_sync()?
|
||||
.into_iter()
|
||||
.map(|ei| ei.id)
|
||||
.collect::<Vec<_>>();
|
||||
assert!(engines.contains(&api::Engine::Ada));
|
||||
assert!(engines.contains(&api::Engine::Babbage));
|
||||
assert!(engines.contains(&api::Engine::Curie));
|
||||
assert!(engines.contains(&api::Engine::Davinci));
|
||||
});
|
||||
|
||||
fn assert_expected_engine_failure<T>(result: Result<T, Error>)
|
||||
where
|
||||
T: std::fmt::Debug,
|
||||
{
|
||||
match result {
|
||||
Err(OpenAIError::APIError(ErrorMessage {
|
||||
Err(Error::APIError(api::ErrorMessage {
|
||||
message,
|
||||
error_type,
|
||||
})) => {
|
||||
@ -615,18 +839,28 @@ mod integration {
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
#[tokio::test]
|
||||
async fn complete_string() -> crate::Result<()> {
|
||||
async_test!(can_get_engine_async, {
|
||||
let client = get_client();
|
||||
client.complete("Hey there").await?;
|
||||
Ok(())
|
||||
}
|
||||
assert_expected_engine_failure(client.engine(api::Engine::Ada).await);
|
||||
});
|
||||
|
||||
#[tokio::test]
|
||||
async fn complete_explicit_params() -> crate::Result<()> {
|
||||
sync_test!(can_get_engine_sync, {
|
||||
let client = get_client();
|
||||
let args = api::CompletionArgsBuilder::default()
|
||||
assert_expected_engine_failure(client.engine_sync(api::Engine::Ada));
|
||||
});
|
||||
|
||||
async_test!(complete_string_async, {
|
||||
let client = get_client();
|
||||
client.complete_prompt("Hey there").await?;
|
||||
});
|
||||
|
||||
sync_test!(complete_string_sync, {
|
||||
let client = get_client();
|
||||
client.complete_prompt_sync("Hey there")?;
|
||||
});
|
||||
|
||||
fn create_args() -> api::CompletionArgs {
|
||||
api::CompletionArgsBuilder::default()
|
||||
.prompt("Once upon a time,")
|
||||
.max_tokens(10)
|
||||
.temperature(0.5)
|
||||
@ -642,17 +876,23 @@ mod integration {
|
||||
"23".into() => 0.0,
|
||||
})
|
||||
.build()
|
||||
.expect("Build should have succeeded");
|
||||
client.complete(args).await?;
|
||||
Ok(())
|
||||
.expect("Bug: build should succeed")
|
||||
}
|
||||
|
||||
#[tokio::test]
|
||||
async fn complete_stop_condition() -> crate::Result<()> {
|
||||
async_test!(complete_explicit_params_async, {
|
||||
let client = get_client();
|
||||
let args = create_args();
|
||||
client.complete_prompt(args).await?;
|
||||
});
|
||||
|
||||
sync_test!(complete_explicit_params_sync, {
|
||||
let client = get_client();
|
||||
let args = create_args();
|
||||
client.complete_prompt_sync(args)?
|
||||
});
|
||||
|
||||
fn stop_condition_args() -> api::CompletionArgs {
|
||||
let mut args = api::CompletionArgs::builder();
|
||||
let completion = args
|
||||
.prompt(
|
||||
args.prompt(
|
||||
r#"
|
||||
Q: Please type `#` now
|
||||
A:"#,
|
||||
@ -662,8 +902,26 @@ A:"#,
|
||||
.top_p(0.0)
|
||||
.max_tokens(100)
|
||||
.stop(vec!["#".into(), "\n".into()])
|
||||
.complete(&client).await?;
|
||||
assert_eq!(completion.choices[0].finish_reason, api::FinishReason::StopSequenceReached);
|
||||
Ok(())
|
||||
.build()
|
||||
.expect("Bug: build should succeed")
|
||||
}
|
||||
|
||||
fn assert_completion_finish_reason(completion: Completion) {
|
||||
assert_eq!(
|
||||
completion.choices[0].finish_reason,
|
||||
api::FinishReason::StopSequenceReached
|
||||
);
|
||||
}
|
||||
|
||||
async_test!(complete_stop_condition_async, {
|
||||
let client = get_client();
|
||||
let args = stop_condition_args();
|
||||
assert_completion_finish_reason(client.complete_prompt(args).await?);
|
||||
});
|
||||
|
||||
sync_test!(complete_stop_condition_sync, {
|
||||
let client = get_client();
|
||||
let args = stop_condition_args();
|
||||
assert_completion_finish_reason(client.complete_prompt_sync(args)?);
|
||||
});
|
||||
}
|
||||
|
Loading…
x
Reference in New Issue
Block a user