diff --git a/portkey_ai/api_resources/apis/vector_stores.py b/portkey_ai/api_resources/apis/vector_stores.py index b9e31127..80fddc13 100644 --- a/portkey_ai/api_resources/apis/vector_stores.py +++ b/portkey_ai/api_resources/apis/vector_stores.py @@ -1,6 +1,7 @@ import json from typing import Any, Dict, Iterable, List, Optional, Union import typing +from portkey_ai._vendor.openai.types.vector_stores import file_batch_create_params from portkey_ai.api_resources.apis.api_resource import APIResource, AsyncAPIResource from portkey_ai.api_resources.client import AsyncPortkey, Portkey from portkey_ai.api_resources.types.shared_types import Metadata @@ -311,6 +312,7 @@ def create( vector_store_id: str, *, file_ids: Union[List[str], Omit] = omit, + files: Union[Iterable[file_batch_create_params.File], Omit] = omit, chunking_strategy: Union[Any, Omit] = omit, **kwargs, ) -> VectorStoreFileBatch: @@ -318,6 +320,7 @@ def create( self.openai_client.with_raw_response.vector_stores.file_batches.create( vector_store_id=vector_store_id, file_ids=file_ids, + files=files, chunking_strategy=chunking_strategy, **kwargs, ) @@ -741,12 +744,14 @@ async def create( vector_store_id: str, *, file_ids: Union[List[str], Omit] = omit, + files: Union[Iterable[file_batch_create_params.File], Omit] = omit, chunking_strategy: Union[Any, Omit] = omit, **kwargs, ) -> VectorStoreFileBatch: response = await self.openai_client.with_raw_response.vector_stores.file_batches.create( # noqa: E501 vector_store_id=vector_store_id, file_ids=file_ids, + files=files, chunking_strategy=chunking_strategy, **kwargs, ) diff --git a/portkey_ai/version.py b/portkey_ai/version.py index 127c148a..3c00bb49 100644 --- a/portkey_ai/version.py +++ b/portkey_ai/version.py @@ -1 +1 @@ -VERSION = "2.1.0" +VERSION = "2.2.0" diff --git a/tests/models.json b/tests/models.json index 0499881a..73033dc7 100644 --- a/tests/models.json +++ b/tests/models.json @@ -23,5 +23,92 @@ "tts-1", "whisper-1" ] + }, + "anyscale": { + "env_variable": "ANYSCALE_API_KEY", + "chat": [ + "meta-llama/Llama-2-7b-chat-hf", + "meta-llama/Llama-2-13b-chat-hf", + "meta-llama/Llama-2-70b-chat-hf", + "mistralai/Mistral-7B-Instruct-v0.1", + "mistralai/Mixtral-8x7B-Instruct-v0.1" + ], + "text": [ + "meta-llama/Llama-2-7b-chat-hf", + "meta-llama/Llama-2-13b-chat-hf", + "meta-llama/Llama-2-70b-chat-hf", + "mistralai/Mistral-7B-Instruct-v0.1", + "mistralai/Mixtral-8x7B-Instruct-v0.1" + ], + "image":[], + "audio":[] + }, + "anthropic": { + "env_variable": "ANTHROPIC_API_KEY", + "chat": [ + "claude-3-opus-20240229", + "claude-3-sonnet-20240229", + "claude-3-haiku-20240307", + "claude-2.1", + "claude-2.0", + "claude-instant-1.2" + ], + "text": [ + "claude-2.1", + "claude-2.0", + "claude-instant-1.2" + ], + "image":[], + "audio":[] + }, + "cohere": { + "env_variable": "COHERE_API_KEY", + "chat": [ + "command-light", + "command-light-nightly", + "command", + "command-nightly" + ], + "text": [ + "command-light", + "command-light-nightly", + "command", + "command-nightly" + ], + "image":[], + "audio":[] + }, + "langchain": { + "env_variable": "OPENAI_API_KEY", + "chat": [ + "gpt-4o", + "gpt-4-0125-preview", + "gpt-4-turbo-preview", + "gpt-4-1106-preview", + "gpt-4-vision-preview", + "gpt-4", + "gpt-4-0613", + "gpt-3.5-turbo", + "gpt-3.5-turbo-0125", + "gpt-3.5-turbo-1106" + ], + "text":[], + "image":[], + "audio":[] + }, + "llamaindex":{ + "env_variable": "LLAMA_INDEX_API_KEY", + "HuggingFaceEmbedding": [ + "sentence-transformers/all-MiniLM-L6-v2", + "sentence-transformers/LaBSE" + ], + "OpenAIEmbedding": [ + "text-embedding-ada-002", + "text-embedding-3-large" + ], + "chat":[], + "text":[], + "image":[], + "audio":[] } } \ No newline at end of file