From 6a47dcf8257852ded2e5151d3254d05e8017da66 Mon Sep 17 00:00:00 2001 From: Wendong-Fan <133094783+Wendong-Fan@users.noreply.github.com> Date: Fri, 25 Oct 2024 13:04:36 +0800 Subject: [PATCH] fix: llama tool call msg & remove samba fast api (#1120) --- .github/ISSUE_TEMPLATE/bug_report.yml | 2 +- README.md | 2 +- camel/__init__.py | 2 +- camel/agents/chat_agent.py | 8 +- camel/configs/__init__.py | 4 - camel/configs/samba_config.py | 39 +------- camel/models/samba_model.py | 2 +- docs/conf.py | 2 +- docs/cookbooks/agents_message.ipynb | 2 +- docs/cookbooks/agents_prompting.ipynb | 2 +- docs/cookbooks/agents_society.ipynb | 2 +- docs/cookbooks/agents_tracking.ipynb | 2 +- docs/cookbooks/agents_with_memory.ipynb | 2 +- docs/cookbooks/agents_with_rag.ipynb | 2 +- docs/cookbooks/agents_with_tools.ipynb | 2 +- docs/cookbooks/create_your_first_agent.ipynb | 2 +- .../create_your_first_agents_society.ipynb | 2 +- .../critic_agents_and_tree_search.ipynb | 2 +- docs/cookbooks/embodied_agents.ipynb | 2 +- docs/cookbooks/knowledge_graph.ipynb | 2 +- docs/cookbooks/model_speed_comparison.ipynb | 2 +- docs/cookbooks/roleplaying_scraper.ipynb | 2 +- docs/cookbooks/task_generation.ipynb | 2 +- docs/cookbooks/video_analysis.ipynb | 2 +- .../cookbooks/workforce_judge_committee.ipynb | 2 +- docs/get_started/installation.md | 2 +- examples/models/samba_model_example.py | 93 ++++++------------- poetry.lock | 24 ++--- pyproject.toml | 2 +- 29 files changed, 73 insertions(+), 143 deletions(-) diff --git a/.github/ISSUE_TEMPLATE/bug_report.yml b/.github/ISSUE_TEMPLATE/bug_report.yml index f3435717d1..26c25a0d8e 100644 --- a/.github/ISSUE_TEMPLATE/bug_report.yml +++ b/.github/ISSUE_TEMPLATE/bug_report.yml @@ -26,7 +26,7 @@ body: attributes: label: What version of camel are you using? description: Run command `python3 -c 'print(__import__("camel").__version__)'` in your shell and paste the output here. - placeholder: E.g., 0.2.5 + placeholder: E.g., 0.2.6 validations: required: true diff --git a/README.md b/README.md index 794b483d73..885b14f9a6 100644 --- a/README.md +++ b/README.md @@ -116,7 +116,7 @@ conda create --name camel python=3.10 conda activate camel # Clone github repo -git clone -b v0.2.5 https://github.com/camel-ai/camel.git +git clone -b v0.2.6 https://github.com/camel-ai/camel.git # Change directory into project directory cd camel diff --git a/camel/__init__.py b/camel/__init__.py index ca65ff1e39..da74eb223e 100644 --- a/camel/__init__.py +++ b/camel/__init__.py @@ -12,7 +12,7 @@ # limitations under the License. # =========== Copyright 2023 @ CAMEL-AI.org. All Rights Reserved. =========== -__version__ = '0.2.5' +__version__ = '0.2.6' __all__ = [ '__version__', diff --git a/camel/agents/chat_agent.py b/camel/agents/chat_agent.py index bfe75a662e..d6e77f50d0 100644 --- a/camel/agents/chat_agent.py +++ b/camel/agents/chat_agent.py @@ -217,6 +217,8 @@ def __init__( self.response_terminators = response_terminators or [] self.init_messages() + self.tool_prompt_added = False + # ruff: noqa: E501 def _generate_tool_prompt(self, tool_schema_list: List[Dict]) -> str: tool_prompts = [] @@ -448,7 +450,10 @@ def step( ) if "llama" in self.model_type.lower(): - if self.model_backend.model_config_dict.get("tools", None): + if ( + self.model_backend.model_config_dict.get("tools", None) + and not self.tool_prompt_added + ): tool_prompt = self._generate_tool_prompt(self.tool_schema_list) tool_sys_msg = BaseMessage.make_assistant_message( @@ -457,6 +462,7 @@ def step( ) self.update_memory(tool_sys_msg, OpenAIBackendRole.SYSTEM) + self.tool_prompt_added = True self.update_memory(input_message, OpenAIBackendRole.USER) diff --git a/camel/configs/__init__.py b/camel/configs/__init__.py index d3e1f9bf83..a7b5f2ce79 100644 --- a/camel/configs/__init__.py +++ b/camel/configs/__init__.py @@ -22,10 +22,8 @@ from .reka_config import REKA_API_PARAMS, RekaConfig from .samba_config import ( SAMBA_CLOUD_API_PARAMS, - SAMBA_FAST_API_PARAMS, SAMBA_VERSE_API_PARAMS, SambaCloudAPIConfig, - SambaFastAPIConfig, SambaVerseAPIConfig, ) from .togetherai_config import TOGETHERAI_API_PARAMS, TogetherAIConfig @@ -54,8 +52,6 @@ 'MISTRAL_API_PARAMS', 'RekaConfig', 'REKA_API_PARAMS', - 'SambaFastAPIConfig', - 'SAMBA_FAST_API_PARAMS', 'SambaVerseAPIConfig', 'SAMBA_VERSE_API_PARAMS', 'SambaCloudAPIConfig', diff --git a/camel/configs/samba_config.py b/camel/configs/samba_config.py index 61a0a1ed89..fd57581f56 100644 --- a/camel/configs/samba_config.py +++ b/camel/configs/samba_config.py @@ -13,7 +13,7 @@ # =========== Copyright 2023 @ CAMEL-AI.org. All Rights Reserved. =========== from __future__ import annotations -from typing import Any, Dict, Optional, Sequence, Union +from typing import Any, Optional, Sequence, Union from pydantic import Field @@ -21,43 +21,6 @@ from camel.types import NOT_GIVEN, NotGiven -class SambaFastAPIConfig(BaseConfig): - r"""Defines the parameters for generating chat completions using the - SambaNova Fast API. - - Args: - max_tokens (Optional[int], optional): the maximum number of tokens to - generate, e.g. 100. - (default: :obj:`2048`) - stop (Optional[Union[str,list[str]]]): Stop generation if this token - is detected. Or if one of these tokens is detected when providing - a string list. - (default: :obj:`None`) - stream (Optional[bool]): If True, partial message deltas will be sent - as data-only server-sent events as they become available. - Currently SambaNova Fast API only support stream mode. - (default: :obj:`True`) - stream_options (Optional[Dict]): Additional options for streaming. - (default: :obj:`{"include_usage": True}`) - """ - - max_tokens: Optional[int] = 2048 - stop: Optional[Union[str, list[str]]] = None - stream: Optional[bool] = True - stream_options: Optional[Dict] = {"include_usage": True} # noqa: RUF012 - - def as_dict(self) -> dict[str, Any]: - config_dict = super().as_dict() - if "tools" in config_dict: - del config_dict["tools"] # SambaNova does not support tool calling - return config_dict - - -SAMBA_FAST_API_PARAMS = { - param for param in SambaFastAPIConfig().model_fields.keys() -} - - class SambaVerseAPIConfig(BaseConfig): r"""Defines the parameters for generating chat completions using the SambaVerse API. diff --git a/camel/models/samba_model.py b/camel/models/samba_model.py index a5bb1bbf7e..fa7a9442a2 100644 --- a/camel/models/samba_model.py +++ b/camel/models/samba_model.py @@ -147,7 +147,7 @@ def check_model_config(self): def run( # type: ignore[misc] self, messages: List[OpenAIMessage] ) -> Union[ChatCompletion, Stream[ChatCompletionChunk]]: - r"""Runs SambaNova's FastAPI service. + r"""Runs SambaNova's service. Args: messages (List[OpenAIMessage]): Message list with the chat history diff --git a/docs/conf.py b/docs/conf.py index 556d2e428f..d8d027bb8b 100644 --- a/docs/conf.py +++ b/docs/conf.py @@ -27,7 +27,7 @@ project = 'CAMEL' copyright = '2024, CAMEL-AI.org' author = 'CAMEL-AI.org' -release = '0.2.5' +release = '0.2.6' html_favicon = ( 'https://raw.githubusercontent.com/camel-ai/camel/master/misc/favicon.png' diff --git a/docs/cookbooks/agents_message.ipynb b/docs/cookbooks/agents_message.ipynb index 47f4810713..ed4800c7b1 100644 --- a/docs/cookbooks/agents_message.ipynb +++ b/docs/cookbooks/agents_message.ipynb @@ -101,7 +101,7 @@ }, "outputs": [], "source": [ - "!pip install \"camel-ai==0.2.5\"" + "!pip install \"camel-ai==0.2.6\"" ] }, { diff --git a/docs/cookbooks/agents_prompting.ipynb b/docs/cookbooks/agents_prompting.ipynb index f26fdcd49d..7e393b2205 100644 --- a/docs/cookbooks/agents_prompting.ipynb +++ b/docs/cookbooks/agents_prompting.ipynb @@ -64,7 +64,7 @@ }, "outputs": [], "source": [ - "!pip install \"camel-ai==0.2.5\"" + "!pip install \"camel-ai==0.2.6\"" ] }, { diff --git a/docs/cookbooks/agents_society.ipynb b/docs/cookbooks/agents_society.ipynb index a48bf47041..a7bac22e6d 100644 --- a/docs/cookbooks/agents_society.ipynb +++ b/docs/cookbooks/agents_society.ipynb @@ -192,7 +192,7 @@ }, "outputs": [], "source": [ - "!pip install \"camel-ai==0.2.5\"" + "!pip install \"camel-ai==0.2.6\"" ] }, { diff --git a/docs/cookbooks/agents_tracking.ipynb b/docs/cookbooks/agents_tracking.ipynb index cc65c9ab63..a2522bf333 100644 --- a/docs/cookbooks/agents_tracking.ipynb +++ b/docs/cookbooks/agents_tracking.ipynb @@ -64,7 +64,7 @@ }, "outputs": [], "source": [ - "%pip install camel-ai[all]==0.2.5\n", + "%pip install camel-ai[all]==0.2.6\n", "%pip install agentops==0.3.10" ] }, diff --git a/docs/cookbooks/agents_with_memory.ipynb b/docs/cookbooks/agents_with_memory.ipynb index 1394e8d90d..334bbf9fd5 100644 --- a/docs/cookbooks/agents_with_memory.ipynb +++ b/docs/cookbooks/agents_with_memory.ipynb @@ -73,7 +73,7 @@ }, "outputs": [], "source": [ - "!pip install \"camel-ai[all]==0.2.5\"" + "!pip install \"camel-ai[all]==0.2.6\"" ] }, { diff --git a/docs/cookbooks/agents_with_rag.ipynb b/docs/cookbooks/agents_with_rag.ipynb index 539adc221b..d0d5c4ad45 100644 --- a/docs/cookbooks/agents_with_rag.ipynb +++ b/docs/cookbooks/agents_with_rag.ipynb @@ -75,7 +75,7 @@ }, "outputs": [], "source": [ - "!pip install \"camel-ai[all]==0.2.5\"" + "!pip install \"camel-ai[all]==0.2.6\"" ] }, { diff --git a/docs/cookbooks/agents_with_tools.ipynb b/docs/cookbooks/agents_with_tools.ipynb index 44e974717d..eae19ddb4f 100644 --- a/docs/cookbooks/agents_with_tools.ipynb +++ b/docs/cookbooks/agents_with_tools.ipynb @@ -77,7 +77,7 @@ }, "outputs": [], "source": [ - "!pip install \"camel-ai[all]==0.2.5\"" + "!pip install \"camel-ai[all]==0.2.6\"" ] }, { diff --git a/docs/cookbooks/create_your_first_agent.ipynb b/docs/cookbooks/create_your_first_agent.ipynb index a5e75b4f95..3eb0977c27 100644 --- a/docs/cookbooks/create_your_first_agent.ipynb +++ b/docs/cookbooks/create_your_first_agent.ipynb @@ -83,7 +83,7 @@ { "cell_type": "code", "source": [ - "!pip install \"camel-ai[all]==0.2.5\"" + "!pip install \"camel-ai[all]==0.2.6\"" ], "metadata": { "id": "UtcC3c-KVZmU" diff --git a/docs/cookbooks/create_your_first_agents_society.ipynb b/docs/cookbooks/create_your_first_agents_society.ipynb index d4aa34c2f6..c9cf04204b 100644 --- a/docs/cookbooks/create_your_first_agents_society.ipynb +++ b/docs/cookbooks/create_your_first_agents_society.ipynb @@ -77,7 +77,7 @@ { "cell_type": "code", "source": [ - "!pip install \"camel-ai==0.2.5\"" + "!pip install \"camel-ai==0.2.6\"" ], "metadata": { "id": "RiwfwyyLYYxo" diff --git a/docs/cookbooks/critic_agents_and_tree_search.ipynb b/docs/cookbooks/critic_agents_and_tree_search.ipynb index 019f89dc2d..6780124551 100644 --- a/docs/cookbooks/critic_agents_and_tree_search.ipynb +++ b/docs/cookbooks/critic_agents_and_tree_search.ipynb @@ -84,7 +84,7 @@ { "cell_type": "code", "source": [ - "%pip install \"camel-ai==0.2.5\"" + "%pip install \"camel-ai==0.2.6\"" ], "metadata": { "id": "UtcC3c-KVZmU" diff --git a/docs/cookbooks/embodied_agents.ipynb b/docs/cookbooks/embodied_agents.ipynb index 44b0f7635c..4c5eb710b6 100644 --- a/docs/cookbooks/embodied_agents.ipynb +++ b/docs/cookbooks/embodied_agents.ipynb @@ -67,7 +67,7 @@ { "cell_type": "code", "source": [ - "%pip install \"camel-ai==0.2.5\"" + "%pip install \"camel-ai==0.2.6\"" ], "metadata": { "id": "UtcC3c-KVZmU" diff --git a/docs/cookbooks/knowledge_graph.ipynb b/docs/cookbooks/knowledge_graph.ipynb index fac042c032..0aa8510a47 100644 --- a/docs/cookbooks/knowledge_graph.ipynb +++ b/docs/cookbooks/knowledge_graph.ipynb @@ -73,7 +73,7 @@ }, "outputs": [], "source": [ - "pip install \"camel-ai[all]==0.2.5\"" + "pip install \"camel-ai[all]==0.2.6\"" ] }, { diff --git a/docs/cookbooks/model_speed_comparison.ipynb b/docs/cookbooks/model_speed_comparison.ipynb index b4327df3d3..06acc253b5 100644 --- a/docs/cookbooks/model_speed_comparison.ipynb +++ b/docs/cookbooks/model_speed_comparison.ipynb @@ -63,7 +63,7 @@ { "cell_type": "code", "source": [ - "!pip install \"camel-ai==0.2.5\"" + "!pip install \"camel-ai==0.2.6\"" ], "metadata": { "id": "UtcC3c-KVZmU" diff --git a/docs/cookbooks/roleplaying_scraper.ipynb b/docs/cookbooks/roleplaying_scraper.ipynb index 4027486586..ce7feb4696 100644 --- a/docs/cookbooks/roleplaying_scraper.ipynb +++ b/docs/cookbooks/roleplaying_scraper.ipynb @@ -89,7 +89,7 @@ }, "outputs": [], "source": [ - "!pip install \"camel-ai[all]==0.2.5\"" + "!pip install \"camel-ai[all]==0.2.6\"" ] }, { diff --git a/docs/cookbooks/task_generation.ipynb b/docs/cookbooks/task_generation.ipynb index 9989b0d30c..ab00c67b35 100644 --- a/docs/cookbooks/task_generation.ipynb +++ b/docs/cookbooks/task_generation.ipynb @@ -62,7 +62,7 @@ }, "outputs": [], "source": [ - "!pip install \"camel-ai==0.2.5\"" + "!pip install \"camel-ai==0.2.6\"" ] }, { diff --git a/docs/cookbooks/video_analysis.ipynb b/docs/cookbooks/video_analysis.ipynb index a9e629f0a2..b028581cc8 100644 --- a/docs/cookbooks/video_analysis.ipynb +++ b/docs/cookbooks/video_analysis.ipynb @@ -28,7 +28,7 @@ }, "outputs": [], "source": [ - "%pip install \"camel-ai[all]==0.2.5\"" + "%pip install \"camel-ai[all]==0.2.6\"" ] }, { diff --git a/docs/cookbooks/workforce_judge_committee.ipynb b/docs/cookbooks/workforce_judge_committee.ipynb index b13cdf0015..f8471a40ab 100644 --- a/docs/cookbooks/workforce_judge_committee.ipynb +++ b/docs/cookbooks/workforce_judge_committee.ipynb @@ -28,7 +28,7 @@ }, "outputs": [], "source": [ - "%pip install \"camel-ai[all]==0.2.5\"" + "%pip install \"camel-ai[all]==0.2.6\"" ] }, { diff --git a/docs/get_started/installation.md b/docs/get_started/installation.md index c6b09cfaa8..495907b2fd 100644 --- a/docs/get_started/installation.md +++ b/docs/get_started/installation.md @@ -60,7 +60,7 @@ conda create --name camel python=3.10 conda activate camel # Clone github repo -git clone -b v0.2.5 https://github.com/camel-ai/camel.git +git clone -b v0.2.6 https://github.com/camel-ai/camel.git # Change directory into project directory cd camel diff --git a/examples/models/samba_model_example.py b/examples/models/samba_model_example.py index 64fd53dada..57eed29d1e 100644 --- a/examples/models/samba_model_example.py +++ b/examples/models/samba_model_example.py @@ -14,29 +14,20 @@ from camel.agents import ChatAgent from camel.configs import ( SambaCloudAPIConfig, - SambaFastAPIConfig, SambaVerseAPIConfig, ) -from camel.messages import BaseMessage from camel.models import ModelFactory from camel.types import ModelPlatformType -samba_fast_api_model = ModelFactory.create( - model_platform=ModelPlatformType.SAMBA, - model_type="llama3-70b", - model_config_dict=SambaFastAPIConfig(max_tokens=800).as_dict(), - api_key="Your SambaNova Fast API Key", - url="https://fast-api.snova.ai/v1/chat/completions", -) +# Define system message +sys_msg = "You are a helpful assistant." + +# Define user message +user_msg = """Say hi to CAMEL AI, one open-source community dedicated to the + study of autonomous and communicative agents.""" -sambaverse_api_model = ModelFactory.create( - model_platform=ModelPlatformType.SAMBA, - model_type="Mistral/Mistral-7B-Instruct-v0.2", - model_config_dict=SambaVerseAPIConfig(max_tokens=800).as_dict(), - api_key="Your SambaVerse API Key", - url="https://sambaverse.sambanova.ai/api/predict", -) +# Use Samba Cloud model samba_cloud_api_model = ModelFactory.create( model_platform=ModelPlatformType.SAMBA, model_type="Meta-Llama-3.1-405B-Instruct", @@ -45,50 +36,41 @@ url="https://api.sambanova.ai/v1", ) - -# Define system message -sys_msg = BaseMessage.make_assistant_message( - role_name="Assistant", - content="You are a helpful assistant.", -) - -# Define user message -user_msg = BaseMessage.make_user_message( - role_name="User", - content="""Say hi to CAMEL AI, one open-source community dedicated to the - study of autonomous and communicative agents.""", -) - # Set agent -camel_agent_fast_api = ChatAgent( - system_message=sys_msg, model=samba_fast_api_model -) - -camel_agent_sambaverse_api = ChatAgent( - system_message=sys_msg, model=sambaverse_api_model -) - camel_agent_samba_cloud_api = ChatAgent( system_message=sys_msg, model=samba_cloud_api_model ) # Get response information -response = camel_agent_fast_api.step(user_msg) +response = camel_agent_samba_cloud_api.step(user_msg) print(response.msgs[0].content) ''' =============================================================================== -Hello CAMEL AI community! - -It's great to connect with a group of innovators and researchers passionate -about autonomous and communicative agents. Your work in this field has the -potential to revolutionize numerous industries and improve human lives. +Hello to the CAMEL AI community. It's great to see open-source communities +like yours working on autonomous and communicative agents, as this field has +the potential to revolutionize many areas of our lives, from customer service +to healthcare and beyond. -What exciting projects are you currently working on? Are there any specific -challenges or areas where I can offer assistance or provide information? I'm -here to help and support your endeavors! +What specific projects or initiatives is the CAMEL AI community currently +working on? Are there any exciting developments or breakthroughs that you'd +like to share? I'm all ears (or rather, all text) and happy to learn more +about your work! =============================================================================== ''' +# Use Samba Verse model +sambaverse_api_model = ModelFactory.create( + model_platform=ModelPlatformType.SAMBA, + model_type="Mistral/Mistral-7B-Instruct-v0.2", + model_config_dict=SambaVerseAPIConfig(max_tokens=800).as_dict(), + api_key="Your SambaVerse API Key", + url="https://sambaverse.sambanova.ai/api/predict", +) + +# Set agent +camel_agent_sambaverse_api = ChatAgent( + system_message=sys_msg, model=sambaverse_api_model +) # Get response information response = camel_agent_sambaverse_api.step(user_msg) @@ -101,20 +83,3 @@ of assistance. =============================================================================== ''' - -# Get response information -response = camel_agent_samba_cloud_api.step(user_msg) -print(response.msgs[0].content) -''' -=============================================================================== -Hello to the CAMEL AI community. It's great to see open-source communities -like yours working on autonomous and communicative agents, as this field has -the potential to revolutionize many areas of our lives, from customer service -to healthcare and beyond. - -What specific projects or initiatives is the CAMEL AI community currently -working on? Are there any exciting developments or breakthroughs that you'd -like to share? I'm all ears (or rather, all text) and happy to learn more -about your work! -=============================================================================== -''' diff --git a/poetry.lock b/poetry.lock index c95fcb8004..efaa1f47c4 100644 --- a/poetry.lock +++ b/poetry.lock @@ -650,13 +650,13 @@ css = ["tinycss2 (>=1.1.0,<1.3)"] [[package]] name = "botocore" -version = "1.35.47" +version = "1.35.48" description = "Low-level, data-driven core of boto 3." optional = true python-versions = ">=3.8" files = [ - {file = "botocore-1.35.47-py3-none-any.whl", hash = "sha256:05f4493119a96799ff84d43e78691efac3177e1aec8840cca99511de940e342a"}, - {file = "botocore-1.35.47.tar.gz", hash = "sha256:f8f703463d3cd8b6abe2bedc443a7ab29f0e2ff1588a2e83164b108748645547"}, + {file = "botocore-1.35.48-py3-none-any.whl", hash = "sha256:34fa25fd717208b05745e60f271a39636108fa87a3512fbca18e7e6f787a3239"}, + {file = "botocore-1.35.48.tar.gz", hash = "sha256:3e766cc251053c9ef98542fdf225381ed58531769c3811a6282bd7247f7e2bdf"}, ] [package.dependencies] @@ -2228,13 +2228,13 @@ tracing = ["opentelemetry-api (>=1.1.0)"] [[package]] name = "google-cloud-vision" -version = "3.7.4" +version = "3.8.0" description = "Google Cloud Vision API client library" optional = true python-versions = ">=3.7" files = [ - {file = "google_cloud_vision-3.7.4-py2.py3-none-any.whl", hash = "sha256:0b956480002545ab8f13d2b4b8f316e9332cdeb6f65f92c0a20d72e9e0df3ad6"}, - {file = "google_cloud_vision-3.7.4.tar.gz", hash = "sha256:80b67f0a2dc587a31d7482d3d2692a773f25fbd09468fd9de45d00b671aad999"}, + {file = "google_cloud_vision-3.8.0-py2.py3-none-any.whl", hash = "sha256:cf63d8dbd1d1b5ead11eb061aeedbe139105029de7558f5bfbcd7b9abeaf6476"}, + {file = "google_cloud_vision-3.8.0.tar.gz", hash = "sha256:591c9d1ca26308f10557fd0b1feccebaa7a65199f7027b77dc7311820400720b"}, ] [package.dependencies] @@ -3448,13 +3448,13 @@ files = [ [[package]] name = "litellm" -version = "1.50.2" +version = "1.50.4" description = "Library to easily interface with LLM API providers" optional = true python-versions = "!=2.7.*,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,!=3.7.*,>=3.8" files = [ - {file = "litellm-1.50.2-py3-none-any.whl", hash = "sha256:99cac60c78037946ab809b7cfbbadad53507bb2db8ae39391b4be215a0869fdd"}, - {file = "litellm-1.50.2.tar.gz", hash = "sha256:b244c9a0e069cc626b85fb9f5cc252114aaff1225500da30ce0940f841aef8ea"}, + {file = "litellm-1.50.4-py3-none-any.whl", hash = "sha256:cc6992275e24a0bbb4a3b377e6842d45a8510fc85d7f255930a64bb872980a36"}, + {file = "litellm-1.50.4.tar.gz", hash = "sha256:a7e68ef614f631b58969c2c7a5154a565ba5974558d437c8cd6c8623654880ea"}, ] [package.dependencies] @@ -8881,13 +8881,13 @@ urllib3 = ">=2" [[package]] name = "types-setuptools" -version = "75.2.0.20241019" +version = "75.2.0.20241025" description = "Typing stubs for setuptools" optional = false python-versions = ">=3.8" files = [ - {file = "types-setuptools-75.2.0.20241019.tar.gz", hash = "sha256:86ea31b5f6df2c6b8f2dc8ae3f72b213607f62549b6fa2ed5866e5299f968694"}, - {file = "types_setuptools-75.2.0.20241019-py3-none-any.whl", hash = "sha256:2e48ff3acd4919471e80d5e3f049cce5c177e108d5d36d2d4cee3fa4d4104258"}, + {file = "types-setuptools-75.2.0.20241025.tar.gz", hash = "sha256:2949913a518d5285ce00a3b7d88961c80a6e72ffb8f3da0a3f5650ea533bd45e"}, + {file = "types_setuptools-75.2.0.20241025-py3-none-any.whl", hash = "sha256:6721ac0f1a620321e2ccd87a9a747c4a383dc381f78d894ce37f2455b45fcf1c"}, ] [[package]] diff --git a/pyproject.toml b/pyproject.toml index 7093a61010..929d6cc0f8 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -4,7 +4,7 @@ build-backend = "poetry.core.masonry.api" [tool.poetry] name = "camel-ai" -version = "0.2.5" +version = "0.2.6" authors = ["CAMEL-AI.org"] description = "Communicative Agents for AI Society Study" readme = "README.md"