diff --git a/README.md b/README.md index cd03ecc..7791b46 100644 --- a/README.md +++ b/README.md @@ -105,7 +105,7 @@ client = ZhipuAiClient(api_key="your-api-key") # Create chat completion response = client.chat.completions.create( - model="glm-4.7", + model="glm-5", messages=[ {"role": "user", "content": "Hello, Z.ai!"} ] @@ -279,7 +279,7 @@ client = ZaiClient(api_key="your-api-key") try: response = client.chat.completions.create( - model="glm-4.7", + model="glm-5", messages=[ {"role": "user", "content": "Hello, Z.ai!"} ] diff --git a/README_CN.md b/README_CN.md index 0c4d08b..4cbd0a5 100644 --- a/README_CN.md +++ b/README_CN.md @@ -107,7 +107,7 @@ client = ZhipuAiClient(api_key="your-api-key") # Create chat completion response = client.chat.completions.create( - model="glm-4.7", + model="glm-5", messages=[ {"role": "user", "content": "Hello, Z.ai!"} ] @@ -286,7 +286,7 @@ client = ZaiClient(api_key="your-api-key") # 请填写您自己的APIKey try: response = client.chat.completions.create( - model="glm-4.7", + model="glm-5", messages=[ {"role": "user", "content": "你好, Z.ai !"} ] diff --git a/examples/basic_usage.py b/examples/basic_usage.py index e0b6a4b..cce0689 100644 --- a/examples/basic_usage.py +++ b/examples/basic_usage.py @@ -6,7 +6,7 @@ def completion(): # Create chat completion response = client.chat.completions.create( - model='glm-4.7', + model='glm-5', messages=[{'role': 'user', 'content': 'Hello, Z.ai!'}], temperature=1.0, ) @@ -19,7 +19,7 @@ def completion_with_stream(): # Create chat completion response = client.chat.completions.create( - model='glm-4.7', + model='glm-5', messages=[ {'role': 'system', 'content': 'You are a helpful assistant.'}, {'role': 'user', 'content': 'Tell me a story about AI.'}, @@ -38,7 +38,7 @@ def completion_with_websearch(): # Create chat completion response = client.chat.completions.create( - model='glm-4.7', + model='glm-5', messages=[ {'role': 'system', 'content': 'You are a helpful assistant.'}, {'role': 'user', 'content': 'What is artificial intelligence?'}, @@ -66,7 +66,7 @@ def completion_with_mcp_server_url(): # Create chat completion with MCP server URL response = client.chat.completions.create( - model='glm-4.7', + model='glm-5', stream=False, messages=[{'role': 'user', 'content': 'Hello, please introduce GPT?'}], tools=[ @@ -95,7 +95,7 @@ def completion_with_mcp_server_label(): # Create chat completion with MCP server label response = client.chat.completions.create( - model='glm-4.7', + model='glm-5', stream=False, messages=[{'role': 'user', 'content': 'Hello, please introduce GPT?'}], tools=[ @@ -217,7 +217,7 @@ def ofZai(): client = ZaiClient() print(client.base_url) response = client.chat.completions.create( - model='glm-4.7', + model='glm-5', messages=[{'role': 'user', 'content': 'Hello, Z.ai!'}], temperature=0.7, ) @@ -227,7 +227,7 @@ def ofZhipu(): client = ZhipuAiClient() print(client.base_url) response = client.chat.completions.create( - model='glm-4.7', + model='glm-5', messages=[{'role': 'user', 'content': 'Hello, Z.ai!'}], temperature=0.7, ) diff --git a/examples/function_call_example.py b/examples/function_call_example.py index 0e01c03..8e661e3 100644 --- a/examples/function_call_example.py +++ b/examples/function_call_example.py @@ -36,7 +36,7 @@ def parse_function_call(model_response, messages): "tool_call_id": tool_call.id }) response = client.chat.completions.create( - model="glm-4.7", # Specify the model name to use + model="glm-5", # Specify the model name to use messages=messages, tools=tools, ) @@ -99,7 +99,7 @@ def parse_function_call(model_response, messages): messages.append({"role": "user", "content": "Help me check the flights from Beijing to Guangzhou on January 23."}) response = client.chat.completions.create( - model="glm-4.7", # Specify the model name to use + model="glm-5", # Specify the model name to use messages=messages, tools=tools, ) @@ -110,7 +110,7 @@ def parse_function_call(model_response, messages): messages.append({"role": "user", "content": "What is the price of flight 8321?"}) response = client.chat.completions.create( - model="glm-4.7", # Specify the model name to use + model="glm-5", # Specify the model name to use messages=messages, tools=tools, ) diff --git a/examples/glm_thinking.py b/examples/glm_thinking.py index f081935..09918e5 100644 --- a/examples/glm_thinking.py +++ b/examples/glm_thinking.py @@ -15,7 +15,7 @@ class ZaiSampler(SamplerBase): def __init__( self, - model: str = "glm-4.7", + model: str = "glm-5", api_key: str = '', system_message: Optional[str] = None, temperature: float = 0.0, diff --git a/examples/stream_tools.py b/examples/stream_tools.py index 37b3a2d..1906f8a 100644 --- a/examples/stream_tools.py +++ b/examples/stream_tools.py @@ -4,7 +4,7 @@ def main(): client = ZhipuAiClient() # create chat completion with tool calls and streaming response = client.chat.completions.create( - model="glm-4.7", + model="glm-5", messages=[ {"role": "user", "content": "How is the weather in Beijing and Shanghai? Please provide the answer in Celsius."}, ],