Skip to content

Commit 05671b1

Browse files
committed
format
1 parent 2134b57 commit 05671b1

18 files changed

Lines changed: 214 additions & 176 deletions

File tree

examples/v1/autogen/example.ipynb

Lines changed: 25 additions & 25 deletions
Original file line numberDiff line numberDiff line change
@@ -46,8 +46,8 @@
4646
"\n",
4747
"from metorial import Metorial\n",
4848
"from metorial.integrations.autogen import (\n",
49-
" create_autogen_tools,\n",
50-
" get_autogen_tool_executor,\n",
49+
" create_autogen_tools,\n",
50+
" get_autogen_tool_executor,\n",
5151
")"
5252
]
5353
},
@@ -67,34 +67,34 @@
6767
"outputs": [],
6868
"source": [
6969
"async def run_agents(query: str):\n",
70-
" async with metorial.provider_session(\n",
71-
" provider=\"openai\",\n",
72-
" server_deployments=[os.getenv(\"EXA_DEPLOYMENT_ID\")],\n",
73-
" ) as session:\n",
74-
" tools = create_autogen_tools(session)\n",
75-
" tool_executor = get_autogen_tool_executor(session)\n",
70+
" async with metorial.provider_session(\n",
71+
" provider=\"openai\",\n",
72+
" server_deployments=[os.getenv(\"EXA_DEPLOYMENT_ID\")],\n",
73+
" ) as session:\n",
74+
" tools = create_autogen_tools(session)\n",
75+
" tool_executor = get_autogen_tool_executor(session)\n",
7676
"\n",
77-
" print(f\"Available tools: {[t['function']['name'] for t in tools]}\")\n",
77+
" print(f\"Available tools: {[t['function']['name'] for t in tools]}\")\n",
7878
"\n",
79-
" llm_config = {\n",
80-
" \"config_list\": [{\"model\": \"gpt-4o\", \"api_key\": os.getenv(\"OPENAI_API_KEY\")}],\n",
81-
" \"tools\": tools,\n",
82-
" }\n",
79+
" llm_config = {\n",
80+
" \"config_list\": [{\"model\": \"gpt-4o\", \"api_key\": os.getenv(\"OPENAI_API_KEY\")}],\n",
81+
" \"tools\": tools,\n",
82+
" }\n",
8383
"\n",
84-
" assistant = AssistantAgent(\n",
85-
" name=\"research_assistant\",\n",
86-
" system_message=\"You are a helpful research assistant.\",\n",
87-
" llm_config=llm_config,\n",
88-
" )\n",
84+
" assistant = AssistantAgent(\n",
85+
" name=\"research_assistant\",\n",
86+
" system_message=\"You are a helpful research assistant.\",\n",
87+
" llm_config=llm_config,\n",
88+
" )\n",
8989
"\n",
90-
" user_proxy = UserProxyAgent(\n",
91-
" name=\"user\",\n",
92-
" human_input_mode=\"NEVER\",\n",
93-
" max_consecutive_auto_reply=5,\n",
94-
" function_map=tool_executor,\n",
95-
" )\n",
90+
" user_proxy = UserProxyAgent(\n",
91+
" name=\"user\",\n",
92+
" human_input_mode=\"NEVER\",\n",
93+
" max_consecutive_auto_reply=5,\n",
94+
" function_map=tool_executor,\n",
95+
" )\n",
9696
"\n",
97-
" await user_proxy.a_initiate_chat(assistant, message=query)"
97+
" await user_proxy.a_initiate_chat(assistant, message=query)"
9898
]
9999
},
100100
{

examples/v1/haystack/example.ipynb

Lines changed: 15 additions & 15 deletions
Original file line numberDiff line numberDiff line change
@@ -67,26 +67,26 @@
6767
"outputs": [],
6868
"source": [
6969
"async def run_pipeline(query: str):\n",
70-
" async with metorial.provider_session(\n",
71-
" provider=\"openai\",\n",
72-
" server_deployments=[os.getenv(\"EXA_DEPLOYMENT_ID\")],\n",
73-
" ) as session:\n",
74-
" tools = create_haystack_tools(session)\n",
70+
" async with metorial.provider_session(\n",
71+
" provider=\"openai\",\n",
72+
" server_deployments=[os.getenv(\"EXA_DEPLOYMENT_ID\")],\n",
73+
" ) as session:\n",
74+
" tools = create_haystack_tools(session)\n",
7575
"\n",
76-
" print(f\"Available tools: {[t.name for t in tools]}\")\n",
76+
" print(f\"Available tools: {[t.name for t in tools]}\")\n",
7777
"\n",
78-
" generator = OpenAIChatGenerator(model=\"gpt-4o\", tools=tools)\n",
79-
" tool_invoker = ToolInvoker(tools=tools)\n",
78+
" generator = OpenAIChatGenerator(model=\"gpt-4o\", tools=tools)\n",
79+
" tool_invoker = ToolInvoker(tools=tools)\n",
8080
"\n",
81-
" pipeline = Pipeline()\n",
82-
" pipeline.add_component(\"generator\", generator)\n",
83-
" pipeline.add_component(\"tool_invoker\", tool_invoker)\n",
84-
" pipeline.connect(\"generator.replies\", \"tool_invoker.messages\")\n",
81+
" pipeline = Pipeline()\n",
82+
" pipeline.add_component(\"generator\", generator)\n",
83+
" pipeline.add_component(\"tool_invoker\", tool_invoker)\n",
84+
" pipeline.connect(\"generator.replies\", \"tool_invoker.messages\")\n",
8585
"\n",
86-
" messages = [ChatMessage.from_user(query)]\n",
87-
" result = pipeline.run({\"generator\": {\"messages\": messages}})\n",
86+
" messages = [ChatMessage.from_user(query)]\n",
87+
" result = pipeline.run({\"generator\": {\"messages\": messages}})\n",
8888
"\n",
89-
" return result[\"tool_invoker\"][\"tool_messages\"]"
89+
" return result[\"tool_invoker\"][\"tool_messages\"]"
9090
]
9191
},
9292
{

examples/v1/langchain/example.ipynb

Lines changed: 18 additions & 16 deletions
Original file line numberDiff line numberDiff line change
@@ -67,26 +67,28 @@
6767
"outputs": [],
6868
"source": [
6969
"async def run_agent(query: str):\n",
70-
" async with metorial.provider_session(\n",
71-
" provider=\"anthropic\",\n",
72-
" server_deployments=[os.getenv(\"EXA_DEPLOYMENT_ID\")],\n",
73-
" ) as session:\n",
74-
" tools = create_langchain_tools(session)\n",
70+
" async with metorial.provider_session(\n",
71+
" provider=\"anthropic\",\n",
72+
" server_deployments=[os.getenv(\"EXA_DEPLOYMENT_ID\")],\n",
73+
" ) as session:\n",
74+
" tools = create_langchain_tools(session)\n",
7575
"\n",
76-
" print(f\"Available tools: {[t.name for t in tools]}\")\n",
76+
" print(f\"Available tools: {[t.name for t in tools]}\")\n",
7777
"\n",
78-
" llm = ChatAnthropic(model=\"claude-sonnet-4-20250514\")\n",
79-
" prompt = ChatPromptTemplate.from_messages([\n",
80-
" (\"system\", \"You are a helpful research assistant.\"),\n",
81-
" (\"human\", \"{input}\"),\n",
82-
" (\"placeholder\", \"{agent_scratchpad}\"),\n",
83-
" ])\n",
78+
" llm = ChatAnthropic(model=\"claude-sonnet-4-20250514\")\n",
79+
" prompt = ChatPromptTemplate.from_messages(\n",
80+
" [\n",
81+
" (\"system\", \"You are a helpful research assistant.\"),\n",
82+
" (\"human\", \"{input}\"),\n",
83+
" (\"placeholder\", \"{agent_scratchpad}\"),\n",
84+
" ]\n",
85+
" )\n",
8486
"\n",
85-
" agent = create_tool_calling_agent(llm, tools, prompt)\n",
86-
" executor = AgentExecutor(agent=agent, tools=tools, verbose=True)\n",
87+
" agent = create_tool_calling_agent(llm, tools, prompt)\n",
88+
" executor = AgentExecutor(agent=agent, tools=tools, verbose=True)\n",
8789
"\n",
88-
" result = await executor.ainvoke({\"input\": query})\n",
89-
" return result[\"output\"]"
90+
" result = await executor.ainvoke({\"input\": query})\n",
91+
" return result[\"output\"]"
9092
]
9193
},
9294
{

examples/v1/langgraph/example.ipynb

Lines changed: 14 additions & 16 deletions
Original file line numberDiff line numberDiff line change
@@ -65,26 +65,24 @@
6565
"outputs": [],
6666
"source": [
6767
"async def run_agent(query: str):\n",
68-
" async with metorial.provider_session(\n",
69-
" provider=\"openai\",\n",
70-
" server_deployments=[os.getenv(\"EXA_DEPLOYMENT_ID\")],\n",
71-
" ) as session:\n",
72-
" tools = create_langgraph_tools(session)\n",
68+
" async with metorial.provider_session(\n",
69+
" provider=\"openai\",\n",
70+
" server_deployments=[os.getenv(\"EXA_DEPLOYMENT_ID\")],\n",
71+
" ) as session:\n",
72+
" tools = create_langgraph_tools(session)\n",
7373
"\n",
74-
" print(f\"Available tools: {[t.name for t in tools]}\")\n",
74+
" print(f\"Available tools: {[t.name for t in tools]}\")\n",
7575
"\n",
76-
" llm = ChatOpenAI(model=\"gpt-4o\")\n",
77-
" agent = create_react_agent(llm, tools)\n",
76+
" llm = ChatOpenAI(model=\"gpt-4o\")\n",
77+
" agent = create_react_agent(llm, tools)\n",
7878
"\n",
79-
" final_response = None\n",
80-
" async for event in agent.astream(\n",
81-
" {\"messages\": [(\"user\", query)]}\n",
82-
" ):\n",
83-
" if \"agent\" in event:\n",
84-
" final_response = event[\"agent\"][\"messages\"][-1].content\n",
85-
" print(f\"Agent: {final_response}\")\n",
79+
" final_response = None\n",
80+
" async for event in agent.astream({\"messages\": [(\"user\", query)]}):\n",
81+
" if \"agent\" in event:\n",
82+
" final_response = event[\"agent\"][\"messages\"][-1].content\n",
83+
" print(f\"Agent: {final_response}\")\n",
8684
"\n",
87-
" return final_response"
85+
" return final_response"
8886
]
8987
},
9088
{

examples/v1/llamaindex/example.ipynb

Lines changed: 10 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -65,19 +65,19 @@
6565
"outputs": [],
6666
"source": [
6767
"async def run_agent(query: str):\n",
68-
" async with metorial.provider_session(\n",
69-
" provider=\"openai\",\n",
70-
" server_deployments=[os.getenv(\"EXA_DEPLOYMENT_ID\")],\n",
71-
" ) as session:\n",
72-
" tools = create_llamaindex_tools(session)\n",
68+
" async with metorial.provider_session(\n",
69+
" provider=\"openai\",\n",
70+
" server_deployments=[os.getenv(\"EXA_DEPLOYMENT_ID\")],\n",
71+
" ) as session:\n",
72+
" tools = create_llamaindex_tools(session)\n",
7373
"\n",
74-
" print(f\"Available tools: {[t.metadata.name for t in tools]}\")\n",
74+
" print(f\"Available tools: {[t.metadata.name for t in tools]}\")\n",
7575
"\n",
76-
" llm = OpenAI(model=\"gpt-4o\")\n",
77-
" agent = ReActAgent.from_tools(tools, llm=llm, verbose=True)\n",
76+
" llm = OpenAI(model=\"gpt-4o\")\n",
77+
" agent = ReActAgent.from_tools(tools, llm=llm, verbose=True)\n",
7878
"\n",
79-
" response = agent.chat(query)\n",
80-
" return response"
79+
" response = agent.chat(query)\n",
80+
" return response"
8181
]
8282
},
8383
{

examples/v1/openai-agents/example.ipynb

Lines changed: 13 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -65,22 +65,22 @@
6565
"outputs": [],
6666
"source": [
6767
"async def run_agent(query: str):\n",
68-
" async with metorial.provider_session(\n",
69-
" provider=\"openai\",\n",
70-
" server_deployments=[os.getenv(\"EXA_DEPLOYMENT_ID\")],\n",
71-
" ) as session:\n",
72-
" tools = create_openai_agent_tools(session)\n",
68+
" async with metorial.provider_session(\n",
69+
" provider=\"openai\",\n",
70+
" server_deployments=[os.getenv(\"EXA_DEPLOYMENT_ID\")],\n",
71+
" ) as session:\n",
72+
" tools = create_openai_agent_tools(session)\n",
7373
"\n",
74-
" print(f\"Available tools: {[t.name for t in tools]}\")\n",
74+
" print(f\"Available tools: {[t.name for t in tools]}\")\n",
7575
"\n",
76-
" agent = Agent(\n",
77-
" name=\"Research Assistant\",\n",
78-
" instructions=\"You are a helpful research assistant. Use the available tools to find information.\",\n",
79-
" tools=tools,\n",
80-
" )\n",
76+
" agent = Agent(\n",
77+
" name=\"Research Assistant\",\n",
78+
" instructions=\"You are a helpful research assistant. Use the available tools to find information.\",\n",
79+
" tools=tools,\n",
80+
" )\n",
8181
"\n",
82-
" result = await Runner.run(agent, query)\n",
83-
" return result.final_output"
82+
" result = await Runner.run(agent, query)\n",
83+
" return result.final_output"
8484
]
8585
},
8686
{

examples/v1/pydantic-ai/example.ipynb

Lines changed: 13 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -65,22 +65,22 @@
6565
"outputs": [],
6666
"source": [
6767
"async def run_agent(query: str):\n",
68-
" async with metorial.provider_session(\n",
69-
" provider=\"openai\",\n",
70-
" server_deployments=[os.getenv(\"EXA_DEPLOYMENT_ID\")],\n",
71-
" ) as session:\n",
72-
" agent = Agent(\n",
73-
" \"openai:gpt-4o\",\n",
74-
" system_prompt=\"You are a helpful research assistant.\",\n",
75-
" )\n",
68+
" async with metorial.provider_session(\n",
69+
" provider=\"openai\",\n",
70+
" server_deployments=[os.getenv(\"EXA_DEPLOYMENT_ID\")],\n",
71+
" ) as session:\n",
72+
" agent = Agent(\n",
73+
" \"openai:gpt-4o\",\n",
74+
" system_prompt=\"You are a helpful research assistant.\",\n",
75+
" )\n",
7676
"\n",
77-
" # Register Metorial tools with the agent\n",
78-
" register_metorial_tools(agent, session)\n",
77+
" # Register Metorial tools with the agent\n",
78+
" register_metorial_tools(agent, session)\n",
7979
"\n",
80-
" print(\"Tools registered with agent\")\n",
80+
" print(\"Tools registered with agent\")\n",
8181
"\n",
82-
" result = await agent.run(query)\n",
83-
" return result.data"
82+
" result = await agent.run(query)\n",
83+
" return result.data"
8484
]
8585
},
8686
{

examples/v1/semantic-kernel/example.ipynb

Lines changed: 23 additions & 23 deletions
Original file line numberDiff line numberDiff line change
@@ -44,7 +44,7 @@
4444
"source": [
4545
"import semantic_kernel as sk\n",
4646
"from semantic_kernel.connectors.ai.function_choice_behavior import (\n",
47-
" FunctionChoiceBehavior,\n",
47+
" FunctionChoiceBehavior,\n",
4848
")\n",
4949
"from semantic_kernel.connectors.ai.open_ai import OpenAIChatCompletion\n",
5050
"from semantic_kernel.contents.chat_history import ChatHistory\n",
@@ -69,33 +69,33 @@
6969
"outputs": [],
7070
"source": [
7171
"async def run_kernel(query: str):\n",
72-
" async with metorial.provider_session(\n",
73-
" provider=\"openai\",\n",
74-
" server_deployments=[os.getenv(\"EXA_DEPLOYMENT_ID\")],\n",
75-
" ) as session:\n",
76-
" kernel = sk.Kernel()\n",
72+
" async with metorial.provider_session(\n",
73+
" provider=\"openai\",\n",
74+
" server_deployments=[os.getenv(\"EXA_DEPLOYMENT_ID\")],\n",
75+
" ) as session:\n",
76+
" kernel = sk.Kernel()\n",
7777
"\n",
78-
" service = OpenAIChatCompletion(\n",
79-
" service_id=\"chat\",\n",
80-
" ai_model_id=\"gpt-4o\",\n",
81-
" )\n",
82-
" kernel.add_service(service)\n",
78+
" service = OpenAIChatCompletion(\n",
79+
" service_id=\"chat\",\n",
80+
" ai_model_id=\"gpt-4o\",\n",
81+
" )\n",
82+
" kernel.add_service(service)\n",
8383
"\n",
84-
" register_metorial_plugin(kernel, session)\n",
85-
" print(\"Metorial plugin registered\")\n",
84+
" register_metorial_plugin(kernel, session)\n",
85+
" print(\"Metorial plugin registered\")\n",
8686
"\n",
87-
" settings = kernel.get_prompt_execution_settings_from_service_id(\"chat\")\n",
88-
" settings.function_choice_behavior = FunctionChoiceBehavior.Auto()\n",
87+
" settings = kernel.get_prompt_execution_settings_from_service_id(\"chat\")\n",
88+
" settings.function_choice_behavior = FunctionChoiceBehavior.Auto()\n",
8989
"\n",
90-
" history = ChatHistory()\n",
91-
" history.add_user_message(query)\n",
90+
" history = ChatHistory()\n",
91+
" history.add_user_message(query)\n",
9292
"\n",
93-
" result = await kernel.invoke_prompt(\n",
94-
" prompt=\"{{$history}}\",\n",
95-
" history=history,\n",
96-
" settings=settings,\n",
97-
" )\n",
98-
" return result"
93+
" result = await kernel.invoke_prompt(\n",
94+
" prompt=\"{{$history}}\",\n",
95+
" history=history,\n",
96+
" settings=settings,\n",
97+
" )\n",
98+
" return result"
9999
]
100100
},
101101
{

examples/v1/smolagents/example.ipynb

Lines changed: 10 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -64,19 +64,19 @@
6464
"outputs": [],
6565
"source": [
6666
"async def run_agent(query: str):\n",
67-
" async with metorial.provider_session(\n",
68-
" provider=\"openai\",\n",
69-
" server_deployments=[os.getenv(\"EXA_DEPLOYMENT_ID\")],\n",
70-
" ) as session:\n",
71-
" tools = create_smolagents_tools(session)\n",
67+
" async with metorial.provider_session(\n",
68+
" provider=\"openai\",\n",
69+
" server_deployments=[os.getenv(\"EXA_DEPLOYMENT_ID\")],\n",
70+
" ) as session:\n",
71+
" tools = create_smolagents_tools(session)\n",
7272
"\n",
73-
" print(f\"Available tools: {[t.name for t in tools]}\")\n",
73+
" print(f\"Available tools: {[t.name for t in tools]}\")\n",
7474
"\n",
75-
" model = LiteLLMModel(model_id=\"gpt-4o\")\n",
76-
" agent = CodeAgent(tools=tools, model=model)\n",
75+
" model = LiteLLMModel(model_id=\"gpt-4o\")\n",
76+
" agent = CodeAgent(tools=tools, model=model)\n",
7777
"\n",
78-
" result = agent.run(query)\n",
79-
" return result"
78+
" result = agent.run(query)\n",
79+
" return result"
8080
]
8181
},
8282
{

0 commit comments

Comments
 (0)