@@ -75,7 +75,7 @@ def query_endpoint_stream(endpoint_name: str, messages: list[dict[str, str]], re
7575
7676def _query_chat_endpoint_stream (endpoint_name : str , messages : list [dict [str , str ]], return_traces : bool ):
7777 """Invoke an endpoint that implements either chat completions or ChatAgent and stream the response"""
78- client = get_deploy_client ("databricks://dogfood " )
78+ client = get_deploy_client ("databricks" )
7979
8080 # Prepare input payload
8181 inputs = {
@@ -94,7 +94,7 @@ def _query_chat_endpoint_stream(endpoint_name: str, messages: list[dict[str, str
9494
9595def _query_responses_endpoint_stream (endpoint_name : str , messages : list [dict [str , str ]], return_traces : bool ):
9696 """Stream responses from agent/v1/responses endpoints using MLflow deployments client."""
97- client = get_deploy_client ("databricks://dogfood " )
97+ client = get_deploy_client ("databricks" )
9898
9999 input_messages = _convert_to_responses_format (messages )
100100
@@ -129,7 +129,7 @@ def _query_chat_endpoint(endpoint_name, messages, return_traces):
129129 if return_traces :
130130 inputs ['databricks_options' ] = {'return_trace' : True }
131131
132- res = get_deploy_client ('databricks://dogfood ' ).predict (
132+ res = get_deploy_client ('databricks' ).predict (
133133 endpoint = endpoint_name ,
134134 inputs = inputs ,
135135 )
@@ -142,7 +142,7 @@ def _query_chat_endpoint(endpoint_name, messages, return_traces):
142142
143143def _query_responses_endpoint (endpoint_name , messages , return_traces ):
144144 """Query agent/v1/responses endpoints using MLflow deployments client."""
145- client = get_deploy_client ("databricks://dogfood " )
145+ client = get_deploy_client ("databricks" )
146146
147147 input_messages = _convert_to_responses_format (messages )
148148
0 commit comments