@@ -187,7 +187,7 @@ def predict(self, feed_data):
187187
188188
189189class PaddlePredictServer (PredictServer ):
190- def __init__ (self , server , config_file , feeds , fetchs , max_failed_times = 3 ):
190+ def __init__ (self , server , config_file , feeds , fetchs , max_failed_times = 2 ):
191191 self ._server = server
192192 self ._config_file = config_file
193193 self ._predict_feed_idxs = []
@@ -295,14 +295,15 @@ def predict(self, feed_data):
295295 def __del__ (self ):
296296 try :
297297 # avoid serving exit bug when hasn't predict
298- if self .client is not None and self ._has_predict :
299- self .client .release ()
298+ #if self.client is not None and self._has_predict:
299+ # self.client.release()
300+ pass
300301 except Exception as e :
301302 logger .critical ('Release client failed with server={}, '
302303 'there may be an unknown error' .format (
303304 self ._server ))
304305 logger .critical ('Exception:\n {}' .format (str (e )))
305- logger .warning ('Stopped predict server={}' .format (self ._server ))
306+ # logger.warning('Stopped predict server={}'.format(self._server))
306307
307308
308309class _TestNopPaddlePredictServer (PaddlePredictServer ):
@@ -362,19 +363,10 @@ def predict_signal_handle(signum, frame):
362363 six .reraise (* sys .exc_info ())
363364
364365
365- def predict_loop (server_item ,
366- working_predict_count ,
367- in_queue ,
368- out_queue ,
369- feeds ,
370- fetchs ,
371- conf_file ,
372- stop_events ,
373- predict_lock ,
374- global_finished_task ,
375- predict_cond ,
376- thread_pool ,
377- max_concurrent = 3 ):
366+ def predict_loop (server_item , working_predict_count , in_queue , out_queue ,
367+ feeds , fetchs , conf_file , stop_events , predict_lock ,
368+ global_finished_task , predict_cond , thread_pool ,
369+ max_concurrent ):
378370 logger .info ('connect server={}' .format (server_item .server ))
379371 predict_server = PaddlePredictServer if _NOP_PREDICT_TEST is False else _TestNopPaddlePredictServer
380372 idx = 0
0 commit comments