@@ -282,8 +282,8 @@ namespace polysolve::nonlinear
282282 stop_watch.start ();
283283
284284 m_logger.debug (
285- " Starting {} with {} solve f₀ ={:g} (stopping criteria: {})" ,
286- descent_strategy_name (), m_line_search->name (), objFunc (x), m_stop.print_message ());
285+ " Starting {} with {} solve {} ={:g} (stopping criteria: {})" ,
286+ descent_strategy_name (), m_line_search->name (), log::f0 (), objFunc (x), m_stop.print_message ());
287287
288288 update_solver_info (objFunc (x));
289289 objFunc.post_step (PostStepData (m_current.iterations , solver_info, x, grad));
@@ -377,17 +377,21 @@ namespace polysolve::nonlinear
377377 {
378378 m_status = Status::NotDescentDirection;
379379 log_and_throw_error (
380- m_logger, " [{}][{}] {} on last strategy (‖Δx‖={:g}; ‖g‖={:g}; Δx⋅g={:g}≥0); stopping" ,
381- current_name, m_line_search->name (), status_message (m_status), delta_x.norm (), compute_grad_norm (x, grad),
382- m_current.xDeltaDotGrad );
380+ m_logger, " [{}][{}] {} on last strategy ({}={:g}; {}={:g}; {}={:g}≥0); stopping" ,
381+ current_name, m_line_search->name (), status_message (m_status),
382+ log::norm (log::delta (" x" )), delta_x.norm (),
383+ log::norm (" g" ), compute_grad_norm (x, grad),
384+ log::delta (" x" ) + log::dot () + " g" , m_current.xDeltaDotGrad );
383385 }
384386 else
385387 {
386388 m_status = Status::Continue;
387389 m_logger.debug (
388- " [{}][{}] {} (‖Δx‖ ={:g}; ‖g‖ ={:g}; Δx⋅g ={:g}≥ 0); reverting to {}" ,
390+ " [{}][{}] {} ({} ={:g}; {} ={:g}; {} ={:g}{} 0); reverting to {}" ,
389391 current_name, m_line_search->name (), status_message (Status::NotDescentDirection),
390- delta_x.norm (), compute_grad_norm (x, grad), m_current.xDeltaDotGrad ,
392+ log::norm (log::delta (" x" )), delta_x.norm (),
393+ log::norm (" g" ), compute_grad_norm (x, grad),
394+ log::delta (" x" ) + log::dot () + " g" , m_current.xDeltaDotGrad , log::ge (),
391395 descent_strategy_name ());
392396 }
393397 continue ;
@@ -403,9 +407,9 @@ namespace polysolve::nonlinear
403407 // --- Variable update ---------------------------------------------
404408
405409 m_logger.trace (
406- " [{}][{}] pre LS iter={:d} f={:g} ‖∇f‖ ={:g}" ,
410+ " [{}][{}] pre LS iter={:d} f={:g} {} ={:g}" ,
407411 descent_strategy_name (), m_line_search->name (),
408- m_current.iterations , energy, m_current.gradNorm );
412+ m_current.iterations , energy, log::norm ( log::grad ( " f " )), m_current.gradNorm );
409413
410414 // Perform a line_search to compute step scale
411415 double rate;
0 commit comments