diff --git a/nifty/minimization/conjugate_gradient.py b/nifty/minimization/conjugate_gradient.py index 8d625605b791c7b0ef376674b6acaffc6724a787..56379c7c5592c6afb2ea160a0906e2e1e70be41d 100644 --- a/nifty/minimization/conjugate_gradient.py +++ b/nifty/minimization/conjugate_gradient.py @@ -158,7 +158,7 @@ class ConjugateGradient(Loggable, object): if self.reset_count is not None: reset += (iteration_number % self.reset_count == 0) if reset: - self.logger.info("Resetting conjugate directions.") + self.logger.info("Computing accurate residuum.") r = b - A(x) else: r -= q * alpha diff --git a/nifty/minimization/line_searching/line_search_strong_wolfe.py b/nifty/minimization/line_searching/line_search_strong_wolfe.py index 61af209c7eb00abd475b6469c23ce8564f86bbde..4864cceb42bcd459d93ba8c65d59eb3753496d12 100644 --- a/nifty/minimization/line_searching/line_search_strong_wolfe.py +++ b/nifty/minimization/line_searching/line_search_strong_wolfe.py @@ -139,7 +139,7 @@ class LineSearchStrongWolfe(LineSearch): phi_alpha1 = le_alpha1.value if (phi_alpha1 > phi_0 + self.c1*alpha1*phiprime_0) or \ - ((phi_alpha1 >= phi_alpha0) and (i > 0)): + ((phi_alpha1 >= phi_alpha0) and (iteration_number > 1)): le_star = self._zoom(alpha0, alpha1, phi_0, phiprime_0, phi_alpha0, phiprime_alpha0, phi_alpha1, le_0)