diff --git a/nifty/minimization/conjugate_gradient.py b/nifty/minimization/conjugate_gradient.py
index 8d625605b791c7b0ef376674b6acaffc6724a787..56379c7c5592c6afb2ea160a0906e2e1e70be41d 100644
--- a/nifty/minimization/conjugate_gradient.py
+++ b/nifty/minimization/conjugate_gradient.py
@@ -158,7 +158,7 @@ class ConjugateGradient(Loggable, object):
             if self.reset_count is not None:
                 reset += (iteration_number % self.reset_count == 0)
             if reset:
-                self.logger.info("Resetting conjugate directions.")
+                self.logger.info("Computing accurate residuum.")
                 r = b - A(x)
             else:
                 r -= q * alpha
diff --git a/nifty/minimization/line_searching/line_search_strong_wolfe.py b/nifty/minimization/line_searching/line_search_strong_wolfe.py
index 61af209c7eb00abd475b6469c23ce8564f86bbde..4864cceb42bcd459d93ba8c65d59eb3753496d12 100644
--- a/nifty/minimization/line_searching/line_search_strong_wolfe.py
+++ b/nifty/minimization/line_searching/line_search_strong_wolfe.py
@@ -139,7 +139,7 @@ class LineSearchStrongWolfe(LineSearch):
             phi_alpha1 = le_alpha1.value
 
             if (phi_alpha1 > phi_0 + self.c1*alpha1*phiprime_0) or \
-               ((phi_alpha1 >= phi_alpha0) and (i > 0)):
+               ((phi_alpha1 >= phi_alpha0) and (iteration_number > 1)):
                 le_star = self._zoom(alpha0, alpha1, phi_0, phiprime_0,
                                      phi_alpha0, phiprime_alpha0, phi_alpha1,
                                      le_0)