I have a question on the convergence of optimization schemes for minimizing smooth, possibly non-convex functions. Can the popular schemes such as Newton, quasi-Newton, and conjugate gradient converge to a non-stationary point even when Strong Wolfe conditions are satisfied at every time step? In other words, are Strong Wolfe conditions (sufficient reduction in function and gradient values) sufficient for ensuring convergence to a stationary point? Are there any simple counter-examples?