In optimization, a descent direction is a vector                               p                ∈                              R                                n                                   that, in the sense below, moves us closer towards a local minimum                                           x                                ∗                                   of our objective function                     f        :                              R                                n                          →                  R                        .
Suppose we are computing                                           x                                ∗                                   by an iterative method, such as line search. We define a descent direction                                           p                                k                          ∈                              R                                n                                   at the                     k                th iterate to be any                                           p                                k                                   such that                     ⟨                              p                                k                          ,        ∇        f        (                              x                                k                          )        ⟩        <        0                , where                     ⟨        ,        ⟩                 denotes the inner product. The motivation for such an approach is that small steps along                                           p                                k                                   guarantee that                               f                         is reduced, by Taylor's theorem.
Using this definition, the negative of a non-zero gradient is always a descent direction, as                     ⟨        −        ∇        f        (                              x                                k                          )        ,        ∇        f        (                              x                                k                          )        ⟩        =        −        ⟨        ∇        f        (                              x                                k                          )        ,        ∇        f        (                              x                                k                          )        ⟩        <        0                .
Numerous methods exist to compute descent directions, all with differing merits. For example, one could use gradient descent or the conjugate gradient method.
More generally, if                     P                 is a positive definite matrix, then                     d        =        −        P        ∇        f        (        x        )                 is a descent direction  at                     x                . This generality is used in preconditioned gradient descent methods.