Conjugate gradient optimization: Difference between revisions

From VASP Wiki
No edit summary
No edit summary
 
Line 13: Line 13:
<references/>
<references/>
----
----
[[Category:Electronic Minimization]][[Category:Electronic Minimization Methods]][[Category:Theory]]
[[Category:Electronic minimization]][[Category:Theory]]

Latest revision as of 10:48, 6 April 2022

Instead of the previous iteration scheme, which is just some kind of Quasi-Newton scheme, it also possible to optimize the expectation value of the Hamiltonian using a successive number of conjugate gradient steps. The first step is equal to the steepest descent step in section Single band steepest descent scheme. In all following steps the preconditioned gradient is conjugated to the previous search direction. The resulting conjugate gradient algorithm is almost as efficient as the algorithm given in Efficient single band eigenvalue-minimization. For further reading see [1][2][3].

References