Skip to content

Bug for magnetic system calculation compiled from ABACUS by icx-toolchain #2889

@QuantumMisaka

Description

@QuantumMisaka

Describe the bug

When running relax job using ABACUS by icx-mkl toolchain, the gradient calculation part seemed to have problem, leading to the calculated force up to a strange value 6.0e+3, and the calculated stress was also blast

# in ABACUS by icx-mkl toolchain
><><><><><><><><><><><><><><><><><><><><><><
TOTAL-STRESS (KBAR):
><><><><><><><><><><><><><><><><><><><><><><
3.689e+05      -6.754e-01     2.270e-01      
-6.754e-01     3.689e+05      2.184e-01      
2.270e-01      2.184e-01      3.689e+05      
TOTAL-PRESSURE: 3.689e+05 KBAR
ETOT DIFF (eV)       : 0.000e+00
LARGEST GRAD (eV/A)  : 6.483e+03
BFGS TRUST (Bohr)    : 5.000e-01
-------------------------------------------
STEP OF ION RELAXATION : 2
-------------------------------------------

After change toolchain to icc-mkl, the force calculation result was normal and relax can converged

# in ABACUS by `icc-mkl` toolchain
 ><><><><><><><><><><><><><><><><><><><><><><
 TOTAL-STRESS (KBAR):
 ><><><><><><><><><><><><><><><><><><><><><><
 -3.510e+02     -6.058e-01     5.005e-01      
 -6.058e-01     -3.467e+02     1.103e+00      
 5.005e-01      1.103e+00      -3.293e+02     
 TOTAL-PRESSURE: -3.423e+02 KBAR
 ETOT DIFF (eV)       : 0.000e+00
 LARGEST GRAD (eV/A)  : 1.894e+00
 BFGS TRUST (Bohr)    : 5.000e-01
 -------------------------------------------
 STEP OF ION RELAXATION : 2
 -------------------------------------------

toolchain method is the same as #2857

calculation example:

Expected behavior

  • icx-mkl toolchain can run relax job properly
  • Since oneapi will update icc to icx, ABACUS compilation by icx should be better than icc, more compilation optimization job should be done in ABACUS itself or in toolchain

To Reproduce

  1. Build abacus by icc-mkl or icx-mkl (better by toolchain method)
  2. Run the job in attached example

Environment

  • Compiler: OneAPI (icc or icx), Intel MKL, Intel MPI
  • OS. Rockylinux 8
  • Hardware: Intel Platinum 8358
  • Parallel Running method: MPI, OMP or MPI-OMP hybrid.

This Problem can also be observed in Bohrium c16-m32 machine by ABACUS compiled by icx toolchain

Additional Context

No response

Metadata

Metadata

Assignees

No one assigned

    Labels

    BugsBugs that only solvable with sufficient knowledge of DFT

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions