The rapid advancement of modern computing has driven extensive research on numerical algorithms for solving large-scale systems of linear equations. Classical methods such as LU decomposition, Jacobi, and Gauss–Seidel have been revisited and optimized to leverage parallel architectures, GPUs, and even quantum platforms. Recent studies demonstrate that optimized algorithms can reduce computation time by more than 50% while maintaining high accuracy in solving high-dimensional problems. LU decomposition, particularly in its parallel and GPU-based implementations, has shown superior performance in batch processing and industrial-scale simulations. Meanwhile, iterative methods such as Jacobi and Gauss–Seidel remain relevant due to their flexibility in numerical modeling, with further developments for block matrix systems, finite element applications, and FPGA architectures. The integration of these enhanced algorithms is not only beneficial for the advancement of scientific software development but also supports practical applications in engineering simulations, large-scale data optimization, and machine learning. Therefore, an integrative review of modern numerical algorithm developments is crucial in bridging the gap between industrial demands and research progress in scientific computing.
Copyrights © 2024