Probing Rate-Dependent Liquid Shear Viscosity Using Combined Machine Learning and Nonequilibrium Molecular Dynamics.
Accurately measuring liquid dynamic viscosity across a wide range of shear rates, from the linear-response to shear-thinning regimes, presents significant experimental challenges due to limitations in resolving high shear rates and controlling thermal effects. In this study, we integrated machine learning (ML) with nonequilibrium molecular dynamics (NEMD) simulations to address these challenges. A supervised artificial neural network (ANN) model was developed to predict viscosity as a function of shear rate, normal pressure, and temperature, effectively capturing the complex interplay among these variables. The model reveals distinct trends in shear viscosity, characterized by the shear-thinning exponent, and highlights nonmonotonic behavior in the radius of gyration components, reflecting molecular morphological changes driven by rate-dependent volume expansion. Notably, temperature effects diminish at higher shear rates, where molecular alignment and spacing dominate the response to shear. By implementing the 'fix npt/sllod' command in LAMMPS, we achieve precise constant-pressure control in NEMD simulations, ensuring accurate representation of system dynamics. This study demonstrates the potential of ML-enhanced NEMD for efficient and accurate viscosity prediction, providing a robust framework for future research in complex fluid dynamics and material design.