Applications of Machine Learning in Fluid Mechanics: Comparison
Please note this is a comparison between Version 1 by Filippos Sofos and Version 2 by Sirius Huang.

The significant growth of artificial intelligence (AI) methods in machine learning (ML) and deep learning (DL) has opened opportunities for fluid dynamics and its applications in science, engineering and medicine. Developing AI methods for fluid dynamics encompass different challenges than applications with massive data, such as the Internet of Things.

  • artificial intelligence
  • machine learning
  • deep learning
  • neural networks
  • fluids

1. Introduction

In the past few years, machine learning (ML) and deep learning (DL) have been pursued in most social, scientific, engineering, and industry branches. Initially established at the point where computer and statistical science meet, ML evolution has been driven by advancements in Artificial Intelligence (AI) and its related algorithms, fortified by the “big data” and access to cost-effective computational architectures. ML-based approaches span fields such as construction engineering, aerospace, biomedicine, materials science, education, financial modelling, and marketing [1][2][1,2].
Machine learning techniques fall into three groups: unsupervised, semi-supervised, and supervised [2][3][2,3]. This text considers DL as a subset of the broader ML field. The supervised approaches rely on labelled data to train the ML algorithm effectively. This is the most common category, with wide applicability in science and technology [4][5][6][7][4,5,6,7]. Unsupervised ML involves extracting features from high-dimensional data sets without the need for pre-labelled training data. Well-established clustering methods, such as K-means, hierarchical agglomerative clustering, and DBSCAN [8][9][8,9], apply here. Furthermore, dimensionality reduction techniques such as proper orthogonal decomposition (POD) [10], principal component analysis (PCA) [11][12][11,12] and dynamic mode decomposition (DMD) [13] emerged from fundamental statistical analysis, and Reduced Order Model techniques have also been bound to unsupervised ML. Meanwhile, semi-supervised learning combines elements from supervised and unsupervised ML and has been successfully incorporated in applications concerning time-dependence and image data [14][15][14,15].
At the heart of ML theories, the algorithms proposed range from linear feed-forward models to complex DL implementations. To distinguish the complexity of various models, the algorithms may fall into shallow or deep ML, with possible combinations. Shallow learning (SL) algorithms, both in classification and regression tasks, involve linear and multiple linear algorithms, their linear variants such as LASSO and Ridge, support vector machines (SVM), stochastic methods based on Gaussian processes and Bayes’ theorem, and decision trees [16]. Ensemble and super-learner methods, combining multiple shallow algorithms in more complex instances, such as the random forest, gradient boosting, and extremely randomized trees, have also increased in popularity in the past years [17].
Notwithstanding its wide acceptance, the applicability of SL is limited to small-to-medium-sized data sets, which might be a problem in demanding fluid mechanics tasks. Thus, the need for more robust methods has emerged, and DL has come to the fore [18]. The idea of DL is based on the multi-layer perceptron, the artificial neural network (ANN) architecture that can be seen as the corresponding artificial mechanism that mimics the biological functions of the neural networks of the human brain [6].
DL models have embedded complex mathematical concepts and advanced computer programming techniques in a great number of interconnected nodes and layers [19], making them an appropriate choice for dealing with applications that involve image and video processing [20][21][22][23][20,21,22,23], speech recognition [24], biological data manipulation [25], and flow reconstruction [26]. Nevertheless, a physical interpretation of the outcome is still lacking, which limits their application in physical sciences, where interpretability has a central role. In such cases, it would be beneficial to adopt models that produce meaningful results, i.e., mathematical expressions, which can spot correlations with existing empirical relations and propose an analytical approach bound to physical laws.
Conventional neural networks operate solely on data and are often considered “black box” models. On the other hand, symbolic regression (SR) is a class of ML algorithms in which mathematical equations bound exclusively to data are derived. In contrast to usual regression procedures, SR searches within a wide mathematical operator set and constants to connect input features to produce meaningful outputs [27] based on evolutionary algorithms. This process of extracting results that can be interpreted and exploited simultaneously has led to the development of numerous genetic programming-based tools that implement SR for disciplines from materials science to construction engineering to medical science [28].
ML methods have been employed in fluid dynamics research but have not been in engineering practice. Computational fluid dynamics (CFD) discretizes space using meshes: as the grid density increases, so does the accuracy, but at the expense of computational cost. Such techniques have benefited the aerospace, automotive and biomedical industries. The challenges are numerical and physical, and several reviews have been written on specific industrial applications or areas of physics [29][30][31][32][29,30,31,32]. However, despite significant advances in the field, there is still a need to improve efficiency and accuracy.
The most critical example is turbulence, a flow regime characterised by velocity and pressure fluctuations, instabilities, and rotating fluid structures (eddies) at vast, macroscopic and molecular scales. Accurately describing transient flow physics using CFD would require many grid points. Such an approach where the Navier–Stokes equations are directly applied is called direct numerical simulation (DNS) and is accompanied by tremendous computational overhead. As such, it is limited to specific problems and requires large high-performance computing (HPC) facilities and long run times. Consequently, reduced turbulent models are often used at the expense of accuracy.
Furthermore, while many such techniques exist, no universal model works consistently across all turbulence flows, so while many computational methods exist for different problems, the computational barrier is often unavoidable. Recent efforts in numerical modelling frequently focus on how to speed up computations, a task that would benefit many industries. A growing school of thought is to couple these numerical methods with data-driven algorithms to improve accuracy and performance. This is where ML becomes relevant.

2. Applications of ML in Fluid Mechanics

As mentioned in the Introduction, a long-standing problem in fluid dynamics is turbulence, a flow regime characterised by unpredictable, time-dependent velocity and pressure fluctuations [33][34][35][36][152,153,154,155]. The use of data-driven methods has increased significantly over the last few years, providing new modelling tools for complex turbulent flows [37][156]. ML has been used to model turbulence, such as the use of ANNs to predict the instantaneous velocity vector field in turbulent wake flows [38][157] and turbulent channel flows [39][40][158,159]. The popularity of many ML methods, especially deep learning, as a promising method for solving complex fluid flow problems is also supported by high-quality open-source data sets published by researchers [41][42][43][160,161,162].
Much effort is currently put into using ML algorithms to improve the accuracy of Reynolds-averaged Navier–Stokes (RANS) models for which GPs have been used to quantify and reduce uncertainties [44][163]. Tracey et al. [45][164] used an ANN that inputs flow properties and reconstructs the SA closure equations. The researchers studied how different components of ML algorithms, such as the choice of the cost function and feature scaling, affected their performance. Other examples include using DL to model the Reynolds stress anisotropy tensor [46][47][165,166]. Bayesian inference has also been used to optimize RANS model coefficients [48][49][167,168], as well as to derive functions that quantify and reduce the gap between model predictions and high-fidelity data, such as DNS data [50][51][52][169,170,171]. Furthermore, a Bayesian neural network (BNN) has been used to improve the predictions of RANS models and to specify the uncertainty associated with the forecast [53][172].
In the field of large eddy simulations (LES), an ANN was used to replace these subgrid models [54][55][173,174]. Convolutional neural networks (CNNs) have also been used to calculate subgrid closure terms [56][175]. Moreover, DL modelling using input vorticity and stream function stencils was implemented to predict a dynamic, time- and space-dependent closure term for the subgrid models [57][176]. More specifically, starting from the detailed DNS simulation output, the main focus was to employ low-resolution LES data to train a NN model and reconstruct fluid dynamics to replace DNS where possible. To this end, many relevant applications have emerged. Bao et al. [58][177] have suggested combining physics-guided, two-component, PDE-based techniques with recurrent temporal data models and super-resolution spatial data methods. An extension of this approach includes a multi-scale temporal path UNet (MST-UNet) model that can reconstruct both temporal and spatial flow information from low-resolution data [58][177].
Fukami et al. [59][178] have applied super-resolution techniques, common in image reconstruction applications, on 2D course-flow fields and managed to reconstruct laminar and turbulent flows. A more recent version of this method employed autoencoders to achieve order reduction for fluid flows [60][179]. A complete super-resolution case study is presented in another work, covering all the details of a CNN-based, super-resolution architecture for turbulent fluid flows with DNS training data for various Reynolds numbers [61][180]. Nevertheless, generative models are often incorporated because DNS data are hard to obtain. In the review of Buzzicotti [62][53], three prevalent models are discussed: variational autoencoders (VAE), GANS, and diffusion models. These models are based on CNNs, and their implementation is focused on reproducing the multiscale and multifrequency nature of fluid dynamics, which is more complex than classical image recognition applications.
Another exciting application of ML in fluid mechanics is related to food processing [63][181]. The ML models aim to optimize process parameters and kinetics for reduced energy consumption, speed up the processing time, and continuously improve product quality. Several food processing operations, including drying, frying, baking, and extrusion, each require the development of ML algorithms and their training and validation against accurate data.
Furthermore, ML has found application in many chemical engineering flows. These applications include transport in porous media, catalytic reactors, batteries, and carbon dioxide storage [64][182]. Many of these problems are linked to energy transition efforts such as batteries, where long-term safety, energy density, and cycle life are particularly interesting. Diverse ML applications in fluid mechanics include nanofluids and ternary hybrid nanofluids [65][183] as well as metallurgy processes [66][184].
Video Production Service