Development of novel computational models based on artificial intelligence technique to predict liquids mixtures separation via vacuum membrane distillation | Scientific Reports
Scientific Reports volume 14, Article number: 24121 (2024) Cite this article
Metrics details
The fundamental objective of this paper is to use Machine Learning (ML) methods for building models on temperature (T) prediction using input features r and z for a membrane separation process. A hybrid model was developed based on computational fluid dynamics (CFD) to simulate the separation process and integrate the results into machine learning models. The CFD simulations were performed to estimate temperature distribution in a vacuum membrane distillation (VMD) process for separation of liquid mixtures. The evaluated ML models include Support Vector Machine (SVM), Elastic Net Regression (ENR), Extremely Randomized Trees (ERT), and Bayesian Ridge Regression (BRR). Performance was improved using Differential Evolution (DE) for hyper-parameter tuning, and model validation was performed using Monte Carlo Cross-Validation. The results clearly indicated the models’ effectiveness in temperature prediction, with SVM outperforming other models in terms of accuracy. The SVM model had a mean R2 value of 0.9969 and a standard deviation of 0.0001, indicating a strong and consistent fit to the membrane data. Furthermore, it exhibited the lowest mean squared error, mean absolute error, and mean absolute percentage error, signifying superior predictive accuracy and reliability. These outcomes highlight the importance of selecting a suitable model and optimizing hyperparameters to guarantee accurate predictions in ML tasks. It demonstrates that using SVM, optimized with DE improves accuracy and consistency for this specific predictive task in membrane separation context.
Separation of liquid solutions can be conducted via membrane systems where different methods have been developed so far such as ultrafiltration, nanofiltration, reverse osmosis, etc. The size of membrane pores and the separation mechanism are different for each membrane separation technique. An important class of membrane separation technique is membrane contactor where these units are used for contacting two phases such as liquid-liquid contact to separation one compound or conduct chemical reaction1,2,3. Membrane contactors can offer various operations due to their unique characteristics in process engineering. Ind fact, various unit operations can be done via membrane contactors such as chemical reactions in membrane contactors where it can be driven in continuous mode of operation4,5.
Combination of membrane contactors with other unit operations can be considered as a novel approach for developing hybrid processing. Such hybrid processes have been developed already such as membrane distillation (MD) where distillation is conducted in continuous flow via membrane contactors6,7,8. For direct-contact membrane distillation, the hot and cold fluids are brought into contact in the membrane contactor and separation occurs in the process.
Other type of MD is vacuum membrane distillation (VMD) which relies on creating pressure difference between the feed and permeate side for separation of compounds based on volatility9. VMD has privilege that it does not need additional cold phase to make driving force between two phases10,11. It can also operate in the ambient condition for separation. The process can be developed and optimized by computational techniques such as computational fluid dynamics (CFD) which is used for numerical solution of transport phenomena in membrane process12,13. Temperature distribution and concentration of solute in the feed can be obtained by this method. Due to the complexity of CFD, other approaches such as machine learning models can be employed for simulation of this process.
MD has been previously simulated using ANFIS (Adaptive Neuro-Fuzzy Inference System) method by using some CFD simulations dataset and the results were promising in terms of accuracy14. Also, concentration distribution of solute in VMD process has been obtained via combination of CFD and ML models15. However, wider analysis is required to explore more machine learning models for analysis of VMD process. Indeed, there is a research gap on development of combined hybrid CFD and machine learning models for VMD process which provides holistic view of the process. This is addressed in this study by developing hybrid CFD-machine learning models for VMD process to predict temperature distribution in the process.
This study explores the effectiveness of different machine learning models in predicting temperature (T) using the input features r and z which are the membrane system’s coordinates. The selected models include Bayesian Ridge Regression (BRR), Extremely Randomized Trees (ERT), Elastic Net Regression (ENR), and Support Vector Machine (SVM). The selected models encompass a blend of linear and non-linear methodologies, enabling a thorough assessment of their predictive performance.
BRR amalgamates Bayesian principles with ridge regression, estimating regression coefficients by integrating prior information and introducing a regularization term to mitigate overfitting. This strategy improves the model’s generalization by harmonizing the model’s fit and complexity via regularization. ERT is an ensemble learning method that constructs multiple unpruned decision trees using random cut-points and the entire training dataset. This technique reduces variance and computational complexity by averaging the predictions from these randomized trees, improving overall prediction accuracy. ENR combines L1 (Lasso) and L2 (Ridge) regularization to handle multicollinearity and promote sparsity. The SVM is a resilient model that effectively identifies the optimal hyperplane in a higher-dimensional space to separate data points efficiently. The primary goal is to maximize the margin between the support vectors and the hyperplane, resulting in minimal training error and enhanced generalization. This methodology is beneficial for both classification and regression assignments. To ensure robust predictions, the study utilized hyper-parameter optimization through Differential Evolution (DE) and model validation via Monte Carlo Cross-Validation (MCCV). DE is an evolutionary optimization technique adept at handling multidimensional data, while MCCV involves multiple random splits of the dataset into training and testing subsets to validate model performance comprehensively. Performance metrics such as R2 score, Mean Squared Error (MSE), Mean Absolute Error (MAE), and Mean Absolute Percentage Error (MAPE) were calculated to assess each model.
The dataset employed in this investigation are more than 13,000 data instances, each characterized by three variables: r, z, and T (temperature of fluid). The initial two coordinates serve as inputs, while T represents the sole output of this analysis. r and z represent the radial and axial coordinates in the membrane system, respectively where the temperature of solution is calculated via CFD simulations.
For data collection, a hollow-fiber membrane contactor for VMD was simulated by utilizing COMSOL package and numerical solution of mass and energy equations. UMFPACK was used as the direct solver in numerical calculations15. The meshed geometry of the model for the feed and membrane sides is illustrated in Fig. 1. As seen, the system is divided into two parts, i.e., feed and membrane. The equations of convection-diffusion for mas transfer and convection-conduction for the heat transfer were solved for the geometry using finite element method. A constant concentration and temperature boundary condition was assumed for the inlet of feed channel, whereas convective flow is assumed for the outlet of feed channel. Separation of an organic compound from aqueous solution was considered in this study. The flow regime was assumed to be laminar due to the low flow rate and small channel dimension in the membrane system14. Once the simulations have been carried out, the temperature distribution entire the feed side was collected for ML modeling. So, the temperature as a function of coordinates was obtained for machine learning modeling in this study. The used equations in this work for CFD simulations are listed in Table 1. It should be noted that the CFD methodology used in this study has been already validated through comparing with experimental data as reported in16,17.
Meshed geometry of VMD process for CFD simulations.
The pair plot depicted in Fig. 2 offers an extensive perspective on the interconnections among the variables. The histograms on the diagonal represent the distribution of each variable, while the scatter plots off the diagonal illustrate the relationships between each pair of variables. The scatter plots indicate a significant negative correlation between z and T, implying that as z increases, T generally decreases. The association between r and T appears to be less pronounced, indicating a relatively weaker relationship.
The box plots in Fig. 3 illustrate the distribution and variability of each variable. For r and z, the data points are tightly clustered around the median, indicating low variability. T, the output variable, also shows a symmetric distribution with a median around 318 K and a slight spread, indicating moderate variability.
Pair plot of the dataset obtained from VMD process.
Box plot of variables extracted from VMD simulation.
The methodology for predicting T based on the input variables r and z involves several key steps: dataset preparation, outlier detection, model selection, hyper-parameter optimization, model validation, and performance evaluation14. As the dataset is complete with no missing values, no preprocessing for missing data was necessary. However, to ensure data quality, outliers were identified and handled using the Elliptic Envelope method (here less than 0.5% of data detected as outliers). The data also normalized using Min-Max scaler. Four ML models were chosen: BRR, ERT, ENR, and SVM. These models were selected to provide a mix of linear and non-linear approaches, allowing a comprehensive evaluation of their predictive capabilities. Hyper-parameters for each model were optimized using DE. Model validation was performed using MCCV. MCCV involves randomly splitting the dataset into training and testing subsets multiple times, training the model on the training subset, and evaluating it on the testing subset. This process was repeated to ensure robust validation of model performance, with mean and standard deviation of the performance metrics being reported. This method guarantees that a single train-test split does not bias the validation results and offers a more complete evaluation of the generalizability of the model18. Here the number of repeats is set to 100 and the train and test subsets are 80% and 20% of the entire dataset. Using R2 score, MSE, MAE, and MAPE, each model’s performance was assessed. Each metric was computed for the models across multiple MCCV iterations to obtain the mean and standard deviation, providing a comprehensive view of the model performance. This systematic approach ensured a thorough evaluation of the selected models, identifying the most accurate and reliable model for predicting temperature based on the input variables r and z. More detailed descriptions of the building blocks that are mentioned here are included in the following subsections. Step-by-step implementation of the proposed approach is shown in Fig. 4. Machine learning algorithms and optimizers implemented using Python programming language.
General methodology designed for this work.
The Elliptic Envelope method is a statistical technique for detecting outliers by assuming that the data follows a multivariate Gaussian distribution. This method constructs an elliptical boundary that encompasses the majority of the data points, effectively capturing the central tendency and dispersion of the dataset. By fitting an ellipse to the data, the method identifies points that lie outside this boundary as outliers19. The process begins by estimating the mean and covariance of the dataset. The Elliptic Envelope then uses these estimates to create an ellipse that encompasses the central data points, defined by a chosen confidence level. Mathematically, it aims to minimize the volume of the ellipse while covering a specified proportion of the data, typically set to 95% or 99%. Data points lying outside this ellipse are flagged as outliers.
DE is recognized as a meta-heuristic evolutionary optimization approach aimed at iteratively enhancing the quality of potential solutions to optimize a given problem. This optimizer can effectively handle multidimensional real-valued data, even when the function to be optimized lacks differentiability. Moreover, it can address issues that are noisy, discontinuous, or dynamic. DE operates by utilizing a population of feasible solutions, combining them through simple mathematical operations to identify the solution that is most optimal for the optimization problem. The algorithm encodes the variables of the problem as a vector of real numbers. The population consists of vectors of length n, corresponding to the parameters in the optimization problem20,21. A vector is represented by \({x}_{g,p}\), where p indicates its index within the population, and g denotes its generation. The components of this vector, \({x}_{g,p,m}\) are bounded within intervals defined by \({x}_{\text{min},m}\) and \({x}_{\text{max},m}\). The DE algorithm involves four stages including initialization, mutation, recombination, and selection22. The algorithm iterates through these stages until a stopping criterion is met, which can be based on the number of generations, time elapsed, or the degree of optimization achieved23,24.
In this study, DE was utilized to optimize the hyperparameters of the machine learning models (SVM, ENR, ERT, BRR). The fitness function which should be maximized is set to R2 score. The algorithm parameters were set as follows:
Population size: 50.
Scaling factor (F): 0.8.
Crossover rate (CR): 0.9.
Number of generations: 100.
The DE algorithm iteratively adjusted the hyperparameters to minimize the objective function, which in this case was the sum of the Root Mean Squared Error (RMSE) and MAE values, considering their standard deviations.
Bayesian statistics and ridge regression form a robust regression data analysis model. This method models the relationship between independent input variables and a continuous dependent response in a flexible and robust manner25. It calculates the coefficients of a linear model through considering prior knowledge about the data using a prior distribution. The posterior distribution is obtained by combining this distribution with the likelihood function. It is used to inform the estimation of coefficients and predictions of data. The technique incorporates a penalty term to reduce overfitting, thereby enhancing generalization. The regression coefficients conform to a normal distribution with a mean of zero and a variance that is determined by the hyper-parameter alpha. The likelihood function is modeled as a normal distribution, with a linear regression determining the mean and an additional hyperparameter, lambda, controlling the variance. The main objective is to infer the most likely values for the regression coefficients beta based on the given data and prior information. Following equation is the posterior distribution of β26:
In this equation, µ represents the mean vector and Σ indicates the covariance matrix of the posterior distribution. The Bayesian formula is utilized to determine these parameters27:
where \(X_{y}^{\prime }\) stands for the transposed input matrix multiplied by the output variable, \(X^{\prime } X\) indicates the transpose of the input variable matrix multiplied by itself, and \(I\) denotes the identity matrix.
ERT is an ensemble learning method introduced by Geurts et al.28 in 2006. Tree-based regressors consist of hierarchical rule sets that predict numerical output values. By averaging the randomized predictions from multiple decision trees, Extra-Trees enhance prediction accuracy while significantly reducing computational complexity29. The extra-trees method relies on the concept of the bias-variance trade-off. The utilization of explicit randomization of cut-points and characteristics, along with ensemble averaging, efficiently diminishes variation to a greater extent compared to less assertive randomization tactics observed in alternative algorithms. To minimize bias, the entire original training set is used instead of bootstrap samples. The computational complexity of growing the trees, assuming balanced trees, is on the order of \({\log}N\) relative to the size of the training sample, similar to other tree-growing algorithms. Moreover, owing to the straightforward nature of the node splitting procedure, the constant factor is anticipated to be significantly lower compared to earlier ensemble methods that optimize cut-points locally.
Elastic Net is a potent technique that synergistically integrates the advantages of L1 (Lasso) and L2 (Ridge) regularization techniques. L1 regularization encourages sparsity by shrinking some coefficients to exactly zero, thereby performing variable selection and enhancing interpretability. Concurrently, L2 regularization mitigates multicollinearity issues and stabilizes the coefficient estimates by shrinking them towards zero without eliminating any variables completely30,31. The Elastic Net model introduces two key hyperparameters: \({\upalpha}\) and \({\uplambda}\). The \({\upalpha}\) parameter controls the mix between L1 and L2 regularization. When \({\upalpha}=1\), the model reduces to Lasso regression, and when \({\upalpha}=0\), it becomes Ridge regression. The \({\uplambda}\) parameter determines the overall strength of the regularization applied to the model, with higher values leading to more regularization. The objective function for Elastic Net combines the penalties from both Lasso and Ridge32:
In this equation, n denotes the observations, \(\:{y}_{i}\) stands for the observed output, and \({X}_{i}\) represents the vector of predictors for the i-th observation. The vector \({\upbeta}\) represents the coefficients to be determined from dataset. The term \({\upalpha}\) balances L1 and L2 penalties, while \({\uplambda}\) denotes the regularization parameter. Finding the optimum values for \({\upalpha}\) and \({\uplambda}\) can be achieved through grid search with cross-validation. This method systematically evaluates different combinations of \({\upalpha}\) and \({\uplambda}\) to identify the best-performing model configuration. This approach ensures that the model is well-tuned and performs optimally on the validation set, thereby enhancing its generalizability to new data. Combining the flexibility of Lasso and the strength of Ridge gives Elastic Net Regularization a well-balanced method for linear regression. For datasets with high-dimensional characteristics or multicollinearity especially it is helpful.
SVM is a popular ML model, particularly adept at fitting data by identifying a line or hyperplane in a higher-dimensional space. The hyperplane is defined by a set of data points known as support vectors, which are the most closely correlated with it. The goal of SVM regression is to find the hyperplane that maximizes the margin between the support vectors and the hyperplane while reducing the training error33,34,35. In SVM regression, we aim to find a function \(f\left( x \right) = w \cdot x + b\) that deviates from the actual target values \({y}_{i}\) by no more than \(\in\), ensuring the function is as flat as possible. The objective is to minimize36:
Considering:
To accommodate errors, slack variables \({{\upxi}}_{i}\) and \({{\upxi}}_{i}^{*}\) are applied in the model, leading to the modified objective36:
According to:
Here, C is a regularization parameter balancing flatness and tolerance for deviations. In SVM regression, the mapping between the input and output spaces is determined by the kernel function. Common kernel functions comprise the radial basis function (RBF), linear, polynomial, and sigmoid. Choosing the kernel function is dependent on the unique problem, the quantity of features, and the noise level within the data37.
The performance of four distinct ML models is evaluated for predicting T based on the input features r and z in the domain of membrane contactor. The models tested were BRR, ERT, ENR, and SVM. DE was used for hyper-parameter optimization, and MCCV was employed for model validation. The evaluation metrics taken into account included the mean R2 score, MSE, MAE, and MAPE, along with their corresponding standard deviations. Table 2 offers a succinct overview of each model’s performance across various metrics.
The results indicate that the SVM performed the best among the other models for fitting the temperature dataset in the MD process. Specifically, the SVM model achieved the highest mean R2 score (0.9969) with the lowest standard deviation (0.0001), indicating a very strong and consistent fit to the data. Additionally, it had the lowest mean MSE (0.03313), MAE (0.15803), and MAPE (4.9774%), demonstrating its superior predictive accuracy and robustness.
The ENR model also performed very well, with a mean R2 score of 0.9912 and significantly lower errors compared to the BRR and ERT models. The ENR model’s mean MSE (0.09349), MAE (0.24718), and MAPE (7.7591%) were higher than those of the SVM but still substantially better than BRR and ET.
The ERT model showed moderate performance, with a mean R2 score of 0.9526 and relatively lower errors compared to BRR but higher than both ENR and SVM. The BRR model exhibited the lowest performance among the models tested, with the highest errors and lowest R2 score.
Figure 5 displays a comparison between the actual values obtained through experimentation (validated CFD simulations) and the values predicted by utilizing all four ML models. Furthermore, Fig. 6 illustrates the distribution of R2 scores obtained through MCCV method.
CFD (Experimental) versus estimated values using BRR, SVM, ERT, and ENR models.
Distribution of R2 scores from Monte Carlo cross-validation (BRR, SVM, ERT, and ENR models).
Overall, the SVM model has superior performance in predicting temperature using the input features r and z. It regularly obtained the greatest R2 scores and the lowest error metrics, demonstrating exceptional predictive accuracy and reliability. The ENR model also demonstrated strong performance and may be regarded as a viable alternative if SVM is deemed less suitable for specific applications due to factors such as interpretability or other reasons. The ERT and BRR models, albeit less precise, can nonetheless be useful in situations when their particular strengths or computational benefits are applicable.
The contour plot of T is given in Fig. 7, using SVM as the most suitable model. Additionally, Fig. 8 presents a 3D depiction of T(K) based on coordinates. Finally, the partial dependencies between the inputs and the temperature T(K) are illustrated in Fig. 9. It is seen that the temperature is decreased along the membrane process which is due to the evaporation of the feed solution which in turn decreases the system’s temperature. In radial direction, a parabolic change for temperature can be observed which could be attributed to the convective term which is dominant in the membrane process in axial direction. This term is mainly due to the fluid flow in axial direction and effect of viscous forces in the solution which causes a parabolic velocity profile in the feed channel of membrane contactor14.
Two-dimensional distribution of T (SVM model).
While the DE algorithm effectively optimized the machine learning models’ hyperparameters, there are potential drawbacks. Overfitting is a concern where models excel on training data but perform poorly on new data. This risk is mitigated by MCCV in this study. DE’s computational costs can be high, especially with large populations and many generations, leading to longer processing times. Efficient computing resources and parallel processing helped to address this. Future study can be performed on development of hybrid optimization techniques to minimize the deviation of ML estimated parameters from the reference data, while avoiding biased models.
3D visualization of temperature as a function of coordinates.
T profiles in radial and axial directions.
This study evaluated the performance of four machine learning models—BRR, ERT, ENR, and SVM—for predicting temperature based on the input variables. The models are trained via dataset obtained from CFD simulations of VMD process for separating liquid mixtures. Using hyper-parameter optimization through DE and model validation via MCCV, each model’s predictive capabilities was assessed. The findings indicated that the SVM model exhibited superior performance compared to the other models in all performance metrics. It attained the highest average R2 score of 0.9969 and the lowest average values for MSE, MAE, and MAPE. The findings emphasize the superior predictive accuracy and robustness of SVM, establishing it as the most dependable model for this particular predictive task. The ENR model exhibited robust performance, displaying noteworthy accuracy and lower error rates in comparison to BRR and ERT. The ERT model demonstrated a moderate level of performance, whereas the BRR model displayed the lowest level of performance among the models that were tested. In summary, this study emphasizes the significance of choosing suitable machine learning models and fine-tuning their hyper-parameters to achieve precise predictions. Subsequent investigations could focus on refining these models and testing their effectiveness on different datasets to confirm their ability to be applied in a wider range of predictive tasks.
The data supporting this study are available when reasonably requested from the corresponding author.
Adaptive neuro-fuzzy inference system
Bayesian ridge regression
Computational fluid dynamics
Crossover rate
Differential evolution
Elastic Net regression
Extremely randomized trees
Membrane distillation
Membrane
Machine learning
Monte Carlo cross-validation
Mean squared error
Mean absolute error
Mean absolute percentage error
Root mean squared error
Support vector machine
Vacuum membrane distillation
Species A
Regularization parameter in SVM
Concentration of species (mol/m3)
Specific heat capacity (J/kg K)
Diffusion coefficient of compound A (m2/s)
Scaling factor
Thermal conductivity (W/m K)
Mass transfer flux of compound A (mol/m2 s)
Radial coordinate (m)
Inner radius of membrane (m)
Coefficient of determination
Temperature (K)
Tube side of membrane
Average fluid velocity in tube side (m/s)
Velocity of fluid (m/s)
Weight parameter
Observed response
Axial coordinate (m)
Regression coefficients
Hyper-parameter controlling the mix between L1 and L2 regularization
Regularization parameter
Mean vector of the posterior distribution
Covariance matrix of the posterior distribution
Slack variable in SVM
Insensitivity parameter in SVM
Lv, Y. T. et al. In situ membrane separation drives nitrous oxide enrichment from nitrate denitrification for energy recovery. J. Water Process. Eng. 59, 105064 (2024).
Article Google Scholar
Ren, K. et al. Fabrication of hollow fiber composite membranes via opposite transmission reaction method for dye/salt separation. J. Hazard. Mater. 475, 134856 (2024).
Article CAS PubMed Google Scholar
Xu, L. et al. Heterogeneous wettability membrane for efficient demulsification and separation of oil-in-water emulsions. Chem. Eng. J. 489, 151466 (2024).
Article CAS Google Scholar
Richard, S. et al. Power-to-ammonia synthesis process with membrane reactors: Techno- economic study. Int. J. Hydrog. Energy 73, 462–474 (2024).
Article CAS Google Scholar
Shigwan, P. & Padhiyar, N. A comparative study of membrane reactor and cylindrical reactor using multi-objective optimization for methane reforming process. Int. J. Hydrog. Energy 77, 639–651 (2024).
Article CAS Google Scholar
Guo, Q. et al. Enhancement and optimization of membrane distillation processes: A systematic review of influential mechanisms, optimization and applications. Desalination 586, 117862 (2024).
Article CAS Google Scholar
Zuo, L. et al. Characterization and prediction modeling of membrane distillation enhanced disc solar still. J. Clean. Prod. 449, 141742 (2024).
Article Google Scholar
Ismael, B. H. et al. Permeation flux prediction of vacuum membrane distillation using hybrid machine learning techniques. Membranes 13(12), 900 (2023).
Article CAS PubMed PubMed Central Google Scholar
Aytaç, E. & Khayet, M. A deep dive into membrane distillation literature with data analysis, bibliometric methods, and machine learning. Desalination 553, 116482 (2023).
Article Google Scholar
Subrahmanya, T. M. et al. Self-surface heating membrane distillation for sustainable production of freshwater: A state of the art overview. Prog. Mater. Sci. 145, 101309 (2024).
Article CAS Google Scholar
Zhu, Y. et al. Review of ammonia recovery and removal from wastewater using hydrophobic membrane distillation and membrane contactor. Sep. Purif. Technol. 328, 125094 (2024).
Article CAS Google Scholar
Baghel, R. et al. CFD modeling of vacuum membrane distillation for removal of Naphthol blue black dye from aqueous solution using COMSOL multiphysics. Chem. Eng. Res. Des. 158, 77–88 (2020).
Article CAS Google Scholar
FarisAbadi, A., Kazemeini, M. & Ekramipooya, A. Investigating a HEX membrane reactor for CO2methanation using a Ni/Al2O3catalyst: A CFD study. Int. J. Hydrog. Energy 48(64), 25075–25091 (2023).
Article ADS CAS Google Scholar
Ye, B. & Zhou, W. Efficiency increment of CFD modeling by using ANFIS artificial intelligence for thermal-based separation modeling. Case Stud. Therm. Eng. 60, 104820 (2024).
Article Google Scholar
Obaidullah, A. J. & Almehizia, A. A. Modeling and validation of purification of pharmaceutical compounds via hybrid processing of vacuum membrane distillation. Sci. Rep. 14(1), 20734 (2024).
Article CAS PubMed PubMed Central Google Scholar
Wu, B. et al. Removal of 1,1,1-trichloroethane from water using a polyvinylidene fluoride hollow fiber membrane module: Vacuum membrane distillation operation. Sep. Purif. Technol. 52(2), 301–309 (2006).
Article CAS Google Scholar
Tahvildari, K. et al. Numerical simulation studies on heat and mass transfer using vacuum membrane distillation. Polym. Eng. Sci. 54(11), 2553–2559 (2014).
Article CAS Google Scholar
Zhang, F. Cross-validation and Regression Analysis in High-Dimensional Sparse Linear Models (Stanford University, 2011).
Usman, N., Utami, E. & Hartanto, A. D. Comparative Analysis of elliptic envelope, isolation forest, one-class SVM, and local outlier factor in detecting earthquakes with status anomaly using outlier. In International Conference on Computer Science, Information Technology and Engineering (ICCoSITE) (IEEE, 2023).
Storn, R. & Price, K. Differential evolution–a simple and efficient heuristic for global optimization over continuous spaces. J. Global Optim. 11, 341–359 (1997).
Article MathSciNet Google Scholar
Kumar, B. V., Oliva, D. & Suganthan, P. N. Differential Evolution: From Theory to Practice Vol. 1009 (Springer, 2022).
García–Nieto, P. J. et al. Hybrid DE optimised kernel SVR–relied techniques to forecast the outlet turbidity and outlet dissolved oxygen in distinct filtration media and micro-irrigation filters. Biosyst. Eng. 243, 42–56 (2024).
Article Google Scholar
Chakraborty, U. K. Advances in Differential Evolution Vol. 143 (Springer Science & Business Media, 2008).
Uludağ, G. & Uyar, A. Ş. Fitness landscape analysis of differential evolution algorithms. In Fifth International Conference on Soft Computing, Computing with Words and Perceptions in System Analysis, Decision and Control (IEEE, 2009).
Bishop, C. M. & Nasrabadi, N. M. Pattern Recognition and Machine Learning Vol. 4 (Springer, 2006).
Williams, P. M. Bayesian regularization and pruning using a Laplace prior. Neural Comput. 7(1), 117–143 (1995).
Article Google Scholar
Kruschke, J. K. Bayesian data analysis. Wiley Interdiscip. Rev. Cogn. Sci. 1(5), 658–676 (2010).
Article PubMed Google Scholar
Geurts, P., Ernst, D. & Wehenkel, L. Extremely randomized trees. Mach. Learn. 63(1), 3–42 (2006).
Article Google Scholar
Kocev, D., Ceci, M. & Stepišnik, T. Ensembles of extremely randomized predictive clustering trees for predicting structured outputs. Mach. Learn. 109, 2213–2241 (2020).
Article MathSciNet Google Scholar
Tay, J. K., Narasimhan, B. & Hastie, T. Elastic net regularization paths for all generalized linear models. J. Stat. Softw. 106 (2023).
Sanejouand, Y. H. Elastic network models: Theoretical and empirical foundations. Biomol. Simul. 601–616. (2013).
Zou, H. & Hastie, T. Regularization and variable selection via the elastic net. J. Royal Stat. Soc. Ser. B Stat. Methodol. 67(2), 301–320 (2005).
Article MathSciNet Google Scholar
Meyer, D., Leisch, F. & Hornik, K. The support vector machine under test. Neurocomputing 55(1–2), 169–186 (2003).
Article Google Scholar
Vapnik, V., Golowich, S. & Smola, A. Support vector method for function approximation, regression estimation and signal processing. Adv. Neural Inf. Process. Syst. 9 (1996).
Begum, M. Y. Advanced modeling based on machine learning for evaluation of drug nanoparticle preparation via green technology: Theoretical assessment of solubility variations. Case Stud. Therm. Eng. 45, 103029 (2023).
Article Google Scholar
Karatzoglou, A., Meyer, D. & Hornik, K. Support vector machines in R. J. Stat. Softw. 15, 1–28 (2006).
Article Google Scholar
Schölkopf, B., Smola, A. J. & Bach, F. Learning with Kernels: Support Vector Machines, Regularization, Optimization, and Beyond (MIT Press, 2002).
Download references
Department of Information Management, School of Big Data and Artificial Intelligence, Guangxi University of Finance and Economics, Nanning, 530003, China
Yanfen Wei
You can also search for this author in PubMed Google Scholar
Y.W.: Conceptualization, Formal analysis, Investigation, Writing - Original Draft, Visualization.
Correspondence to Yanfen Wei.
The authors declare no competing interests.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Open Access This article is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License, which permits any non-commercial use, sharing, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if you modified the licensed material. You do not have permission under this licence to share adapted material derived from this article or parts of it. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by-nc-nd/4.0/.
Reprints and permissions
Wei, Y. Development of novel computational models based on artificial intelligence technique to predict liquids mixtures separation via vacuum membrane distillation. Sci Rep 14, 24121 (2024). https://doi.org/10.1038/s41598-024-75074-0
Download citation
Received: 04 July 2024
Accepted: 01 October 2024
Published: 15 October 2024
DOI: https://doi.org/10.1038/s41598-024-75074-0
Anyone you share the following link with will be able to read this content:
Sorry, a shareable link is not currently available for this article.
Provided by the Springer Nature SharedIt content-sharing initiative