Altmetrics
Downloads
220
Views
146
Comments
0
A peer-reviewed article of this preprint also exists.
Submitted:
11 December 2023
Posted:
13 December 2023
You are already at the latest version
Automated F.E. Tool | Operation | Tool Tested On | Developer | Paper |
---|---|---|---|---|
ExploreKit | Feature generation & ranking | DT, SVM, RF | UC Berkeley | [125] |
One Button Machine | Feature discovery in relational DBs | RF, XGBOOST | IBM | [126] |
AutoLearn | Feature generation & selection | kNN, LR, SVM, RF, Adaboost, NN, DT | IIIT | [127] |
GeP Feature Construction | Feature generation from GeP on DTs | kNN, DT, Naive Bayes | Wellington Uni. | [128] |
Cognito | Feature generation & selection | N/A | IBM | [129] |
RLFE | Feature generation & selection | RF | IBM | [130] |
LFE | Feature transformation | LR, RF | IBM | [131] |
AHO | Automated Hyperparameter Optimization |
AIC | Akaike Information Criterion |
ALA | Adaptive Linear Approximation |
API | Application Programming Interface |
AUC | Area Under Curve |
Auto-WEKA | Automatic Model Selection & Hyperparameter Optimization |
BDW | Best Daubechies Wavelet Coefficients |
BFC | Best Fourier Coefficients |
BIC | Bayesian Information Criterion |
BOHB | Bayesian Optimization Hyperband |
CART | Classification and Regression Tree |
CASH | Combined Algorithm Selection & Hyperparameter optimization |
CI/CD | Continuous Integration Continuous Delivery or Deployment |
CMA-ES | Covariance Matrix Adaption Evolutionary Strategy |
CoD | Curse of Dimensionality |
CV | Cross-Validation |
DB | Database |
DDoS | Distributed Denial-of-Service |
DevOps | Development Operations |
DFS | Deep Feature Synthesis |
DGUFS | Dependence Guided Unsupervised Feature Selection |
DNN | Deep Neural Network |
DT | Decision Tree |
ELT | Extract, Load, Transform |
ETL | Extract, Transform, Load |
EUFS | Embedded Unsupervised Feature Selection |
FIR | Finite Impulse Response |
FN | False Negative |
FP | False Positive |
FSFS | Feature Selection with Feature Similarity |
GeP | Genetic Programming |
GP | Gaussian Process |
HVAC | Heating Ventilation and Air Conditioning |
IARPA | Intelligence Advanced Research Projects Activity |
KDD | Knowledge Discovery from Data |
kNN | k-Nearest Neighbors |
LARS | Lasso Regression |
LBFGS | Broyden-Fletcher-Goldfarb-Shanno |
LDS | Linear Discriminant Analysis |
LLC-fs | Local Learning-based Clustering with feature selection |
LR | Logistic Regression |
MAE | Mean Absolute Error |
MAPE | Mean Absolute Percentage Error |
MCFS | Multi-Cluster Feature Selection |
MDI | Mean Decrease in Impurity |
MDL | Minimum Description Length |
MICI | Maximal Information Compression Index |
ML | Machine Learning |
MLOps | Machine Learning Operations |
MRDTL | Multi-Relational Decision Tree Learning |
MSE | Mean Squared Error |
NAS | Neural Automated Search |
NDFS | Non-negative Discriminative Feature Selection |
NN | Neural Network |
NNI | Neural Network Intelligence |
OLS | Ordinary Least Squares |
OoB | Out-of-Bag |
OOP | Object Oriented Programming |
PoLP | Principle of Least Privilege |
PCA | Principal Component Analysis |
REFSVM | Recursive Feature Elimination Support Vector Machines |
RF | Random Forest |
RICA | Reconstruction Independent Component Analysis |
RMSE | Root Mean Squared Error |
ROC | Receiver Operating Characteristic |
RRFS | Relevance Redundancy Feature Selection |
SLA | Service Level Agreement |
SQL | Structured Query Language |
SRM | Structural Risk Minimization |
SS-SFS | Simplified Silhouette Sequential Forward Selection |
SVD | Singular Value Decomposition |
SVM | Support Vector Machines |
TDD | Test-Driven Development |
TN | True Negative |
TOC | Total Operating Characteristic |
TP | True Positive |
TPOT | Tree-based Pipeline Optimization Tool |
UDFS | Unsupervised Discriminative Feature Selection |
VPC | Virtual Private Cloud |
Scaling Method | Scaled Feature | Scaling Effect | ML Algorithm/Model |
---|---|---|---|
Min-Max | k-Means, kNN, SVM | ||
Standardization (z-score) | , | Linear/Logistic Regression, NN | |
l2-Normalization | Vector Space Model |
Encoding Method | Original Feature | Transformed Features | Result |
---|---|---|---|
Ordinal Encoding | string1, string2, ... | 1, 2, ... | Nonordinal categorical data becomes ordinal |
One-Hot Encoding | string1, string2, ... | 001, 010, ... | k features for k categories, only one bit is "on" |
Dummy Encoding | string1, string2, ... | 001, 010, ..., (000) | features for k categories, reference category is 0 |
Effect Encoding | string1, string2, ... | 001, 010, ..., (-1-1-1) | k features for k categories, reference category is -1 |
ML Model/Algorithm | Parametric | Linear | Train, Test, Space Complexity | Paper |
---|---|---|---|---|
Ordinary Least Squares (OLS) | ✓ | ✓ | , , | [148] |
Kernel Ridge Regression | ✓ | ✓ | , -, | [149] |
Lasso Regression (LARS) | ✓ | ✓ | , -, - | [150] |
Elastic Net | ✓ | ✓ | , -, - | [108] |
Logistic Regression | ✓ | ✓ | , , | [151] |
GP Regression | ✗ | ✗ | , -, | [152] |
Multi-Layer Perceptron | ✓ | ✗ | ★ | [153,154] |
RNN/LSTM | ✓ | ✗ | ★ | - |
CNN | ✓ | ✗ | ★ | - |
Transformers | ✓ | ✗ | ★ | - |
Radial Basis Function NN | ✗ | ✗ | ★ | - |
DNN | ✓ | ✗ | ★ | - |
Naive Bayes Classifier | ✓ | ✓ | , , | [155] |
Bayesian Network | ✗ | ✗ | ★ | [156] |
Bayesian Belief Network | ✓ | ✗ | ★ | - |
SVM | ✗ | ✓ | , , | [157] |
PCA | ✗ | ✓ | , -, | [158] |
kNN | ✗ | ✗ | , , | [159,160] |
CART | ✗ | ✗ | , , | [161] |
RF | ✗ | ✗ | , , | [162] |
Gradient Boost Decision Tree | ✗ | ✗ | , , | [163] |
LDA | ✓ | ✓ | , -, , | [164] |
CV Category | Specific CV Method | Result |
---|---|---|
Exhaustive CV | · Leave-p-out CV | models trained |
· Leave-one-out CV | models trained | |
Non-Exhaustive CV | · k-fold CV | k models trained |
· Holdout | 1 model trained | |
· Repeated Random Sub-Sampling | k models trained | |
Validation (a.k.a. Monte Carlo CV) | ||
Nested CV | · k*l-fold CV | models trained |
· k-fold CV with validation and test set | k models trained with test set |
Performance Index | Formula | Purpose |
---|---|---|
Mean Squared Error (MSE) | Regression | |
Root Mean Squared Error (RMSE) | Regression | |
Mean Absolute Error (MAE) | Regression | |
Mean Absolute Percentage Error (MAPE) | Regression | |
Coefficient of Determination () | Regression | |
Adjusted Coefficient of Determination (A-) | Regression | |
Confusion Matrix | TP, TN, FP, FN | Classification |
Accuracy | Classification | |
Balanced Accuracy | Classification | |
Missclassification | Classification | |
F1-Score | Classification | |
F-Score | Classification | |
Receiver Operating Characteristic (ROC) | Graphical | Classification |
Area Under Curve (AUC) | Graphical | Classification |
Total Operating Characteristic (TOC) | Graphical | Classification |
Method | Approach to Speed-Up | Paper |
---|---|---|
Lower fidelity estimates | Less epochs, data subsets, downscaled models/data, etc. | [215,216,220,221,222,223] |
Learning curve extrapolation | Performance extrapolated after few epochs | [224,225,226,227] |
Weight inheritance/network morphisms | Models warm-started with inherited weights | [228,229,230,231,232] |
One-Shot models/weight sharing | One-shot model’s weights shared across architectures | [233,234,235,236,237,238] |
Software | Problem Automated | AutoML Method | Paper |
---|---|---|---|
Auto-WEKA | CASH | Bayesian optimization | [248] |
Auto-WEKA 2.0 | CASH with parallel runs | Bayesian optimization | [249] |
Hyperopt-Sklearn | Space search of random hyperparameters | Bayesian otimization | [250] |
Auto-Sklearn | Improved CASH with algorithm ensembles | Bayesian optimization | [251,252] |
TPOT | Classification with FE | GeP | [253] |
Auto-Net | Automates DNN tuning | Bayesian optimization | [36] |
Auto-Net 2.0 | Automates DNN tuning | BOHB | [36] |
Automatic Statistician | Automates data science | Various | [36] |
AutoPytorch | Algo. selection, ensemble constr., hyperpar. tuning | Bayesian opt., meta-learn. | [254] |
AutoKeras | NAS, hyperpar. tuning in DNN | Bayesian opt. guides network morphism | [255] |
NNI | NAS, hyperpar. tuning, model compression, FE | One-shot modes, etc. | [256,257] |
TPOT | Hyperpar. tuning, model selection | GeP | [253] |
AutoGluon | Hyperpar. tuning | - | [258] |
H2O | DE, FE, hyperpar. tuning, ensemble model selection | Random grid search, Bayesian opt. | [259] |
FEDOT | Hyperparameter tuning | Evolutionary algorithms | - |
Auto-Sklearn 2 | Model selection | Meta-learning, bandit strategy | [252] |
FLAML | Algorithm selection, hyperpar. tuning | Search strategies | [260] |
AutoGluon-TS | Ensemble constr. for time-series forecasting | Probabilistic time-series | [261] |
AutoTS | Time-series data analysis | Various | [262] |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 MDPI (Basel, Switzerland) unless otherwise stated