Escolar Documentos
Profissional Documentos
Cultura Documentos
POLYTECHNICAL SCHOOL
INDUSTRIAL AND SYSTEMS ENGINEERING GRADUATE PROGRAM (PPGEPS)
CURITIBA
2023
MARCELO FELICIANO FILHO
CURITIBA
2023
Dados da Catalogação na Publicação
Pontifícia Universidade Católica do Paraná
Sistema Integrado de Bibliotecas – SIBI/PUCPR
Biblioteca Central
Sônia Maria Magalhães da Silva – CRB 9/1191
TERMO DE APROVAÇÃO
Dissertação aprovada como requisito parcial para obtenção do grau de Mestre no Curso de Mestrado em
Engenharia de Produção e Sistemas, Programa de Pós-Graduação em Engenharia de Produção e
Sistemas, da Escola Politécnica da Pontifícia Universidade Católica do Paraná, pela seguinte banca
examinadora:
_________________________________
Presidente da Banca
Prof. Dr. Gilberto Reynoso Meza
(Orientador)
_________________________________
Prof. Dr. André Schneider de Oliveira
(Membro Externo)
_________________________________
Prof. Dr. Roberto Zanetti Freire
(Membro Interno)
EXAMINING COMMITTEE
_____________________________________
Gilberto Reynoso Meza, Ph.D.
Advisor
(PPGEPS/PUCPR)
_____________________________________
André Schneider de Oliveira, Ph.D.
External Examiner Member
(UTFPR)
_____________________________________
Roberto Zanetti Freire, Ph.D.
External Examiner Member
(UTFPR)
This study was financed in part by the Conselho Nacional de Desenvolvimento Científico e
Tecnológico (CNPq), the Fundação Araucária (FAPPR) - Brazil Finance Codes:
310079/2019-5-PQ2, 437105/2018-0-Univ, 51432/2018-PPP, PRONEX-042/2018.
“I think it’s very important to have a
feedback loop, where you’re constantly
thinking about what you’ve done and how
you could do it better. I think that’s the
best advice: constantly think about how
you could be doing things better and
question yourself”.
Elon Musk.
ABSTRACT
The advent of the Fourth Industrial Revolution (Industry 4.0) necessitates innovative
solutions to tackle new challenges. A pertinent example is the analysis of the Electronic
Flight Control System (EFCS) benchmarking introduced by Airbus at the 2020
International Federation of Automatic Control (IFAC) World Congress. This research
focuses on Oscillatory Failure Cases (OFCs), which are crucial in the structural design
of commercial aircraft. The identification of OFCs in real-time is pivotal for enhancing
cost-efficiency, energy conservation, and flight reliability. Data-driven Soft Sensors
(SS) and machine learning (ML) algorithms have emerged as effective tools for
identifying OFCs. However, their performance significantly improves with the
integration of Feature Engineering (FEn) techniques. This study aims to develop an
FEn framework that optimizes SS performance in EFCS for real-time OFC detection.
The methodology encompasses a Systematic Literature Review (SLR) driven by four
fundamental questions. The SLR process involved an extensive analysis, starting with
2,153 papers and culminating in a detailed Content Analysis (CA) of 59 selected
papers. This SLR provides insights into the current state of SS and FEn, underlying
mathematical concepts, and answers to the four critical questions. It offers a
comprehensive overview of SS applications across seven global sectors, their
correlation with Industry 4.0, and the utilization of ML and FEn in SS implementations.
The benchmark is thoroughly detailed, including Simulink™ diagrams and specific
requirements. Initial results using Support Vector Machine (SVM), Decision Tree (DT),
and Multi-Layer Perceptron (MLP) algorithms for SS implementation in the benchmark
showed that, without FEn, the accuracy rates for identifying OFCs were below 50%.
This finding underscores the necessity of incorporating FEn. The study concludes that
SS implementation augmented with FEn considerably improves real-time OFC
detection. This discovery leads to potential advancements in FEn techniques,
indicating a promising area for future research. The research also includes the
development and presentation of an FEn framework, leading to the publication of two
papers detailing the findings from the SLR and the FEn framework. Notably, applying
a Feature Reduction technique reduced the computational cost by 75% while
enhancing performance by 70%. Therefore, this research offers valuable insights into
feature engineering within soft sensors, especially in applications like EFCS, thereby
enriching the academic and scientific community's understanding of this field.
1 INTRODUCTION .......................................................................................... 18
1.1 MOTIVATION: THE DEMAND FOR SOFT SENSORS ................................ 19
1.2 OBJECTIVES ............................................................................................... 20
1.2.1 General Objective ....................................................................................... 20
1.2.2 Specific Objectives ..................................................................................... 20
1.3 JUSTIFICATION ........................................................................................... 20
1.4 TOOLS AND METHODS .............................................................................. 22
1.4.1 Software Required ...................................................................................... 22
1.4.2 Hardware Employed ................................................................................... 23
1.5 RESEARCH IMPACTS ................................................................................. 23
1.6 DOCUMENT STRUCTURE .......................................................................... 24
2 THE SYSTEMATIC LITERATURE REVIEW ................................................ 26
2.1 SYSTEMATIC LITERATURE REVIEW: THE STATE OF ART ..................... 26
2.2 SLR PROCEDURES: THE TEN STEPS TO SYSTEMATIC RESEARCH .... 27
2.2.1 Procedure One: Research Areas and Theme ........................................... 27
2.2.2 Procedure Two: Qualitative Literature Data Review ................................ 28
2.2.3 Procedure Three: The Research Guideline Questions ............................ 28
2.2.4 Procedure Four: The Most Important Keywords for Research ............... 29
2.2.5 Procedure Five: Inclusion and Exclusion (I/E) Criteria Determination .. 30
2.2.6 Procedure Six: The Survey for Papers in Databases .............................. 31
2.2.7 Procedure Seven: To apply the Inclusion and Exclusion Criteria .......... 32
2.2.8 Procedure Eight: To Define a Classification Criteria ............................... 34
2.2.9 Procedure Nine: Appling the Classification Criteria................................ 34
2.2.10 Procedure Ten: The Content Analysis of Included Papers ..................... 38
3 RESEARCH FINDINGS ............................................................................... 46
3.1 SOFT SENSORS: STATE OF ART .............................................................. 46
3.1.1 Model-Driven Soft Sensors ........................................................................ 49
3.1.1.1 Phenomenological Modelling ........................................................................ 49
3.1.1.2 Kalman Filter ................................................................................................ 50
3.1.2 Data-Driven Soft Sensors .......................................................................... 50
3.2 MACHINE LEARNING TECHNIQUES ......................................................... 52
3.2.1 Support Vector Machine ............................................................................ 52
3.2.2 Deep Learning ............................................................................................. 54
3.2.3 Fuzzy systems ............................................................................................ 55
3.2.4 Decision Tree .............................................................................................. 56
3.2.5 Random Forest ........................................................................................... 58
3.2.6 Genetic Algorithm....................................................................................... 59
3.2.7 Gradient Boosting Machine (GBM) ........................................................... 61
3.2.8 XGBoost ...................................................................................................... 61
3.3 MATHEMATICAL BACKGROUND ............................................................... 62
3.3.1 Time Series ................................................................................................. 62
3.3.2 Classification Task ..................................................................................... 63
3.3.3 Learning Phase ........................................................................................... 64
3.3.4 Performance Analysis Techniques and Mathematical Background ...... 65
3.4 FEATURE ENGINEERING ........................................................................... 66
3.4.1 Feature Engineering: The State of The Art ............................................... 66
3.5 Q.01 MAIN APPLICATION AREAS FOR SS ................................................ 67
3.5.1 Industrial Applications ............................................................................... 67
3.5.2 Soft Sensors Applied to Aeronautics Solutions ...................................... 69
3.5.3 The Employment of SS in the Quimiometrics Industry ........................... 69
3.5.4 The Cloud Computing Solutions Based on SS ........................................ 71
3.5.5 Soft Sensors: Enhancing Health and Care Solutions.............................. 72
3.5.6 Building and Household applications ....................................................... 73
3.5.7 The General Applications for SS ............................................................... 73
3.6 Q.02: THE RELATIONSHIP BETWEEN SS AND INDUSTRY 4.0 ............... 74
3.6.1 Soft Sensors in I4.0 Scenario .................................................................... 74
3.6.2 Soft Sensors Employed in Smart Factories ............................................. 76
3.7 Q.03: FEATURE ENGINEERING AND ML APPLIED TO SS ....................... 76
3.7.1 Feature Engineering Employment to SS Development ........................... 76
3.7.2 The Machine Learning Approaches for Soft Sensors ............................. 78
3.8 Q.04: THE METHODS FOR FEATURE ENGINEERING IN SS ................... 81
3.8.1 Feature Engineering Enhancing Soft Sensors ......................................... 81
3.8.2 Hyperparameter Tunning in Soft Sensors Implementation .................... 84
3.9 SYSTEMATIC LITERATURE REVIEW SUMMARIZED RESULTS .............. 84
4 BENCHMARK DISCUSSION ....................................................................... 86
4.1 AIRBUS: OFC X IFAC – THE BENCHMARK ............................................... 86
4.2 THE MODEL SYSTEM: DIAGRAMS AND CODE ........................................ 87
4.2.1 Flight Trajectory Angle Control Module ................................................... 88
4.2.2 Load Factor Control Module ...................................................................... 89
4.2.3 Detection Surface Servo Command Simulator (Real Servo)................... 91
4.2.4 Aircraft Turbulence Dynamics Simulator ................................................. 93
5 RESULTS: THE CRITICAL ANALYSIS ....................................................... 95
5.1 SOFT SENSOR: THE SOFTWARE DEVELOPMENT .................................. 95
5.2 MATHLAB® AND PYTHON INTEGRATION ................................................ 98
5.3 BENCHMARK TESTING ............................................................................ 100
5.4 THE ML METHODS APPLICATION IN SS DEVELOPMENT RESULTS ... 101
5.5 THE DEMAND FOR FEATURE ENGINEERING IN SS ............................. 103
5.6 FRAMEWORK PROPOSE FOR SOLVING THE CHALLENGE ................. 105
5.6.1 Introduction to the Framework ................................................................ 106
5.6.2 Components of the Framework ............................................................... 106
5.6.3 Implementation Strategy .......................................................................... 107
5.6.4 Anticipated Challenges and Solutions ................................................... 107
5.6.5 Expected Outcomes and Impact ............................................................. 108
6 FEATURE ENGINEERING FRAMEWORK RESULTS .............................. 109
6.1 EMPLOYING FEATURE ENGINEERING TO THE DATASET ................... 109
6.1.1 Explaining the Outputted code ................................................................ 110
6.2 TRAINING ML ALGORITHMS WITH THE FEATURED DATASET ............ 112
6.3 IDENTIFYING OFCS WITH FEN IN THE ACQUIRED DATA ..................... 114
6.3.1 Explaining the Identification Process Code ........................................... 115
6.4 PERFORMANCE ANALYSIS ..................................................................... 116
6.4.1 Performance Metrics Calculation ............................................................ 117
6.4.2 Model Performance and Comparison ..................................................... 118
6.4.3 Performance Improvement Assessment ................................................ 119
6.4.4 Confusion Matrix Interpretation for Machine Learning methods ......... 119
6.5 INTELLIGENT SYSTEM FRAMEWORK FOR SS PERFORMANCE
ENHANCEMENT ..................................................................................................... 125
6.5.1 Framework Code Overview ...................................................................... 125
6.5.2 Framework Architecture .......................................................................... 126
6.5.3 Advantages of the Proposed Framework ............................................... 128
6.5.4 Future Directions and Considerations .................................................... 129
6.6 FEATURE REDUCTION TECHNIQUES IN SOFT SENSOR
PERFORMANCE EVALUATION ............................................................................. 129
6.6.1 Feature Reduction as a Key for Performance Enhancement ................ 129
6.6.2 The Imperative Nature of Feature Reduction ......................................... 130
6.6.3 The Role of Feature Reduction in Soft Sensor Performance ................ 131
6.6.4 Implementing Feature Reduction for Performance Evaluation ............ 133
6.6.5 Feature Landscape and Model Mastery .................................................. 134
7 DISCUSSION ............................................................................................. 141
7.1 REVIEW OF FINDINGS ............................................................................. 141
7.1.1 ML Models and Performance ................................................................... 141
7.1.2 Feature Engineering and Enhancement of Decision Tree Model ......... 141
7.1.3 Improvement in Predictive Performance ................................................ 141
7.1.4 Comparative Evaluation ........................................................................... 142
7.1.5 Interpretation and Insight ........................................................................ 142
7.2 EVALUATION OF THE DECISION TREE MODEL .................................... 142
7.2.1 Accuracy of the Model ............................................................................. 142
7.2.2 Analysis of Confusion Matrix .................................................................. 143
7.2.3 Precision, Recall, and F1-Score .............................................................. 143
7.2.4 Model Interpretability ............................................................................... 143
7.2.5 Comparative Analysis with Industry Standards..................................... 143
7.2.6 Implications for the Field ......................................................................... 144
7.3 SIGNIFICANCE OF THE RESEARCH ....................................................... 144
7.3.1 Academic Significance ............................................................................. 144
7.3.2 Industrial Significance ............................................................................. 145
7.3.3 Broader Implications ................................................................................ 145
7.4 LIMITATIONS ............................................................................................. 145
8 CONCLUSION............................................................................................ 147
REFERENCES ........................................................................................................ 150
ATTACHMENT A .................................................................................................... 158
ATTACHMENT B .................................................................................................... 160
ATTACHMENT C .................................................................................................... 161
ATTACHMENT D .................................................................................................... 163
ATTACHMENT E .................................................................................................... 165
ATTACHMENT F .................................................................................................... 168
18
1 INTRODUCTION
During the fourth industrial revolution, also known as Industry 4.0, there was a
burgeoning drive for innovation and resource optimization. In collaboration with data
microprocessing, Soft Sensors (SS) have surfaced as potential problem solvers. These
technologies are employed in prediction models, real-time control, and system scaling,
among other applications (L. Fortuna, S. Graziani, and A. Rizzo et al., 2014). SS, or
virtual sensors, use Machine Learning (ML) techniques to process real-time data,
facilitating informed decision-making. As a result, their application is becoming
increasingly invaluable across numerous industrial sectors.
Another general application of SS is measuring costly intricate variables using
existing methods. The acquisition of such data can involve multiple sensors, expert
input, or extended processing time in outdated software (F. Souza, A. Francisco, and
R. Araújo et al., 2016). Finally, the large-scale deployment of SS in various production
lines justifies their importance in indicating Key Performance Indicators (KPIs), which
allow departments to track performance and identify areas requiring improvement (D.
Parmenter, 2010).
In high-level control systems, one crucial characteristic is the ability to perform
data acquisition, considering all system actuators and sensors. This process involves
linear actuators, servo-controlled joints, and a network of highly precise analog sensors
that feed back into a closed control loop. Such principles underpin the case study from
AIRBUS presented at IFAC (International Federation of Automatic Control)
(Engelbrecht and Goupil, 2020).
According to IFAC's case study presenters, SS also play a pivotal role in
developing intelligent products, such as flight control systems that employ a variety of
embedded sensors for controlling altitude, speed, and trajectory. These systems
demand smarter algorithms to enhance fault detection modules, increase the
robustness of established systems, and avert potential catastrophes. In addressing
these challenges, this case study exemplifies the role of control and automation
engineering: to transmute industrial issues into engineering problems.
The increased use of sensors in the industry has catalyzed the need for artificial
intelligence techniques, particularly machine learning, to treat data effectively and add
value to Big Data infrastructure (W. Lee, G. Mendis, and J. Sutherland, 2019). By
19
The authors (J. Engelbrecht and P. Goupil, 2020) emphasize the pivotal role of
the Electronic Flight Control System (EFCS) in an aircraft. The EFCS, tasked with
regulating attitude, speed, and trajectory, comprises an intricate network of
components, including wiring, probes, actuators, numerical buses, power sources, and
sensors. This system facilitates communication between the cockpit and the aircraft's
movable parts and control surfaces. In this context, the EFCS's consistent availability,
even under fault conditions, is paramount, making fault detection a stringent aspect of
aircraft design. A case in point is the Oscillatory Failure Case (OFC), a failure type
resulting from weight-saving techniques that can adversely affect the aircraft's
structure and robustness (J. Engelbrecht and P. Goupil, 2020).
To meet this stringent standard, (V. Ribeiro, R. Kagami, and G. Reynoso-Meza,
2020) devised a data-driven detection model utilizing the Decision Tree (DT) algorithm
across various scenarios. In addition, they noted the complexity of the problem and
employed signal processing techniques for filtering and extracting features from two
signals – one related to control action and the other to the feedback sensor. Therefore,
when coupled with the DT method, feature engineering could enhance fault detection
within the FCS, improving the aircraft's weight performance.
This lays the groundwork for the primary research question driving this
investigation: How can the performance of a Soft Sensor for Oscillatory Failure
Anomaly detection in an aircraft's Flight Control System (FCS) be improved
using Feature Engineering?
20
1.2 OBJECTIVES
This section presents the research's general and specific objectives for feature
engineering applied to Soft Sensors performance improvement to identify OFCs.
The general objective demands the achievement of each one of the specifically
defined objectives listed:
a) To survey the most suitable papers related to the general objective theme
by developing a Systematic Literature Review. Which guide questions
will be defined to guide the exploratory research;
b) To integrate the MATLAB® and SimuLink™ benchmark with Python;
c) To implement a Python, embed Soft Sensor and ML classes (SVM, MLP,
and DT) to process the benchmark data in real-time;
d) To propose a framework to employ Feature Engineering to improve the
results obtained by B;
e) To conduct a discussion around this implementation and propose an
Intelligent System framework to improve SS performance;
f) To compose a critical analysis with research results, propose a final
dissertation project guideline.
1.3 JUSTIFICATION
“Further studies can be carried out with the dual objective of reducing the
minimal OFC detectable amplitude and avoiding using a system model. A
One-Class Support Vector Machine (OC-SVM) technique might be used
directly on flight data to define a suitable test statistic”.
In their third related work, (S. Urbano, E. Chaumette, P. Goupil, et al., 2018)
employed an industrial Airbus desktop simulator to aid a Monte Carlo test. They
observed performance degradation as turbulence levels escalated. They further
22
underscored the need for additional research into threshold tuning as part of their
Monte Carlo test campaign.
Research by (A. Zolghadri, J. Cieslak, D. Efimov, et al., 2015) delved into
conventional design methods and advanced model-based techniques for failure
detection in Flight Control Systems (FCS). They contended that while model-based
techniques cannot entirely supplant the redundancy of physical sensors in aircraft and
aerospace systems, they can significantly bolster fault detection performance when
adequately harnessed.
A related paper (R. Cordeiro, J. Azinheira, and A. Moutinho, 2020) suggested
that their proposed Failure Detection and Isolation System (FDIS) could be further
enhanced as applied to a Boeing 747 aircraft simulator. In addition, they proposed the
inclusion of a Supervisory Loop (SL) to interpret the results of Kalman Filters, thus
enabling diagnosis and decision-making features through an additional Feed-Forward
Differential.
In conclusion, the justification for this dissertation stems from the exigency to
enhance Machine Learning methods' performance within the Electronic Flight Control
System in the presented benchmark. It posits that Feature Engineering can play a
pivotal role in improving the performance of the data-driven approach taken by Soft
Sensors.
This case study demands many tools and methods to fulfill the proposed
requirements and the main research question shown in problematization (section 1.1).
Due to this reason, this subsection will be divided into the software required and the
hardware employed to achieve the proposed goals. Such information is relevant for
the scientific community to reproduce its features and validate the solution's reliability
and robustness.
The hardware employed for this benchmark solution is a Ryzen 5 3600 Mhz
processor, with a six-core CPU processor and 12 logic processors, with 48.0 Gb of
RAM installed, a 500Gb SSD, and a 12Gb video card GPU model RTX 3060.
Nonetheless, it was all assembled in ASRock B450M Steel Legend motherboard and
employed Windows 10 as an Operational System (OS).
This research can impact many technological scenarios, for example, the
industrial, chemometrics, and other engineering or computer science fields. Once its
core is related to a real-world problem benchmark modeled in Simulink, with control
24
This dissertation document starts with an introduction in chapter one and related
themes such as problematization, the specific objectives of this research, and the
general. Then, the justification of soft sensors-based research and the methodology
adopted.
Chapter two presents the Systematic Literature Review procedures by
explaining its definition and all steps to achieve the conducted research. Besides that,
the main questions for this research will be explored, such as inclusion and exclusion
25
criteria, and research data will be presented in graphics and tables. In addition, the
content Analysis heads a discussion about survey discoveries.
After exploring the research guidelines, their results will be presented in chapter
three, starting with soft sensors” state of the art, their main applications to solve
engineering problems, and other demands according to the guideline questions
defined at the SLR.
Chapter four discusses the case study and the achieved solutions that the
course of research might find. Chapter five presents the results of employed ML
methods to the SS application over the benchmark. Chapter six presents the Feature
Engineering Framework structure and its applications over the benchmark modifying
the number of features to evaluate the accuracy.
Therefore, chapter seven conducts a concise discussion about each results
theme in this work. Finally, chapter eight presents assignment plans for the remainder
of this research with a schedule.
26
This chapter presents the Systematic Literature Review (SLR), which explores
papers on the Soft Sensors field and areas related to Intelligent Systems, Industry 4.0,
and Feature Engineering. At the end of this survey, a Component Analysis (CA) will be
conducted to present the main findings of SLR.
According to (Palmatier, Houston, and Hulland, 2018), the SLR is guided by ten
procedures that can be summarized in Figure 1:
To this extent, the survey will be conducted according to the guidelines in Figure
1 to reach the defined objectives. Nevertheless, technical subjects must define every
procedure to reduce bias and evidence the most relevant files for the research fields.
As presented by (Palmatier, Houston, and Hulland, 2018), “The author sets clear
objectives for the review and articulates the specific research questions or hypotheses
that will be investigated.” The defined objectives and motivation were the guidelines
for defining the theme and research areas for this research.
Towards the first step, define the research areas around the chosen theme for
the survey. For example, chapter one shows that soft sensors and feature engineering
are the main research themes. Thus, the related research areas are Machine Learning,
Multi-Objective optimization, Soft Sensors Implementation, Industry 4.0, and Intelligent
Systems.
28
“Recently, there has been a move from such scales to more qualitative quality
measures for different study designs. As well as critical appraisal, sub-group
analyses can be used to determine whether the meta-analysis results are
altered by removing specific studies or groups of studies. If the results from all
sub-group analyses are consistent, then the analysis results are more likely to
be found to be robust”.
As presented in item 1.1, the problem this research is facing is a practical solution
for an engineering problem, using a Soft Sensor (SS) for feature engineering
applications. Therefore, this experiment aims to improve the SS performance, and
every paper related to this focus will be analyzed and included or not in the primary
survey.
The general and specific objectives guide the main research questions and are
essential for defining the search aim and reaching the work’s objectives (Nightingale,
2009). Based on that, this survey will be modeled by the four questions presented in
table 1:
Q.01 Which are the main application areas for SS in general engineering?
What is the relationship between Intelligent Systems and SS in an Industry
Q.02
4.0 scenario?
Which feature engineering or ML techniques are employed in SS
Q.03
development?
Which are the possible methods for performing Feature Engineering in Soft
Q.04
Sensors Intelligent System applications?
Source: The Author, 2022.
These four questions in table 1 are key to reading the papers and searching for
the appropriate information for this research. Each included paper will be subjected to
answer or provide related information. It is crucial to notice that the soft sensor
definition is not one of the guideline questions. However, such information will be
collected to lead the SS state of art section.
29
In agreement with (D. Moher et al., 2015), after defining the research Theme
and Area, the qualitative data, and the key questions, the three most relevant keywords
and their two main synonyms are defined. Afterward, they will be combined and used
to search papers in databases. In this research case, the main keywords can be
summarized in table 2:
Such keywords will be searched at CAPES and Science Direct Databases while
browsing paper’s titles are analyzed according to procedure five. With such filters,
selecting the maximum combination of possible related papers with the most relevant
contributions to this research is possible.
After that procedure, objective inclusion and exclusion criteria for desired papers
are adopted for the seventh procedure. Then, it is the moment to start surveying for
keywords in Table 3 and condensing all data in the sixty steps.
After defining the combination between the main keywords in Table 3, the sixty
procedure starts by searching for each one at CAPES and Science Direct Databases.
Finally, the search results with open access and open archive filter only are
summarized in Table 5:
number of related articles was the combination between “Soft Sensors” AND “Feature
Discovery”, which resulted in only one paper being found. Such quantitative analysis
is shown in figure 2:
1200
NUMBER OF PAPERS
1000
800
600
400
200
0
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16
KEYWORDS INDEX
The most results were found when combining “Feature Engineering” and
“Machine Learning” because the search engine looks for the engineering word inside
the paper. For this reason, the following procedures are essential to filter and correct
such misunderstandings.
After browsing for more than two thousandth papers, this procedure significantly
reduces the total number of papers, selecting them based on the Inclusion and
Exclusion criteria presented in table 4. Hence, table 6 shows the relationship of
selected papers by keyword combination:
After applying the I/E criteria, the number of papers was reduced to 109, only
with filters in search engines and title reading techniques. Nevertheless, the
synthesizing capacity of the seventh procedure was reduced by 95% of the amount of
found papers. This result is shown in figure 3:
10
8
6
4
2
0
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16
KEYWORD INDEX
This procedure does not require deep analysis in papers. Only title reading
techniques and filtering in the search engine are enough. In such a manner, some
34
articles could not match the research, so procedure eight defines the classification
criteria for selected articles.
The eighth procedure defines the Classification Criteria for found papers and
aims to add another relevant filtering for selected papers. These criteria must follow
the defined objectives and ensure that papers are related to the central research
questions and are relevant enough to read thoroughly. The classification process is
based on the paper’s title, keywords, and abstract reading.
Such procedure filtered the 109 articles found in previous steps using the
classification criteria defined in the past procedure. According to (D. Moher et al.,
2015), the process consists of reading the paper’s abstracts and classifying them into
defined criteria.
These criteria are defined by technical details for selecting the relevant articles
that accomplish objectives and solve problems defined in the introduction. For
example, suppose that the article presents a case study that can contribute to this
research goal of applying a framework for SS Hyperparameter Tuning. In that case,
the article can be classified as “A.” On the other hand, if the contribution to the research
is less significant and has only theoretical information, it will be classified as a “C” class
article, and so on. These criteria are presented in Board 1:
Criteria A1 is based on the demand to find SS-related papers with research. The
second classification level can be developed for paper clustering related to intelligent
systems containing SS. The articles which have no classification in either cluster
35
criteria are excluded from this survey, as they might not contribute to answering the
key questions. Hence, Board 2 shows 59 articles that meet the established criteria:
7 A two-step multivariate statistical learning approach for batch process soft sensing X
19 Decision Tree for Oscillatory Failure Case Detection in a Flight Control System X
CRITERIA
ID Article's Title
A B C
Development of an intelligent tool condition monitoring system to identify
22 X
manufacturing tradeoffs and optimal machining conditions
Distributed estimation over a low-cost sensor network: A Review of the state of
23 X
the art
Encoding and exploring latent design space of optimal material structures via a
24 X
VAE-LSTM model
Energy consumption prediction by using machine learning for smart building:
25 X
Case Study in Malaysia
Evaluation of machine learning for sensor-less detection and classification of
26 X
faults in electromechanical drive systems
Fermentation 4.0, a case study on computer vision, soft sensor, connectivity, and
27 X
control applied to the fermentation of a thraustochytrid
FIEMA, a system of fuzzy inference and emission analytics for sustainability-
28 X
oriented chemical process design
29 Flexible, wearable biosensors for digital health X
Genetic programming-based symbolic regression for goal-oriented dimension
30 X
reduction
Hyperparameter tuning to optimize implementations of denoising autoencoders
31 X
for imputation of missing spatial-temporal data.
Industry 4.0 based process data analytics platform: A waste-to-energy plant case
32 X
study
Industry 4.0 in Action: Digitalization of Continuous Process Manufacturing for
33 X
Formulated Products
34 IoT-based Indoor Occupancy Estimation Using Edge Computing X
Laundry fabric classification in vertical axis washing machines using data-driven
35 X
soft sensors
Machine learning based adaptive soft sensor for flash point inference in a refinery
36 X
real-time process
Machine learning based identification of energy states of metal-cutting machine
37 X
tools using load profiles
38 Machine learning for biochemical engineering: A review X
MANU-ML: Methodology for the application of machine learning in manufacturing
39 X
processes
Model stacking to improve prediction and variable importance robustness for soft
40 X
sensor development.
Moving towards an era of hybrid modelling: advantages and challenges of
41 coupling mechanistic and data-driven models for upstream pharmaceutical X
bioprocesses
42 Neuro-fuzzy Soft Sensor Estimator for Benzene Toluene Distillation Column X
Online Parameterization of a Milling Force Model using an Intelligent System
43 X
Architecture and Bayesian Optimization
PLS-based soft-sensor to predict ammonium concentration evolution in hollow
44 X
fiber membrane contactors for nitrogen recovery
Prediction of sorption-enhanced steam methane reforming products from machine
45 X
learning based soft-sensor models
Predictive maintenance enabled by machine learning: Use cases and challenges
46 X
in the automotive industry
37
CRITERIA
ID Article's Title
A B C
Predictive maintenance on sensorized stamping presses by time series
47 X
segmentation, anomaly detection, and classification algorithms
Predictive model-based quality inspection using Machine Learning and Edge
48 X
Cloud Computing
Process PLS: Incorporating substantive knowledge into the predictive modelling
49 X
of multiblock, multistep, multidimensional and multicollinear process data
Proposition of the methodology for Data Acquisition, Analysis and Visualization in
50 X
support of Industry 4.0
Radiomics and Artificial Intelligence for Biomarker and Prediction Model
51 X
Development in Oncology
52 Self-healing sensorized soft robots X
Soft sensor of bath temperature in an electric arc furnace based on a data-driven
53 X
Takagi–Sugeno fuzzy model
STLF-Net: Two-stream deep network for short-term load forecasting in residential
54 X
buildings
Technical Note describing the joint Airbus-Stellenbosch University Industrial
55 X
Benchmark on Fault Detection
The biological transformation of industrial manufacturing – Technologies, status
56 X
and Scenarios for a sustainable future of the German manufacturing industry
Towards an intelligent linear winding process through sensor integration and
57 X
machine learning techniques
Understanding chemical production processes by using PLS path model
58 X
parameters as soft sensors
Using a support vector machine for building a quality prediction model for a center-
59 X
less honing process
Source: The Author, 2022.
B, 17%
Classified
A, 23%
C, 13%
Not Related,
A B C Not Related
Source: The author, 2022.
This figure shows that the inclusion and exclusion criteria cannot filter every
relevant paper for an SLR. Nevertheless, classification criteria open the discussion
about content analysis, which is the next and last step for this SLR.
After applying the research methodologies for an SLR, the most relevant data
will be extracted from every selected and classified article presented on Board 2. Their
ID will be used to identify them in Board 3, which summarizes their main contributions
and which guideline questions they answer:
RSL QUESTION
ID Main Contribution to this SLR RELATION
01 02 03 04
This paper explores biomedical datasets that use the NER module of
the Spark NLP library. They require no handcrafted features or task-
8 X X X
specific resources and achieve state-of-the-art scores on popular
biomedical datasets and clinical concept extraction challenges.
RSL QUESTION
ID Main Contribution to this SLR RELATION
01 02 03 04
Five algorithms are proposed based on the features that reflect their
strengths to calculate the rating of batters, bowlers, batting all-
13 rounders, bowling all-rounders, and wicketkeepers. X X X
CS-PSO hybridization is a feature optimization strategy to eliminate
redundant, irrelevant, and noisy features.
RSL QUESTION
RELATION
ID Main Contribution to this SLR
01 02 03 04
RSL QUESTION
RELATION
ID Main Contribution to this SLR
01 02 03 04
The work studied data-driven soft sensors in the case study to predict
syngas heating value and hot flue gas temperature. The neural
32 network-based NARX model demonstrated better performance among X X X
the studied data-driven methods. Besides that, it presents the data-
driven soft sensors as valuable tools for predictive data analytics.
This article presents a combined solution that aligns with the concepts
of Industry 4.0 by providing a digital twin, cloud integration, and
33 sophisticated statistical, hybrid, and mechanistic models. The models X X X
are used for soft sensors, Model Predictive Control, and Optimisation
algorithms to predict and control product Quality Attributes.
RSL QUESTION
RELATION
ID Main Contribution to this SLR
01 02 03 04
In this study, two soft sensor models were developed and used to
predict and estimate variables that would be difficult to measure
45 directly. Both artificial neural networks and random forest models were X X X
developed as soft sensor prediction models. Besides that, it brings
feature selection relevant contributions.
RSL QUESTION
RELATION
ID Main Contribution to this SLR
01 02 03 04
50
NUMBER OF ANSWERS
40
30
Q.01, 54
Q.03, 49
20
Q.02, 27
10 Q.04, 23
0
Q.01 Q.02 Q.03 Q.04
QUESTIONS
3 RESEARCH FINDINGS
This chapter presents the findings of SLR in the Soft Sensors field of study,
Machine Learning (ML) processes. First, however, the benchmark technical details will
be explored in the next chapter. Then, starting with the state of the art of Soft Sensors
in 3.1, the exploration of some relevant ML techniques, for example, Support Vector
Machine (SVM), Decision Tree (DT), Deep Learning (DL) methods, and others in 3.2.
In addition, the mathematical background for Time-Series, Classification Task, and
Learning Phase are discussed in 3.3. Finally, feature engineering is the theme of sub-
chapters 3.4, 3.5, 3.6, 3.7, and 3.8, respectively, presenting the table’s 1 guideline
questions answers.
Soft Sensors can be defined as inference tools that process sensor data in real-
time to, with this information, estimate other more complex variables to measure, such
as data from a statistical laboratory test, as presented (Souza, Araújo, and Mendes,
2016). The intelligence of these sensors is based on algorithms and machine learning
techniques for data mining and improving the quality of information collected by
sensors, eliminating outliers, or condensing information with mathematical models.
For (Jalee and Aparna, 2016), the origin of the term Soft Sensor derives from
the junction between “software” and “sensor”. These models were developed through
computing processing hardware information already presented in supervisory systems,
and technicians evaluated alarms to make decisions. However, previously
immeasurable variables can be estimated based on secondary variables read by the
sensors using ML or DL techniques.
Such tools help to construct intelligent products by allowing them to make real-
time decisions. They can be classified into two groups concerning how the data is
treated (Maggipinto et al., 2019):
a) Model-Driven, in which data is acquired in real-time to feed predictive models,
promoting quick decision-making;
b) Data-Driven, in which statistical models are built through a robust database
already obtained during tests, is an intelligent sensor that employs most
machine learning techniques.
47
Aiming to map those areas (Kadlec and Gabrys, 2009) defined a hierarchical
tree of the machine learning methodologies used by each of these strands, as shown
in Figure 6:
The main techniques are raised and categorized according to the authors'
studies. Therefore, the two ways of implementing SS will be presented in the following
subsections. Nevertheless, table 10 summarizes every definition for SS made by each
one of the papers, following the publication time order presented in Board 4:
Soft sensors are models that can provide accurate estimations in real-time for
4 2016 these hard-to-measure parameters without the financial investment and
maintenance requirements, using the relationships with conventional sensor data.
a ‘soft sensor’ infers from measurable quantities (furnace temperature and
6 2019 transport time) on the estimated microstructural state as a function of the process
setting (strain rates and pause times).
process manufacturers rely on soft sensors, which can model data collected from
16 2020
conventional measurements and used to predict key variables.
However, using already existing sensors’ signals as an input for deep neural
networks to infer the desired data rather than measure it directly could be a viable
20 2020
alternative [11]. These so-called soft or virtual sensors could satisfy data needs
using the same hardware.
48
With this new soft sensor, it is now possible to monitor moisture across the six
33 2021 chambers in real time while using the single NIR Moisture sensor to measure
moisture at the endpoint before feeding the granules to the tablet press.
A Soft Sensor (SS) [4] is a technology that allows for estimating the value of a
quantity that is too costly or impossible to measure from indirect sensor
35 2021
measurements, making it well-suited for the typology detection task. They can be
divided into Model-Driven or Data-Driven.
Based on machine learning techniques, soft sensors can infer the value of a
certain magnitude from the indirect measurement of other magnitudes. In other
36 2021 words, a data-driven soft sensor is an inference scheme capable of learning
certain multi-parametric and highly non-linear causality relationships from a
historical data set.
integration with mass balance equations for soft-sensor development [168]. These
hybrid models often show higher predictive power and data efficiency than purely
38 2021
physical or data-driven models and are robust to small datasets with low quality
(e.g., noisy data).
Soft sensors can be broadly categorized based on the type of model they utilize:
mechanistic, which uses first-principles to develop a description of the process;
data-driven, which uses historical process data combined with ML algorithms to
build a model; and hybrid, which combines the two. First-principle models are
40 2021 desirable but are limited by the necessity of adequate knowledge of the
underlying process mechanisms and usually do not account for uncertainties.
Instead, data-driven methods need only historically processed data and, as such,
have been widely explored in academia and industry in processes where apriori
knowledge is not available.
Soft sensors measure the unmeasured quantity (primary variable) from the
measured quantity (secondary variable). For example, temperature, pressure,
liquid levels, etc., are the sensing variables in the process or chemical industry
[5]. Two types of soft sensors are used, i.e., model-driven and data-driven soft
42 2022 sensors. Model-driven soft sensors, also called the phenomenological model, are
based on the first principle model, whereas data-driven soft sensors are based on
measured data within plants. Data-driven soft sensors achieved popularity
compared with model-driven ones since they mainly depend on the actual
process and can represent it more accurately.
A soft sensor is computer software that maps the values from the
Input variables to predict the output variable/s. Note that primary variables (mainly
44 2022
nutrient and organic concentration) are traditionally measured in the laboratory
and, thus, are characterized by time-delayed responses.
49
Relating this cost to the performance of the batch in terms of process variables
49 2022 will result in a better understanding of the batch variations. It can even result in a
soft sensor that can predict the cost for a running batch in real-time.
The power of using a combination of variables as soft sensors
In production, processes are thoroughly established. Using model parameters as
58 2022
soft sensors may provide much more information about the actions to take when
something goes wrong.
Source: The author, 2022.
Thus, (Kadlec and Gabrys, 2009) define the FPM (First Principle Model)
approach as a phenomenological model in which models are defined based on base
equations and mathematical descriptions of the systems studied. The focus is on
steady-state analysis that does not consider disturbances caused by adverse
conditions in the ideal model. However, the researchers point out that with the increase
in instrumentation in industrial plants, these models lost their place in the market to SS
based directly on data, showing more excellent reliability.
50
Moreover, (Zambonin et al., 2019) highlight that SSs are statistical technologies
that transform low-cost data into complex or high-cost information, improving process
performance as corrective actions are taken in real-time. The authors emphasize the
demand for machine learning techniques supervised by neural networks. The most
51
each new value read. Hence, the average is normalized, simple predictions can be
made, and the method can work to treat noises.
“The comparative analysis of different MLAs shows that the Support Vector
Machine (SVM) outperforms other Machine Learning (ML) models…
Furthermore, the sensitivity analysis performed in this study illustrates that the
SVM is less sensitive to the number of samples and mislabeling in the model
training than other MLAs (Machine Learning Algorithms)”.
Hence, according to (C. Chang and C. Lin., 2001), who developed the LIBSVM,
the method presents a wide range of applications. For example, solving SVM
optimization problems, theoretical convergence, multi-class classification, probability
estimation, and parameter selection.
Another relevant information is the main equations related to the method as a
supervised algorithm used for classification and regression. Its primary objective,
defined by the equation 1 as the objective function and the equation 2 as the constraint:
𝑛
1
𝑚𝑖𝑛𝜔,𝛽,𝜀 ‖𝜔²‖ + 𝐶 ∑ 𝜀𝑖 (1)
2
𝑖=1
𝑦𝑖 (𝜔 ∙ 𝑥𝑖 + 𝑏) ≥ 1 − 𝜀𝑖 , 𝜀𝑖 ≥ 0 (2)
This equation finds a hyperplane characterized by the weight vector (𝜔) and
bias (b) that maximizes the separation margin between two classes. Meanwhile, the
1
term (2 ‖𝜔²‖) aims to maximize this margin, while (𝐶 ∑𝑛𝑖=1 𝜀𝑖 ) penalizes
misclassifications, with the constant (C) determining the trade-off between margin
maximization and misclassification penalty. Therefore, the constraints (𝑦𝑖 (𝜔 ∙ 𝑥𝑖 + 𝑏) ≥
1 − 𝜀𝑖 ) ensure data points (𝑥𝑖 ) of class (𝑦𝑖 ) lie on the correct side of the margin. Then,
the slack variables (𝑥𝑖 ) are introduced to allow certain points to lie inside the margin or
be misclassified, providing flexibility to the model for better generalization.
Finally, the main advantages of using SVM for a data-driven approach using SS
are the memory efficiency, the possibility to implement high dimensional spaces, the
versatility in implementing many kernel functions, and many samples.
54
According to the Data Science Academy (2018), deep learning has extended
what was known until the 2000s as machine learning to a new level due to the growth
of computational capacity and new artificial intelligence techniques. Presenting more
satisfactory results in increasingly complex challenges in robotic computing and AI.
What differentiates ML from DL is the complexity of neural networks. While in the first,
the networks have fewer neurons and thus a smaller amount of data processing, while
the second has deeper layers of learning, demanding parallel processing, those
differences between simple ANN and DLNN can be observed in Figure 9:
As presented in Figure 4, neural networks have three layers: input, output, and
intermediate layers called hidden layers. The processing occurs intensively in these
layers, and the number of iterations is high. For example, in the output layer, the input
data were mathematically processed by the condensed neurons and transformed into
information relevant to decision-making.
According to (P. Bezak, P. Bozek, and Y. Nikitin, 2014), “Deep learning methods
have the capability of recognizing or predicting a large set of patterns by learning
sparse features of a small set of patterns.” Hence, they can be applied even in data-
less scenarios.
According to (M. Lei et al., 2019), DL algorithms are based on neural networks
and generally process data from different types of sensors. Whose focus is the
classification of information based on data previously obtained in a system, then
intelligent decisions can be made through these datasets.
55
Besides that, this subset of machine learning techniques utilizes artificial neural
networks, especially deep architectures with many layers. In this sense, a basic
component of DL is the artificial neuron, represented by the equation 3:
𝑦 = 𝑓(𝑊 ∙ 𝑥 + 𝑏) (3)
In this equation, (𝑥) is the input vector, (𝑊) is the weight vector, (𝑏) is the bias,
and (𝑓) is the activation function which introduces non-linearity, enabling the network
to model complex patterns. Deep learning models are composed of multiple such
neurons organized in layers. The depth, or the number of layers, and the non-linear
transformations allow these models to learn and represent intricate patterns from vast
amounts of data. The training process involves adjusting the weights ( 𝑊) and biases
( 𝑏) using algorithms like backpropagation to minimize a loss function, which measures
the discrepancy between the predicted and actual outputs.
According to (J. Jang, 1991), fuzzy logic rules originate from the description of
the behavior of the systems. Based on the premises, each new rule is produced by the
combination of rules (“I”, “J”, etc...), and, at the end of the process, the system output
(“Z”) presents the weighting of all established rules. The author also presents the
ANFIS topology as shown in Figure 10:
Thus, five layers build this network, and the values of the assumptions are
determined according to the accurate modeling situations. Then, the second layer
performs the product between assumptions and delivers the result to the third layer.
Which calculates the rate at which the obtained weights are triggered up to the fourth
layer. In this step, the consequence parameters act on the factors obtained during the
process to perform the sum of the results in the last layer and emit the output signal of
this neural network. Thus, the model uses data to train and improve itself.
According to (H. Pacco, 2022), “Fuzzy Logic is a method of reasoning based on
approximation and assumptions that resembles the human reasoning model,” which
allows the Boolean decision-making algorithm based on the input layer. Then, the main
equations
|𝑆𝑣|
𝐼𝐺(𝑆, 𝐴) = 𝐻(𝑆) − ∑ 𝐻(𝑆𝑣 ) (5)
|𝑆|
𝜗𝜖𝐷𝐴
Where (𝑆) is a set of samples, (𝐴) is an attribute, (𝑝𝑖 ) and (𝑝𝑖 ) are the proportions
of positive and negative samples. The goal is to find the attribute that returns the
highest information gain, and this attribute is the decision boundary. Thus, the process
repeats with the sub-sets of data until the tree reaches its termination conditions.
Therefore, the decision tree provides a flowchart-like structure where each
internal node denotes a test on an attribute, each branch signifies the outcome of that
test, and each leaf node holds a class label. In the equation 5, the (𝑆) is the total set of
node samples, (𝐷𝐴 ) represents the subset of the dataset (𝐷) where the attribute (𝐴)
takes on a specific value (𝜗). At a glance, it's essentially a filtering of the main dataset
based on a particular attribute value. So, when computing something for (𝐷𝐴 ), the focus
is the data where the attribute (𝐴) is equal to (𝜗). This subset is used in decision trees
to determine the quality of a split based on an attribute's value. Therefore, the attribute
(𝐴) that yields the maximum information gain is selected to make the split. This process
is applied to each child node until a stopping criterion is met.
Further, figure 11 shows an example of a simple decision tree adapted from the
introduction of DTs presented by (J. Quinlan, 1986):
This classical problem presents the weather on Saturday morning, in which the
attributes are the outlook, with sunny, overcast, and rain attributes; the humidity, which
can be high or regular; and the windy, which is Boolean.
After defining those variables, there are two defined classes in the dataset, the
Positive instances (P) and the Negative Instances (N), defined by the author’s
judgment about the time (in real applications, it will be defined by data features).
Nevertheless, the cited authors also pointed out their main advantages: DTs
can handle multi-output problems, use the white box models approach, and require
little data preparation. Another advantage is their ability to manipulate categorical or
numerical data, performing well even when outlier data violate the proper model.
Finally, it is possible to validate such models using statistical data from the dataset,
matching its reliability.
Their limitations relate to over-complex datasets, which could generate
overfitting once their predictions are not continuous. Hence, extrapolation is not an
expected feature in those models. According to (J. Quinlan, 1986), “the iterative
framework cannot be guaranteed to converge on a final tree unless the window can
grow to include the entire training set.”
In this equation, 𝐹𝑚−1 (𝑥) refers to the ensemble model constructed until the
𝑚 − 1𝑡ℎ stage. The parameter 𝛼 is the learning rate, dictating the step size at each
iteration in the search space. The sum encompasses all regions, denoted by 𝐽, where
𝑅𝑗𝑚 represents the j-th region of the m-th tree. The term 𝛾𝑗𝑚 is the output value for this
specific region, while the indicator function 𝐼 ensures that the sum is taken over regions
where the condition inside the parentheses is met. Therefore, (Friedman, J.H, 2001)
provided a comprehensive exploration of the Gradient Boosting Machine and its
potential applications in various domains.
3.2.8 XGBoost
𝑛 𝑇
𝑂𝑏𝑗(𝜃) = ∑ 𝐿(𝑦𝑖 , 𝑦̂𝑖 ) + ∑ Ω(𝑓𝑖 ) (7)
𝑢 𝑗=1
Here, the term 𝐿 signifies the loss function, responsible for gauging the disparity
between the real value 𝑦𝑖 and its corresponding prediction𝑦̂. The component Ω(𝑓𝑖 )
denotes the regularization term associated with the j-th tree, a crucial element
distinguishing XGBoost from traditional gradient-boosting methods. This regularization
term is further detailed in equation 8:
1 𝑇
Ω(𝑓) = 𝛾𝑇 + 𝜆 ∑ 𝜔𝑗2 (8)
2 𝑗=1
In the above expression, 𝑇 stands for the number of leaves present in the tree.
The regularization parameters 𝛾 and 𝜆 are employed to control the complexity of the
model. Specifically, 𝜔𝑖 represents the score designated to the j-th leaf of the tree.
Meanwhile, (T. Chen, and C. Guestrin, 2016) work on XGBoost in 2016 thoroughly
illustrates the underlying mechanisms and the inherent advantages of this gradient-
boosting system.
The mathematical approach section presents three related concepts and a brief
state of the art: the Time series in 3.3.1, the Classification task in 3.3.2, the Learning
phase in 3.3.3, and the main performance analysis as (F1), accuracy, precision, and
recall.
“With emerging concepts like cloud computing and big data and their vast
applications in recent years, research has been increased on unsupervised
solutions like clustering algorithms to extract knowledge from this avalanche
of data. Clustering time-series data has been used in diverse scientific areas
to discover patterns that empower data analysts to extract valuable
63
information from complex and massive datasets. The time-series data is one
of the popular data types in clustering problems and is broadly used from gene
expression data in biology to stock market analysis in finance”.
Hence, the time series approach is present in signal processing, data mining,
pattern recognition, control engineering, ML clustering, classification, anomaly
detection, forecasting, and other relevant applications. However, they are not present
only in engineering. Other general areas, such as economics, biology, mathematics,
physics, medicine, and others, can employ them to present data behavior in function
of time.
An example of a time series applied to the AIRBUS benchmark simulation is
presented in figure 14, in which the time measured starts from zero to ten seconds,
measures the airplane’s rod sensor deflection in rad, and presents the command of
the FCS:
TP
𝑃𝑟𝑒𝑐𝑖𝑠𝑖𝑜𝑛 = (13)
(𝑇𝑃 + 𝐹𝑃)
TP
𝑅𝑒𝑐𝑎𝑙𝑙 = (14)
(𝑇𝑃 + 𝐹𝑁)
66
Finally, the F1 Score is the harmonic mean of precision and recall and provides
a balance between them. It's particularly useful when the class distribution is
imbalanced. The F1 Score is the best metric to use if you have an uneven class
distribution and if the cost of false positives and false negatives are roughly equivalent.
Then, it provides a balance between precision and recall, especially important for
imbalanced datasets as stated by (D. Lewis, 1991). The equation 15 shows the F1
Score:
𝑃𝑟𝑒𝑐𝑖𝑠𝑖𝑜𝑛 × 𝑅𝑒𝑐𝑎𝑙𝑙
𝐹1 𝑆𝑐𝑜𝑟𝑒 = 2 × (15)
𝑃𝑟𝑒𝑐𝑖𝑠𝑖𝑜𝑛 + 𝑅𝑒𝑐𝑎𝑙𝑙
This section explores the Continuous Numeric Data and Categorical Data
approaches for Feature Engineering (FEn) and their state-of-the-art in 3.4.1.
The FEn can be defined as a set of data filtering procedures that integrate expert
knowledge from the theme domain to transform, integrate, and adjust features to
increase ML predictor algorithms correlation, as defined by (A. Gal-Tzur, S. Bekhor,
and Y. Barsky, 2022). Moreover, according to (Z. Qadir, S.I. Khan, and E. Khalaji et
al., 2021), the FEn eliminates low-quality data. Furthermore, it selects the most crucial
features to reduce computational costs and minimize error.
According to (R. Yao, N. Wang, Z. Liu, et al., 2021), FEn is one of the most
appropriate data processing steps that extracts the main features from datasets.
Meanwhile, (N. Mapes, C. Rodriguez, and P. Chowriappa et al., 2019) extracted
features from a comprehensive dataset with thirty-nine-dimensional features, reducing
the number of features by combining them into new features.
In (F. Chiarello, P. Belingheri, and G. Fantoni, 2021) study, the FEn was
classified as a data analysis process that depends on the context to identify meaningful
feature representations to increase the accuracy of ML systems. Moreover, a
knowledge-driven approach, FEn, is presented by (Z.H. Janjua, D. Kerins, and B.
O’Flynn et al., 2022) as sensitive task specialists perform before applying ML
techniques. In addition, this process was employed in medical applications for blood
67
This section presents the main applications in which Soft Sensors were
employed in SLR papers; 54 of 59 papers presented contributions for answering this
question. Such a section is subdivided into Industrial Applications, Aeronautics,
Quimiometrics, Cloud Computing solutions, health and care solutions, building and
household applications, and general applications.
There are applications related to car body parts drawing strokes, cited by (R.
Meyes, J. Donauer, and A. Schmeing, et al., 2019), using the SS approach with strain
gauge and laser sensors to digitalize the metal sheet retraction points. The authors
used an RNN architecture to draw the 3D model based on time series data.
In property control of hot forming-based grain size for steel, the authors (M.
Bambach, M. Imram, and I. Sizova et al., 2021) showed an SS implementation as a
surrogate model. That improved the accuracy of a cost function to measure the
distance of predicted and measured domain boundaries. Another industrial application
is cited by (M. Tabba, A. Brahmi, and B. Chouri et al., 2021) for a PLC controlling
system for industrial digitalization over level detection in a complex scenario using a
set of connected sensors.
In manual screwing manufacturing, Soft Sensors can detect activities
characterized by body posture and arm and hand activity (L. Günther, S. Kärcher, and
68
T. Bauernhansl, 2019). Although, for (N. Tvenge, O. Ogorodnyk, and N. Østbø et al.,
2020), SS is a crucial component for real-world digitalization, composing the Digital
Twin scenario, being more than a simple model, it enables decision-making around
further actions of a modeled system.
Another SS Industrial machinery application is shown by (B. Maschler, S.
Ganssloser, A. Hablizel, et al., 2021) by measuring the cylinder pressure to calculate
other relevant combustion parameters in large engines to reach additional
maintenance requirements. Therefore, they indicate future works: “To facilitate the
mentioned optimization procedures in smaller engines, too, the use of virtual cylinder
pressure sensors is a promising option.”
For (W. Lee, G. Mendis, and J. Sutherland, 2019), an SVM can be trained using
sensor signals as input and outputting the tool wear once the real-time multi-sensor
dataset is employed to predict the tool wear. Furthermore, according to (S. He, H. Shin,
S. Xu, et al., 2020), these applications can utilize low-cost sensors, enabling scalability
features.
Material engineering can filter noisy SS data models on materials structure
Variational autoencoders for Long short-term memory (VAE-LSTM) approach to
optimization, according to (A. Lew and M. Buehler, 2021). Although an
electromechanical fault detection benchmark was presented by (T. Grüner, F. Böllhoff,
and R. Meisetschläger et al., 2020), using a data set to generate the model and predict
the faults based on indirect sensors as model input.
The authors (M. Barton and B. Lennox, 2022) articulate ensemble methods to
apply SS in industrial scenarios for predictive performance. In the automotive industry,
(A. Theissler, J. Pérez-Velázquez, M. Kettelgerdes, et al., 2021) presented fault
detection and predictive maintenance for autonomous vehicles.
The SS applied to the stamping press is explained by (D. Coelho, D. Costa, E.
Rocha, et al., 2022), using Long Short-Term Memory (LSTM) networks to predict
failure in metal stamping processes. Furthermore, the study by (S. Shafiq, E.
Szczerbicki, E. Sanin, et al., 2019) presented SS for data acquisition and visualization
supporting I4.0 in different machining scenarios.
Self-healing robots are the focus of (E. Roels, S. Terryn, J. Brancart, et al.,
2022) study, applying SS to recover the system in case of severe damage. They use
self-healing materials such as polymer networks, novel elastomeric, and SS-
69
The Aeronautics industry can be improved due to the use of “sensors and
software to monitor multiple aspects of aerospace vehicles” (K. Ranasinghe, R.
Sabatini, and A. Gardi et al., 2021). The authors have shown a National Aeronautics
and Space Administration (NASA) application for SS over Vehicle Health Monitoring
(VHM), which provided the vehicle’s failure prognostic or diagnostic management for
predictive maintenance.
Another aerospace application for SS is presented by (V. Henrique, R. Massao,
and G. Reynoso-Meza, 2021) for Oscillatory Failure Case detection using this
technology on the EFCS. The inputs are the command control current and the
feedback signal sensor from the built-in rod sensor. They developed a Simulink design
for failure detection.
The technical note presented by IFAC to solve the AIRBUS benchmark (J.
Engelbrecht and P. Goupil, 2020) details the FCS of a commercial aircraft and its
sensors, systems, power sources, wiring, and many other movable parts.
for nonlinear soft sensing modeling has been employed in the recent years, particularly
regarding the prediction accuracy and saving of computational costs.”
As described by (G. van Kollenburg, J. van Es, J. Gerretzen, et al., 2020), the
SS application uses historical process data to improve the chemical process conditions
and control laws using the PLS-Path Modelling.
A work by (H. Paggi, J. Soriano, and V. Rampérez et al., 2013) defines Wireless
System Networks (WSN) as interconnected low-cost sensors. Such WSNs are
implemented in health monitoring, military target tracking, animal monitoring, smart
homes, environmental control systems, and other applications, composing a so-called
“Intelligent Space” (IS).
Image processing solutions are present in almost every Cloud Computing (CC)
platform. Then, a Deep Learning approach can solve remote-sensing applications (L.
Ma, Y. Liu, X. Zhang, et al., 2019). Furthermore, other applications are presented by
this author, for example, fusion, segmentation, change detection, and registration.
An Industry 4.0 case study employing soft sensors in a waste-to-energy plant
benchmark is presented by (J. Kabugo, S. Jämsä-Jounela, and R. Schiemann et al.,
2020). CPS uses an IIoT-based data platform to detect temperature faults and correct
the PID gains instead of only displaying alarms. Another I4.0 application using real-
time SS was conducted by (D. Ntamo, E. Lopez-Montero, and J. Mack et al., 2022) to
monitor the end-point moisture using a data-driven approach.
Another CC low-cost solution for occupancy estimation was developed by (K.
Rastogi and D. Lohani, 2019) to enable frequent sampling and reduce the need for
human involvement. In this sense, (B. Schumucker, F. Trautwein, and R. Hartl et al.,
2022) describe a cloud-based environment to measure and simulate curring forces in
workpiece quality, monitoring the tool’s current level in real-time, enabling intelligent
decision-making.
In the study case presented by (J. Schimitt, J. Bönig, and T. Borggräfe et al.,
2020), a SS is employed for quality inspection using ML and Edge Cloud Computing
(ECC) to identify defects in the automobile and electronics industry.
72
An application for Soft Sensors working with ML and Cuckoo Search and
Particle Swarm Optimization (CS-PSO) to quantify performance metrics for athletes
based on data was presented by (M. Ishi, J. Patil, and V. Patil, 2022).
74
This section presents the relationship between Soft Sensors and Industry 4.0 in
the 3.6.1 sub-section and Smart Factories in 3.6.2, present in 27 of the 59 SLR-
selected papers.
This section presents the leading Feature Engineering (FEn) and ML techniques
employed in Soft Sensors implementations in 49 of 59 researched papers. The sub-
section 3.7.1 presents the FEn employment found in 20 papers. Meanwhile, 3.7.2
handles the ML approaches cited by 46 papers.
2019) for FEn, in which “CNN is characterized by scale invariance and can capture
local dependencies in data.”
Some main benefits of employing FEn to SS ML methods were declared by (M.
Ishi, J. Patil, and V. Patil, 2022). Among them are the dimensional reduction of large
datasets, the selection of significant features to enhance classification performance,
and the reduction of computational costs. In addition, they cited that “Metaheuristic
algorithms are recognized as a viable approach for addressing feature optimization
problems.”
The FEn can reduce the dimensionality of an SS model dataset, according to
(O. Fisher, N. Watson, and J. Escrig et al., 2020). For example, implementing FEn and
a PCA in their study reduced the dataset's dimensionality from 26 features to 9.
Another detailed FEn study for SS dataset feature reduction was carried out by (V.
Henrique, R. Massao, and G. Reynoso-Meza, 2021), in which ten features were
created for the input signal and processed by:
Machine (ELM) and has significantly improved its performance. In the (M. Maggipinto,
E. Pesavento, and F. Altinier et al., 2019), FEn is used to extract features from complex
data to feed ML algorithms. In this case, more than fifty features were extracted.
Thus, the FEn data preprocessing using probabilistic data cleaning and PCA for
SS implementation to feed an RF algorithm is displayed in (I. Mendia, S. Gil-López,
and I. Landa-Torres, et al., 2022). However, it is stated by (L. Petruschke, J. Walther,
and M. Burkhardt et al., 2021) that the DL approach can overcome FEn techniques
due to them inheriting the classification tasks.
A PLS model implemented in SS development by (D. Aguado, G. Noriega-
Hevia, and J. Ferrer et al., 2022) showed the importance of FEn in feature extraction
for predicting ammonia concentration using different pH features combined. In
addition, the Out-Of-Bag concept is explained by (P. Nkulikiyinka, Y. Yan, and F. Güleç
et al., 2020) to refer to some samples not used for fitting one single DT in the RF for
SS implementation.
The FEn requirements for extracting new features after the training process are
evidenced by (D. Coelho, D. Costa, and E. Rocha et al., 2022) and (R. Forghani, P.
Savadjiev, and A. Chatterjee et al., 2019). Once the FEn's main objective is to improve
the model with optimal parameters only, the (A. Gejji, S. Shukla, and S. Pimparkar et
al., 2020) work explored the use of FEn for SS dataset dimensionality reduction using
ML methods as an example: ANN, DLNN, DT, SVM, ensemble method and logical
regression.
Neighbors (KNN), Decision Tree (DT), Logistic Regression (LR), Deep Neural Network
(DNN), and Convolutional Neural Network (CNN).”
Some data-driven statistical methods were proposed by (M. Zaghloul and G.
Achari, 2022) to identify faults based on the SS approach employing the PCA and ANN
to control the process. Moreover, the PCA technique was embedded in SS in a DL
deep layer of data reduction, providing gains in predictive performance, according to
(M. Abdar, F. Pourpanah, and S. Hussain et al., 2022). Another PCA implementation
for SS is shown in (A. Hicks, M. Johnston, and M. Mowbray et al., 2021) to reduce the
dimensionality of time-series datasets by identifying correlations between output and
input layer variables.
In (M. Bambach, M. Imram, and I. Sizova et al., 2021) paper, an ANN was used
as an SS to predict the final boundary deformation sequence in a steel forming
operation. Although the cloud platforms play a crucial role in ML solutions for SS, as
shown by (V. Kocaman and D. Talby, 2022), the Google Cloud Platform and AWS
present contributions for extracting relevant medical information from biosensor data.
The study employed an ANN for an SS data-driven approach in data prediction (B.
Negash, L. Tufa, and R. Marappagounder et al., 2016).
An approach for Structure Health Monitoring was carried out by (S. Baduge, S.
Thilakarathna, and J. Perera et al., 2022), using different sensors (e.g., acoustic
sensors, electromagnetic devices, and Accelerometers) for an SS implementation.
Such features were developed using the following ML techniques: “ANN, DL, Support
Vector Machine (SVM), Principal Component Analysis (PCA), k-Nearest Neighbor
(KNN), and low-rank matrix decomposition.” In addition, many other techniques can be
employed for fault detection using SS, according to (T. Grüner, F. Böllhoff, and R.
Meisetschläger et al., 2020):
“The following ML methods are utilized for the classification of the normal and
fault states in the underlying data set of measurements of the motor current:
Traditional methods: KNN, SVM; Ensemble methods: Random forests;
extreme gradient boosting machines (XGBoost); Deep Learning: Two fully-
connected feed-forward ANNs with three (ANN-3) and 20 (ANN-20) hidden
layers. For the implementation, the Python libraries scikit-learn (KNN, SVM,
Random forests), XGBoost2, and Tensorflow3 (ANNs) are used.”
"In particular, artificial neural networks and their subclass of convolutional neural
networks are frequently applied.” Moreover, the implementation of PLS is considered
by (G. van Kollenburg, R. Bouman, and T. Offermans et al., 2021) as an easy-to-
explain technique for SS models in an industry environment.
An SS onboard system was presented by (K. Ranasinghe, R. Sabatini, and A.
Gardi et al., 2021), in which ML algorithms were employed for fault diagnostic. The
main ML methods used in their study are: “SVM, GPR, neural networks, Markov Chain,
fuzzy logic, and Monte Carlo.” The data-driven approach for SS (cited in (O. Fisher, N.
Watson, and J. Escrig et al., 2020) implemented an ANN to model prediction without
initial system knowledge.
The paper (T. Krivec, J. Kocijan, and M. Perne et al., 2021) presented an
alternative for statistical ML methods, the non-parametric and probabilistic Gaussian
Process (GP) model, for an SS data-driven forecast implementation. Meanwhile, the
paper by (V. Henrique, R. Massao, and G. Reynoso-Meza, 2021) employed a DT
algorithm for the SS model in the fault detection application model.
Three steps are presented by (B. Maschler, S. Ganssloser, and A. Hablizel et
al., 2021) for implementing an SS using the MLP approach. These steps consist of
data preprocessing, choosing the ANN type, and structuring it to estimate the engine
parameters using the data acquired from the sensor. A set of ML techniques was cited
by (W. Lee, G. Mendis, and J. Sutherland, 2019) for implementing SS in the data-
driven model, for example, SVM, ANN, FS, and RF. Nevertheless, their main results
were achieved using SVM to maximize the kernel function.
Two categories for multi-sensor fusion ML algorithm were described by (S. He,
H. Shin, and S. Xu et al., 2020): State Vector Fusion and Information Vector Fusion.
The first focuses on local estimations over the sensor network, while the second “refers
to direct or indirect exchanges of local measurements among sensor nodes.”
Moreover, such contribution is relevant for SS network implementation, which can be
employed for large-scale, low-cost solutions.
The study by (A. Guzman-Urbina, K. Ouchi, and H. Ohno et al., 2022) showed
that the FIEMA approach for data-driven SS presented higher accuracy over SVM and
ANN ML methods. According to (P. Zhu, H. Peng, and A. Rwei, 2022): “Current
machine learning and big-data analytical techniques rely on high-quality data for
81
algorithm training as well as data analysis, highlighting the importance of signal fidelity
for wearable sensors.”
The Microsoft Azure IoT platform was presented by (J. Kabugo, S. Jämsä-
Jounela, and R. Schiemann et al., 2020) for data-driven SS development using the
available ML techniques Azure. In addition, another SS implementation using ML is
shown (D. Ntamo, E. Lopez-Montero, and J. Mack et al., 2022). These algorithms were
employed to predict and estimate hard-to-measure variables.
The SS design is cited as an inferential sensor by (M. Barton and B. Lennox,
2022), using the decision tree ML algorithm to fit and train the model for low-bias in a
high-variance dataset. Meanwhile, in the paper presented by (E. Jalee and K. Aparna,
2016) found in the literature review on the employment of ANFIS, SVM, PLS, Kalman
Filters, ANN, and Fuzzy Logic with GA for SS implementations, the authors stated:
This section explores the found methods for Feature Engineering in 3.8.1 and
their differences compared to Hyperparameter Tuning (HT) in 3.8.2. Those
contributions are present in 23 of the 59 researched papers.
The study (R. Meyes, J. Donauer, and A. Schmeing et al., 2019) performed the
FEn classification task of Soft Sensors signals, using the feature vector raw signal and
its information in the frequency domain to make predictions over the time-series data.
82
In addition, for a CNN implementation, the hybrid bidirectional LSTM was employed by
(V. Kocaman and D. Talby, 2022) to eliminate further FEn procedures.
According to (L. Günther, S. Kärcher, and T. Bauernhansl, 2019), Feature
Engineering procedures are considered one of the most critical procedures in ML
projects once they avoid the dependency on human experience in feature selection or
extraction in SS development. Besides this application, (S. Baduge, S. Thilakarathna,
and J. Perera et al., 2022) demonstrate that FEn improves the SS versatility and
accuracy in ML model development.
In data-driven SS applications, (D. Aguado, G. Noriega-Hevia, and J. Ferrer et
al., 2022) proposed the FEn to extract features from pH and their direct interactions
with input variables. They affirm that “Feature extraction based on the technical
knowledge of the process was key to make the development of a reliable data-driven
PLS soft-sensor possible.” Another data-driven system studied by (O. Fisher, N.
Watson, and J. Escrig et al., 2020) presents three observations of FEn processes
employed in SS development, which are ensuring the model boundaries testing, fitting,
and predicting capacities, accomplishing any temporal variation in the system in
collected data, and distributing data between the defined boundaries.
According to the author, these procedures must be followed because it is crucial
“to ensure the model is capable of fitting data and making predictions throughout the
system.” Nevertheless, in (M. Ishi, J. Patil, and V. Patil, 2022) research, many MOO
algorithms were employed to reduce the number of hyperparameters in ML techniques
for an SS implementation.
An FCS application for SS enhancement using FEn was conducted by (V.
Henrique, R. Massao, and G. Reynoso-Meza, 2021) to lead the final prediction using
input features, following these three steps: data acquisition, model training, and model
validation. Finally, a FEn application for SS with embedded ML methods was presented
by (L. Ma, Y. Liu, and X. Zhang et al., 2019) in CNN and SVM classifiers for the remote-
sensing classifier.
Based on (M. Shapi, N. Ramil, and L. Awalin, 2021), the FEn approach was by
inputting different sets of features in the ML technique to enhance the energy
consumption for a data-driven predictive SS model. According to (T. Grüner, F.
Böllhoff, and R. Meisetschläger et al., 2020), the FEn could transform a time series
83
The SLR conduction presented 59 relevant papers to this research field and
answered the guideline questions proposed in table 1. The process started by defining
the ten procedures explored in section 2.2, with relevant study areas being the first.
Then, the research guideline questions were proposed, and the keywords were defined
in step one, then mixed to generate research strings to browse over the Science Direct
engine.
The first search resulted in 2,153 papers, and by applying the Inclusion and
Exclusion Criteria proposed in step seven in table 4, this number was reduced to 109
papers. Hence, the classification criteria browsed their abstracts following the table’s
85
seven criteria, and only 59 articles remained in the leading research. After this research
over every paper, each one of the questions was answered and contributed to the
global understanding of SS, ML techniques, Industry 4.0, and Feature Engineering.
Figure 17 presents the general frequency percentage of each question:
Q.04
15%
Q.01
35%
Q.03
32%
Q.02
18%
These papers were read and submitted to table 1 questions, their main
contribution, and if they presented any relevant Soft Sensor definition that could
improve the State of the Art. Notably, only 28.81% of papers presented some SS
definition, and less than 40% explored feature engineering at the required level.
Nevertheless, the contribution of each paper achieves the letter “C” specific
objective by presenting relevant information about research areas and presenting each
paper’s primary information to substantiate the SLR.
86
4 BENCHMARK DISCUSSION
According to the technical note of the IFAC benchmark, the flight control
systems (Flight Control System - FCS) are the most important in operation, being
responsible for the control of altitude, trajectory, and speed. In addition, such systems
are among the numerous controllers of the pilot” panel and the actuators, comprising
all the actuator and sensor systems in the avionics system. The great advantage of
this embedded technology is that it allows the application of advanced electronic
control loops on surfaces and must be available for use under any circumstances.
Because of this, fault detection is a critical aspect, given the impact of these
disturbances on aircraft structural modeling. Thus, the mobile mechanism responsible
for detecting oscillatory faults (OFC) can be seen in the illustration in figure 18:
The figure above shows that the actuator's surface receives analog signals from
the FCC (Flight Control Computer) increased in K units by a proportional block, and
the rod sensor (Rod Sensor) feeds back to the system, closing the loop of control.
According to the authors, such a system must be able to detect low-frequency faults
(below 20Hz). However, other frequencies can be studied in some instances, so only
disturbances located on moving surfaces should be considered.
87
Due to these oscillations' nature, the researchers point to the existence of OFCs
of a “liquid” character that occur when a sinusoidal signal is added to the servo-
controlled signal or “solid” when the sinusoidal signal overlaps the nominal signal.
Thus, it is possible to determine the OFC detection methodology for each described
case. Furthermore, the design requirements of this benchmark are defined by the
authors and summarized in table 7:
Fault signals must be detected with at least three oscillation periods regardless
3
of the OFC frequency.
4 Liquid or solid flaws must be detected.
5 Control and measurement signals from the sensor must be detected.
The fault detection system must not produce false alarms under the following
circumstances:
6 a) Normal flight with or without turbulence of any level;
b) Control input in step, sinusoidal or chirp-type signals (which increase in
frequency as a function of time).
Source: The Author, 2022.
According to the benchmark report, this module will receive the settings
informed by the user about the flight path angle and define, through a switch, the
aircraft path control mode depending on the mode chosen by the user, among them:
“FPA_CONTROL”, ”NZ_STEP”, “NZ_SIN” and “NZ_CHIRP”. Then, respectively being
converted to numbers depending on the selection mode in the Aircraft class, this
diagram is presented in figure 20:
89
In this way, the control signal is selected between control by the Function Point
Analysis – FPA, control by unit step signals, control with proportional sine signal, or
control by the signal of the Chirp type whose frequency increases with time. With this,
the signals are sent to the load factor control block.
In the last sub-chapter, the term “load factor” was presented, which can be
understood as the total ratio between the force imposed by the air resistance and the
force proportional to the weight of the aircraft. Therefore, this variable is proportional
to the speed and the flight angle, according to the private pilot ground school (2006).
Therefore, this variable impacts the flight dynamics and can be changed through the
deflection of the airplane's actuators.
On the other hand, this module commands the control surface deflection by
measuring the load factor (first feedback) and gyroscope measurements to measure
the angulation rate of each axis (second feedback) to control the flight angle. With that
defined, figure 21 shows the load control block:
90
The diagram in figure 21 shows that the controller receives a command signal
from the previous block in a saturator. This signal is distributed to a proportional gain
block and another branch to a proportional-integral block, and both are added to two
feedbacks (Figure 18). Such signals enter the load control module's transfer function
to command the servo simulator block to the wobble detection surface. It is worth
noting that the “Nz_cmd” block stores the control signals sent by the angle control
module after saturation, that is, the signal that arrives in the proportional and
proportional-integral gain systems.
91
These block models the behavior of the control surface servo system, employing
an actuator and a rod-type sensor to measure the deflection. The control command is
received by the “delta_des” variable and sent to the workspace by the 2D array variable
defined as “dx_comm”. The block output is the estimated deflection on the control
surface measured by the rod-type sensor. Finally, the “Real Servo” module is fed back
with the “delta” signal from the exact measurement of the rod sensor deflection without
considering the sensor noise. To illustrate such situations, figure 22 below shows the
Simulink diagram connecting all the variables and blocks described:
It is shown in figure 22 that the “servo” block sends an array to the workspace
containing the control command, the deflection measured by the sensor, and the
deflection measurement without the sensor noise simulation. This module is detailed
in the diagram in figure 23:
92
The diagram in figure 23 presents that the control signal from the load factor
controller passes through a saturator and a change rate limiter, keeping it within
acceptable limits for real situations. After these steps, the rod sensor angle signal is
converted to position and inserted into input 1 of the “plant 1” block and, in the second
input, the “delta” output feedback. The diagram of the “plant 1” module, responsible for
simulating the dynamics of the sensor positioning system as a function of the control
loop performance and the noise-free feedback “p_des” can be seen in figure 24:
Figure 24 shows the blocks that emulated the flight system identification under
adverse conditions in the Simulink model. The “F_aero” signal is convoluted with the
last corrected sensor position signal and transformed into a binary signal, which can
be 1 or -1 depending on the sensor reading. According to the benchmark organizers,
the oscillatory fault can be produced by the output of the servo command and feeding
93
the closed positioning control loop. Such modeling is presented in Figure 18. The
authors obtained the simulated position of the rod sensor “Rod_Sensor” and “rod_pos”
are sensor positions. However, one is given without noise (ideal situation) to feed back
the main simulation loop and the other with noise, simulating a natural process.
Meanwhile, the module that simulates the turbulence of the system is presented
in the Simulink diagram in figure 25:
In this step, a selector receives the turbulence mode selected by the user, which
can be without turbulence (number zero), with slight turbulence (number one),
moderate turbulence (number two), or severe turbulence (number three). These
variables contain the “.mat” extension because they are stored in MatLab® files
containing real turbulence data series classified in the groups presented above. Then,
the signals are divided into three in demultiplexer, and two are sent to the diagram in
figure 18. Meanwhile, the other signals have amplitudes multiplied by a gain “1/V_trim”,
a turbulence parameter, and a deflection angle parameter. Then, the received is
converted from radians to degrees; in this way, the signal is again multiplexed and sent
to a closed loop with other gains and an integrator for reconditioning the turbulence
signal through the Von Karman turbulence model, natively available by MatLab®. With
the turbulence generation block explained, the final part of the diagram can be
94
presented, in which the data is filtered and sent to the workspace, as shown in Figure
26:
At the end of the process, the data is received by a demultiplexer block, which
transforms a batch of data into parallel signals that, in turn, are sent to the process
loop feedback, filters, or directly sent to the work area, thus becoming available for use
or query.
95
Based on RSL, this work resulted in a Soft Sensor implementation on the fault
detection benchmark presented. Hence, this critical analysis will be divided into five
parts. First, software development for Soft Sensor in 5.2, MATLAB® integration with
Python language in 5.3, the results found by each of the applied methods (SVM, MLP,
and DT) in the 5.4 section, and the demand for feature engineering into this benchmark
solution in 5.5.
The Soft Sensor software development started in the final control and
automation engineering project, presented by (M. Feliciano and G. Reynoso-Meza,
2020), with a different purpose: to be a commercial software that provides real-time
graphs and detailed datasheets with acquired data. Due to this reason, it presents a
Graphical User Interface (GUI), which can be seen in figure 27:
Figure 27 GUI allows the user to configure the parameters for simulation and
set the folder where the models of the Simulink™ or data acquisition file are located.
Then, presents to the user the steps performed, plotting the graphs in real-time or
simulating the real-time process in the case of emulation. The benchmark button, in
96
addition to asking the user the directory where the data is, also enables the 'start'
button and a combo box to choose the case study.
In the next step, the user defines the OFC source, divided into four groups:
'current' or 'cs_current' to measure sensor amplitude and variance (bias) in mA
(milliamps) or 'sensor' or 'cs_sensor' to measure these quantities in millimeters. With
the source defined, the types of OFCs must be provided, among them: 'none'
(parameterizes some variables to null), 'liquid' (when a sinusoidal signal is added to
the servo-controlled signal) or 'solid' (when a sinusoid is superimposed the nominal
signal of the disturbance). The user must then choose the type of turbulence that can
be classified between: 'none', which does not generate turbulence; 'light' which is slight
turbulence; 'moderate' which causes moderate turbulence in the simulation; or 'severe'
which will generate turbulence severe in the emulated system.
After defining the turbulence, the user chooses the type of control to be applied
to the plant, the first of which is the FPA control mode 'FPA_Control', which works with
an analysis function of past points to correct the current output in the closed loop. The
'NZ_STEP' type works with unit step control, 'NZ_Chirp' uses a sine signal with variable
frequency, and 'NZ_Sine,' whose control signal is given through a sinusoid. It is worth
mentioning that all control signals act together with the blocks shown in figure 20.
With this configured, the user must set the wave amplitude, the sensor error or
variance, and the OFC frequency that will be worked on in the case study (in actual
cases, this parameter will be provided by itself). Hence, the user proceeds to the final
parameterization phase, where the benchmark dataset and the training method are
defined. Such methods are Support Vector Machine (SVM), Decision Tree (DT), or
Multi-Layer Perceptron (representative of the neural network family). Before starting,
the user must provide the total simulation time or, if it is data acquisition, the software
interprets it as infinite. With all these variables defined, table 8 can be presented, which
explains each parameter:
97
sensor
Rod sensor in millimeters.
cs_sensor
OFC Source
current
Rod sensor with mA measure.
cs_current
none No OFC.
OFC with sine signal added to the servo-
OFC Type liquid
controlled signal
The sinusoidal signal overrides the
solid
nominal signal of the disturbance.
none No turbulence
Once the parameters are converted to start the simulation in the specific case
study of this project, Python starts the MatLab® API, and it is possible to execute
commands through the code inside a virtual workspace, declaring all the variables
used in the simulation within Simulink. MathWorks developed the used library, and it
can be installed by downloading their engine API, not via “pip install.”
The average computational cost to open the MATLAB® API on the computer
proposed in subchapter 1.4.2 is approximately 5 seconds, and that of emulating the
case study plant in Simulink is around 3 seconds for processes of 5 to 60 seconds in
duration in simulation.
The data loading function sends commands through the API and creates a
virtual environment that emulates the avionics system on Simulink™. Therefore, it is
necessary to employ techniques to emulate the obtained datasets being used in the
case of real-time data acquisition, in which case the graph construction functions
remain. However, the inputs will be received directly from MATLAB®. The complete
running process of this software is presented in the diagram in figure 28:
Meanwhile, the communication with Python occurs throughout API, and the data flows
from simulation directly to the ML methods implemented.
In this case study, SVM, DT, and MLP methods were implemented using the
sklearn library. However, many other ML or optimization methods can be employed.
Python language was chosen due to its versatility to be employed on Cloud Computing
applications and aeronautics solutions, as shown in 3.5.4 and 3.5.2, related to SS
implementations. Hence, the processed data from the process simulation is stored in
a Maria DB database instance, which runs locally. However, it can run in the cloud and
be processed remotely in real-time.
The developed interface presents user graphs, as shown in figure 28, using
simulation data after FEn and ML application over a time series model to identify and
display the possible Oscillatory Failure Cases. These are presented as red crosses in
figure 29:
In this phase, the results produced by the software will be presented. These
results include a report on the steps performed by the software, a summary of general
data, a report highlighting only the failures that have occurred, and an exported chart
in image format. The experiment will be conducted by analyzing three variations of
parameters for each of the classifiers mentioned above, as shown in Table 9:
From table 9, one can infer that the oscillation source is kept constant as it would
not be meaningful to change the source while observing how the methods evaluate
failures and changing the control method would not be viable either. However, other
variables compose the different scenarios. For example, the OFC type influences the
process by introducing sinusoidal signals into the response simulation, causing more
unstable scenarios. Turbulence also increases the possibility of failures.
The parameter of oscillation amplitude affects the accuracy of the Machine
Learning (ML) method in measuring the signals. The higher the amplitude, the better
the response from the method should be. As the scenario becomes more unstable, the
oscillation amplitude increases, making it more imprecise. This scenario challenges
the methods by evaluating them in emulated scenarios in MATLAB® Simulink.
Lastly, the oscillation frequency decreases as the scenario worsens. Therefore,
low amplitudes are more critical when it comes to measuring oscillatory failures, as
explained by the benchmark proposers.
101
The main results obtained by the software were developed in response to the
objective of generating this project to present the best among the three implemented
methods (SVM, DT, and MLP). Therefore, the control surface deflection amplitude and
sensor amplitude will be evaluated in degrees per each scenario and ML method as
the identified failure percentage, without considering error type 1 and error type 2,
which will be considered only in the last column.
For that, four scenarios were emulated in each of the methods implemented in
the software. The main variables as the maximum amplitude of the control command
and the sensor deflection response, the number of identified failures, and the
performance in the confusion matrix will be aspects discussed in the analysis of table
10 to determine the best identification method given by the confusion matrix trace:
As evidenced in table 10, the three methods were tested in the four scenarios
and observed in the confusion matrices condensed in the table above. Furthermore,
the amplitude columns of the control command and sensor reading range are
considered to ensure that the scenarios are not biased. Moreover, the fifth column
contains the percentage of failures identified in the 400 data emulated in each test,
which is 10 seconds of simulation. Finally, the performance according to the confusion
matrix is presented in the last column based on the matrix trace. The construction of
this column is given through the data considered ideal, obtained by the Decision Tree
classifier implemented in MatLab® with the help of specific toolboxes.
Regarding the methods discussed in table 13, starting with MPL, it is clear that this
method was the most assertive because 100% of the analyzed data were identified as
failures. However, this is due to the bias of considering all data failures. This fact can
cause the user to be unreliable in the software because, as 100% of the data are given
as failures, there is no way to identify which ones are and which are not in real
situations, with this method becoming unfeasible for this project.
On the other hand, the SVM method considers most data as non-faults, even
when they are present. Unfortunately, fault tolerance presents one of the greatest
dangers in the aerospace engineering scenarios where this case study takes place.
Thus, the number of false negatives is alarming and causes the impossibility of this
method for application in the benchmark. Furthermore, the collision with the error in a
classification method cannot be considered.
However, the Decision Tree method, implemented in Python, proved reliable
since it identified OFCs consistent with the reference used in practically all scenarios.
It presented a few false negatives, representing the failures where they were. Another
essential characteristic is its low accuracy (Table 13) for identifying false positives,
pointing out flaws in data that did not fail for OFC classification.
Thus, with the results arranged for analysis, considering the weight of the
average of false positives (error type I) obtained at 20%, this data is relevant because
it presents the number of failures identified but did not exist. While the weight of false
negatives (error type II) was assigned to 50%, as it is more critical to identify faults than
not, the higher weight is considered in this scenario. The remaining 30% refers to the
average precision of the methods for each scenario, so the weighting field is obtained
103
by adding the three weights weighted by the percentages described above. Therefore,
table 10 shows each one of the methods, summarizing what is described in table 11:
Support Vector
2° 5.00% 52.44% 46.31%
Machine
Multi-Layer
3° 43.94% 0.00% 57.81%
Perceptron
Source: M. Feliciano, G. Reynoso-Meza, 2020.
As can be seen from table 11, due to the SVM having identified few failures in
every scenario, its number of false positives is minimal, and the opposite is true for the
MLP method, which identified all data as failures, not scoring false negatives. However,
the DT method presented many false positives and a good slice of false negatives.
Finally, as a result of this analysis, it can be said that the only machine learning method
among the three analyzed (DT, SVM, and MLP) that is reliable for application to the
case study proposed by AirBus to IFAC is the DT.
With the exception that it is possible to work on feature engineering (According
to Annex C) of the learning dataset so that the classification presents a sharper
performance, this feature may result from future work. Although the software in the
MVP version will still have the other classifiers in this way due to the possibility of
acquiring data in real-time, the methods that did not present satisfactory performance
may be re-evaluated for other processes.
After evaluating the results presented in table 14 and having the (V. Ribeiro, R.
Kagami, and G. Reynoso-Meza, 2020) conclusion about DT results over this
benchmark, feature engineering employment became a growing demand. Hence, they
provided a block diagram for Simulink™ implementation in the discrete domain, using
the Z transformation to improve the feature selection in this study case, as shown in
figure 30:
104
The diagram depicts a signal processing system with multiple parallel branches
for data analysis. The primary input undergoes a "Moving Average" operation to
smooth out data. It then passes through various delay elements, which vary in the
range presented in equation 16:
[ 𝑧 −20 , 𝑧 −1 ] (16)
Post-delay, data streams are combined with the primary data and processed
using "Moving RMS" to measure the data's magnitude. Outputs a to j result from these
processes, each corresponding to specific delay and combination operations.
Additionally, there are inputs c and d, with c subjected to a "Moving Average" and d to
a "Moving Standard Deviation", assessing the data's average and variability,
respectively. The system's design indicates an emphasis on trend identification and
variability assessment.
In this sense, at least ten resources are created for each input signal, totaling a
window with forty data processed by blocks of moving average, moving variance,
moving average square, and even zero detection. The data production block capable
of training the model uses the methods of the block above. It guarantees the quality of
the information acquired by some genuine sensor, which may be of the rod type, and
the techniques were responsible for guaranteeing a reliable training base.
data's variability, which can often reveal valuable insights about the underlying
patterns.
The moving average square was employed to highlight any trends in the
dispersion or scatter of the data. On the other hand, zero detection was used to find
instances where the signal crosses the zero level, which can often indicate significant
events in time-series data, such as control signals or sensor readings.
Applying these techniques generated at least ten additional features for each
input signal. This process effectively enhanced the original dataset, rendering it more
conducive to learning by the machine learning model. Notably, each newly generated
feature played a crucial role in capturing the inherent patterns and relationships
embedded within the data. This comprehensive set of features significantly fortified the
model's predictive capabilities, enabling it to make more precise and accurate
predictions.
The engineered features were seamlessly integrated with the original dataset,
creating an enhanced dataset comprising 50 features. Including these engineered
features augmented the dataset's information content, potentially improving the
models' performance and their ability to extract meaningful insights from the data. This
new dataset was expected to improve the model's performance, leading to more
accurate and reliable predictions of OFCs.
This systematic feature engineering approach emphasized data preparation's
importance in machine learning tasks. By transforming and enriching the original
dataset, it was anticipated that the enhanced features would significantly improve the
machine learning model's ability to detect and predict Oscillatory Failure Cases
accurately. The subsequent steps would then evaluate the impact of these engineered
features on the model's performance using the code in Attachment A.
Initially, two datasets, 'X' and 'Y' were loaded from their separate CSV files using
the panda's library's read_csv function. The dataset 'X' comprised 40 features, and 'Y'
was the output or target variable.
Subsequently, for the sake of simplicity and ease of reference, the columns in
dataset 'X' were renamed to 'f-0', 'f-1', ..., 'f-39', while the output column in the 'Y'
dataset was renamed to 'Output'. As mentioned above, the renaming was executed
through Python list comprehension, which was quite efficient and saved considerable
computational resources.
The data were then scrutinized for any missing values. When data were missing,
imputation was used to handle these inconsistencies. In this case, the mean value of
the respective column was used to replace any missing values.
The dataset was normalized after imputation using the `StandardScaler` from
the sklearn library. This step was critical to ensure that all the features were on a similar
scale, optimizing the machine learning model's performance by preventing any single
feature from dominating others due to its scale.
Once the normalization was performed, datasets 'X' and 'Y' were combined into
a single dataframe. This was done to simplify the feature engineering and
transformation process, allowing operations to be performed on the entire data
simultaneously.
The next significant step involved splitting the dataset into multiple "sessions" to
facilitate profound feature synthesis. To ensure each session is unique, the
`split_into_sessions` function was adjusted to assign a unique session ID to each row.
After that, deep feature synthesis was conducted with the help of the
Featuretools library. The entity set was created, and each session was added as an
entity to the entity set. Finally, the `dfs` function automatically creates new features
using specified aggregation primitives.
In conclusion, the successful execution of this section was pivotal in extracting
more complex, high-level features from the raw data, significantly contributing to the
enhancement of the machine learning model's performance. The entire process
underlines the importance of data preprocessing and feature engineering in machine
learning and data science projects.
112
pipeline stage. Each method within the class represents a step in model training and
evaluation.
The method `split_data()` was employed to partition the dataset into features
(`X`) and targets (`Y`). A conventional train-test split was conducted, designating 80%
of the data for training and the remaining 20% for testing. Ensuring a balance between
training and testing data is crucial for creating an effective model. Too little training
data may lead to underfitting, while excessive training data at the expense of testing
data may result in an overfit model.
The `select_model()` method selected a random forest classifier as the
predictive model. Random forest was chosen due to its robustness and generalization
capability, being an ensemble method that operates by constructing multiple decision
trees at training time and outputting the class that is the mode of the classes of the
individual trees.
The `train_model()` method facilitated the training of the selected model using
the training data. This involves feeding the model with the feature variables and their
corresponding target values to learn the underlying patterns in the data.
Following model training, the model's performance was evaluated using the test
data that was initially set aside. The `evaluate_model()` method predicted the target
values for the feature variables in the test data and compared them to the actual target
values. The results were then printed to provide insight into the model's predictive
accuracy and the precision, recall, and F1 score for each class. Table 12 shows the
results of the code:
As can be seen from table 12, 'support' refers to the number of actual
occurrences of the class in the specified dataset. For instance, when evaluating a
binary classifier's performance, support gives the number of samples of the proper
response in that class.
114
The 'support' metric can provide valuable context for understanding the other
metrics in the classification report. For example, a high precision or recall value may
not be as meaningful if the support is shallow, as the metric is calculated over fewer
instances.
As per the results obtained, the model's accuracy is approximately 0.70,
suggesting that 70% of the predictions made by the model on the test data are correct.
The classification report reveals that the precision, recall, and F1 scores for both
classes (0 and 1) are also around 0.70. These values indicate a balanced model
performance for both classes, suggesting that the model performs similarly well in
predicting both classes.
While the model's performance seems satisfactory, additional steps can be
taken to improve its predictive capabilities. For instance, hyperparameter tuning can
be performed to optimize the model's performance. Additionally, other models could
be tested and compared to find the one that best fits the data. The following stages of
this research could delve into these advanced techniques for model optimization.
In conclusion, training the DT model with the enhanced dataset by Feature
Engineering represented a critical phase in this investigation. By using the newly
created features, the model was expected to have an improved ability to identify and
predict Oscillatory Failure Cases accurately. The subsequent sections will further
elaborate on the DT model's performance after employing feature engineering to the
dataset.
Once the Decision Tree (DT) model was appropriately trained with the featured
dataset, the next step was to deploy it to identify Oscillatory Failure Cases (OFCs) in
the newly acquired data. This process involved feeding the model with the data
acquired from the simulated environment, enriched by the Feature Engineering
process, and observing the prediction results.
The enriched dataset comprised numerous features from the primary sensor
readings, contributing to a more comprehensive understanding of the system's state.
These new features, including statistical attributes like moving averages, variance, and
zero detection, were designed to capture essential aspects of the sensor readings that
might indicate an OFC's presence.
115
Once these enriched datasets were prepared and structured adequately, they
were presented to the trained DT model. The model then traversed its learned decision
paths, basing its decision on the thresholds and rules learned during the training phase.
The DT model delivered a classification output for each instance in the dataset.
This output was either a prediction of the presence or absence of an OFC based on
the inherent patterns learned from the featured dataset. The model's results were then
compared to the actual state of the system to evaluate the DT model's effectiveness
at identifying OFCs.
Identifying OFCs in the acquired data with the trained DT model represented a
crucial validation of the developed model. Moreover, it served as a testament to the
effectiveness of feature engineering in enhancing the model's predictive capabilities.
The subsequent analysis would shed more light on the DT model's performance,
emphasizing the impact of feature engineering on the system's predictive performance.
The performance of the Decision Tree (DT) model, after the training process
with the featured dataset and deployed for Oscillatory Failure Cases (OFCs)
identification, was analyzed. The main objective was to evaluate its accuracy,
precision, recall, and F1 score, which were critical indicators of its prediction capability
and the overall efficiency of the process.
Accuracy was first assessed, representing the proportion of total predictions the
model made correctly, both for the presence and absence of OFCs. Such a process
gave a broad overview of how well the DT model performed overall in classifying the
acquired data.
Precision and recall, on the other hand, provided more detail on the model's
performance. For example, precision quantified how many predicted OFCs were
117
actual, highlighting the model's ability to avoid false positives. Recall, alternatively,
quantified how many actual OFCs were correctly identified by the model, focusing on
the model's ability to detect true positives and avoid false negatives.
Finally, the F1 score was calculated to give a balanced measure of the model's
precision and recall. This metric is beneficial when the cost of false positives and false
negatives is roughly equivalent, as it ensures that both aspects are considered in
evaluating the model's performance.
The DT model's performance metrics were then compared against the original
model (without feature engineering) and the industry standards to gauge the overall
effectiveness of the applied feature engineering process. This comparison assessed
how much improvement was obtained through feature engineering, validating its
application in this context.
It was also necessary to conduct further analysis by diving deeper into each
OFC scenario (Ideal, Light, Moderate, Severe). This was to ascertain the model's
performance under different conditions, as the model must maintain high performance
consistently across various scenarios.
The performance analysis served as a crucial step in assessing the benefits of
feature engineering on the DT model's effectiveness. In addition, it helped determine
whether the Feature Engineering process contributed to improving the model's ability
to detect OFCs accurately and would inform future directions for further improving the
model.
ratio of correct predictions against all predictions. Precision offers insight into the
model's ability to avoid false positives, as it quantifies the proportion of predicted
positives that were indeed positive. Recall, or sensitivity, measures the proportion of
actual positives correctly classified, providing insight into the model's ability to detect
all potential positive cases. Lastly, the F1-score, calculated as the harmonic mean of
precision and recall, offers a balanced view of these two metrics.
After the metrics are calculated and stored in the `performance_metrics`
dictionary, they are presented in a neat, tabular format using the `display_metrics`
method. This method transforms the dictionary into a pandas DataFrame, facilitating
an easy-to-interpret view of the model's performance metrics.
Further extending the evaluation, the `plot_confusion_matrix` method is
employed. This method visualizes the model's performance using a confusion matrix,
a tabular layout representing the instances in predicted classes against those in actual
classes. When the `normalize` parameter is set to True, the confusion matrix presents
proportions rather than absolute counts, aiding in the interpretability of the results.
The `PerformanceAnalysis` class facilitates in-depth analysis of the trained
decision tree model's performance, highlighting its strengths and weaknesses. The
calculated metrics and confusion matrix offer valuable insights into the model's
predictive ability, providing the necessary understanding for future enhancements.
The calculated performance metrics were then analyzed and compared with the
original model (without feature engineering) and the industry standards to evaluate the
improvements gained through feature engineering.
Comparing the accuracy of the DT model after feature engineering with the
original model gave an exact measure of how much the predictive ability had improved.
Similarly, the Precision, Recall, and F1-score metrics provided detailed insights into
the model's improved performance regarding false positives and negatives and the
balance between Precision and Recall.
Moreover, comparing these performance metrics with industry standards gave
a relative measure of how well the DT model performed in the industry context. It also
highlighted the DT model's suitability for real-world applications and its
competitiveness with existing solutions.
119
In the top-left corner of the matrix, the value is approximately 0.69. This value
indicates that around 69% of the instances in class 0 (True Negative) were correctly
predicted by the XGBoost model as class 0. Conversely, the top-right corner of the
matrix shows a value of approximately 0.31, suggesting that around 31% of the
instances in class 0 were incorrectly predicted as class 1 (False Positive).
Therefore, moving to the bottom row of the matrix, the bottom-left corner has a
value of around 0.32. This value represents the False Negative rate, meaning about
32% of the instances in class 1 were incorrectly classified as class 0. The bottom-right
corner of the matrix has a value of around 0.68, indicating that roughly 68% of the
instances in class 1 were correctly classified as class 1 (True Positive) by the model.
Furthermore, the first image also mentions additional performance metrics -
Accuracy, Precision, Recall, and F1-score. For the XGBoost model, the accuracy is
around 0.684, which means about 68.4% of the total instances were correctly
classified. The Precision is the ratio of true positive predictions to the total positive
predictions, around 0.684. The Recall, the ratio of true positive predictions to the total
122
actual positives, is also around 0.684. Lastly, the F1-score, the harmonic mean of
precision and recall, is approximately 0.681. This score better measures the incorrectly
classified cases than the Accuracy Metric, especially when the class distribution is
uneven.
These results provide a comprehensive view of how well the XGBoost model
performed for this classification task.
Therefore, the Random Forest technique results are also depicted through a
confusion matrix in figure 33, which helps to evaluate its performance in classification
tasks. In this confusion matrix, the x-axis represents the predicted labels, while the y-
axis signifies the true labels. The values within the matrix demonstrate the proportions
of instances classified correctly or incorrectly.
In the top-left corner of the matrix, there is a value of about 0.70, suggesting that
approximately 70% of the instances that belong to class 0 were accurately predicted
as class 0 by the Random Forest model (True Negatives). On the other hand, the top-
123
right corner shows a value close to 0.30, indicating that around 30% of the instances
in class 0 were incorrectly predicted as class 1 (False Positives).
Meanwhile, examining the bottom row of the matrix in Figure 33, the bottom-left
corner holds a value near 0.30, representing the False Negative rate, meaning that
about 30% of instances in class 1 were incorrectly classified as class 0. The bottom-
right corner of the matrix contains a value of approximately 0.70, denoting that roughly
70% of instances in class 1 were accurately classified as class 1 by the model (True
Positives).
Additionally, the image provides other performance metrics - Accuracy,
Precision, Recall, and F1-score. The accuracy of the Random Forest model is around
0.697, meaning that about 69.7% of the total instances were correctly classified. The
Precision, which reflects the proportion of true positive predictions to the total positive
predictions, is roughly 0.697. The Recall, or the proportion of true positive predictions
to the total actual positives, is also about 0.697. Finally, the F1-score, the harmonic
mean of precision and recall, is around 0.697. This metric provides a more balanced
measure of the incorrectly classified cases than the Accuracy Metric, especially when
the class distribution is uneven.
These results provide a thorough understanding of the Random Forest model's
performance in this classification task.
Meanwhile, the gradient-boosting technique results are presented through
another confusion matrix in figure 34. Similar to the previous examples, the x-axis
represents the predicted labels, and the y-axis represents the true labels. The matrix’s
values display the proportions of instances that were classified correctly or incorrectly.
In the top-left corner of the matrix, there is a value close to 0.71, indicating that
approximately 71% of the instances in class 0 were correctly predicted as class 0 by
the Gradient Boosting model (True Negatives). The top-right corner, with a value
around 0.29, shows that about 29% of the instances in class 0 were incorrectly
predicted as class 1 (False Positives). As can be observed:
124
Based on this graph, moving to the first row, the value from row 1 and column 0
has a value near 0.50, representing the False Negative rate, indicating that about 50%
of instances in class 1 were incorrectly classified as class 0. The bottom-right corner
of the matrix, which typically shows True Positives, is not visible in this case. However,
since the total must be 100%, it can be inferred that this value would be around 0.50
(100% - 50%).
Additional performance metrics include Accuracy, Precision, Recall, and F1-
score. The accuracy of the Gradient Boosting model is around 0.606, which means
about 60.6% of the total instances were correctly classified. The Precision is the ratio
of true positive predictions to the total positive predictions, which is about 0.665. The
Recall, the ratio of accurate optimistic predictions to the total actual positives, is about
0.50. Finally, the F1-score, the harmonic mean of precision and recall, is roughly 0.600.
These results provide insight into the Gradient Boosting model's performance
for this classification task. It is noticeable that the model has room for improvement,
especially regarding the high False Negative rate.
125
This method offered nuanced insights into the model's performance, not just in
general accuracy but in the context of different misclassifications.
Such precise knowledge of false positives and negatives was invaluable in
understanding the model's behavior in identifying Oscillatory Failure Cases, a crucial
detail considering the potentially high costs of misclassification in this context. The
confusion matrix, thus, served as a beneficial tool for evaluating the model's
robustness and adaptability to handle complex classifications.
Finally, in evaluating the four machine learning methods for the classification
task, Random Forest emerged as the top performer with an accuracy of approximately
69.7% and an F1-score of around 0.697. The Decision Tree method closely trailed,
boasting an accuracy of roughly 69.6% and an F1-score of about 0.698, making it
almost on par with Random Forest and a viable alternative. Meanwhile XGBoost,
although a potent model, ranked third with an accuracy of around 68.4% and an F1-
score of approximately 0.681. Gradient Boosting lagged behind the rest, manifesting
the weakest performance with an accuracy of about 60.6% and an F1-score near
0.600. The subtle differences between Random Forest and Decision Tree make them
both strong contenders, while Gradient Boosting would necessitate further optimization
for this particular task.
simulate and analyze the performance of Soft Sensors in detecting Oscillatory Failure
Anomalies within an aircraft's FCS.
The MATLAB engine is imported into Python, allowing for seamless interaction
between the two environments. A named tuple structure, `ParameterScenario`, is
defined to encapsulate various parameters related to the simulation scenarios. Four
distinct scenarios, namely ideal, light, moderate, and severe, are then defined using
this structure. These scenarios represent different conditions or states of the FCS,
each with its unique set of parameters.
A function, `test_matlab`, is introduced to simulate the Soft Sensor's
performance in MATLAB using the predefined scenarios. This function initializes the
MATLAB engine, sets the necessary paths, and loads the required variables. It then
configures the simulation parameters based on the provided scenario and initiates the
simulation. The function returns the simulation results, which include the desired
control input, measured control input, time, and whether an oscillatory failure anomaly
was detected.
Another function, `plot_data`, is designed to visualize the simulation results. It
converts the MATLAB engine object into a NumPy array and plots the desired and
measured control inputs against time. Points where oscillatory failure anomalies are
detected are highlighted in the plot.
Finally, the code iterates over the predefined scenarios simulates each scenario
using the `test_matlab` function, and appends the results to a list. The results for each
scenario are printed for further analysis.
The foundation of the proposed framework is built upon the integration achieved
between MATLAB® and SimuLink™ benchmark with Python. This integrated platform
serves as the base for the framework, which operates in several layers:
The Data Acquisition Layer is responsible for collecting real-time data from the
FCS and directing it to subsequent layers. Its primary function is to ensure the
uninterrupted transfer of information while maintaining data integrity.
The Feature Engineering Layer processes raw data to extract or construct the
most relevant features for anomaly detection. The transformation and selection
127
strategies applied aim to enhance the clarity and distinguishability of patterns indicating
an oscillatory failure.
The Feedback Loop is a distinctive feature of this framework, emphasizing its
ability to learn and adapt. As anomalies are detected (or missed), these instances are
logged, and the system continuously refines its algorithms, ensuring that the Soft
Sensor becomes increasingly proficient at its task. This complete process is presented
in figure 35:
The Intelligent System framework aims to combine the best of both worlds: Soft
Sensors' precision and machine learning's adaptability. By doing so, it promises
several advantages:
The power of feature engineering gives enhanced detection rates; the system
boasts an advanced ability to pinpoint anomalies with unmatched precision. Therefore,
this heightened accuracy diminishes the chances of false positives and significantly
curtails false negatives, ensuring that potential threats are neither overlooked nor over-
reported. This dual advantage becomes instrumental in enhancing the reliability of the
system.
Real-time Processing: In the demanding and unpredictable world of aircraft
FCS, where split-second decisions can make a vast difference, this system delivers
instantaneous feedback. Such real-time responsiveness is not merely an added
feature but a non-negotiable necessity, making the system an invaluable asset in the
intricate flight control ecosystem.
Scalability: As technology advances and the FCS grows more complex with the
addition of new data streams or sensors, a system's ability to adapt becomes
paramount. The presented framework does not just adapt; it thrives in such evolving
129
While the framework provides a solid foundation for Soft Sensor enhancement,
viewing it as an evolving entity is crucial. Future research could incorporate more
sophisticated machine learning algorithms, refine feature engineering techniques
based on new findings, or integrate more robust feedback mechanisms.
In conclusion, the Intelligent System framework proposed herein represents a
holistic approach to tackling the challenges posed by Oscillatory Failure Anomaly
detection in aircraft's FCS. Its modular design and emphasis on continuous
improvement offer a promising path forward for researchers and aerospace engineers.
lens that refines this vast data landscape, bringing into focus the pivotal attributes that
drive decision-making.
Feature reduction involves various techniques that distill large datasets,
stripping away the redundant or irrelevant noise and preserving only the most
significant variables. This streamlined subset of attributes not only boosts
computational efficiency but also often enhances the underlying predictive power of
the model.
In the Soft Sensor domain, especially within the complex realm of an aircraft's
Flight Control System (FCS), the art and science of feature reduction assume
paramount importance. With the high stakes, the system demands nothing less than
the most potent combination of features. This section sheds light on the imperative
nature of feature reduction, its role in Soft Sensor performance, the methodologies
employed, and its broader impact on the field.
underlying system dynamics ensures robust and adaptive models, echoing the proper
patterns without redundancy.
Enhanced Interpretability often distinguishes a good model from a great one. If
convoluted by a web of redundant features, the nuances of model dynamics can
remain elusive. A refined feature set, on the other hand, offers a lucid window into the
intricacies of the model, demystifying relationships and offering invaluable insights.
Such transparency and clarity become instrumental for academicians and
practitioners, fostering more profound understanding and enabling informed decision-
making.
Noise Reduction emerges as a beacon of clarity in the cacophony of data. More
often than not, extraneous features cloud judgments, masking genuine patterns
beneath layers of irrelevant information. The art of feature reduction, in this context,
transforms into a science of clarity, meticulously weeding out the superfluous and
spotlighting the essential, ensuring that the decision-making machinery remains
attuned to the crux of the data.
Resource Optimization goes beyond computation, resonating with the tangible
facets of data storage and transfer. While potentially adding informational value, each
feature also demands storage space and bandwidth share. An optimized feature set,
therefore, not only enhances model efficacy but also paves the way for efficient and
cost-effective operations, optimizing resources without sacrificing performance.
Navigating the complexities of the FCS underscores the indispensable role of
feature reduction. Within this intricate dance of myriad system variables, where
precision is non-negotiable and the stakes are astronomically high, feature reduction
stands tall as a guiding force. Its myriad advantages collectively champion its cause,
reinforcing its stature as an essential tool in the study and practical implementation of
Soft Sensors, both within the confines of the FCS and in broader horizons.
Soft sensors delicately thread the complex interplay of raw data and algorithmic
prowess, embodying the quintessence of digital alchemy. Entrusted with the
monumental task of mimicking tangible measurements, especially in contexts where
traditional sensors falter in feasibility or reliability, soft sensors' performance is
inexorably linked to the quality and relevance of their foundational features. This
132
soft sensors' dexterity and molds them into resilient, efficient, and precise instruments.
Within the intricate choreography of systems like the FCS and beyond, feature
reduction and soft sensor excellence dance in harmonious synchrony, weaving the
tapestry of precision and reliability.
only primes the path for optimized model performances but also fosters a deeper
understanding and more informed decision-making in the vast universe of machine
learning.
To complement the data in the table, a graph depicting the relationship between
the number of features and accuracy offers visual insights. Peaks on this graph in
Figure 36 highlight the regions of optimal performance, pointing to areas where the
feature count harmonizes with model accuracy:
As can be seen, the graph depicts the performance of the Decision Tree model
in terms of accuracy as the number of features varies from 5 to 40. Initially, the
accuracy rapidly increases until around 21 features. After 21 features, the accuracy
plateaus, remaining roughly constant at approximately 0.696, indicating that using
136
more than 21 features does not yield any further improvement in model accuracy, and
an optimal feature set for this model is likely to be found within the first 21 features.
It required around 12 minutes and 50 seconds to simulate all the possible
combinations and evaluate their performance with the hardware described in section
1.4.2.
However, a crucial note for those looking to delve more deeply: the complexity
of this benchmarking process demands time and patience. Even with high-
performance computers available, several hours were spent waiting for the code to
complete its thorough exploration of the feature landscape. Researchers aiming to
replicate or extend this study should anticipate this time investment and plan their
resources accordingly.
Therefore, the result for the Gradient Boosting model in the same simulation
flow is presented in figure 37:
The graph for the Gradient Boosting model portrays an initially volatile accuracy
as the number of features increases from 5 to around 25, indicating the model is
attempting to ascertain the optimal combinations and interactions of the features. As
the number of features approaches 25, the accuracy progression stabilizes and follows
137
a more consistent upward trend. Notably, the accuracy reaches its peak around 35
features, nearing 0.60, after which it begins to plateau. This plateau suggests that
additional features beyond this point do not significantly contribute to improving the
model's performance.
Compared to XGBoost, Gradient Boosting took slightly longer to evaluate, with
a total duration of approximately 3 hours, 52 minutes, and 23 seconds. Considering
the time invested and the progression in accuracy, Gradient Boosting appears to
benefit from more features than XGBoost. However, when selecting a model, it is
important to weigh the marginal gains in accuracy against the computational cost and
complexity, especially as the number of features increases. Balancing performance
and efficiency is crucial in practical applications.
Therefore, the results for Gradient Boosting in the same simulation flow are
presented in figure 38:
The Random Forest graph displays the model's accuracy as the number of
features increases from 1 to 40. In the beginning, there is a sharp increase in accuracy,
which reaches a plateau relatively quickly, around ten features. Subsequently, the
139
accuracy fluctuates moderately within a narrow range, oscillating around the 0.697
mark. No discernible upward or downward trend exists beyond this point, suggesting
that Random Forest’s performance stabilizes early with a smaller subset of features.
Notably, Random Forest took significantly longer to evaluate than the other
models, with a total time of approximately 6 hours and 5 minutes. This extended
duration could be attributed to the inherent complexity of the Random Forest algorithm,
which builds multiple decision trees during training. Though Random Forest's accuracy
plateaus relatively early, it performs well, making it a robust model. However, the
computational cost can be a consideration for applications that require efficiency or
have resource constraints.
In light of the empirical evaluation of the four distinct algorithms – Decision Tree,
Random Forest, XGBoost, and Gradient Boosting – it is imperative to analyze the
results within the framework of model accuracy juxtaposed with computational
efficiency. This analysis is deducting the most appropriate machine-learning technique
for the dataset.
The Decision Tree algorithm exhibited remarkable performance, achieving an
accuracy plateau of approximately 0.696. Decision Tree's computational complexity is
significantly lower than the ensemble methods. Hence, this can be attributed to
Decision Trees not relying on constructing multiple base learners, as with Random
Forest and boosting methods.
Random Forest, an ensemble technique that constructs many decision trees,
registered a comparable accuracy level of around 0.697. However, it incurred a
substantially higher computational cost, clocking in at approximately 6 hours and 5
minutes. The increased complexity and computational time can be ascribed to the
inherent nature of the Random Forest algorithm, which necessitates the construction
and amalgamation of multiple decision trees.
XGBoost, a gradient-boosting algorithm, exhibited a notable accuracy level of
approximately 0.68. The algorithm’s efficacy in progressively correcting errors through
the sequential construction of decision trees is evident, albeit with diminishing returns
beyond certain features. The computational time for XGBoost was approximately 3
hours, 47 minutes, and 14 seconds.
Gradient Boosting achieved the lowest accuracy, capping at around 0.60. Its
computational time was the second highest, taking roughly 3 hours, 52 minutes, and
140
7 DISCUSSION
The essence of this research was thoroughly analyzed in this chapter, and
conclusions were derived from the comprehensive investigation carried out during this
study. The primary objective of this research was the exploration of the development
of a framework handling the application of Machine Learning (ML) techniques, to detect
and predict Oscillatory Failure Cases (OFCs) in a simulated Aerospatiale environment.
In this section, a more detailed examination of the key findings from this
research is undertaken.
When various machine learning models were applied, it was noted that the
Decision Tree algorithm demonstrated the most promising results. While the other
algorithms, such as Random Forest and Support Vector Machines, did exhibit
acceptable performances, they did not outperform the Decision Tree model in terms of
ease of interpretation, computational efficiency, and overall accuracy. Thus, the
Decision Tree would be the model of focus for this study.
Applying Feature Engineering techniques was the next significant phase of this
research. The initial 40-feature dataset was augmented to a more comprehensive 50-
feature dataset. It was discovered that creating these additional features had a
transformative impact on the model’s predictive performance. These newly
incorporated features, derived from the original dataset, enhanced the inherent
patterns pivotal to the prediction task, thus optimizing the Decision Tree algorithm's
ability to detect and predict OFCs accurately.
Employing the enriched 50-feature dataset, the retrained Decision Tree model
demonstrated a considerable advancement in its predictive performance. It was found
142
that the model's accuracy had improved substantially, reaching approximately 70%
correctness in its predictions. This result is significant in its relative improvement over
the initial model and potential implications for the field. This performance highlights the
importance of feature engineering in machine learning and its potential to improve
model performance.
The metrics derived from evaluating the Decision Tree model's performance
were compared with the initial models' results and examined in the context of industry
standards. This comparative analysis brought forth the effectiveness of the applied
feature engineering process in improving the model's performance. These results, in
turn, supported the argument that the techniques employed in this research could
apply to similar industrial prediction tasks.
Beyond the improved performance metrics, the findings from this research also
provided more profound insight into the nature of the data and the oscillatory failure
cases. By employing the Decision Tree algorithm and through feature engineering, the
study unearthed complex, non-linear relationships between different features that
would have been challenging to detect using more traditional statistical approaches.
Identifying these relationships could have far-reaching implications for understanding,
predicting, and preventing OFCs in a simulated industrial environment.
70% accuracy rate is substantial, future work might focus on approaches to enhance
this accuracy further.
One of the standout attributes of the Decision Tree model is its interpretability.
It was found that the model allowed a transparent understanding of the decision-
making process in predicting OFCs. This ease of interpretation could be valuable in a
practical, industrial context where comprehending the reasoning behind predictions
may be as crucial as the predictions themselves.
When the performance of the Decision Tree model was compared with industry
standards, it was found to hold up commendably. The model’s accuracy, precision,
recall, and F1-score all performed at a level that is in line with, if not superior to,
comparable models employed in the industry. Hence, the approaches and techniques
144
employed in this research could have broader applicability and potential benefits
beyond the specific context of this study.
The findings of this evaluation could have significant implications for the field.
The enhanced performance of the Decision Tree model in predicting OFCs highlights
the potential of machine learning, specifically feature engineering techniques, in
tackling complex, real-world industrial problems. The lessons learned could inspire
similar approaches in other related domains, promoting a more extensive application
of these techniques.
This section elaborates on the significance of the research and its potential
implications for academia and industry.
The theoretical implications of the research are multifold. This study adds to the
body of knowledge by providing evidence of the effectiveness of feature engineering
and machine learning techniques in predicting oscillatory failure cases (OFCs). It offers
an in-depth exploration of how an optimized Decision Tree model can be utilized
effectively in this domain.
Furthermore, this research can be a valuable reference for future researchers
aiming to apply similar methods in related fields. This study's findings and methodology
provide a rich ground for potential replication and validation in other contexts.
It’s also important to highlight that the framework showed that using fewer
features can output the same performance, reducing the computational costs.
The evaluation of the Decision Tree model’s performance might spur further
research. Given that the model achieved an accuracy rate of approximately 70%, it
suggests room for improvement. Future studies could explore how other machine
learning techniques, such as ensemble or deep learning, might increase this accuracy.
145
The results of this research are poised to bear far-reaching implications. They
could instigate reformative policy changes concerning the administration of
maintenance and failure prediction strategies within the aerospace industry. Moreover,
successfully applying the Decision Tree model might catalyze other industries to
embrace similar machine-learning methodologies for predicting and preempting
operational failures.
In essence, the research's monumental significance resides in its capability to
redefine academic comprehension and industrial methodologies pertinent to
Oscillatory Failure Cases (OFCs). The revelations gleaned from this study hold the
potential to drive further research, igniting innovation in academia and industry alike.
7.4 LIMITATIONS
This section provides a deeper look at the limitations identified in the research.
It is essential to recognize the limitations of any research study to fully understand the
context in which the findings should be interpreted.
Three significant limitations can be identified in the current study. The first
limitation pertains to the dataset used for the analysis. Although the dataset was
146
substantial and contained a wealth of information, it was limited to one specific type of
industrial machinery. As a result, the conclusions drawn in this study may not be
generalizable to other types of machinery or other industries.
Secondly, this research solely focused on the Decision Tree model. Despite its
demonstrable performance in predicting OFCs, other machine learning algorithms
such as Random Forest, Neural Networks, or Support Vector Machines might offer
different or more accurate results.
Lastly, the accuracy rate achieved, while impressive, is not perfect, indicating
that unknown factors influencing the OFCs that the current model has not accounted.
147
8 CONCLUSION
The journey of this dissertation began with a focused quest: to unravel the
capabilities of machine learning, particularly emphasizing the Decision Tree model, in
the context of predicting Oscillatory Failure Cases (OFCs) within aerospace
engineering. The need for this exploration was grounded in the tangible challenges
posed by OFCs, with implications spanning both financial and operational domains in
aerospace operations.
This lays the groundwork for the primary research question driving this
investigation: How can the performance of a Soft Sensor for Oscillatory Failure
Anomaly detection in an aircraft's Flight Control System (FCS) be improved
using Feature Engineering?
The foundation of this research was defined by its objectives, aiming to pioneer
a transformative feature engineering framework. This framework sought to refine the
performance of a Soft Sensor within an aircraft's Flight Control System. In response to
the primary research question, the answer emerged through a meticulously conducted
Systematic Literature Review, the study's various phases, and the empirical results.
Feature engineering, when applied judiciously, had a transformative effect on the Soft
Sensor's performance, enabling it to predict OFCs with heightened accuracy and
precision.
Guided by the literature review, the study identified and highlighted significant
gaps, especially in the realm of applying machine learning for OFC predictive
maintenance. A significant milestone in this journey was the successful integration of
the MATLAB® and SimuLink™ benchmarks with Python. This integration, facilitating
real-time data processing, harnessed the Python-embedded Soft Sensor and an array
of ML classes. This step laid the groundwork for subsequent advancements, most
notably the introduction of innovative feature engineering methodologies. These
methodologies showcased their transformative potential, marking a substantial
enhancement in the predictive outcomes observed in the preliminary research stages.
It’s important to highlight this step will be summarized in a concise paper for an
international journal publication.
Transitioning to the empirical facet, the research delved deep into a curated
dataset, rich with historical OFC incidents. This exploration led to the construction and
rigorous evaluation of the Decision Tree model. The model, in its final avatar, exhibited
148
REFERENCES
A. Blažič, I. Škrjanc, V. Logar. Soft sensor of bath temperature in an electric arc furnace
based on a data-driven Takagi–Sugeno fuzzy model. Applied Soft Computing, v.
113, n. 1, p. 1-11, 2021.
A. Gal-Tzur, S. Bekhor, Y. Barsky. Feature engineering methodology for congestion
forecasting. Journal of Traffic and Transportation Engineering (English Edition), v. 1,
n. 1, p, 1-14, 2022.
A. Gejji, S. Shukla, S. Pimparkar, et al. Using a support vector machine for building a
quality prediction model for center-less honing process. Procedia Manufacturing, v.
46, n. 2019, p. 600-607, 2020.
A. Guzman-urbina, K. Ouchi, H. Ohno, et al. FIEMA, a system of fuzzy inference and
emission analytics for sustainability-oriented chemical process design. Applied Soft
Computing, v. 126, n. 1, p. 1-16, 2022.
A. Hicks, M. Johnston, M. Mowbray, et al. A two-step multivariate statistical learning
approach for batch process soft sensing. Digital Chemical Engineering, v. 1, n.
October, p. 1-8, 2021.
A. Lew, M. Buehler. Encoding and exploring latent design space of optimal material
structures via a VAE-LSTM model. Forces in Mechanics, v. 5, n. 1, p. 1-8, 2021.
A. Mayr, D. Kißkalt, A. Lomakin, et al. Towards an intelligent linear winding process
through sensor integration and machine learning techniques. Procedia CIRP, v. 96, n.
1, p. 80-85, 2020.
A, Nightingale. A guide to systematic literature reviews, Surg, v. 27, n. 1, p. 381–384,
2009.
A. Theissler, J. Pérez-Velázquez, M. Kettelgerdes, et al. Predictive maintenance
enabled by machine learning: Use cases and challenges in the automotive industry.
Reliability Engineering and System Safety, v. 215, n. 1, p. 1-21, 2021.
A. Tsopanoglou, I. Jiménez del Val. Moving towards an era of hybrid modelling:
advantages and challenges of coupling mechanistic and data-driven models for
upstream pharmaceutical bioprocesses. Current Opinion in Chemical Engineering,
v. 32, n. 1, p. 1-8, 2021.
A. Zolghadri, J. Cieslak, D. Efimov, et al. Signal and model-based fault detection for
aircraft systems. IFAC-PapersOnLine, v. 28, n. 21, p, 1096-1101, 2015.
B. Maschler, S. Ganssloser, A. Hablizel, et al. Deep learning based soft sensors for
industrial machinery. Procedia CIRP, v. 99, n. 1, p. 662-667, 2021.
B. Negash, L. Tufa, R. Marappagounder, et al. Conceptual Framework for Using
System Identification in Reservoir Production Forecasting. Procedia Engineering, v.
148, n. 1, p. 878-886, 148.
151
L. Ma, Y. Liu, X. Zhang, et al. Deep learning in remote sensing applications: A meta-
analysis and review. ISPRS Journal of Photogrammetry and Remote Sensing, v. 152,
n. March, p. 166-177, 2019.
L. Petruschke, J. Walther, M. Burkhardt, et al. Machine learning based identification of
energy states of metal cutting machine tools using load profiles. Procedia CIRP, v.
104, n. 1, p. 357-362, 2021.
L. Ramona, et al. PHOTONAI—A Python API for Rapid Machine Learning Model
Development. PLoS One, vol. 16, no. 7, Public Library of Science, p. 25-62, 2021.
M. Abdar, F. Pourpanah, S. Hussain, et al. A review of uncertainty quantification in
deep learning: Techniques, applications and challenges. Information Fusion, v. 76,
n. 1, p. 243-297, 2022.
M. Abdel-Basset, H. Hawash, K. Sallam, et al. STLF-Net: Two-stream deep network
for short-term load forecasting in residential buildings. Journal of King Saud
University - Computer and Information Sciences, v. 34, n. 7, p. 4296-4311, 2022.
M. Bambach, M. Imram, I. Sizova, et al. A soft sensor for property control in multi-stage
hot forming based on a level set formulation of grain size evolution and machine
learning. Advances in Industrial and Manufacturing Engineering, v. 2, n. February,
p. 1-13, 2021.
M. Barton, B. Lennox. Model stacking to improve prediction and variable importance
robustness for soft sensor development. Digital Chemical Engineering, v. 3, n.
February, p. 1-13, 2022.
M. Ishi, J. Patil, V. Patil. An efficient team prediction for one day international matches
using a hybrid approach of CS-PSO and machine learning algorithms. Array, v. 14, n.
February, p. 1-12, 2022.
M. Feliciano, G. Reynoso-Meza. Soft Sensors: Virtual Sensors Applied to
Engineering problems. Final Project (Control and Automation Engineering Degree)
– Polytechnic School, Pontifical Catholic University of Parana (PUCPR). Curitiba, PR
– Brasil, p.170. 2020.
M. Maggipinto, E. Pesavento, F. Altinier, et al. Laundry fabric classification in vertical
axis washing machines using data-driven soft sensors. Energies, v. 12, n. 21, p. 1-14,
2019.
M. Mowbray, T. Savage, C. Wu, et al. Machine learning for biochemical engineering:
A review. Biochemical Engineering Journal, v. 172, n. May, p. 1-22, 2021.
M. Shapi, N. Ramil, L. Awalin. Energy consumption prediction by using machine
learning for smart building: Case study in Malaysia. Developments in the Built
Environment, v. 5, n. November, p. 1-14, 2021.
M. Siddiqi, B. Jiang, R. Asadi, et al. Hyperparameter tuning to optimize
implementations of denoising autoencoders for imputation of missing Spatio-temporal
data. Procedia Computer Science, v. 184, n. 2020, p. 107-114, 2021.
155
ATTACHMENT A
import pandas as pd
from sklearn.preprocessing import StandardScaler
import featuretools as ft
from featuretools.primitives import Mean, Sum, Std, Max, Min
class FeatureEngineeringOFC:
def __init__(self, X_file, Y_file):
self.X_file = X_file
self.Y_file = Y_file
self.dataset = None
def load_data(self):
# Load the dataset from the CSV files
dataframe_X = pd.read_csv(self.X_file)
dataframe_Y = pd.read_csv(self.Y_file)
return dataframe_X, dataframe_Y
target_dataframe_name="session_0",
agg_primitives=[Mean, Sum, Std,
Max, Min],
verbose=True)
return features
def execute(self):
df_X, df_Y = self.load_data()
df_X, df_Y = self.rename_columns(df_X, df_Y)
df_X = self.handle_missing_values(df_X)
df_X = self.normalize(df_X)
self.join_dataframes(df_X, df_Y)
ATTACHMENT B
class ModelSelectionAndEvaluation:
def __init__(self, dataset):
self.dataset = dataset
self.X = None
self.Y = None
self.X_train = None
self.X_test = None
self.Y_train = None
self.Y_test = None
self.model = None
def split_data(self):
self.X = self.dataset.drop('Output', axis=1)
self.Y = self.dataset['Output']
self.X_train, self.X_test, self.Y_train, self.Y_test =
train_test_split(self.X, self.Y, test_size=0.2, random_state=42)
def select_model(self):
self.model = RandomForestClassifier()
def train_model(self):
self.model.fit(self.X_train, self.Y_train)
def evaluate_model(self):
predictions = self.model.predict(self.X_test)
print("Model Accuracy: ", accuracy_score(self.Y_test,
predictions))
print("\nClassification Report:\n",
classification_report(self.Y_test, predictions))
def execute(self):
self.split_data()
self.select_model()
self.train_model()
self.evaluate_model()
ATTACHMENT C
input_data.append(features)
output_data.append(is_failure)
def extract_features(window):
mean = np.mean(window)
std = np.std(window)
skewness = skew(window)
return [mean, std, skewness]
# Prepare the input and output datasets using sliding time windows
input_data, output_data = load_csv_data(file_x, file_y) #
prepare_datasets(data_x, data_y)
# Split the data into training and testing sets (70% training, 30%
testing)
X_train, X_test, y_train, y_test = train_test_split(scaled_input_data,
output_data, test_size=0.3, random_state=42)
dt = DecisionTreeClassifier()
dt.fit(X_train, y_train)
163
ATTACHMENT D
class PerformanceAnalysis:
"""
A class used to analyze the performance of a classification model.
...
Attributes
----------
model : sklearn.base.ClassifierMixin
A trained model of ClassifierMixin type or any of its subclasses.
X_test : array-like
Test input data.
y_test : array-like
True labels for X_test.
y_pred : array-like
Predicted labels for X_test.
performance_metrics : dict
A dictionary to store performance metrics.
Methods
-------
calculate_metrics():
Calculates performance metrics.
display_metrics() -> pd.DataFrame:
Returns performance metrics as a pandas DataFrame.
plot_confusion_matrix(normalize=False):
Plots the confusion matrix.
"""
def calculate_metrics(self):
"""Calculates accuracy, precision, recall, and F1-score of the
model."""
self.performance_metrics['Accuracy'] =
accuracy_score(self.y_test, self.y_pred)
self.performance_metrics['Precision'] =
precision_score(self.y_test, self.y_pred, average='macro')
self.performance_metrics['Recall'] = recall_score(self.y_test,
self.y_pred, average='macro')
164
self.performance_metrics['F1_score'] = f1_score(self.y_test,
self.y_pred, average='macro')
plt.figure(figsize=(10, 7))
sns.set(font_scale=1.4)
sns.heatmap(df_cm, annot=True, annot_kws={"size": 16}, fmt='.2f'
if normalize else 'd')
plt.ylabel('True label')
plt.xlabel('Predicted label')
plt.show()
ATTACHMENT E
Returns
-------
None.
"""
benchmark_path =
r"C:\Master_Degree_Dissertation\Master_Degree_Code\SOFT SENSORS"
system_path = r"C:\Master_Degree_Dissertation\Master_Degree_Code\SOFT
166
SENSORS\ofc_benchmark.slx"
bench = "ofc_benchmark_acquire"
# Starting the benchmark and loads the main variables to the API
console
system.eval("ofc_benchmark_init", nargout=0)
system.eval(f"simulation.setSimulinkModel('{bench}');", nargout=0)
return process
delta_des = data[:, 0]
delta_meas = data[:, 1]
time = data[:, 2]
ofc_detected = data[:, 3]
plt.xlabel('Time (s)')
plt.ylabel('Control Input')
plt.title(f'{scenario.capitalize()} Scenario')
plt.legend()
plt.show()
data_list = []
for scenario, parameters in parameters_scenario.items():
result = test_matlab(parameters)
data_list.append((scenario, result))
ATTACHMENT F
class FeatureReductionPerformance:
"""
A class to evaluate a classification model's performance using
reduced features.
"""
Parameters:
- model: Classifier model to evaluate.
- X_train: Training data features.
- y_train: Training data labels.
- X_test: Test data features.
- y_test: Test data labels.
"""
self.model = model
self.X_train_original = X_train
self.y_train = y_train
self.X_test_original = X_test
self.y_test = y_test
self.results = {}
self.combinations = {}
Parameters:
- X_test_reduced: Test data with reduced feature set.
Returns:
- performance_metrics: Dictionary containing accuracy, precision,
recall, and F1_score.
"""
y_pred = model_clone.predict(X_test_reduced)
performance_metrics = {
'Accuracy': accuracy_score(self.y_test, y_pred),
'Precision': precision_score(self.y_test, y_pred,
average='macro'),
'Recall': recall_score(self.y_test, y_pred, average='macro'),
'F1_score': f1_score(self.y_test, y_pred, average='macro')
}
169
return performance_metrics
def _evaluate(subset):
"""
Evaluates the model's performance for a given subset of
features.
Parameters:
- Subset: List of indices representing the subset of features
to evaluate.
Returns:
- key: String representing the feature subset, with indices
separated by commas.
- performance_metrics: Dictionary containing performance
metrics for the given feature subset.
"""
X_train_reduced = self.X_train_original[:, np.array(subset)]
X_test_reduced = self.X_test_original[:, np.array(subset)]
model_clone = clone(self.model)
model_clone.fit(X_train_reduced, self.y_train)
performance_metrics = self.calculate_metrics(model_clone,
X_test_reduced)
key = ','.join(map(str, subset))
return key, performance_metrics
def plot_performance_against_features(self):
"""
Plot the model's accuracy against the number of features used.
"""
feature_counts = [3*len(key.split(',')) for key in
self.results.keys()]
accuracies = [val['Accuracy'] for val in self.results.values()]
plt.figure(figsize=(10, 6))
plt.plot(feature_counts, accuracies, marker='o')
plt.xlabel('Number of Features')
plt.ylabel('Accuracy')
plt.title('Model Accuracy vs. Number of Features')
plt.grid(True)
plt.show()