The fourth industrial revolution (Industry 4.0) brought significant changes in manufacturing, driven by technologies like artificial intelligence (AI), Internet of Things (IoT), 5G, robotics, and big data analytics. For industries to remain competitive, the primary goals must be the improvement of the efficiency and safety of machinery, the reduction of production costs, and the enhancement of product quality. Predictive maintenance (PdM) utilizes historical data and AI models to diagnose equipment’s health and predict the remaining useful life (RUL), providing critical insights for machinery effectiveness and product manufacturing. This prediction is a critical strategy to maximize the useful life of equipment, especially in large-scale and important infostructures. This study focuses on developing an unsupervised machine state-classification solution utilizing real-world industrial measurements collected from a pneumatic pressing machine. Unsupervised machine learning (ML) models were tested to diagnose and output the working state of the pressing machine at each given point (offline, idle, pressing, defective). Our research contributes to extracting valuable insights regarding real-world industrial settings for PdM and production efficiency using unsupervised ML, promoting operation safety, cost reduction, and productivity enhancement in modern industries. The fourth industrial revolution (Industry 4.0) resulted in the convergence of new technologies such as artificial intelligence (AI), the Internet of Things (IoT), 5G technology, robotics, and big data analytics, which brought significant changes in the enhancement of production processes and smart manufacturing. Modern applications will focus on sustainability and resilience, increasing safety, reliability, interconnection with human operators, and the industry’s overall productivity. Additionally, a significant improvement in sensing technologies and cyber-physical systems enables the monitoring and collection of raw heterogeneous data in high velocity, variability, veracity, volume, and value (the Big Data five Vs) [ 1]. All of these aspects provide a prosperous ground for the implementation of predictive maintenance (PdM), monitoring critical process parameters, and predicting the health state or the remaining useful life (RUL) of machinery equipment [ 2, 3, 4]. Predictive maintenance (PdM) relies on the use of historical data [ 1] to provide predictions about the production process, information about the health state of the equipment, and prognoses regarding the remaining useful life (RUL) of the machinery, leading to more efficient maintenance scheduling and management [ 5, 6]. PdM can reduce overall maintenance costs by 20% and reduce machine downtime by 5% [ 7]. Furthermore, according to [ 8], PdM can eliminate breakdowns by 70% to 75%, returning ten times the investment by increasing production by 20% to 25%. However, embracing the PdM approach is challenging. Some drawbacks include collecting significant failure data, handling unstructured raw data from heterogeneous sources, and the periodic optimization and re-training of PdM models with recent historical data [ 8, 9]. Ref. [ 10] proposed a PdM ML solution for the RUL prediction of smelting pots. In their study, the authors overcame the drawback of collecting high-volume input measurements by treating the RUL prediction as a classification task, i.e., by calculating malfunction likelihoods, data availability and prediction performances were increased. For PdM solutions, supervised ML algorithms are commonly used, but unsupervised algorithms can play a valuable role in prediction and final decisions. Despite their significant advantages and essential role in new-era industries, ML techniques still present some limitations, especially in industrial environments. The main limitations of supervised ML algorithms include the following: the need for a large amount of labeled data, which, in real-world industries, may denote a challenging and expensive approach [ 11]. Moreover, the potential of model overfitting is evident, where algorithms that have trained in one industrial environment may lack the ability to be generalized [ 12]. The limitations of unsupervised ML algorithms include the following: communication errors and sensors fault inputs are commonly found, leading to noisy data that can degrade the performance and accuracy of the algorithm [ 12], and in many cases, unsupervised techniques are computationally expensive, especially with the vast amount of data that industries generate [ 13]. Our study aims to provide a classification solution regarding heavy machinery in certain industries while evaluating the performance of established and state-of-the-art unsupervised machine learning (ML) algorithms. The output will extract crucial information regarding the working state of the equipment and the diagnosis of a pre-failure condition in the context of PdM. A data analysis procedure will be conducted following data collection. Data analysis techniques are vital when dealing with the heterogeneous data that we usually encounter in industrial environments. Hence, this research aims to develop a PdM model based on real-world data collected from a pneumatic pressing machine. The data refer to four stages of the pressing machine—specifically, offline mode, idle mode, pressing mode and defective. Our research is structured in two distinct stages. The first stage focuses on condition-based maintenance (CBM) approaches to diagnose the working state of the machine, while the second stage, which is presented in this paper, focuses on state classification and is developed based on historical data. Concerning the decision-making process, we examined typical AI models, such as K-means, DBSCAN, the Gaussian mixture model, hierarchical clustering, mean shift and spectral clustering, which were evaluated based on visualization, silhouette score, homogeneity and completeness. The rationale behind selecting these algorithms is that they are among the most well-known and extensively utilized in the field [ 14, 15, 16]. Thus, it was essential to have a direct comparison between them. The results will be compared to provide valuable information about the health state of the equipment. This paper is structured as follows: Section 2 provides a literature background; Section 3 describes the experimental system and presents the data acquisition methods and the characteristics of the collected data; Section 4 explains the models and the metrics employed; Section 5 showcases and discusses the outputted results; and Section 6 offers a brief conclusion and suggests directions for future research. Building on the foundational understanding of PdM outlined in the introduction, the esteemed literature is now directed toward optimal maintenance methods as businesses seek to remain competitive, resilient, and sustainable [ 17]. Over the years, the methods used in equipment maintenance have evolved to improve time utilization, minimize cost, and enhance product quality [ 18]. An initial literature analysis indicates that numerous definitions exist for the term maintenance [ 19, 20]. Based on the Swedish standard SS-EN 13306 [ 21] maintenance denotes a ‘Combination of all technical, administrative and managerial actions during the life cycle of an item intended to retain it in, or restore it to, a state in which it can perform the required function’. Three main maintenance categories were identified: corrective, preventive, and predictive [ 22, 23, 24]. However, an extension of PdM (prescriptive maintenance) arises, answering the questions, ‘What should be done for maintenance?’ and ‘How should it be done?’ [ 25]. More specifically, corrective maintenance (also known as run-to-failure maintenance) is a maintenance strategy that focuses on repairing equipment or system components after the malfunction has occurred and on restoring the equipment to a functional state [ 26, 27]. Ref. [ 28] proposed that replacements or repairs are performed only when the critical component is fully worn out and corrective maintenance fails. However, despite the merits of corrective maintenance, the malfunction of the system can lead to unexpected downtime and unwanted events, and, more importantly, it endangers the operators’ safety and the system’s reliability. On the other hand, preventive maintenance depicts the maintenance strategy performed periodically in specific scheduled time sequences regardless of whether the system is malfunctioning, reducing the probability of failure [ 29]. While this approach is more effective in preventing failures compared to corrective maintenance, it often leads to unnecessary corrective actions, causing increased operational costs and machinery downtime [ 22, 23, 30]. However, implementing PdM is being considered to enhance operational safety and productivity, minimize costs, and reduce downtime. [ 24] stated that PdM utilizes the current health status of a given critical component to predict its future condition and plan maintenance actions. The author of [ 29] defined PdM as historic data-based models that identify trends, behavior patterns, and correlations in failures based on statistical or ML models. If not replaced or repaired in the appropriate time frame, the deterioration of a system’s components will eventually lead to malfunction. This technique utilizes historical data acquired from sensor measurements that are used as input features in data-driven algorithms. This makes it possible to predict the health state of the system and the necessity of maintenance actions in the early stages or to determine the exact RUL of a component [ 30, 31]. Hence, the main categories of PdM approaches can be divided into condition-based maintenance (CBM) and prognostics and health management (PHM). Condition-based maintenance (CBM) is a decision-making method and a domain of predictive maintenance that utilizes real-time machinery data to diagnose the health state of the components. The decision to perform maintenance is made based on the overall observation of the system’s condition [ 27, 31]. Condition monitoring is divided into two categories: online and offline monitoring. Offline monitoring, also referred to as periodic condition monitoring, is about analyzing data from sensors at specific time intervals. Data analysis occurs in a laboratory environment away from the machine [ 32]. On the other hand, in online monitoring, also called continuous condition monitoring, the data are continuously collected, analyzed, and compared with the acceptable threshold [ 32, 33]. Ref. [ 27] introduced condition-based approaches to investigate and compare various maintenance techniques in the automotive industry. Additionally, the design and implementation of a SCADA system with a CBM approach was proposed, which utilizes IoT and ML techniques to improve the detection of failures [ 34]. Similarly, a CBM approach combining ML techniques and augmented reality was developed [ 35]. Prognostics and health management (PHM) is the case when predictive maintenance utilizes the current health status of a given critical component to predict its future condition and plan maintenance actions [ 27]. The reliability of prognostics and health management systems relies upon an accurate understanding of critical components’ degradation states to diagnose the current health state and predict RUL [ 31]. Various approaches and indicators depict and visualize the RUL of a component or machinery. Despite the differences among manufacturing sectors, the most common indicators are vibrations, temperatures, acoustic emissions, currents, pressures, rotational speed, historical failures, and exact times of measurement [ 8, 10, 36]. An open-source ecosystem combining industrial cyber-physical systems, IoT, and AI technologies to manage production processes efficiently and identify the health state of the machines, predicting their RUL, was suggested [ 37]. Finally, a study regarding the RUL prediction of slow speed bearings was proposed Aye and Heyns in 2018 [ 38]. They acquired historical data from an experimental setup simulating actual working conditions to induce bearing malfunction and predict the RUL with optimal accuracy. In the promising research field of prescriptive maintenance, the topic of interest is the control of the occurrence of a specific event, improving and optimizing upcoming maintenance procedures [ 25, 36]. Prescriptive maintenance, an extension of PdM, uses critical components predicted in a PdM model as an input. Using statistical methods or ML approaches, it determines the efficacy and optimizes key variables and hyperparameters of the experimented system. In [ 39], the authors developed a prescriptive application to predict the load vs. displacement curve of targeted cotton fiber/polypropylene, commonly used in the aerospace and automotive industries. The author of [ 37] proposed an intelligent computational model for prescriptive decisions in an industrial wastewater treatment plant. In the era of condition-based monitoring and predictive maintenance, the driving force behind innovation is AI, and especially ML [ 38, 40]. ML can be categorized into three main categories: supervised learning, unsupervised learning, and reinforcement learning [ 40]. In this research, the applied technique was unsupervised learning, which was implemented for the state classification and productivity identification of a single pneumatic pressing machine. Unsupervised learning was implemented on unlabeled data, which aims to recognize and describe hidden patterns within them. The lack of categorized data means that the evaluation of the results is not as straightforward as in the supervised learning techniques. Unsupervised algorithms can perform more complex tasks, but the output is more unpredictable [ 40]. Our research explores real-world use cases to provide a deeper understanding of how unsupervised AI models can be used and how to extract valuable insights from often unpredictable outputs. The use case was implemented on a pneumatic pressing machine used by a company that manufactures heat exchangers. Specifically, the press punches holes in the metal frame according to the customer’s requirements ( Figure 1). The correct operation of the machine was investigated and studied for the PdM application. The machine has three stages of operation. The pressing phase is when the processing of the metal is taking place; the idle phase is when the machine is waiting to start; and the offline phase is when the machine is turned off. However, based on the input of the on-field operators, it is observed that, in some cases, the initial metal frame is not properly placed, causing the machine to press defective parts. If the malfunction is not identified by the machine operator early enough to stop the pressing machine and position the frame appropriately, a significant amount of metal waste will be generated. Thus, an additional contribution of our PdM approach is to use unsupervised ML classification algorithms for identifying problems at an early stage and reduce metal waste, increasing the effectiveness, efficiency and profitability of the pressing process. Real-time data were collected using the Smart Condition Monitoring (SCM) Kit of Mitsubishi Electric Corporation, Tokyo, Japan, which is equipped with a FAG Smart Check sensor. The Mitsubishi SCM kit is a powerful, modular plug-and-play system that can easily be integrated into every existing application that can produce vibrations during its working state, e.g., motors, pumps, etc., in order to analyze machine data and parameters. The FAG Smart Check sensor is a vibration monitoring system that can measure, capture, and analyze signals and switch outputs or display the machine status on LEDs depending on the alarm limits [ 41, 42]. Figure 2 shows the installation of the sensor, and Figure 3 depicts the proposed system architecture. Following the description of the general system architecture and its components, we elaborate on the data acquisition for the study. Python programming language and the PyModbus library were implemented to connect the sensor kit and the analysis software. More specifically, the collected data were received through the Mitsubishi PLC Software (GX Works 3) (mitsubishielectric, available online: mitsubishielectric.com/fa/products/cnt/plceng/smerit/gx_works3/ last accessed on 10 October 2024) and an industrial communication network, Modbus TCP/IP, via Python. In the specific case study, there were six channels of SLMP communication for the connection between PLC and sensors. Through the ‘PyModbus’ library (pymodbus, available online: pymodbus.readthedocs.io last accessed on 10 October 2024), the software connected to the hardware system, revealing the necessary IP address and allowing access to TCP communication. During the interconnection and real-time measurements, the PLC transmitted 16-bit data in a string form. Converting from 16 bits to 32 bits is necessary for the values to be readable and printable as float numbers. To better organize and analyze the data and to make them easier to study, the data were stored in a dataframe through the Pandas library. Creating a data structure with detailed dates and organized rows and columns enabled the visualization of incoming real-time data in a comprehensible format. Finally, the measurements were stored in a CSV file for further processing. This presents a comprehensive analysis of collected measurements and the characteristics of each feature. The dataset contains historical data of measurements alongside the timestamp of each collected instance. It consists of five columns: velocity-ISO, demodulation, RMS-acceleration, peak-to-peak acceleration, temperature, and timestamp. Temperature: Measured in Celsius (C), temperature is an essential indicator of the health state of machines. In most cases, before a machine fails, the temperature rises. So, an abnormal temperature value can be a sign of incoming failure [ 43, 44]. Velocity-ISO: Refers to the rotational speed of the impeller of the inductive motor. The unit of measurement is mm/s. It can be a helpful indicator because defects on the impeller, material faults, and blockages of the flow during the operation can change the speed of rotation [ 45]. RMS-demodulation: Demodulation can identify faults like bearing defects or improper gear meshing. It can also help us to understand the lubricant’s effectiveness [ 46, 47]. Acceleration: Acceleration is the rate at which velocity changes. Its unit is m/sec2, and it is one of the most reliable indicators of vibration severity, especially in high-frequency vibration [ 48, 49]. For acceleration to be more useful in terms of information, it must be in the frequency spectrum and not in the time spectrum. For this reason, a fast Fourier transform (FFT) was applied. In this dataset, there are two types of acceleration: RMS (root-mean-square) acceleration and peak-to-peak acceleration. The incoming measurements can depict inconsistencies like missing values and outliers. These errors may rely upon the nature of the machinery or on sensor faults like noise. It is necessary to address these issues before feeding the data into the prediction algorithms [ 50, 51, 52]. The Mitsubishi Smart Condition Monitoring Kit collected 6472 rows, representing three operational working days of measurements, which depict the aforementioned input features. The scheduling of the experiment process differed because the process was handled manually by an operator. Approximately seven shape variations were extracted, where the throughput of each variation differed from 80 to 120 pieces. The added value of this research was to implement unsupervised ML algorithms to cluster the different working states of the machinery, i.e., offline, idle, pressing and defective faces. The collection of operating measurements will enable the operator to recognize early malfunctions and the supervisor to output the overall failures and working capacity of each piece of machinery. This delves into the specifics of the ML models utilizing the collected measurements. ML is a subdivision of AI that entails the development of algorithms to acquire knowledge and improve performance automatically by analyzing data and prior experiences. Various sophisticated algorithms in the literature are considered state-of-the-art tools for solving different problems. K-means, DBSCAN, the Gaussian mixture model, hierarchical clustering, mean shift and spectral clustering are among these. It is important to note that selecting the appropriate algorithm depends on the type of problem and the nature of the available data [ 53]. For example, K-means is better for smaller, well-separated datasets when the number of clusters is known [ 54, 55]. On the other hand, DBSCAN can be used when the number of clusters is unknown and the data have similar density. The Gaussian mixture model can be used when it is assumed that data are Gaussian-distributed and data clusters have different sizes and correlations. Hierarchical clustering can be used when a smaller dataset is available [ 14, 15, 16]. Mean shift can be used when the clusters are dense and well-separated and their number is unknown at the beginning [ 56, 57]. Finally, spectral clustering is useful when large datasets are available but the data are not easily separable [ 57, 58]. Hence, a deep understanding of the underlying principles and techniques is necessary to implement and utilize these state-of-the-art algorithms effectively. K-means is one of the most known clustering algorithms. The primary purpose of the algorithm is to separate N observations into K clusters. In a simplified manner, the algorithm works by creating the starting clusters, repositioning every point to the nearest center, redefining the centers based on calculating the mean of the points in the cluster, and repeating the same process. Because the algorithm is affected by the starting values, it is necessary to execute the test many times. In this way, the influence of the starting center values decreases [ 40, 59, 60]. DBSCAN was developed in 1996 as the first density-based clustering algorithm [ 61, 62]. Being a density-based algorithm, DBSCAN groups points based on density, which is defined as the closeness of a point to the neighboring points. Furthermore, the algorithm considers Epsilon and MinPts as key parameters. Epsilon is the maximum distance for points to be considered neighbors, and MinPts is the minimum number of points within Epsilon to identify a core point [ 61, 62, 63]. The algorithm starts from a random point in the dataset. If this point is a core point, meaning that it has a minimum number of points (MinPts) in a reachable distance (Epsilon), it becomes the starting point for a cluster and finds all the points that are in a reachable area and adds them to the cluster. The cluster is marked as complete when all the reachable points have been included [ 61]. Some of the main advantages of the DBSCAN over other algorithms is that it can effectively identify outliers and handle large datasets [ 62, 64]. The Gaussian mixture model (GMM) is a well-known probabilistic model that assumes all the data are generated from a mixture of Gaussian distributions. Regarding data clustering, each distribution represents a data cluster in the overall data. A brief overview of the creation of a distribution is as follows: taking the number of distributions to include in the model as an input and specifying the needed parameters with an initial guess, an expectation-maximization algorithm is employed to refine these parameters. This iterative process continues until the parameters have stabilized. One of the characteristics that make the Gaussian mixture model useful is that, unlike other clustering methods that assign a data point to a single cluster, the GMM calculates the possibility that a data point belongs to every cluster [ 40, 65, 66, 67]. Hierarchical clustering extends beyond simple clustering. It creates a multi-branched hierarchical tree by creating clusters at different levels. There are two types of hierarchical clustering. The first type is agglomerative, which assumes that every data point is a cluster of its own and merges with others based on similarities until one single cluster is created. The second type is divisive, which is the opposite of agglomerative. It assumes all the data points are in a single big cluster and progressively splits that cluster into smaller ones until every data point is a cluster of its own. Hierarchical clustering has some advantages over other clustering algorithms, such as the visualization of the results using the dendrogram output, the flexibility in adjusting the number of clusters, and its higher efficiency in more complex clusters [ 68, 69, 70]. Mean shift is a clustering technique that it can be used even if the number of the clusters is unknown. The main characteristic of mean shift clustering is that every data point in the dataset is a starting point for an iterative process. The algorithm based on the mean of neighbor points moves every data point to a higher-density area. Every high-density area has a local maximum, named a mode. The mode denotes the center of every cluster, and every point moves towards it utilizing a parameter called bandwidth [ 71]. Some of the advantages of mean shift are that is able to model complex clusters, it is not easily affected by outliers, and it has shown promising results in low-dimensional problems [ 56, 72, 73, 74]. Spectral clustering makes use of the Laplacian spectrum by constructing a Laplacian matrix [ 75]. Some of the advantages of the algorithm are that it clusters data points using similarity, which is less restrictive than parameter distance, utilized by the majority of clustering techniques. Furthermore, spectral clustering does not make assumptions about cluster shapes and has the flexibility to find clusters with ‘random’ shapes. However, despite its merits, spectral clustering is computationally expensive compered to center-based algorithms, is highly error-prone regarding outliers, and has lower performance accuracy when handling large data inputs [ 58, 76]—it cannot handle large datasets without using approximation. Outliers are a problem for the algorithm, so is preferred to handle them at an early stage [ 77]. The evaluation of predictive models is crucial in assessing their performance and determining their effectiveness in real-world applications [ 78]. One of the primary methods used to evaluate the model in clustering algorithms is to visualize the clusters. Thus, our research uses scatter plots to visualize and explain the model’s effectiveness. In addition to the visualization, we calculate metrics such as silhouette score, homogeneity, and completeness. The silhouette score is also known as the silhouette coefficient. It measures the quality by calculating the compactness of clusters and the separation among them. The values that can be taken are between −1 and +1. A score equal to +1 means that the clusters are clearly distinguished, a score of 0 means that the distance between clusters is not significant, and a score of −1 means that the clusters have been assigned incorrectly. A higher silhouette score means that a point is more poorly matched to the neighbor clusters but better matched to its own cluster. The score is calculated for each data point from the following type: S(i)=(b(i)−a(i))max(a(i),b(i))" role="presentation"> S(i)=(b(i)−a(i))max(a(i),b(i)) S ( i ) = ( b ( i ) − a ( i ) ) max ( a ( i ) , b ( i ) ) (1) where a(i) is the average distance from a data point to the other data points in the same cluster, and b(i) is the smallest average distance from a data point to a data point in a different cluster [ 79, 80]. Homogeneity measures how well a clustering result satisfies the condition that every cluster contains only data points from the same class. It can take values from 0 to 1, where 1 means perfectly homogeneous labeling [ 81, 82]. Completeness measures whether all data-point members of a class are elements of the same cluster. The completeness metric is satisfied if the data points in a cluster are all in the same class. Completeness can take values between 0 and 1, where 1 indicates perfectly complete labeling [ 83, 84]. A pneumatic pressing machine is a crucial component in the metal industry. This research focuses on its application in forming holes in metal sheets within an industry specializing in manufacturing heat exchangers, where precision in metal forming is important. As mentioned before, the machine operates in four distinct states: offline (deactivated), idle (paused and awaiting activation), pressing (actively forming the metal sheet), and defective (the metal sheet extracted during the working operation is defective). Data collection was conducted over three days, with four hours of measurements each day, resulting in a total amount of 6472 measurements. The overall purpose of this research is to investigate and evaluate the performance of various unsupervised ML algorithms on real-world datasets. While unsupervised ML has shown promise in theoretical studies, there is a lack of in-depth analysis of its effectiveness when applied in real-world data, especially regarding PdM, where the most studies are focused on supervised algorithms. This study aims to bridge this gap by comparing different algorithms with different means of data handling. Additionally, it demonstrates that unsupervised ML is a valuable tool for state classification, revealing its potential for pattern and anomaly recognition, which can prevent defective parts and equipment failures. Thus, practitioners can use this finding to select the most appropriate technique, algorithm and feature combination for their specific application. K-means, DBSCAN, the Gaussian mixture model, hierarchical clustering, mean shift and spectral clustering were used to diagnose the pressing machine’s state and work cycle. Different approaches and clustering techniques were tested. To justify the selection of features, a heatmap was generated for better visualization ( Figure 4). Then, the tested algorithms were categorized into two groups based on the number of input features. The first group used velocity_iso as the input feature, and the second used velocity_iso and acceleration. The following heatmap shows that the velocity_Iso and acceleration exhibit a strong correlation with each other, in contrast to temperature, which would not be so helpful as it has a negative correlation. Testing the algorithms with different numbers of input features is beneficial for several reasons. Firstly, testing with varying features helps to understand algorithm behavior, including how the algorithm adapts to different data complexities [ 85]. Secondly, in almost every real-world scenario, the datasets have many features, some of which might be irrelevant, so testing a few features at a time helps to identify the most important ones [ 86]. Thirdly, using too many features may lead to noise learning and overfitting the algorithm [ 87]. Lastly, if the algorithm can perform well with different number of features, it is more likely to generalize better to unseen data [ 88]. Following the selection of appropriate input features, namely velocity (value_ISO) and acceleration (value_ACC), the algorithms were tested for machine working-state clustering. For experimental purposes, scaled and unscaled input data were tested. In general, using scaled data is the most common practice. Some important reasons for this include that scaling helps algorithms converge faster and be more stable during the training phase and ensures that all features contribute equally to the distance calculation; also, there is a visualization benefit using scaled data [ 89, 90, 91]. However, on the other hand, using unscaled data prevents the amplification and impact of the outliers. Some inherently robust algorithms like decision trees can have better results with unscaled data. Finally, there are cases where preserving the original unit, such as physical measurements like temperature and weight, is necessary [ 92]. The evaluation was conducted using silhouette score, homogeneity, and completeness metrics. The metrics are presented in tables, and the clusters were visualized using the Matplotlib Python library. As the following metric tables and plots indicate, when only one feature was used as an input in the algorithms, regardless of whether the data were scaled or unscaled, the performance was decreased. In contrast, the inclusion of the second feature led to an improvement in performance across all the evaluated metrics and plots. The optimal plot is shown in Figure 5. The optimal unsupervised classification accuracy is depicted using the DBSCAN algorithm with two features as input, and the collected measurements are scaled. For comparison purposes, Figure 6 depicts the outputs of the DBSCAN algorithm utilizing same input features in an unscaled dataset of measurements, extracting less accurate classification results However, in the two pairs of tables and plots ( Figure 7, Table 1 and Figure 8, Table 2), where the algorithm uses only one input feature, the performance is notably poor. This is particularly evident in models like the Gaussian mixture model, where the clusters overlap. In both experimental setups—one feature with scaled data ( Figure 7) and one with unscaled data ( Figure 8)—the results show inaccurate classification performance, highlighting the importance of more appropriate feature selection utilizing additional features to assist the prediction models. The next pairs, utilizing two input features, demonstrate better overall results. Specifically, the results show that when the data were scaled, the performance was increased. The best model was for DBSCAN, where the clusters have high density and clear separation. There is great importance in data scaling ( Figure 9, Table 3), as in the experimental setup including two features with unscaled data ( Figure 10, Table 4), the results indicated that the classification performances were inaccurate. The results confirmed that unsupervised ML is effective in clustering machine operational states, particularly when using appropriate features such as velocity and acceleration. The DBSCAN algorithm, in particular, outperformed others when it used two input features and scaled data, offering a clear separation between the machine’s different operational phases. This finding reinforces the importance of feature selection and data scaling for the success of ML models. Moreover, the study showed that unsupervised learning can be a valuable tool for addressing PdM challenges when labeled data are not available. Feature selection played a critical role in improving model performance. Testing different combinations of features revealed that including multiple, correlated features such as velocity and acceleration led to better clustering results. Conversely, using only one feature resulted in poor performance. Additionally, scaling the input data proved crucial for improving the model’s stability and efficiency, especially in algorithms like DBSCAN and K-means. Scaling helped the models converge faster and ensured more accurate classification. Among the algorithms tested, DBSCAN stood out due to its ability to handle noise and identify well-separated clusters. Its density-based approach proved to be well-suited to the structure of the dataset, which required distinguishing between dense clusters representing the machine’s operational states. Other algorithms, such as K-means and the Gaussian mixture model, were less effective, particularly when fewer features were used or when the data were not scaled. This study demonstrates the potential of unsupervised learning techniques in real-world industrial applications, particularly for PdM. Although supervised models are typically preferred for such tasks, this research highlights that unsupervised models can provide valuable insights, especially when labeled data are limited or unavailable. The developed unsupervised model can be replicated and applied to machines with the same work processes throughout the industry. This enables an efficient and effective recognition of working states, and consequently the diagnosis of defective states, without the necessity of further technological analysis and additional resources for each respective similar procedure. That being so, our model indicates repeatability and has been trained to intelligently recognize and diagnose different states. This methodology not only us allows to identify obvious thresholding relationships that may fall into descriptive statistics but also to address instances that would typically require inferential statistics or complex correlations between variables. Therefore, we do not provide explicit thresholds but a model capable of identifying patterns and effectively handling complex scenarios. Additionally, the overall optimal approach, which we proposed and proved through experimentation, involved scaled data and appropriate multidimensional features. The presented results provide guidance for similar applications where the output data may not be directly accessible, demonstrating the approach’s high accuracy and effectiveness. This research explored the use of unsupervised machine learning (ML) models to classify the operational states of a pneumatic pressing machine, with the goal of enhancing productivity and minimizing downtimes. By analyzing vibration data, the study successfully identified the distinct working phases of the machine (offline, idle, pressing, defective) and proposed a framework for improving operational efficiency through state classification. The results indicated that the DBSCAN algorithm outputted the most accurate classification outputs, using a scaled dataset of velocity and acceleration measurements as input features, precisely classifying the machine’s different operational phases. This approach contributes to enhancing predictive maintenance (PdM) approaches, which are essential for reducing costs and enhancing machinery longevity in modern industrial environments. By clustering the four states of the machine (offline, idle, pressing, defected), it was possible to identify, using unsupervised ML, the working states of the machinery and introduce a promising solution to diagnose (in the context of condition-based maintenance) the defective state of the production process in the early stages. Although this study focused on a pneumatic pressing machine, the methodology can be extended to a wide range of industrial machinery. Future research could explore how these ML techniques perform across different types of equipment, industries, and operating conditions, thus generalizing the findings to a broader range of applications. In conclusion, this study demonstrates that unsupervised ML is a powerful tool for state classification in industrial contexts, offering significant potential for improving PdM strategies. In future work, incorporating supervised learning models could further enhance the predictive capabilities of this system. By leveraging labeled data on machine failures and operational faults, it could lead to more accurate predictions of the machine’s remaining useful life (RUL) and provide more actionable insights for maintenance teams. Another promising direction is the integration of anomaly-detection techniques alongside clustering algorithms. This hybrid approach could improve the identification of operational anomalies, which may indicate early signs of machine wear or potential failure, allowing for more proactive maintenance strategies. The implementation of real-time monitoring and feedback systems could also enhance the practical application of this research. By continuously analyzing vibration and acceleration data in real time, maintenance schedules could be dynamically adjusted based on the current condition of the machine. Such systems could issue alerts when irregular operational patterns are detected, helping to prevent unexpected breakdowns. In this scope, as a future extension, we propose the development of a real-time monitoring system interface showcasing the outcomes of working-state classification and potential alarms to the machine operator and technical managers, increasing the practical applicability of our research. As industries aim to achieve societal goals, respect the environmental boundaries of our planet, and place the well-being of industrial workers at the core of production processes [ 93], integrating advanced ML techniques into maintenance workflows will be crucial for enhancing operational efficiency, resilience and sustainability. Conceptualization, A.K., P.M., M.K., C.B. and D.B.; Methodology, A.K., P.M., M.K., C.B. and D.B.; Software, A.K., P.M. and M.K.; Validation, A.K., P.M., M.K., C.B., D.B. and E.D.; Formal analysis, A.K., P.M. and M.K.; Resources, A.K. and P.M., C.B. and D.B.; Writing—original draft, A.K., P.M. and M.K.; Writing—review & editing, C.B. and D.B. All authors have read and agreed to the published version of the manuscript. This research received no external funding. The data presented in this study are not publicly available and are available on request from the corresponding author due to privacy restrictions. We sincerely thank FRIGOPLAST S.A and Vasileios-Christos Vasileiou, who generously provided knowledge, expertise and hardware equipment for the implementation of this research. FRIGOPLAST S.A is a technical manufacturer of heat exchangers with vast expertise in evaporators, condensers, axial fan heaters and related items. Additionally, we would like to thank UTECO S.A. team, who generously provided the Mitsubishi Smart Condition Monitoring Kit for data collection purposes and the implementation of our use case. UTECO S.A. is a technical equipment provider in the field of automation with vast expertise in industrial, marine applications and infrastructure projects. Author Evangelos Diamantis was employed by the company UTECO S.A. Industrial and Marine Automation Systems. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest. Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. Abstract The fourth industrial revolution (Industry 4.0) brought significant changes in manufacturing, driven by technologies like artificial intelligence (AI), Internet of Things (IoT), 5G, robotics, and big data analytics. For industries to remain competitive, the primary goals must be the improvement of the efficiency and safety of machinery, the reduction of production costs, and the enhancement of product quality. Predictive maintenance (PdM) utilizes historical data and AI models to diagnose equipment’s health and predict the remaining useful life (RUL), providing critical insights for machinery effectiveness and product manufacturing. This prediction is a critical strategy to maximize the useful life of equipment, especially in large-scale and important infostructures. This study focuses on developing an unsupervised machine state-classification solution utilizing real-world industrial measurements collected from a pneumatic pressing machine. Unsupervised machine learning (ML) models were tested to diagnose and output the working state of the pressing machine at each given point (offline, idle, pressing, defective). Our research contributes to extracting valuable insights regarding real-world industrial settings for PdM and production efficiency using unsupervised ML, promoting operation safety, cost reduction, and productivity enhancement in modern industries. Keywords: predictive maintenance; Industry 4.0; artificial intelligence; unsupervised machine learning; DBSCAN
周老师: 13321314106
王老师: 17793132604
邮箱号码: lub@licp.cas.cn