<?xml version="1.0" encoding="utf-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.0 20120330//EN" "JATS-journalpublishing1.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">INFORMATICA</journal-id>
<journal-title-group><journal-title>Informatica</journal-title></journal-title-group>
<issn pub-type="epub">1822-8844</issn><issn pub-type="ppub">0868-4952</issn><issn-l>0868-4952</issn-l>
<publisher>
<publisher-name>Vilnius University</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">INFOR578</article-id>
<article-id pub-id-type="doi">10.15388/24-INFOR578</article-id>
<article-categories><subj-group subj-group-type="heading">
<subject>Research Article</subject></subj-group></article-categories>
<title-group>
<article-title>Data Augmentation for Classification of Multi-Domain Tension Signals</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name><surname>Žvirblis</surname><given-names>Tadas</given-names></name><email xlink:href="tadas.zvirblis@mf.vu.lt">tadas.zvirblis@mf.vu.lt</email><xref ref-type="aff" rid="j_infor578_aff_001">1</xref><xref ref-type="corresp" rid="cor1">∗</xref><bio>
<p><bold>T. Žvirblis</bold> received his PhD in technology sciences from the Faculty of Mechanics, Vilnius Gediminas Technical University (Lithuania), in 2022. He is currently employed as a senior researcher and a postdoctoral fellow at the Institute of Data Science and Digital Technologies, Vilnius University as well as an associate professor at the Faculty of Medicine, Vilnius University (Lithuania). His research interests include applied statistics, artificial intelligence, neural networks, data mining methods, and biostatistics. He is the author of 40 articles published in scientific journals and 25 works in conference proceedings.</p></bio>
</contrib>
<contrib contrib-type="author">
<name><surname>Pikšrys</surname><given-names>Armantas</given-names></name><email xlink:href="armantas.piksrys@mif.stud.vu.lt">armantas.piksrys@mif.stud.vu.lt</email><xref ref-type="aff" rid="j_infor578_aff_002">2</xref><bio>
<p><bold>A. Pikšrys</bold> is MSc student at the Faculty of Mathematics and Informatics, Vilnius University (Lithuania). His research focuses on the application of machine and deep neural learning models.</p></bio>
</contrib>
<contrib contrib-type="author">
<name><surname>Bzinkowski</surname><given-names>Damian</given-names></name><email xlink:href="damianbzinkowski@gmail.com">damianbzinkowski@gmail.com</email><xref ref-type="aff" rid="j_infor578_aff_003">3</xref><bio>
<p><bold>D. Bzinkowski</bold> is PhD student at the Faculty of Mechanical Engineering at Radom University (Poland). His main research interest is diagnostics and optimization of production processes.</p></bio>
</contrib>
<contrib contrib-type="author">
<name><surname>Rucki</surname><given-names>Mirosław</given-names></name><email xlink:href="m.rucki@uthrad.pl">m.rucki@uthrad.pl</email><xref ref-type="aff" rid="j_infor578_aff_003">3</xref><bio>
<p><bold>M. Rucki</bold> received his PhD degree and habilitation in mechanical engineering at Poznan University of Technology (Poland) and the title of full professor at VSB-Technical University Ostrava (Czech Republic). At present, he is with Casimir Pulaski Radom University (Poland). His main research interest is metrology, especially the measurement systems related to the industrial applications. Apart from that, he has got PhD degrees in humanities (Aramaic literature) and social sciences (family sciences).</p></bio>
</contrib>
<contrib contrib-type="author">
<name><surname>Kilikevičius</surname><given-names>Artūras</given-names></name><email xlink:href="arturas.kilikevicius@vilniustech.lt">arturas.kilikevicius@vilniustech.lt</email><xref ref-type="aff" rid="j_infor578_aff_004">4</xref><bio>
<p><bold>A. Kilikevičius</bold> received his PhD degree in technological sciences (measurement engineering) at Vilnius Gediminas Technical University (Lithuania). Currently, he is a director and a chief research fellow at the Institute of Mechanical Science at Vilnius Gediminas Technical University (Vilnius Tech). Also, he is a teaching professor at the Faculty of Mechanics, as well as a Chairman of PhD defense council in mechanical engineering at Vilnius Tech (Lithuania). A. Kilikevičius is the author of more than 200 scientific articles, co-author of more than 10 technologies (in the fields of environmental protection and precision mechanics), some of which are patented.</p></bio>
</contrib>
<contrib contrib-type="author">
<name><surname>Kurasova</surname><given-names>Olga</given-names></name><email xlink:href="olga.kurasova@mif.vu.lt">olga.kurasova@mif.vu.lt</email><xref ref-type="aff" rid="j_infor578_aff_001">1</xref><bio>
<p><bold>O. Kurasova</bold> received a PhD degree in computer science from the Institute of Mathematics and Informatics, Vytautas Magnus University (Lithuania) in 2005. She is currently employed as a principal researcher and a professor at the Institute of Data Science and Digital Technologies, Vilnius University (Lithuania). Her research interests include data mining methods, optimization theory and applications, artificial intelligence, neural networks, visualization of multidimensional data, multiple criteria decision support, parallel computing, and image processing. She is the author of more than 100 scientific publications.</p></bio>
</contrib>
<aff id="j_infor578_aff_001"><label>1</label>Institute of Data Science and Digital Technologies, <institution>Vilnius University</institution>, <country>Lithuania</country></aff>
<aff id="j_infor578_aff_002"><label>2</label>Institute of Computer Science, <institution>Vilnius University</institution>, <country>Lithuania</country></aff>
<aff id="j_infor578_aff_003"><label>3</label>Faculty of Mechanical Engineering, <institution>Kazimierz Pułaski University of Technology and Humanities in Radom</institution>, <country>Poland</country></aff>
<aff id="j_infor578_aff_004"><label>4</label>Institute of Mechanical Science, <institution>Vilnius Gediminas Technical University</institution>, <country>Lithuania</country></aff>
</contrib-group>
<author-notes>
<corresp id="cor1"><label>∗</label>Corresponding author.</corresp>
</author-notes>
<pub-date pub-type="ppub"><year>2024</year></pub-date><pub-date pub-type="epub"><day>19</day><month>11</month><year>2024</year></pub-date><volume>35</volume><issue>4</issue><fpage>883</fpage><lpage>908</lpage><history><date date-type="received"><month>8</month><year>2024</year></date><date date-type="accepted"><month>11</month><year>2024</year></date></history>
<permissions><copyright-statement>© 2024 Vilnius University</copyright-statement><copyright-year>2024</copyright-year>
<license license-type="open-access" xlink:href="http://creativecommons.org/licenses/by/4.0/">
<license-p>Open access article under the <ext-link ext-link-type="uri" xlink:href="http://creativecommons.org/licenses/by/4.0/">CC BY</ext-link> license.</license-p></license></permissions>
<abstract>
<p>There are different deep neural network (DNN) architectures and methods for performing augmentation on time series data, but not all the methods can be adapted for specific datasets. This article explores the development of deep learning models for time series, applies data augmentation methods to conveyor belt (CB) tension signal data and investigates the influence of these methods on the accuracy of CB state classification. CB systems are one of the essential elements of production processes, enabling smooth transportation of various industrial items, therefore its analysis is highly important. For the purpose of this work, multi-domain tension data signals from five different CB load weight conditions (0.5 kg, 1 kg, 2 kg, 3 kg, 5 kg) and one damaged belt condition were collected and analysed. Four DNN models based on fully convolutional network (FCN), convolutional neural network combined with long short-term memory (CNN-LSTM) model, residual network (ResNet), and InceptionTime architectures were developed and applied to classification of CB states. Different time series augmentations, such as random Laplace noise, drifted Gaussian noise, uniform noise, and magnitude warping, were applied to collected data during the study. Furthermore, new CB tension signals were generated using a TimeVAE model. The study has shown that DNN models based on FCN, ResNet, and InceptionTime architectures are able to classify CB states accurately. The research has also shown that various data augmentation methods can improve the accuracy of the above-mentioned models, for example, the combined addition of random Laplace and drifted Gaussian noise improved FCN model’s baseline (without augmentation) classification accuracy with 2.0 s-length signals by 4.5% to 92.6% ± 1.54%. FCN model demonstrated the best accuracy and classification performance despite its lowest amount of trainable parameters, thus demonstrating the importance of selecting and optimizing the right architecture when developing models for specific tasks.</p>
</abstract>
<kwd-group>
<label>Key words</label>
<kwd>fully convolutional network</kwd>
<kwd>convolutional neural network</kwd>
<kwd>long short-term memory model</kwd>
<kwd>residual networks</kwd>
<kwd>inception networks</kwd>
<kwd>data augmentation</kwd>
<kwd>sliding window</kwd>
<kwd>magnitude warping</kwd>
<kwd>variational autoencoder</kwd>
<kwd>conveyor belt tension signals</kwd>
</kwd-group>
</article-meta>
</front>
<body>
<sec id="j_infor578_s_001">
<label>1</label>
<title>Introduction</title>
<p>With the rapid development of deep learning (DL), the issue of insufficient datasets has become increasingly prominent. The performance of deep neural networks (DNN) is particularly dependent on the quantity, quality, and variety of training data (Sarker, <xref ref-type="bibr" rid="j_infor578_ref_019">2021</xref>). This problem is particularly evident in mechanical engineering, where machine learning (ML) is widely used. In experimental scenarios, obtaining sufficient data to train robust models can be difficult and expensive. The reliance on large datasets for training DL models highlights the need for obtaining sufficient data while lowering data collection costs.</p>
<p>Our research paper focuses on applying DL models to the classification of conveyor belt (CB) states (damaged CB and loaded with 0.5 kg, 1 kg, 2 kg, 3 kg, or 5 kg), specifically using belt tension time series signals. DL methods for the classification of CB states using images have been widely researched already, but the methods for classifying CB tension signals remain limited. Previous studies have demonstrated the potential of DL models like LSTM for this purpose (Žvirblis <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor578_ref_027">2022</xref>).</p>
<p>In the industrial sector, CB systems is one of the essential elements of production processes, enabling smooth transportation of various items. Depending on specific industrial application, CB systems must meet certain criteria and requirements, such as sterility in the food industry (Klištincová <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor578_ref_015">2024</xref>) or high wear resistance and durability (Bortnowski <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor578_ref_003">2022b</xref>). The reliability and efficiency of these conveyors are important to optimize work processes and avoid unplanned stops. An integral aspect of the maintenance of conveyor systems is monitoring their operational status to ensure correct functioning of the system and timely detection of potential faults (Dąbek <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor578_ref_005">2023</xref>). Traditional CB monitoring methods, such as manual, spectral, or radiographic damage detection (Li <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor578_ref_016">2011</xref>), are usually too expensive or require a lot of manual labour and are prone to human error.</p>
<p>Monitoring the status of CB systems is a critical aspect of their operational efficiency and safety. In the past, classification tasks in this area were performed using conventional ML algorithms and shallow models such as logistic regression and decision trees (Andrejiova <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor578_ref_001">2021</xref>). Over the last decade, DL methods have been applied increasingly widely because of their higher accuracy and efficiency. Santos <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor578_ref_018">2020</xref>) introduced binary classification models which used CB images. Classification was performed using deep convolutional neural networks (CNN) such as the visual geometry group (VGG) network, residual network (ResNet), and densely connected convolutional network (DenseNet). The highest average classification accuracy (89.8%) for particular data was achieved using DenseNet model. Zhang <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor578_ref_026">2021</xref>) performed a detailed analysis of ML algorithms and a comparison of DL models such as region-based CNN (R-CNN), single-shot detector (SSD), receptive field block net (RFBNet), Yolov3, and Yolov4 for the classification of CB damage images. Improved by the latter authors, Yolov3 architecture achieved an average classification accuracy of 97.3% for four damage classes.</p>
<p>Recent research has further revealed the application potential of DNNs in CB monitoring systems. Wang <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor578_ref_023">2023</xref>) presented a computer vision model capable of identifying CB defects with 94% accuracy, but this model was very sensitive to environmental effects and image quality. In another study, Bortnowski <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor578_ref_002">2022a</xref>) presented a long short-term memory (LSTM) autoencoder for automatization of damage detection by using recorded CB vibration signals. However, this model was not adapted to detect different types of CB damage. In addition, the vibration signals used in the study may be volatile due to various factors, such as load conditions or conveyor operating speed, which may affect the accuracy of the monitoring system.</p>
<p>CB tension signal data was applied to train various ML and DL models by Žvirblis <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor578_ref_027">2022</xref>) to determine the minimum signal length while maintaining high classification accuracy. However, in the study, DL models were only applied to classify two CB states (loaded with a 2 kg weight and unloaded). The authors did not include the detection of CB damage. Also, in the study, the initial dataset of CB tension signals was insufficient to train the models, so the authors performed two data augmentations methods like addition of random Laplace noise and drifted Gaussian noise. However, the main aim of the above-mentioned study was to develop high-accuracy classification models.</p>
<p>Achieving high classification accuracy with certain DL models requires collecting a sufficiently large and diverse dataset. Collecting large amounts of CB tension time series data can be difficult and expensive, therefore data augmentation methods can be used to increase the amount and variety of data. Data augmentation involves the creation of new data that is modified or synthesized from the original dataset. This enables the model to better generalize data and recognize features in unseen data. Data augmentation techniques are widely studied in computer vision and natural language processing, but their application to time series data is still being developed.</p>
<p>For time series data, traditional image augmentation methods such as scaling, rotating, or cropping often are not suitable due to the time dependence of time series data. Improper time series augmentations can negatively affect the accuracy and robustness of the model in real-world scenarios. In research works, some of the most effective time series data augmentation methods were the application of a sliding window, the addition of noise, and the synthesis of data using variational autoencoders (VAE) (Kingma and Welling, <xref ref-type="bibr" rid="j_infor578_ref_014">2019</xref>) or generative adversarial networks (GAN) (Goodfellow <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor578_ref_007">2014</xref>).</p>
<p>Data augmentation methods in DL are used to find effective strategies for improvement of the accuracy and robustness of models while having limited or unbalanced datasets. Recently, some scientific works related to data augmentation were published (Chlap <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor578_ref_004">2021</xref>; Wang <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor578_ref_022">2017a</xref>). However, most of these works provided applications for the areas related to image augmentation in computer vision models.</p>
<p>Further below, this work reviews various time series data augmentation methods that could augment the time series dataset and potentially improve the classification accuracy and robustness of DL models.</p>
<p>Raw time series data usually forms one long time series. The sliding window method can be used to generate more data for training. This augmentation method was used by Žvirblis <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor578_ref_027">2022</xref>) in a study where the initial CB tension signal data was divided into two stages. In the first stage, a sliding window method was used to divide the original signal into smaller signals and thereby expand the dataset. One of the disadvantages of a sliding window method is that dividing the original signal into smaller signals can cut off important features of time series data. Because of this, a DL model may not learn properly to classify small-window-size time series data.</p>
<p>Additional noise in the data simulates real data, as real equipment such as signal sensors can introduce noise into the observations. Therefore, adding noise makes DL models more robust to small variations in the data. Laplace, drifted Gaussian, and uniform noise are commonly used for augmentation of time series data (Um <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor578_ref_021">2017</xref>; Iwana and Uchida, <xref ref-type="bibr" rid="j_infor578_ref_013">2021</xref>; Žvirblis <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor578_ref_027">2022</xref>). However, adding too much noise to the original dataset can hinder the DL model’s ability to extract features from the signal. For this reason, it is important to study the influence of different amounts of noise on the model’s classification accuracy.</p>
<p>In augmentation of time series data, scaling means changing the amplitude or size of the original time series. Scaling augmentation methods for time series include magnitude warping, time warping, window warping, and frequency warping (Iwana and Uchida, <xref ref-type="bibr" rid="j_infor578_ref_013">2021</xref>; Um <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor578_ref_021">2017</xref>). However, these augmentations can over-distort important features of time series data, so it is important to choose the methods appropriate for a specific dataset.</p>
<p>VAEs can be used for augmentation of time series data as well (Kingma and Welling, <xref ref-type="bibr" rid="j_infor578_ref_014">2019</xref>; Goubeaud <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor578_ref_008">2021</xref>). Desai <italic>et al.</italic> (<xref ref-type="bibr" rid="j_infor578_ref_006">2021</xref>) presented the architecture of time series data variational autoencoder (TimeVAE), which was compared with other time series synthesis models such as TimeGAN (Yoon <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor578_ref_025">2019</xref>). On the average, the presented TimeVAE architecture showed higher accuracy of time series data synthesis than other models, especially with small data sets.</p>
<p>GANs is another widely researched framework for data augmentation and data synthesis, including time series data (Goodfellow <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor578_ref_007">2014</xref>). The architecture of GANs learns the distribution of data by extracting key features of the data. A trained generator of this architecture can then synthesize completely new data.</p>
<p>Currently, there are many different uses of GAN architectures for time series data (Huang <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor578_ref_011">2023</xref>; Iglesias <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor578_ref_012">2023</xref>). TimeGAN, one of the most widely used GAN architectures for time series, adds two new embedder and recovery networks to the conventional generator and discriminator architecture (Yoon <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor578_ref_025">2019</xref>). These new embedding and retrieval networks form an autoencoder in TimeGAN architecture that aims to learn the time dependence and key features of the data. In TimeGAN architecture, the autoencoder uses a recovery loss function, which aims to ensure that the network can accurately recover the original time series data from the latent space.</p>
<p>Conditional GANs contribute to traditional architectures by incorporating conditional information into the training process. This allows the network to be trained to generate more accurate data based on specific inputs, such as classes of data. The conditional GAN time series architecture (TSGAN) has achieved higher accuracy in synthesizing time series data for classification tasks than other GAN architectures (Smith and Smith, <xref ref-type="bibr" rid="j_infor578_ref_020">2020</xref>). TSGAN architecture was tested on 70 datasets and compared with Wasserstein Generative Adversarial Network (WGAN) architecture (Gregor Hartmann <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor578_ref_009">2018</xref>). The accuracy of the data synthesized by TSGAN architecture was higher than that of WGAN architecture by about 11% on the average.</p>
<p>There are different DNN architectures and methods for performing augmentations on time series data, but not all the methods can be adapted for specific datasets. This work aims to develop DL models for time series, apply data augmentation methods to CB tension signal data and investigate the influence of these methods on the accuracy of CB state classification.</p>
<p>The rest of the paper is organized as follows. Materials and methods are described in Section <xref rid="j_infor578_s_002">2</xref>. The main study results are provided in Section <xref rid="j_infor578_s_011">3</xref>. Conclusions close the paper in Section <xref rid="j_infor578_s_020">4</xref>.</p>
</sec>
<sec id="j_infor578_s_002" sec-type="materials|methods">
<label>2</label>
<title>Materials and Methods</title>
<p>The aim of the study on the classification of load and defect states of a CB is to not only compare the classification accuracy of different DNN models but also to compare how the accuracy of the models is affected by different data augmentation methods. This chapter examines selected time series data augmentation methods that will be applied to CB tension signals as well as selected DL models, accuracy metrics, and data acquisition.</p>
<sec id="j_infor578_s_003">
<label>2.1</label>
<title>Experimental Design and Data Collection</title>
<p>The test stand, on which the measurements were carried out, was a CB model, shown in Fig. <xref rid="j_infor578_fig_001">1</xref>. Its supporting structure consists of four self-aligning ball bearing units and two drums are embedded in the inner raceways: drive and return, on which strain gauges T1, T2, and T3 are located. The housings were connected by threaded rods and bolted with nuts, and sets of lenticular washers were used between the surfaces of the bearing housings on both sides to compensate for curvature.</p>
<fig id="j_infor578_fig_001">
<label>Fig. 1</label>
<caption>
<p>CB model.</p>
</caption>
<graphic xlink:href="infor578_g001.jpg"/>
</fig>
<p>The system uses strain gauges, in which the resistance <inline-formula id="j_infor578_ineq_001"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">T</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${R_{T}}$]]></tex-math></alternatives></inline-formula> depends directly on the resultant belt force <inline-formula id="j_infor578_ineq_002"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[${F_{n}}$]]></tex-math></alternatives></inline-formula>. During the dynamic testing of the CB (Fig. <xref rid="j_infor578_fig_002">2</xref>), the speed of the drive drum is set at <italic>v</italic> and the recorded waveforms of the signals from the strain gauges depend on the pre-tension of the belt, longitudinal damage UW, transverse damage UP, and the load <italic>m</italic>. The strain gauge sensors have nonlinear characteristics, and there is a nonlinear dependence of the readings depending on where the belt is pressed on the belt strain gauge which girds the shaft. The analog-to-digital converter (ADC) electronic circuit receives the analog data and converts it to digital, and then sends it via Bluetooth transmission to a computer. At the stage of converting analog values to digital, the signal is discretized and its values are measured at a frequency of 200 Hz. The signal, which goes to the computer in digital form, is represented in analog-to-digital units (ADUs) and the acquired values are subject to rounding and linearization.</p>
<fig id="j_infor578_fig_002">
<label>Fig. 2</label>
<caption>
<p>CB condition monitoring system.</p>
</caption>
<graphic xlink:href="infor578_g002.jpg"/>
</fig>
<fig id="j_infor578_fig_003">
<label>Fig. 3</label>
<caption>
<p>CB damage diagram. Here UW I is a longitudinal cut of 50 mm, UW II is a longitudinal cut of 70 mm, UW III is a longitudinal cut of 45 mm with a depth of 1 mm, UW IV is a longitudinal cut of 50 mm with a depth of 1.5 mm and UP I is a cross cut of 10 mm.</p>
</caption>
<graphic xlink:href="infor578_g003.jpg"/>
</fig>
<p>The main purpose of data collection was to observe and analyse the influence of various belt loads and defects on CB tension signals. The observations were collected by using three strain gauge sensors, which were placed in parallel at different sections of the CB (top, middle, and bottom) to fully record the strain signal of the CB. For this reason, further data collection, analysis, and model building were done on multi-domain data input structure. Observations were carried out in two stages. In the first stage, observations were made by using the conveyor loaded with one of five different weights: 0.5 kg, 1 kg, 2 kg, 3 kg, or 5 kg. Each weight category was designed to simulate different loading conditions on the CB. In the second stage, CB was intentionally damaged in certain places to simulate defects in real conditions as shown in Fig. <xref rid="j_infor578_fig_003">3</xref>. During the second stage of observations, CB tension signals were recorded as in the case of belt damage without any weight load.</p>
<fig id="j_infor578_fig_004">
<label>Fig. 4</label>
<caption>
<p>Sensor signal values in different conditions of the conveyor belt: a) first lower sensor; b) second middle sensor; c) third upper sensor.</p>
</caption>
<graphic xlink:href="infor578_g004.jpg"/>
</fig>
<fig id="j_infor578_fig_005">
<label>Fig. 5</label>
<caption>
<p>The scheme of experiment.</p>
</caption>
<graphic xlink:href="infor578_g005.jpg"/>
</fig>
<p>To simulate different rotational speeds of the CB, three different revolutions per minute (RPM) speeds were chosen: 159, 318, and 540. The CB speed of 159 RPM corresponded to a linear speed of 0.5 m/s, 318 RPM corresponded to 1 m/s, and 540 RPM corresponded to 1.7 m/s. Each observation was performed 9 times (3 times for each RPM). All the observations lasted about 8 seconds each on the average.</p>
<p>To gain more insight into the collected data, the amplitude means of the tension signals under different weight loads and defect states were calculated and presented in Fig. <xref rid="j_infor578_fig_004">4</xref>. The damaged CB signal has the highest average amplitude of all the load conditions, therefore it is clearly identifiable. Higher amplitude can be explained by the presence of defects and irregularities on the surface of the CB, which cause larger fluctuations in the tension signals. It can also be observed that the average tension amplitude of the maximum weight load of 5 kg was the lowest while that of 0.5 kg was the highest as compared with other weight loads.</p>
<p>Augmentation of the collected data and other steps of the experiment are shown in Fig. <xref rid="j_infor578_fig_005">5</xref>.</p>
</sec>
<sec id="j_infor578_s_004">
<label>2.2</label>
<title>Data Augmentation Techniques</title>
<p>Testing a wide range of data augmentation techniques is crucial in DL because different augmentations can have varied and sometimes profound impacts on model performance, generalization, and robustness. Data augmentation helps to prevent overfitting by exposing the model to diverse data transformations, which simulates the variability it may encounter in real-world scenarios. Different data types or domains may benefit from specific augmentations. For instance, noise injections can be useful for vibration or sound data. Testing various augmentations helps identify those that address domain-specific challenges effectively, enhancing the model’s adaptability.</p>
<p>We applied a range of data augmentation techniques, both traditional and advanced. These included:</p>
<list>
<list-item id="j_infor578_li_001">
<label>•</label>
<p>Basic augmentations: sliding window.</p>
</list-item>
<list-item id="j_infor578_li_002">
<label>•</label>
<p>Advanced augmentations: random Laplace noise, drifted Gaussian noise, uniform noise, and magnitude warping.</p>
</list-item>
<list-item id="j_infor578_li_003">
<label>•</label>
<p>Generative augmentations: variational autoencoders.</p>
</list-item>
</list>
<p>A sliding window splits the data into more time series of smaller sizes. The sliding window formula used in this work is: 
<disp-formula id="j_infor578_eq_001">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:mi mathvariant="italic">W</mml:mi>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo fence="true" stretchy="false">]</mml:mo>
<mml:mo>=</mml:mo>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">m</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo fence="true" stretchy="false">]</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mspace width="1em"/>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">k</mml:mi>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">m</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ W[i]=[1+i,m+i],\hspace{1em}i=0,1,2,\dots ,k-m,\]]]></tex-math></alternatives>
</disp-formula> 
where a time series of size <italic>k</italic> is split into multiple time series of length <italic>m</italic>. Here, <inline-formula id="j_infor578_ineq_003"><alternatives><mml:math>
<mml:mi mathvariant="italic">W</mml:mi>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi>
<mml:mo fence="true" stretchy="false">]</mml:mo></mml:math><tex-math><![CDATA[$W[i]$]]></tex-math></alternatives></inline-formula> represents the <italic>i</italic>-th window, starting with index <inline-formula id="j_infor578_ineq_004"><alternatives><mml:math>
<mml:mn>1</mml:mn>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi></mml:math><tex-math><![CDATA[$1+i$]]></tex-math></alternatives></inline-formula> and ending with index <inline-formula id="j_infor578_ineq_005"><alternatives><mml:math>
<mml:mi mathvariant="italic">m</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="italic">i</mml:mi></mml:math><tex-math><![CDATA[$m+i$]]></tex-math></alternatives></inline-formula>.</p>
<p>To expand the dataset of CB signals and to standardize their length, signals of all original observations were divided into 0.5 s (200 points), 1.0 s (400 points), and 1.5 s (600 points) signals. The step of each signal’s sliding window was 100% of the window size itself. For example, a step of 0.5 s-length signal is 0.5 s (200 points). This was done to create enough of different signals for training the models and to avoid over-fitting the models.</p>
<p>Random Laplace noise is based on sampling random values from a Laplace distribution. This distribution is characterized by the Laplace probability density function: 
<disp-formula id="j_infor578_eq_002">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">f</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">e</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>−</mml:mo><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">μ</mml:mi>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {f_{1}}(x)=\frac{1}{2\sigma }{e^{-\frac{|x-\mu |}{\sigma }}},\]]]></tex-math></alternatives>
</disp-formula> 
where <italic>μ</italic> is the mean of the distribution, and <italic>σ</italic> is the scale parameter controlling the width of the distribution.</p>
<p>Drifted Gaussian noise adds a random value from a Gaussian (normal) distribution to each point in the signal. Gaussian distribution is characterized by Gaussian probability density function: 
<disp-formula id="j_infor578_eq_003">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">f</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
<mml:msqrt>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mi mathvariant="italic">π</mml:mi>
</mml:mrow>
</mml:msqrt>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">e</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>−</mml:mo><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:msup>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo><mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">μ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:msup>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {f_{2}}(x)=\frac{1}{\sigma \sqrt{2\pi }}{e^{-\frac{1}{2}{(\frac{x-\mu }{\sigma })^{2}}}},\]]]></tex-math></alternatives>
</disp-formula> 
where <italic>μ</italic> is the mean of the distribution and <italic>σ</italic> is the standard deviation.</p>
<p>Uniform noise generates a value from a uniform distribution. The uniform distribution is characterized by the density function that represents the equal likelihood of any value within the specified interval <inline-formula id="j_infor578_ineq_006"><alternatives><mml:math>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mo>;</mml:mo>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo fence="true" stretchy="false">]</mml:mo></mml:math><tex-math><![CDATA[$[a;b]$]]></tex-math></alternatives></inline-formula>: 
<disp-formula id="j_infor578_eq_004">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:mi mathvariant="italic">f</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mfenced separators="" open="{" close="">
<mml:mrow>
<mml:mtable columnspacing="4.0pt" equalrows="false" columnlines="none" equalcolumns="false" columnalign="left left">
<mml:mtr>
<mml:mtd class="array">
<mml:mstyle displaystyle="false">
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">a</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mspace width="1em"/>
</mml:mtd>
<mml:mtd class="array">
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mo>⩽</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo>⩽</mml:mo>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd class="array">
<mml:mn>0</mml:mn>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mspace width="1em"/>
<mml:mspace width="1em"/>
</mml:mtd>
<mml:mtd class="array">
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal">&lt;</mml:mo>
<mml:mi mathvariant="italic">a</mml:mi>
<mml:mspace width="2.5pt"/>
<mml:mtext>or</mml:mtext>
<mml:mspace width="2.5pt"/>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal">&gt;</mml:mo>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mfenced>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ f(x)=\left\{\begin{array}{l@{\hskip4.0pt}l}\frac{1}{b-a},\hspace{1em}& a\leqslant x\leqslant b,\\ {} 0,\hspace{1em}\hspace{1em}& x\lt a\hspace{2.5pt}\text{or}\hspace{2.5pt}x\gt b.\end{array}\right.\]]]></tex-math></alternatives>
</disp-formula> 
where <italic>a</italic> and <italic>b</italic> are the lowest and highest value of <italic>x</italic>, respectively.</p>
<p>Magnitude warping for time series data involves the random scaling of certain segments of the data. To perform the deformations, nodes <inline-formula id="j_infor578_ineq_007"><alternatives><mml:math>
<mml:mi mathvariant="italic">u</mml:mi>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">u</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">u</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">u</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub></mml:math><tex-math><![CDATA[$u={u_{1}},{u_{2}},\dots ,{u_{i}}$]]></tex-math></alternatives></inline-formula> are generated randomly from a Gaussian distribution. The scaling is then defined by cubic spline interpolation of the nodes <inline-formula id="j_infor578_ineq_008"><alternatives><mml:math>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$S(x)$]]></tex-math></alternatives></inline-formula> (Iglesias <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor578_ref_012">2023</xref>). The magnitude warping function is represented by the formula: 
<disp-formula id="j_infor578_eq_005">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">α</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
</mml:msubsup>
<mml:mo>=</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">T</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">T</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo fence="true" stretchy="false">}</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {x_{0}^{(\alpha )}}=\{{\alpha _{1}}{x_{1}},\dots ,{\alpha _{t}}{x_{t}},\dots ,{\alpha _{T}}{x_{T}}\},\]]]></tex-math></alternatives>
</disp-formula> 
where <inline-formula id="j_infor578_ineq_009"><alternatives><mml:math>
<mml:mi mathvariant="italic">α</mml:mi>
<mml:mo>=</mml:mo>
<mml:mo fence="true" stretchy="false">{</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mo>…</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">α</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo fence="true" stretchy="false">}</mml:mo>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$\alpha =\{{\alpha _{1}},{\alpha _{2}},\dots ,{\alpha _{i}}\}=S(x)$]]></tex-math></alternatives></inline-formula>, <inline-formula id="j_infor578_ineq_010"><alternatives><mml:math>
<mml:mi mathvariant="italic">S</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[$S(x)$]]></tex-math></alternatives></inline-formula> is cubic spline interpolation of the knots.</p>
<p>TimeVAE architecture is one more of the methods for data augmentation of CB tension signals (Desai <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor578_ref_006">2021</xref>). TimeVAE architecture is trained using the evidence lower bound (ELBO) loss function: 
<disp-formula id="j_infor578_eq_006">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
<mml:mi mathvariant="italic">ϕ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mo>−</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">E</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">q</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">ϕ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">z</mml:mi>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
</mml:msub>
<mml:mo fence="true" maxsize="1.19em" minsize="1.19em">[</mml:mo>
<mml:mo movablelimits="false">log</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">z</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo fence="true" maxsize="1.19em" minsize="1.19em">]</mml:mo>
<mml:mo>+</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">D</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">KL</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">q</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">ϕ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">z</mml:mi>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo maxsize="1.19em" minsize="1.19em" stretchy="true">|</mml:mo>
<mml:mo maxsize="1.19em" minsize="1.19em" stretchy="true">|</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">z</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal" fence="true" maxsize="1.19em" minsize="1.19em">)</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ {L_{\theta ,\phi }}=-{\mathbb{E}_{{q_{\phi }}(z|x)}}\big[\log {p_{\theta }}(x|z)\big]+{D_{\mathrm{KL}}}\big({q_{\phi }}(z|x)\big|\big|{p_{\theta }}(z)\big),\]]]></tex-math></alternatives>
</disp-formula> 
where the first term <inline-formula id="j_infor578_ineq_011"><alternatives><mml:math>
<mml:mo>−</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">E</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">q</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">ϕ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">z</mml:mi>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
</mml:mrow>
</mml:msub>
<mml:mo fence="true" stretchy="false">[</mml:mo>
<mml:mo movablelimits="false">log</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">z</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo fence="true" stretchy="false">]</mml:mo></mml:math><tex-math><![CDATA[$-{\mathbb{E}_{{q_{\phi }}(z|x)}}[\log {p_{\theta }}(x|z)]$]]></tex-math></alternatives></inline-formula> is the reconstruction loss, which measures how accurately the model reconstructs the input data. It includes the log-likelihood for the variable <italic>z</italic> drawn from the distribution <inline-formula id="j_infor578_ineq_012"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">q</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">ϕ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">z</mml:mi>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[${q_{\phi }}(z|x)$]]></tex-math></alternatives></inline-formula>, where <inline-formula id="j_infor578_ineq_013"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">q</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">ϕ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">z</mml:mi>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[${q_{\phi }}(z|x)$]]></tex-math></alternatives></inline-formula> is the encoded latent space for the variable <italic>x</italic>. The second term <inline-formula id="j_infor578_ineq_014"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">D</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">KL</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">q</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">ϕ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">z</mml:mi>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo stretchy="false">|</mml:mo>
<mml:mo stretchy="false">|</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">z</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[${D_{\mathrm{KL}}}({q_{\phi }}(z|x)||{p_{\theta }}(z))$]]></tex-math></alternatives></inline-formula> is the Kullback-Leibler deviation between <inline-formula id="j_infor578_ineq_015"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">q</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">ϕ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">z</mml:mi>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[${q_{\phi }}(z|x)$]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_infor578_ineq_016"><alternatives><mml:math>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="italic">p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">θ</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">z</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo></mml:math><tex-math><![CDATA[${p_{\theta }}(z)$]]></tex-math></alternatives></inline-formula> distributions. This regularization term is meant to ensure that the learned latent space remains similar to the prior distribution. In TimeVAE architecture, the variable <italic>z</italic> is taken from a Gaussian distribution and passed to the decoder, thus making the VAE decoder generative.</p>
<p>The encoder passes the input through a one-dimensional convolutional layer with ReLU activation function. The input is flattened and then connected to the output from the encoder’s fully connected layer. The encoder output parameters are used for constructing Gaussian distribution from which the variable <italic>z</italic> is derived. This variable is then passed to the decoder, which consists of fully connected, convolution, and time-distributed concatenation layers. The output data from the decoder is the same shape as the input data.</p>
</sec>
<sec id="j_infor578_s_005">
<label>2.3</label>
<title>Deep Learning Models Architectures</title>
<p>Multiple DL models were tested in order to ensure the broad applicability of findings, including fully convolutional network (FCN), convolutional neural network combined with long short-term memory network (CNN-LSTM), residual network (ResNet), and inception network (InceptionTime). All the models were built to classify six CB states, which included five load conditions (0.5 kg, 1 kg, 2 kg, 3 kg, and 5 kg) and one damaged-belt condition.</p>
<p>All the models were built using the Python programming language with TensorFlow and Keras DL libraries. The Kaggle platform was used for executing the experiments, and the <italic>NVIDIA Tesla P100</italic> graphics processor was used for model training. In addition, each model was trained using the cross-entropy loss function and the Adam optimizer with a training rate of <inline-formula id="j_infor578_ineq_017"><alternatives><mml:math>
<mml:mi mathvariant="italic">η</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>0.001</mml:mn></mml:math><tex-math><![CDATA[$\eta =0.001$]]></tex-math></alternatives></inline-formula>. The following subsections present detailed architecture and parameter configuration of each constructed model.</p>
<sec id="j_infor578_s_006">
<label>2.3.1</label>
<title>FCN Model</title>
<p>FCN is an architecture based on deep CNNs that were originally developed for image segmentation (Long <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor578_ref_017">2015</xref>). In the case of time series, FCN architecture can be used for feature extraction. In the output layer, the classification can be performed using either the exponential normalization (softmax) or the sigmoid activation function (Wang <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor578_ref_024">2017b</xref>). The basic block of FCN architecture consists of a convolutional layer, followed by a batch normalization layer and a rectified linear unit (ReLU) activation layer. During the training, the batch normalization layer accelerates gradient convergence and improves the model’s robustness. Batch normalization is given by the following formula: 
<disp-formula id="j_infor578_eq_007">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:mi mathvariant="normal">BN</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo>−</mml:mo>
<mml:mi mathvariant="italic">μ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msqrt>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="italic">σ</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="italic">ϵ</mml:mi>
</mml:mrow>
</mml:msqrt>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>×</mml:mo>
<mml:mi mathvariant="italic">γ</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="italic">β</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ \mathrm{BN}(x)=\frac{x-\mu }{\sqrt{{\sigma ^{2}}+\epsilon }}\times \gamma +\beta ,\]]]></tex-math></alternatives>
</disp-formula> 
where <italic>x</italic> is the input, <italic>μ</italic> is the mini-batch mean, <italic>σ</italic> is the mini-batch standard deviation, <italic>ϵ</italic> is the numerical stability constant, <italic>γ</italic> is the learned scale parameter, and <italic>β</italic> is the learned shift parameter.</p>
<p>The mathematical expression of FCN architecture block is given in formulas (<xref rid="j_infor578_eq_008">1</xref>), (<xref rid="j_infor578_eq_009">2</xref>), and (<xref rid="j_infor578_eq_010">3</xref>): <disp-formula-group id="j_infor578_dg_001">
<disp-formula id="j_infor578_eq_008">
<label>(1)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right left" columnspacing="0pt">
<mml:mtr>
<mml:mtd class="align-odd"/>
<mml:mtd class="align-even">
<mml:mi mathvariant="italic">z</mml:mi>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="italic">W</mml:mi>
<mml:mo>∗</mml:mo>
<mml:mi mathvariant="italic">x</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="italic">b</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[\begin{aligned}{}& z=W\ast x+b,\end{aligned}\]]]></tex-math></alternatives>
</disp-formula>
<disp-formula id="j_infor578_eq_009">
<label>(2)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right left" columnspacing="0pt">
<mml:mtr>
<mml:mtd class="align-odd"/>
<mml:mtd class="align-even">
<mml:mi mathvariant="italic">y</mml:mi>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="normal">BN</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">z</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[\begin{aligned}{}& y=\mathrm{BN}(z),\end{aligned}\]]]></tex-math></alternatives>
</disp-formula>
<disp-formula id="j_infor578_eq_010">
<label>(3)</label><alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right left" columnspacing="0pt">
<mml:mtr>
<mml:mtd class="align-odd"/>
<mml:mtd class="align-even">
<mml:mi mathvariant="italic">h</mml:mi>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="normal">ReLU</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">(</mml:mo>
<mml:mi mathvariant="italic">y</mml:mi>
<mml:mo mathvariant="normal" fence="true" stretchy="false">)</mml:mo>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[\begin{aligned}{}& h=\mathrm{ReLU}(y),\end{aligned}\]]]></tex-math></alternatives>
</disp-formula>
</disp-formula-group> where ∗ denotes the convolution operation, <italic>x</italic> is the input data, <italic>W</italic> is the convolution layer kernel, <italic>b</italic> is the bias, BN denotes the batch normalization operation, and ReLU is rectified linear unit operation.</p>
<p>The final FCN is formed by concatenating three convolutional blocks. After applying these blocks, the extracted data features are passed to the global average pooling (GAP) layer, which is responsible for reducing the feature map size (Hsiao <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor578_ref_010">2019</xref>). GAP layer is superior to the traditional fully connected layer because it significantly reduces the number of weights and helps the model avoid over-training. The final layer consists of the softmax or sigmoid activation function.</p>
<p>The architecture of the built FCN model is shown in Fig. <xref rid="j_infor578_fig_006">6</xref>. The developed model is composed of three one-dimensional convolutional layers with the number of filters 64, 128, and 64, respectively. The filter size of each convolutional layer was <inline-formula id="j_infor578_ineq_018"><alternatives><mml:math>
<mml:mn>3</mml:mn>
<mml:mo>×</mml:mo>
<mml:mn>1</mml:mn></mml:math><tex-math><![CDATA[$3\times 1$]]></tex-math></alternatives></inline-formula>. The selection of 64, 128, and 64 filters was made based on findings in previous studies that demonstrated good accuracy in time series classification tasks using similar FCN model (Wang <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor578_ref_024">2017b</xref>). The final layers consist of a GAP layer and a fully connected layer with a softmax activation function. The resulting FCN model was the smallest among all the models built in this study as it consisted of only 26437 trainable parameters.</p>
<fig id="j_infor578_fig_006">
<label>Fig. 6</label>
<caption>
<p>FCN model architecture.</p>
</caption>
<graphic xlink:href="infor578_g006.jpg"/>
</fig>
</sec>
<sec id="j_infor578_s_007">
<label>2.3.2</label>
<title>CNN-LSTM Model</title>
<p>The architecture and parameter configuration of the second hybrid CNN-LSTM model is shown in Fig. <xref rid="j_infor578_fig_007">7</xref>. The developed model is composed of four different one-dimensional convolutional blocks and two LSTM layers. The number of convolutional block filters in the model decreases from 512 to 8, respectively. Convolutional blocks use batch normalization, ReLU activation function, average pooling, and dropout layers. The number of convolutional filters was empirically tested with various configurations, and the chosen structure provided optimal classification accuracy. The convolutional blocks are followed by two LSTM layers, each composed of 16 cells. The final layer is fully connected and utilizes the softmax activation function. The trainable parameters of the created CNN-LSTM model ranged from 226.077 with 0.5 s-length signals to 245.533 with 2.0 s-length signals.</p>
<fig id="j_infor578_fig_007">
<label>Fig. 7</label>
<caption>
<p>CNN-LSTM model architecture.</p>
</caption>
<graphic xlink:href="infor578_g007.jpg"/>
</fig>
</sec>
<sec id="j_infor578_s_008">
<label>2.3.3</label>
<title>ResNet Model</title>
<p>The architecture and parameter configuration of ResNet model is shown in Fig. <xref rid="j_infor578_fig_008">8</xref>. The model consists of three residual blocks. In the first residual block, each of one-dimensional convolutions has 64 filters and in the second and third, 128. The selected filter sizes were based on previous studies that applied ResNet architecture for time series classification (Wang <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor578_ref_024">2017b</xref>). In the architecture diagram of the model, arrows with a plus sign represent skip connections. The last two layers of the model consist of a GAP layer and a fully connected layer with a softmax activation function. The developed ResNet model had the largest number of parameters among all the built models and consisted of 508.357 trainable parameters.</p>
</sec>
<sec id="j_infor578_s_009">
<label>2.3.4</label>
<title>InceptionTime Model</title>
<p>The architecture and parameter configuration of the InceptionTime model with six inception modules is shown in Fig. <xref rid="j_infor578_fig_009">9</xref>. Every third module is connected by ResNet skip connections. The final two layers of the model include a GAP layer and a fully connected layer with a softmax activation function. The InceptionTime model has a total of 427.685 trainable parameters.</p>
<fig id="j_infor578_fig_008">
<label>Fig. 8</label>
<caption>
<p>ResNet model architecture.</p>
</caption>
<graphic xlink:href="infor578_g008.jpg"/>
</fig>
<fig id="j_infor578_fig_009">
<label>Fig. 9</label>
<caption>
<p>Inception network InceptionTime model architecture.</p>
</caption>
<graphic xlink:href="infor578_g009.jpg"/>
</fig>
</sec>
</sec>
<sec id="j_infor578_s_010">
<label>2.4</label>
<title>Classification Accuracy Metrics</title>
<p>Overall classification accuracy measures overall accuracy of the model. It is the ratio of correct guesses to all guesses. The overall accuracy in percent is represented by the formula: 
<disp-formula id="j_infor578_eq_011">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:mi mathvariant="normal">Accuracy</mml:mi>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="normal">TP</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="normal">TN</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">TP</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="normal">FN</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="normal">FP</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="normal">TN</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>·</mml:mo>
<mml:mn>100</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ \mathrm{Accuracy}=\frac{\mathrm{TP}+\mathrm{TN}}{\mathrm{TP}+\mathrm{FN}+\mathrm{FP}+\mathrm{TN}}\cdot 100\% ,\]]]></tex-math></alternatives>
</disp-formula> 
where TP represents true positives, TN represents true negatives, FP represents false positives, and FN represents false negative model results.</p>
<p>The accuracy of each experiment was measured more than once, so it is important to estimate the error in order to evaluate the data from these samples of different accuracy. By calculating the standard error (SE) of the mean, it is possible to assess the extent to which the sample is representative of the population and draw reasonable conclusions from it. SE was calculated by the formula: 
<disp-formula id="j_infor578_eq_012">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:mi mathvariant="normal">SE</mml:mi>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="italic">s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msqrt>
<mml:mrow>
<mml:mi mathvariant="italic">n</mml:mi>
</mml:mrow>
</mml:msqrt>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ \mathrm{SE}=\frac{s}{\sqrt{n}},\]]]></tex-math></alternatives>
</disp-formula> 
where <italic>s</italic> is the sample standard deviation and <italic>n</italic> is the sample size.</p>
<p>Classification accuracy alone can lead to misinterpretation of results when the dataset is unbalanced. For this reason, depending on the classification results of the classification model, recall and precision metrics of the model can also be calculated. These metrics are represented as percentages: 
<disp-formula id="j_infor578_eq_013">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true" columnalign="right left" columnspacing="0pt">
<mml:mtr>
<mml:mtd class="align-odd"/>
<mml:mtd class="align-even">
<mml:mi mathvariant="normal">Recall</mml:mi>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="normal">TP</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">TP</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="normal">FN</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>·</mml:mo>
<mml:mn>100</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd class="align-odd"/>
<mml:mtd class="align-even">
<mml:mi mathvariant="normal">Precision</mml:mi>
<mml:mo>=</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="normal">TP</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">TP</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="normal">FP</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>·</mml:mo>
<mml:mn>100</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[\begin{aligned}{}& \mathrm{Recall}=\frac{\mathrm{TP}}{\mathrm{TP}+\mathrm{FN}}\cdot 100\% ,\\ {} & \mathrm{Precision}=\frac{\mathrm{TP}}{\mathrm{TP}+\mathrm{FP}}\cdot 100\% ,\end{aligned}\]]]></tex-math></alternatives>
</disp-formula> 
where recall shows whether the model can predict all the instances of different classes and precision shows how often the positive predictions are correct.</p>
<p>In addition, F1-score, which is a weighted average of recall and precision, can be used to determine the classification accuracy in the case of unbalanced datasets. This metric is expressed as a percentage in the formula: 
<disp-formula id="j_infor578_eq_014">
<alternatives><mml:math display="block">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:mi mathvariant="normal">F</mml:mi>
<mml:mn mathvariant="normal">1</mml:mn>
<mml:mo>=</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo>·</mml:mo><mml:mstyle displaystyle="true">
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="normal">Recall</mml:mi>
<mml:mo>·</mml:mo>
<mml:mi mathvariant="normal">Precision</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">Recall</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="normal">Precision</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mstyle>
<mml:mo>·</mml:mo>
<mml:mn>100</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi>
<mml:mo mathvariant="normal">,</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable></mml:math><tex-math><![CDATA[\[ \mathrm{F}\mathrm{1}=2\cdot \frac{\mathrm{Recall}\cdot \mathrm{Precision}}{\mathrm{Recall}+\mathrm{Precision}}\cdot 100\% ,\]]]></tex-math></alternatives>
</disp-formula> 
where Recall is the proportion of actual positives that were correctly identified and Precision is the proportion of predictions that were correct.</p>
</sec>
</sec>
<sec id="j_infor578_s_011">
<label>3</label>
<title>Results and Discussion</title>
<p>In this section, a study on the accomplished classification of CB load and defect states is provided. First, DNN models were built. Various time series data augmentation methods were then applied to the dataset and comparisons of the original and augmented signals were made. Finally, a study of conveyor state classification was carried out by using the constructed DNN models and various data augmentation methods. The results of all the studied methods were compared and the conclusions were drawn. The DNN models, augmentation methods, and data used in the study are available in the GitHub repository.<xref ref-type="fn" rid="j_infor578_fn_001">1</xref><fn id="j_infor578_fn_001"><label><sup>1</sup></label>
<p><uri>https://github.com/ArmantasPik/Conveyor-belt-state-classification/</uri></p></fn></p>
<sec id="j_infor578_s_012">
<label>3.1</label>
<title>Application of Data Augmentations</title>
<p>Data augmentation was performed on the CB tension signals before analysing the CB state classification accuracy. All the applied data augmentation methods are presented in this section.</p>
<table-wrap id="j_infor578_tab_001">
<label>Table 1</label>
<caption>
<p>Sliding window signals.</p>
</caption>
<table>
<thead>
<tr>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Signal length (s)</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Data points</td>
<td style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Signal count</td>
</tr>
</thead>
<tbody>
<tr>
<td style="vertical-align: top; text-align: left">0.5</td>
<td style="vertical-align: top; text-align: left">200</td>
<td style="vertical-align: top; text-align: left">1.058</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">1.0</td>
<td style="vertical-align: top; text-align: left">400</td>
<td style="vertical-align: top; text-align: left">514</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">1.5</td>
<td style="vertical-align: top; text-align: left">600</td>
<td style="vertical-align: top; text-align: left">319</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">2.0</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">800</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">233</td>
</tr>
</tbody>
</table>
</table-wrap>
<sec id="j_infor578_s_013">
<label>3.1.1</label>
<title>Sliding Window</title>
<p>A sliding window was applied to the raw tension signals to generate datasets of four different lengths: 0.5 s (200 points), 1.0 s (400 points), 1.5 s (600 points), and 2.0 s (800 points). The step size of each sliding window was 100% of the signal’s length. The number of signals for each length is presented in Table <xref rid="j_infor578_tab_001">1</xref>. The shortest-signal (0.5 s) dataset was the biggest (1.058), while that of the longest signal (2.0 s) was the smallest (233). These signals of different lengths were used in further time series data augmentation and classification accuracy studies. Additionally, the signals were normalized before applying data augmentations.</p>
<fig id="j_infor578_fig_010">
<label>Fig. 10</label>
<caption>
<p>Comparison of different scale random Laplace noise with the original signal.</p>
</caption>
<graphic xlink:href="infor578_g010.jpg"/>
</fig>
</sec>
<sec id="j_infor578_s_014">
<label>3.1.2</label>
<title>Noise Addition</title>
<p>Various types of noise such as random Laplace, drifted Gaussian, or uniform noises were added to the processed tension signal data. The scale of the noise was determined by the standard deviation (std) of the original signal. The different scales of noise were generated by scaling the standard deviation of the original signal to fractions of its value: std/100, std/50, and std/20. This scaling adjusts the noise magnitude, with smaller proportions (e.g., std/100) representing lower noise levels and larger proportions (e.g., std/20) introducing higher noise, all in proportion to the original signal’s variability. Figure <xref rid="j_infor578_fig_010">10</xref> presents a comparison between the original tension signal and the one augmented with random Laplace noise at different scales. As can be seen, the signal becomes progressively noisier as the noise scale increases from std/100 to std/20, demonstrating how higher magnitudes introduce more fluctuations while preserving the overall structure of the original signal.</p>
<fig id="j_infor578_fig_011">
<label>Fig. 11</label>
<caption>
<p>Comparison of random Laplace and drifted Gaussian noise combinations with the original signal.</p>
</caption>
<graphic xlink:href="infor578_g011.jpg"/>
</fig>
<p>Noise addition data augmentations were performed by using various noise combinations. Random Laplace noise was combined with drifted Gaussian noise and uniform noise, each with a magnitude of std/100. Figure <xref rid="j_infor578_fig_011">11</xref> compares the original signal with the combination of random Laplace and drifted Gaussian noise. The use of both Laplace and drifted Gaussian noise introduces a combination of sharp, sudden deviations (characteristic for Laplace noise) and smoother, gradual deviations (characteristic for Gaussian noise). The comparison shows that the augmented signal retains the major trends of the original signal, while the addition of noise results in slight amplitude variations.</p>
</sec>
<sec id="j_infor578_s_015">
<label>3.1.3</label>
<title>Magnitude Warping</title>
<p>Magnitude warping was applied to the processed tension signals. Deformation parameter <inline-formula id="j_infor578_ineq_019"><alternatives><mml:math>
<mml:mi mathvariant="italic">σ</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>0.2</mml:mn></mml:math><tex-math><![CDATA[$\sigma =0.2$]]></tex-math></alternatives></inline-formula> and 4 nodes were used for the deformation function. The comparison of the original and deformed signal is shown in Fig. <xref rid="j_infor578_fig_012">12</xref>. The warped signal showed noticeable changes in amplitude as compared to the original signal. Despite these amplitude modifications, the main features and overall trends of the original signal remained intact, making this method suitable for creating variations of the original data without altering the core structure.</p>
<fig id="j_infor578_fig_012">
<label>Fig. 12</label>
<caption>
<p>Comparison of magnitude warping with the original signal.</p>
</caption>
<graphic xlink:href="infor578_g012.jpg"/>
</fig>
</sec>
<sec id="j_infor578_s_016">
<label>3.1.4</label>
<title>Variational Autoencoder Signal Generation</title>
<fig id="j_infor578_fig_013">
<label>Fig. 13</label>
<caption>
<p>Comparison of original and reconstructed signals.</p>
</caption>
<graphic xlink:href="infor578_g013.jpg"/>
</fig>
<p>Processed tension signals were used to train TimeVAE model with three hidden layers containing 50, 100, and 200 neurons, respectively, and 20 neurons in the latent space. The selection of 50, 100, and 200 neurons for the hidden layers was based on the architecture proposed in the original study on variational autoencoders for time series generation (Desai <italic>et al.</italic>, <xref ref-type="bibr" rid="j_infor578_ref_006">2021</xref>). These parameters were empirically tested to increase classification accuracy on the conveyor belt dataset by minimizing reconstruction loss. Similarly, for the latent space, 20 neurons were selected as it provided sufficient compression of the signals while maintaining important features. The model was trained for 200 epochs. Proper parameter configuration was crucial as the error of reconstructed and generated signals depended on these parameters. After training, random new signals were generated for classification accuracy study.</p>
<p>The comparison of the original signals and the signals reconstructed by VAE is shown in Fig. <xref rid="j_infor578_fig_013">13</xref> where the signals are converted into 2D spectrograms. As the graph shows, reconstructed data almost perfectly matches the original data.</p>
<p>A comparison of original and randomly generated signal is presented in Fig. <xref rid="j_infor578_fig_014">14</xref>. It can be seen that the amplitude and frequency of the generated signal are similar to that of the real signal, but the generated signal has less noise.</p>
<fig id="j_infor578_fig_014">
<label>Fig. 14</label>
<caption>
<p>CB tension signals: a) original signal; b) TimeVAE generated signal.</p>
</caption>
<graphic xlink:href="infor578_g014.jpg"/>
</fig>
</sec>
</sec>
<sec id="j_infor578_s_017">
<label>3.2</label>
<title>Classification Accuracy Study</title>
<p>The study on the classification accuracy of CB states was carried out in two stages. In the first stage, different DNN models were assessed by classifying the signals processed by a sliding window of varying lengths (0.5 s, 1.0 s, 1.5 s, 2.0 s) without applying any additional data augmentation methods. The second stage focused on examining the impact of different data augmentation methods on the classification accuracy of DNN models, specifically using 2.0 s-length signals. The 2.0 s-length signals were selected for this stage because their dataset was the smallest.</p>
<sec id="j_infor578_s_018">
<label>3.2.1</label>
<title>DNN Models’ Study</title>
<p>For each signal length (0.5 s, 1.0 s, 1.5 s, 2.0 s), the CB dataset was divided into five parts, ensuring class balance in each fold. Balanced 5-fold cross-validation was used to construct the training and testing sets. This approach involved using 80% of the data (four out of five folds) for training and 20% (the remaining fold) for testing. Training and testing sets were switched five times for each signal length. The final accuracy was calculated by averaging the results of these five training runs and determining the standard error of the mean. This method was chosen to provide a more reliable estimate of the model classification accuracy, particularly given the small dataset sizes.</p>
<table-wrap id="j_infor578_tab_002">
<label>Table 2</label>
<caption>
<p>Augmentation-free multi-class (0.5 kg, 1 kg, 2 kg, 3 kg, 5 kg, and a class for a damaged belt) classification accuracy of CB states of DNN models.</p>
</caption>
<table>
<thead>
<tr>
<td rowspan="2" style="vertical-align: middle; text-align: left; border-top: solid thin; border-bottom: solid thin">Signal length, s</td>
<td colspan="4" style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Accuracy ± SE, %</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">CNN-LSTM</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">FCN</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">ResNet</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">InceptionTime</td>
</tr>
</thead>
<tbody>
<tr>
<td style="vertical-align: top; text-align: left">0.5</td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_020"><alternatives><mml:math>
<mml:mn>74.3</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>0.82</mml:mn></mml:math><tex-math><![CDATA[$74.3\pm 0.82$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_021"><alternatives><mml:math>
<mml:mn>76.4</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.14</mml:mn></mml:math><tex-math><![CDATA[$76.4\pm 1.14$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_022"><alternatives><mml:math>
<mml:mn>80.6</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.27</mml:mn></mml:math><tex-math><![CDATA[$80.6\pm 1.27$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_023"><alternatives><mml:math>
<mml:mn>79.3</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>0.71</mml:mn></mml:math><tex-math><![CDATA[$79.3\pm 0.71$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">1.0</td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_024"><alternatives><mml:math>
<mml:mn>65.5</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.08</mml:mn></mml:math><tex-math><![CDATA[$65.5\pm 1.08$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_025"><alternatives><mml:math>
<mml:mn>81.5</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.29</mml:mn></mml:math><tex-math><![CDATA[$81.5\pm 1.29$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_026"><alternatives><mml:math>
<mml:mn>85.4</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.26</mml:mn></mml:math><tex-math><![CDATA[$85.4\pm 1.26$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_027"><alternatives><mml:math>
<mml:mn>84.9</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>0.59</mml:mn></mml:math><tex-math><![CDATA[$84.9\pm 0.59$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">1.5</td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_028"><alternatives><mml:math>
<mml:mn>66.3</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>2.47</mml:mn></mml:math><tex-math><![CDATA[$66.3\pm 2.47$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_029"><alternatives><mml:math>
<mml:mn>83.7</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.51</mml:mn></mml:math><tex-math><![CDATA[$83.7\pm 1.51$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_030"><alternatives><mml:math>
<mml:mn>84.8</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.09</mml:mn></mml:math><tex-math><![CDATA[$84.8\pm 1.09$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_031"><alternatives><mml:math>
<mml:mn>84.8</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.52</mml:mn></mml:math><tex-math><![CDATA[$84.8\pm 1.52$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">2.0</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><inline-formula id="j_infor578_ineq_032"><alternatives><mml:math>
<mml:mn>64.3</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.58</mml:mn></mml:math><tex-math><![CDATA[$64.3\pm 1.58$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><inline-formula id="j_infor578_ineq_033"><alternatives><mml:math>
<mml:mn>88.1</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.48</mml:mn></mml:math><tex-math><![CDATA[$88.1\pm 1.48$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><inline-formula id="j_infor578_ineq_034"><alternatives><mml:math>
<mml:mn>91.1</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.09</mml:mn></mml:math><tex-math><![CDATA[$91.1\pm 1.09$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><inline-formula id="j_infor578_ineq_035"><alternatives><mml:math>
<mml:mn>89.1</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.54</mml:mn></mml:math><tex-math><![CDATA[$89.1\pm 1.54$]]></tex-math></alternatives></inline-formula></td>
</tr>
</tbody>
</table>
</table-wrap>
<p>A total of 6 different classes were assigned to the tension signals based on their load and defect states: 0.5 kg, 1 kg, 2 kg, 3 kg, 5 kg, and a class for a belt with damage. A categorical cross-entropy loss function was used for all classification models, with a softmax activation function applied to the output layer. Each model was trained five times for 300 epochs per experiment, using a batch size of 16 signals per iteration. Before training, all training sets of different lengths were aligned to ensure that each set consisted of 2000 signals. To achieve this, the training sets in the first stage were multiplied without augmentations to reach the required 2000 signals.</p>
<p>The classification accuracy results of CB states in the first stage, using various DNN models without augmentations, are presented in Table <xref rid="j_infor578_tab_002">2</xref>. According to the results, ResNet model outperformed other models in all signal lengths and had the highest classification accuracy, which reached <inline-formula id="j_infor578_ineq_036"><alternatives><mml:math>
<mml:mn>91.1</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi>
<mml:mo>±</mml:mo>
<mml:mn>1.09</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$91.1\% \pm 1.09\% $]]></tex-math></alternatives></inline-formula> with 2.0 s-length signals. ResNet and InceptionTime models have demonstrated similar accuracies for all the signal lengths, differing by only one or two percentage points. FCN model also performed competitively, achieving an accuracy of <inline-formula id="j_infor578_ineq_037"><alternatives><mml:math>
<mml:mn>88.1</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi>
<mml:mo>±</mml:mo>
<mml:mn>1.48</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$88.1\% \pm 1.48\% $]]></tex-math></alternatives></inline-formula> with 2.0 s-length signals, which was only slightly lower than that of ResNet and InceptionTime models. The accuracy of CNN-LSTM model decreased in line with increasing signal length, indicating that it performed comparatively poorly for this classification task. In addition, the accuracy of all other models improved with increasing signal length, except of CNN-LSTM model.</p>
<table-wrap id="j_infor578_tab_003">
<label>Table 3</label>
<caption>
<p>Augmentation-free multi-class (0.5 kg, 1 kg, 2 kg, 3 kg, 5 kg, and a class for a damaged belt) classification recall of CB states of DNN models for each class.</p>
</caption>
<table>
<thead>
<tr>
<td rowspan="2" style="vertical-align: middle; text-align: left; border-top: solid thin; border-bottom: solid thin">Model</td>
<td rowspan="2" style="vertical-align: middle; text-align: left; border-top: solid thin; border-bottom: solid thin">Signal length, s</td>
<td colspan="6" style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Recall, %</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">0.5 kg</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">1 kg</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">2 kg</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">3 kg</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">5 kg</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">With damage</td>
</tr>
</thead>
<tbody>
<tr>
<td style="vertical-align: top; text-align: left">CNN-LSTM</td>
<td style="vertical-align: top; text-align: left">0.5</td>
<td style="vertical-align: top; text-align: left">71.6</td>
<td style="vertical-align: top; text-align: left">61.0</td>
<td style="vertical-align: top; text-align: left">51.6</td>
<td style="vertical-align: top; text-align: left">64.2</td>
<td style="vertical-align: top; text-align: left">74.3</td>
<td style="vertical-align: top; text-align: left">98.2</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">1.0</td>
<td style="vertical-align: top; text-align: left">59.8</td>
<td style="vertical-align: top; text-align: left">28.6</td>
<td style="vertical-align: top; text-align: left">42.8</td>
<td style="vertical-align: top; text-align: left">65.5</td>
<td style="vertical-align: top; text-align: left">69.4</td>
<td style="vertical-align: top; text-align: left">96.8</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">1.5</td>
<td style="vertical-align: top; text-align: left">69.4</td>
<td style="vertical-align: top; text-align: left">38.0</td>
<td style="vertical-align: top; text-align: left">47.2</td>
<td style="vertical-align: top; text-align: left">58.6</td>
<td style="vertical-align: top; text-align: left">66.3</td>
<td style="vertical-align: top; text-align: left">94.0</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">2.0</td>
<td style="vertical-align: top; text-align: left">60.8</td>
<td style="vertical-align: top; text-align: left">43.4</td>
<td style="vertical-align: top; text-align: left">43.6</td>
<td style="vertical-align: top; text-align: left">57.0</td>
<td style="vertical-align: top; text-align: left">66.9</td>
<td style="vertical-align: top; text-align: left">96.8</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">FCN</td>
<td style="vertical-align: top; text-align: left">0.5</td>
<td style="vertical-align: top; text-align: left">77.2</td>
<td style="vertical-align: top; text-align: left">62.6</td>
<td style="vertical-align: top; text-align: left">56.2</td>
<td style="vertical-align: top; text-align: left">64.8</td>
<td style="vertical-align: top; text-align: left">76.4</td>
<td style="vertical-align: top; text-align: left">98.6</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">1.0</td>
<td style="vertical-align: top; text-align: left">78.8</td>
<td style="vertical-align: top; text-align: left">84.0</td>
<td style="vertical-align: top; text-align: left">51.6</td>
<td style="vertical-align: top; text-align: left">76.0</td>
<td style="vertical-align: top; text-align: left">81.5</td>
<td style="vertical-align: top; text-align: left">100</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">1.5</td>
<td style="vertical-align: top; text-align: left">95.6</td>
<td style="vertical-align: top; text-align: left">64.4</td>
<td style="vertical-align: top; text-align: left">69.4</td>
<td style="vertical-align: top; text-align: left">72.4</td>
<td style="vertical-align: top; text-align: left">81.6</td>
<td style="vertical-align: top; text-align: left">97.6</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">2.0</td>
<td style="vertical-align: top; text-align: left">96.6</td>
<td style="vertical-align: top; text-align: left">80.0</td>
<td style="vertical-align: top; text-align: left">70.0</td>
<td style="vertical-align: top; text-align: left">88.8</td>
<td style="vertical-align: top; text-align: left">90.4</td>
<td style="vertical-align: top; text-align: left">98.4</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">ResNet</td>
<td style="vertical-align: top; text-align: left">0.5</td>
<td style="vertical-align: top; text-align: left">83.8</td>
<td style="vertical-align: top; text-align: left">72.6</td>
<td style="vertical-align: top; text-align: left">55.6</td>
<td style="vertical-align: top; text-align: left">73.2</td>
<td style="vertical-align: top; text-align: left">80.6</td>
<td style="vertical-align: top; text-align: left">99.2</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">1.0</td>
<td style="vertical-align: top; text-align: left">91.6</td>
<td style="vertical-align: top; text-align: left">75.4</td>
<td style="vertical-align: top; text-align: left">61.0</td>
<td style="vertical-align: top; text-align: left">85.2</td>
<td style="vertical-align: top; text-align: left">85.4</td>
<td style="vertical-align: top; text-align: left">100</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">1.5</td>
<td style="vertical-align: top; text-align: left">88.4</td>
<td style="vertical-align: top; text-align: left">75.2</td>
<td style="vertical-align: top; text-align: left">67.6</td>
<td style="vertical-align: top; text-align: left">81.6</td>
<td style="vertical-align: top; text-align: left">85.9</td>
<td style="vertical-align: top; text-align: left">100</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">2.0</td>
<td style="vertical-align: top; text-align: left">100</td>
<td style="vertical-align: top; text-align: left">79.0</td>
<td style="vertical-align: top; text-align: left">72.2</td>
<td style="vertical-align: top; text-align: left">91.1</td>
<td style="vertical-align: top; text-align: left">97.2</td>
<td style="vertical-align: top; text-align: left">100</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">InceptionTime</td>
<td style="vertical-align: top; text-align: left">0.5</td>
<td style="vertical-align: top; text-align: left">85.8</td>
<td style="vertical-align: top; text-align: left">64.4</td>
<td style="vertical-align: top; text-align: left">57.8</td>
<td style="vertical-align: top; text-align: left">69.4</td>
<td style="vertical-align: top; text-align: left">79.3</td>
<td style="vertical-align: top; text-align: left">99.2</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">1.0</td>
<td style="vertical-align: top; text-align: left">88.6</td>
<td style="vertical-align: top; text-align: left">82.6</td>
<td style="vertical-align: top; text-align: left">64.6</td>
<td style="vertical-align: top; text-align: left">74.6</td>
<td style="vertical-align: top; text-align: left">84.9</td>
<td style="vertical-align: top; text-align: left">100</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">1.5</td>
<td style="vertical-align: top; text-align: left">93.2</td>
<td style="vertical-align: top; text-align: left">68.3</td>
<td style="vertical-align: top; text-align: left">75.0</td>
<td style="vertical-align: top; text-align: left">82.0</td>
<td style="vertical-align: top; text-align: left">84.8</td>
<td style="vertical-align: top; text-align: left">100</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"/>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">2.0</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">93.4</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">85.0</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">66.2</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">90.1</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">91.6</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">100</td>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap id="j_infor578_tab_004">
<label>Table 4</label>
<caption>
<p>Augmentation-free multi-class (0.5 kg, 1 kg, 2 kg, 3 kg, 5 kg, and a class for a damaged belt) classification precision of CB states of DNN models for each class.</p>
</caption>
<table>
<thead>
<tr>
<td rowspan="2" style="vertical-align: middle; text-align: left; border-top: solid thin; border-bottom: solid thin">Model</td>
<td rowspan="2" style="vertical-align: middle; text-align: left; border-top: solid thin; border-bottom: solid thin">Signal length, s</td>
<td colspan="6" style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Precision, %</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">0.5 kg</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">1 kg</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">2 kg</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">3 kg</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">5 kg</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">With damage</td>
</tr>
</thead>
<tbody>
<tr>
<td style="vertical-align: top; text-align: left">CNN-LSTM</td>
<td style="vertical-align: top; text-align: left">0.5</td>
<td style="vertical-align: top; text-align: left">66.4</td>
<td style="vertical-align: top; text-align: left">55.0</td>
<td style="vertical-align: top; text-align: left">53.8</td>
<td style="vertical-align: top; text-align: left">74.8</td>
<td style="vertical-align: top; text-align: left">75.0</td>
<td style="vertical-align: top; text-align: left">98.8</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">1.0</td>
<td style="vertical-align: top; text-align: left">49.6</td>
<td style="vertical-align: top; text-align: left">32.2</td>
<td style="vertical-align: top; text-align: left">48.0</td>
<td style="vertical-align: top; text-align: left">66.4</td>
<td style="vertical-align: top; text-align: left">72.4</td>
<td style="vertical-align: top; text-align: left">98.0</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">1.5</td>
<td style="vertical-align: top; text-align: left">59.4</td>
<td style="vertical-align: top; text-align: left">44.6</td>
<td style="vertical-align: top; text-align: left">39.4</td>
<td style="vertical-align: top; text-align: left">65.8</td>
<td style="vertical-align: top; text-align: left">67.5</td>
<td style="vertical-align: top; text-align: left">100</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">2.0</td>
<td style="vertical-align: top; text-align: left">51.8</td>
<td style="vertical-align: top; text-align: left">40.6</td>
<td style="vertical-align: top; text-align: left">41.8</td>
<td style="vertical-align: top; text-align: left">66.3</td>
<td style="vertical-align: top; text-align: left">68.6</td>
<td style="vertical-align: top; text-align: left">100</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">FCN</td>
<td style="vertical-align: top; text-align: left">0.5</td>
<td style="vertical-align: top; text-align: left">67.6</td>
<td style="vertical-align: top; text-align: left">63.8</td>
<td style="vertical-align: top; text-align: left">56.6</td>
<td style="vertical-align: top; text-align: left">72.6</td>
<td style="vertical-align: top; text-align: left">76.7</td>
<td style="vertical-align: top; text-align: left">99.0</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">1.0</td>
<td style="vertical-align: top; text-align: left">77.2</td>
<td style="vertical-align: top; text-align: left">66.0</td>
<td style="vertical-align: top; text-align: left">67.0</td>
<td style="vertical-align: top; text-align: left">83.2</td>
<td style="vertical-align: top; text-align: left">88.8</td>
<td style="vertical-align: top; text-align: left">100</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">1.5</td>
<td style="vertical-align: top; text-align: left">77.2</td>
<td style="vertical-align: top; text-align: left">78.2</td>
<td style="vertical-align: top; text-align: left">69.4</td>
<td style="vertical-align: top; text-align: left">84.0</td>
<td style="vertical-align: top; text-align: left">90.2</td>
<td style="vertical-align: top; text-align: left">100</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">2.0</td>
<td style="vertical-align: top; text-align: left">87.0</td>
<td style="vertical-align: top; text-align: left">79.4</td>
<td style="vertical-align: top; text-align: left">83.8</td>
<td style="vertical-align: top; text-align: left">88.1</td>
<td style="vertical-align: top; text-align: left">95.0</td>
<td style="vertical-align: top; text-align: left">98.6</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">ResNet</td>
<td style="vertical-align: top; text-align: left">0.5</td>
<td style="vertical-align: top; text-align: left">73.8</td>
<td style="vertical-align: top; text-align: left">71.8</td>
<td style="vertical-align: top; text-align: left">66.8</td>
<td style="vertical-align: top; text-align: left">73.6</td>
<td style="vertical-align: top; text-align: left">80.8</td>
<td style="vertical-align: top; text-align: left">98.4</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">1.0</td>
<td style="vertical-align: top; text-align: left">80.2</td>
<td style="vertical-align: top; text-align: left">75.0</td>
<td style="vertical-align: top; text-align: left">72.6</td>
<td style="vertical-align: top; text-align: left">85.4</td>
<td style="vertical-align: top; text-align: left">85.5</td>
<td style="vertical-align: top; text-align: left">100</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">1.5</td>
<td style="vertical-align: top; text-align: left">77.6</td>
<td style="vertical-align: top; text-align: left">70.2</td>
<td style="vertical-align: top; text-align: left">76.8</td>
<td style="vertical-align: top; text-align: left">85.9</td>
<td style="vertical-align: top; text-align: left">92.0</td>
<td style="vertical-align: top; text-align: left">100</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">2.0</td>
<td style="vertical-align: top; text-align: left">97.2</td>
<td style="vertical-align: top; text-align: left">83.8</td>
<td style="vertical-align: top; text-align: left">78.4</td>
<td style="vertical-align: top; text-align: left">91.7</td>
<td style="vertical-align: top; text-align: left">93.2</td>
<td style="vertical-align: top; text-align: left">100</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">InceptionTime</td>
<td style="vertical-align: top; text-align: left">0.5</td>
<td style="vertical-align: top; text-align: left">69.6</td>
<td style="vertical-align: top; text-align: left">73.0</td>
<td style="vertical-align: top; text-align: left">62.6</td>
<td style="vertical-align: top; text-align: left">72.0</td>
<td style="vertical-align: top; text-align: left">79.6</td>
<td style="vertical-align: top; text-align: left">99.6</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">1.0</td>
<td style="vertical-align: top; text-align: left">77.8</td>
<td style="vertical-align: top; text-align: left">86.2</td>
<td style="vertical-align: top; text-align: left">67.2</td>
<td style="vertical-align: top; text-align: left">82.2</td>
<td style="vertical-align: top; text-align: left">85.5</td>
<td style="vertical-align: top; text-align: left">100</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">1.5</td>
<td style="vertical-align: top; text-align: left">73.6</td>
<td style="vertical-align: top; text-align: left">81.8</td>
<td style="vertical-align: top; text-align: left">74.5</td>
<td style="vertical-align: top; text-align: left">85.6</td>
<td style="vertical-align: top; text-align: left">94.0</td>
<td style="vertical-align: top; text-align: left">100</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"/>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">2.0</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">95.6</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">71.6</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">78.6</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">89.1</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">97.2</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">100</td>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap id="j_infor578_tab_005">
<label>Table 5</label>
<caption>
<p>Augmentation-free multi-class (0.5 kg, 1 kg, 2 kg, 3 kg, 5 kg, and a class for a damaged belt) classification F1-score of CB states of DNN models for each class.</p>
</caption>
<table>
<thead>
<tr>
<td rowspan="2" style="vertical-align: middle; text-align: left; border-top: solid thin; border-bottom: solid thin">Model</td>
<td rowspan="2" style="vertical-align: middle; text-align: left; border-top: solid thin; border-bottom: solid thin">Signal length, s</td>
<td colspan="6" style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">F1-score, %</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">0.5 kg</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">1 kg</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">2 kg</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">3 kg</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">5 kg</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">With damage</td>
</tr>
</thead>
<tbody>
<tr>
<td style="vertical-align: top; text-align: left">CNN-LSTM</td>
<td style="vertical-align: top; text-align: left">0.5</td>
<td style="vertical-align: top; text-align: left">68.6</td>
<td style="vertical-align: top; text-align: left">57.6</td>
<td style="vertical-align: top; text-align: left">52.6</td>
<td style="vertical-align: top; text-align: left">68.6</td>
<td style="vertical-align: top; text-align: left">74.3</td>
<td style="vertical-align: top; text-align: left">98.6</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">1.0</td>
<td style="vertical-align: top; text-align: left">53.6</td>
<td style="vertical-align: top; text-align: left">29.8</td>
<td style="vertical-align: top; text-align: left">45.0</td>
<td style="vertical-align: top; text-align: left">65.6</td>
<td style="vertical-align: top; text-align: left">70.8</td>
<td style="vertical-align: top; text-align: left">97.2</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">1.5</td>
<td style="vertical-align: top; text-align: left">63.4</td>
<td style="vertical-align: top; text-align: left">39.8</td>
<td style="vertical-align: top; text-align: left">42.2</td>
<td style="vertical-align: top; text-align: left">61.0</td>
<td style="vertical-align: top; text-align: left">66.1</td>
<td style="vertical-align: top; text-align: left">97.0</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">2.0</td>
<td style="vertical-align: top; text-align: left">55.0</td>
<td style="vertical-align: top; text-align: left">41.2</td>
<td style="vertical-align: top; text-align: left">41.8</td>
<td style="vertical-align: top; text-align: left">62.2</td>
<td style="vertical-align: top; text-align: left">65.9</td>
<td style="vertical-align: top; text-align: left">98.4</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">FCN</td>
<td style="vertical-align: top; text-align: left">0.5</td>
<td style="vertical-align: top; text-align: left">71.8</td>
<td style="vertical-align: top; text-align: left">63.0</td>
<td style="vertical-align: top; text-align: left">56.0</td>
<td style="vertical-align: top; text-align: left">68.6</td>
<td style="vertical-align: top; text-align: left">76.2</td>
<td style="vertical-align: top; text-align: left">98.6</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">1.0</td>
<td style="vertical-align: top; text-align: left">76.4</td>
<td style="vertical-align: top; text-align: left">72.4</td>
<td style="vertical-align: top; text-align: left">58.2</td>
<td style="vertical-align: top; text-align: left">80.8</td>
<td style="vertical-align: top; text-align: left">81.3</td>
<td style="vertical-align: top; text-align: left">100</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">1.5</td>
<td style="vertical-align: top; text-align: left">84.6</td>
<td style="vertical-align: top; text-align: left">69.8</td>
<td style="vertical-align: top; text-align: left">68.6</td>
<td style="vertical-align: top; text-align: left">84.2</td>
<td style="vertical-align: top; text-align: left">85.0</td>
<td style="vertical-align: top; text-align: left">98.8</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">2.0</td>
<td style="vertical-align: top; text-align: left">90.8</td>
<td style="vertical-align: top; text-align: left">78.2</td>
<td style="vertical-align: top; text-align: left">73.2</td>
<td style="vertical-align: top; text-align: left">88.1</td>
<td style="vertical-align: top; text-align: left">91.2</td>
<td style="vertical-align: top; text-align: left">98.6</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">ResNet</td>
<td style="vertical-align: top; text-align: left">0.5</td>
<td style="vertical-align: top; text-align: left">77.8</td>
<td style="vertical-align: top; text-align: left">71.8</td>
<td style="vertical-align: top; text-align: left">60.2</td>
<td style="vertical-align: top; text-align: left">73.0</td>
<td style="vertical-align: top; text-align: left">80.2</td>
<td style="vertical-align: top; text-align: left">98.8</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">1.0</td>
<td style="vertical-align: top; text-align: left">85.6</td>
<td style="vertical-align: top; text-align: left">75.0</td>
<td style="vertical-align: top; text-align: left">65.6</td>
<td style="vertical-align: top; text-align: left">84.6</td>
<td style="vertical-align: top; text-align: left">85.0</td>
<td style="vertical-align: top; text-align: left">100</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">1.5</td>
<td style="vertical-align: top; text-align: left">82.2</td>
<td style="vertical-align: top; text-align: left">72.0</td>
<td style="vertical-align: top; text-align: left">70.2</td>
<td style="vertical-align: top; text-align: left">84.6</td>
<td style="vertical-align: top; text-align: left">86.0</td>
<td style="vertical-align: top; text-align: left">100</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">2.0</td>
<td style="vertical-align: top; text-align: left">98.4</td>
<td style="vertical-align: top; text-align: left">80.4</td>
<td style="vertical-align: top; text-align: left">73.6</td>
<td style="vertical-align: top; text-align: left">90.8</td>
<td style="vertical-align: top; text-align: left">94.6</td>
<td style="vertical-align: top; text-align: left">100</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">InceptionTime</td>
<td style="vertical-align: top; text-align: left">0.5</td>
<td style="vertical-align: top; text-align: left">76.6</td>
<td style="vertical-align: top; text-align: left">68.0</td>
<td style="vertical-align: top; text-align: left">59.8</td>
<td style="vertical-align: top; text-align: left">70.8</td>
<td style="vertical-align: top; text-align: left">79.1</td>
<td style="vertical-align: top; text-align: left">99.4</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">1.0</td>
<td style="vertical-align: top; text-align: left">82.2</td>
<td style="vertical-align: top; text-align: left">83.2</td>
<td style="vertical-align: top; text-align: left">64.0</td>
<td style="vertical-align: top; text-align: left">78.2</td>
<td style="vertical-align: top; text-align: left">84.5</td>
<td style="vertical-align: top; text-align: left">100</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left"/>
<td style="vertical-align: top; text-align: left">1.5</td>
<td style="vertical-align: top; text-align: left">81.6</td>
<td style="vertical-align: top; text-align: left">73.8</td>
<td style="vertical-align: top; text-align: left">74.2</td>
<td style="vertical-align: top; text-align: left">84.5</td>
<td style="vertical-align: top; text-align: left">87.2</td>
<td style="vertical-align: top; text-align: left">100</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"/>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">2.0</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">93.6</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">77.2</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">71.6</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">89.1</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">94.0</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">100</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The results of classification recall, precision, and F1 statistics for all the models of CB states for each class are presented in Tables <xref rid="j_infor578_tab_003">3</xref>, <xref rid="j_infor578_tab_004">4</xref>, and <xref rid="j_infor578_tab_005">5</xref>. The results show that all the models classified the damaged CB class with the highest accuracy, for example, F1 statistics of ResNet and InceptionTime models for 1.0 s, 1.5 s and 2.0 s signals reached 100% accuracy for this class. Different load states were classified worse by all the models. All the models were the most sensitive to 0.5 kg, 5 kg, and damaged-belt states, possibly because these signals stood out the most, while the 1 kg, 2 kg, and 3 kg signals were relatively similar, bringing more difficulty in their classification.</p>
<p>The training times of DNN models for various CB signal lengths are presented in Table <xref rid="j_infor578_tab_006">6</xref>. The signal length refers to the duration of the input CB signal (in seconds), and the epoch training time is the time (in seconds) taken by each model to complete one training epoch. The table shows that FCN model required the shortest training time for all the signal lengths. Meanwhile, ResNet and InceptionTime models took the longest training time. For example, for 2.0 s signals, training time of ResNet was 3.4 s and that of InceptionTime 3.6 s, while that of FCN took 0.8 s only. On the average, training time of InceptionTime model took several fractions of a second longer than that of ResNet, making it the slowest to train of all the studied models. Longer training time of ResNet and InceptionTime models likely occurred due to larger number of their parameters.</p>
<table-wrap id="j_infor578_tab_006">
<label>Table 6</label>
<caption>
<p>Single epoch training time of DNN models.</p>
</caption>
<table>
<thead>
<tr>
<td rowspan="2" style="vertical-align: middle; text-align: left; border-top: solid thin; border-bottom: solid thin">Signal length, s</td>
<td colspan="4" style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Epoch training time, s</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">CNN-LSTM</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">FCN</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">ResNet</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">InceptionTime</td>
</tr>
</thead>
<tbody>
<tr>
<td style="vertical-align: top; text-align: left">0.5</td>
<td style="vertical-align: top; text-align: left">0.6</td>
<td style="vertical-align: top; text-align: left">0.3</td>
<td style="vertical-align: top; text-align: left">0.7</td>
<td style="vertical-align: top; text-align: left">0.9</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">1.0</td>
<td style="vertical-align: top; text-align: left">1.0</td>
<td style="vertical-align: top; text-align: left">0.5</td>
<td style="vertical-align: top; text-align: left">1.8</td>
<td style="vertical-align: top; text-align: left">2.1</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">1.5</td>
<td style="vertical-align: top; text-align: left">1.2</td>
<td style="vertical-align: top; text-align: left">0.7</td>
<td style="vertical-align: top; text-align: left">2.8</td>
<td style="vertical-align: top; text-align: left">2.7</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">2.0</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">1.5</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">0.8</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">3.4</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">3.6</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The first-stage study demonstrated that without any data augmentations, ResNet model was the most accurate for CB state classification. Also, FCN model’s classification accuracy was only slightly lower than that of ResNet and InceptionTime models. FCN model had the advantage of being the fastest to train and requiring fewer computing resources as compared to other models. The next stage of the research explores the impact of various data augmentation methods on the classification accuracy of DNN models by artificially augmenting CB signal dataset.</p>
</sec>
<sec id="j_infor578_s_019">
<label>3.2.2</label>
<title>Data Augmentation Impact</title>
<p>In the second stage of the research, a study of data augmentation methods was carried out with the purpose of comparing the impact of various augmentation methods on DNN models in the classification of CB states. The study used the 2.0 s sliding-window processed signals because this dataset was the smallest, consisting of only 233 signals.</p>
<p>A balanced 5-fold cross-validation was also used for each experiment. The hyperparameters for training the models were the same as in the first stage, that is, in each experiment, the models were trained 5 times for 300 epochs, where the batch size of one iteration was 16 signals. However, at this stage, various data augmentations were applied to artificially increase the training sets, increasing the size of each of them to 2000 signals.</p>
<table-wrap id="j_infor578_tab_007">
<label>Table 7</label>
<caption>
<p>Multi-class (0.5 kg, 1 kg, 2 kg, 3 kg, 5 kg, and a class for a damaged belt) classification accuracy for all used augmentation methods.</p>
</caption>
<table>
<thead>
<tr>
<td rowspan="2" style="vertical-align: middle; text-align: left; border-top: solid thin; border-bottom: solid thin">Augmentation method</td>
<td colspan="4" style="vertical-align: top; text-align: left; border-top: solid thin; border-bottom: solid thin">Accuracy ±  SE, %</td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">CNN-LSTM</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">FCN</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">ResNet</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">InceptionTime</td>
</tr>
</thead>
<tbody>
<tr>
<td style="vertical-align: top; text-align: left">No augmentations</td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_038"><alternatives><mml:math>
<mml:mn>64.3</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.58</mml:mn></mml:math><tex-math><![CDATA[$64.3\pm 1.58$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_039"><alternatives><mml:math>
<mml:mn>88.1</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.48</mml:mn></mml:math><tex-math><![CDATA[$88.1\pm 1.48$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_040"><alternatives><mml:math>
<mml:mn>91.1</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.09</mml:mn></mml:math><tex-math><![CDATA[$91.1\pm 1.09$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_041"><alternatives><mml:math>
<mml:mn>89.1</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.54</mml:mn></mml:math><tex-math><![CDATA[$89.1\pm 1.54$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Laplace (std/100)</td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_042"><alternatives><mml:math>
<mml:mn>62.4</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>4.00</mml:mn></mml:math><tex-math><![CDATA[$62.4\pm 4.00$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_043"><alternatives><mml:math>
<mml:mn>89.1</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.35</mml:mn></mml:math><tex-math><![CDATA[$89.1\pm 1.35$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_044"><alternatives><mml:math>
<mml:mn>91.6</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.08</mml:mn></mml:math><tex-math><![CDATA[$91.6\pm 1.08$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_045"><alternatives><mml:math>
<mml:mn>92.1</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.32</mml:mn></mml:math><tex-math><![CDATA[$92.1\pm 1.32$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Laplace (std/50)</td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_046"><alternatives><mml:math>
<mml:mn>61.4</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.71</mml:mn></mml:math><tex-math><![CDATA[$61.4\pm 1.71$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_047"><alternatives><mml:math>
<mml:mn>89.6</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.91</mml:mn></mml:math><tex-math><![CDATA[$89.6\pm 1.91$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_048"><alternatives><mml:math>
<mml:mn>91.1</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>0.90</mml:mn></mml:math><tex-math><![CDATA[$91.1\pm 0.90$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_049"><alternatives><mml:math>
<mml:mn>90.6</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.12</mml:mn></mml:math><tex-math><![CDATA[$90.6\pm 1.12$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Laplace (std/20)</td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_050"><alternatives><mml:math>
<mml:mn>59.9</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.68</mml:mn></mml:math><tex-math><![CDATA[$59.9\pm 1.68$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_051"><alternatives><mml:math>
<mml:mn>91.1</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.11</mml:mn></mml:math><tex-math><![CDATA[$91.1\pm 1.11$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_052"><alternatives><mml:math>
<mml:mn>91.6</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>0.86</mml:mn></mml:math><tex-math><![CDATA[$91.6\pm 0.86$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_053"><alternatives><mml:math>
<mml:mn>92.4</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>0.72</mml:mn></mml:math><tex-math><![CDATA[$92.4\pm 0.72$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Drifted Gaussian (std/100)</td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_054"><alternatives><mml:math>
<mml:mn>61.9</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>3.04</mml:mn></mml:math><tex-math><![CDATA[$61.9\pm 3.04$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_055"><alternatives><mml:math>
<mml:mn>91.1</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.82</mml:mn></mml:math><tex-math><![CDATA[$91.1\pm 1.82$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_056"><alternatives><mml:math>
<mml:mn>90.1</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>0.71</mml:mn></mml:math><tex-math><![CDATA[$90.1\pm 0.71$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_057"><alternatives><mml:math>
<mml:mn>91.1</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.68</mml:mn></mml:math><tex-math><![CDATA[$91.1\pm 1.68$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Drifted Gaussian (std/50)</td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_058"><alternatives><mml:math>
<mml:mn>64.4</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.54</mml:mn></mml:math><tex-math><![CDATA[$64.4\pm 1.54$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_059"><alternatives><mml:math>
<mml:mn>88.1</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>2.87</mml:mn></mml:math><tex-math><![CDATA[$88.1\pm 2.87$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_060"><alternatives><mml:math>
<mml:mn>92.1</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.47</mml:mn></mml:math><tex-math><![CDATA[$92.1\pm 1.47$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_061"><alternatives><mml:math>
<mml:mn>91.6</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.14</mml:mn></mml:math><tex-math><![CDATA[$91.6\pm 1.14$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Drifted Gaussian (std/20)</td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_062"><alternatives><mml:math>
<mml:mn>61.9</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.61</mml:mn></mml:math><tex-math><![CDATA[$61.9\pm 1.61$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_063"><alternatives><mml:math>
<mml:mn>92.5</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.86</mml:mn></mml:math><tex-math><![CDATA[$92.5\pm 1.86$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_064"><alternatives><mml:math>
<mml:mn>90.1</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>0.94</mml:mn></mml:math><tex-math><![CDATA[$90.1\pm 0.94$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_065"><alternatives><mml:math>
<mml:mn>89.5</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.08</mml:mn></mml:math><tex-math><![CDATA[$89.5\pm 1.08$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Uniform (std/100)</td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_066"><alternatives><mml:math>
<mml:mn>62.9</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>2.20</mml:mn></mml:math><tex-math><![CDATA[$62.9\pm 2.20$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_067"><alternatives><mml:math>
<mml:mn>89.1</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.91</mml:mn></mml:math><tex-math><![CDATA[$89.1\pm 1.91$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_068"><alternatives><mml:math>
<mml:mn>90.1</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.44</mml:mn></mml:math><tex-math><![CDATA[$90.1\pm 1.44$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_069"><alternatives><mml:math>
<mml:mn>91.1</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.09</mml:mn></mml:math><tex-math><![CDATA[$91.1\pm 1.09$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Uniform (std/50)</td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_070"><alternatives><mml:math>
<mml:mn>59.9</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.30</mml:mn></mml:math><tex-math><![CDATA[$59.9\pm 1.30$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_071"><alternatives><mml:math>
<mml:mn>89.6</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.27</mml:mn></mml:math><tex-math><![CDATA[$89.6\pm 1.27$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_072"><alternatives><mml:math>
<mml:mn>91.1</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.14</mml:mn></mml:math><tex-math><![CDATA[$91.1\pm 1.14$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_073"><alternatives><mml:math>
<mml:mn>90.6</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>0.47</mml:mn></mml:math><tex-math><![CDATA[$90.6\pm 0.47$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Uniform (std/20)</td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_074"><alternatives><mml:math>
<mml:mn>59.9</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>2.71</mml:mn></mml:math><tex-math><![CDATA[$59.9\pm 2.71$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_075"><alternatives><mml:math>
<mml:mn>89.1</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.49</mml:mn></mml:math><tex-math><![CDATA[$89.1\pm 1.49$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_076"><alternatives><mml:math>
<mml:mn>89.1</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.54</mml:mn></mml:math><tex-math><![CDATA[$89.1\pm 1.54$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_077"><alternatives><mml:math>
<mml:mn>89.2</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>2.09</mml:mn></mml:math><tex-math><![CDATA[$89.2\pm 2.09$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Laplace and drifted Gaussian</td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_078"><alternatives><mml:math>
<mml:mn>63.4</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>2.56</mml:mn></mml:math><tex-math><![CDATA[$63.4\pm 2.56$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_079"><alternatives><mml:math>
<mml:mn>92.6</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.54</mml:mn></mml:math><tex-math><![CDATA[$92.6\pm 1.54$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_080"><alternatives><mml:math>
<mml:mn>91.1</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>0.92</mml:mn></mml:math><tex-math><![CDATA[$91.1\pm 0.92$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_081"><alternatives><mml:math>
<mml:mn>90.6</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>0.85</mml:mn></mml:math><tex-math><![CDATA[$90.6\pm 0.85$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Laplace and uniform noise</td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_082"><alternatives><mml:math>
<mml:mn>62.9</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.97</mml:mn></mml:math><tex-math><![CDATA[$62.9\pm 1.97$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_083"><alternatives><mml:math>
<mml:mn>89.1</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.51</mml:mn></mml:math><tex-math><![CDATA[$89.1\pm 1.51$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_084"><alternatives><mml:math>
<mml:mn>91.1</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>0.87</mml:mn></mml:math><tex-math><![CDATA[$91.1\pm 0.87$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_085"><alternatives><mml:math>
<mml:mn>92.5</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>0.71</mml:mn></mml:math><tex-math><![CDATA[$92.5\pm 0.71$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left">Magnitude warping</td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_086"><alternatives><mml:math>
<mml:mn>59.4</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>2.13</mml:mn></mml:math><tex-math><![CDATA[$59.4\pm 2.13$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_087"><alternatives><mml:math>
<mml:mn>88.1</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.93</mml:mn></mml:math><tex-math><![CDATA[$88.1\pm 1.93$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_088"><alternatives><mml:math>
<mml:mn>90.6</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>2.19</mml:mn></mml:math><tex-math><![CDATA[$90.6\pm 2.19$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left"><inline-formula id="j_infor578_ineq_089"><alternatives><mml:math>
<mml:mn>90.1</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>1.41</mml:mn></mml:math><tex-math><![CDATA[$90.1\pm 1.41$]]></tex-math></alternatives></inline-formula></td>
</tr>
<tr>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin">TimeVAE</td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><inline-formula id="j_infor578_ineq_090"><alternatives><mml:math>
<mml:mn>75.7</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>0.88</mml:mn></mml:math><tex-math><![CDATA[$75.7\pm 0.88$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><inline-formula id="j_infor578_ineq_091"><alternatives><mml:math>
<mml:mn>77.2</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>2.75</mml:mn></mml:math><tex-math><![CDATA[$77.2\pm 2.75$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><inline-formula id="j_infor578_ineq_092"><alternatives><mml:math>
<mml:mn>90.1</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>2.93</mml:mn></mml:math><tex-math><![CDATA[$90.1\pm 2.93$]]></tex-math></alternatives></inline-formula></td>
<td style="vertical-align: top; text-align: left; border-bottom: solid thin"><inline-formula id="j_infor578_ineq_093"><alternatives><mml:math>
<mml:mn>88.6</mml:mn>
<mml:mo>±</mml:mo>
<mml:mn>2.90</mml:mn></mml:math><tex-math><![CDATA[$88.6\pm 2.90$]]></tex-math></alternatives></inline-formula></td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The classification accuracy results of each experiment for all DNN models using various data augmentations are presented in Table <xref rid="j_infor578_tab_007">7</xref>. The results also show the baseline classification accuracy of each model without augmentations, which were obtained in the first phase of the study by classifying 2.0 s-length signals.</p>
<p>The results show that each data augmentation affected the classification accuracy differently, depending on the model and the augmentation method itself. The best classification accuracy results after applying data augmentations were achieved by using FCN, ResNet, and InceptionTime models.</p>
<p>Adding Laplace noise increased the accuracy of FCN, ResNet, and InceptionTime models across all the noise levels, but decreased the accuracy of CNN-LSTM model. The addition of drifted Gaussian noise notably increased the baseline classification accuracy of FCN model by 4.4% to <inline-formula id="j_infor578_ineq_094"><alternatives><mml:math>
<mml:mn>92.5</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi>
<mml:mo>±</mml:mo>
<mml:mn>1.86</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$92.5\% \pm 1.86\% $]]></tex-math></alternatives></inline-formula> with std/20 noise. It also increased the accuracy of ResNet model with std/50 noise by 1% to <inline-formula id="j_infor578_ineq_095"><alternatives><mml:math>
<mml:mn>92.1</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi>
<mml:mo>±</mml:mo>
<mml:mn>1.47</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$92.1\% \pm 1.47\% $]]></tex-math></alternatives></inline-formula>, and InceptionTime model by 2% to <inline-formula id="j_infor578_ineq_096"><alternatives><mml:math>
<mml:mn>91.1</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi>
<mml:mo>±</mml:mo>
<mml:mn>1.68</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$91.1\% \pm 1.68\% $]]></tex-math></alternatives></inline-formula> with std/100 noise. Adding uniform noise resulted on the average in worse classification results for all the models as compared to adding Laplace or drifted Gaussian noise.</p>
<p>The best accuracy result of the entire study, which increased the baseline accuracy of FCN model by 4.5% to <inline-formula id="j_infor578_ineq_097"><alternatives><mml:math>
<mml:mn>92.6</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi>
<mml:mo>±</mml:mo>
<mml:mn>1.54</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$92.6\% \pm 1.54\% $]]></tex-math></alternatives></inline-formula>, was achieved by training the model on data augmented with combined Laplace and drifted Gaussian noise. Adding Laplace and uniform noise increased the accuracy of InceptionTime model by 3.4% to <inline-formula id="j_infor578_ineq_098"><alternatives><mml:math>
<mml:mn>92.5</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi>
<mml:mo>±</mml:mo>
<mml:mn>0.71</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$92.5\% \pm 0.71\% $]]></tex-math></alternatives></inline-formula>. All the experiments involving combined noise used noise with a size of std/100.</p>
<p>Magnitude warping augmented data increased the accuracy of only InceptionTime model by 1% to <inline-formula id="j_infor578_ineq_099"><alternatives><mml:math>
<mml:mn>90.1</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi>
<mml:mo>±</mml:mo>
<mml:mn>1.41</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$90.1\% \pm 1.41\% $]]></tex-math></alternatives></inline-formula>. Augmenting signals with TimeVAE increased the accuracy of only CNN-LSTM model, which reached <inline-formula id="j_infor578_ineq_100"><alternatives><mml:math>
<mml:mn>75.7</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi>
<mml:mo>±</mml:mo>
<mml:mn>0.71</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$75.7\% \pm 0.71\% $]]></tex-math></alternatives></inline-formula>, but was still lower as compared to other models.</p>
<p>CNN-LSTM model generally experienced a negative impact from most data augmentations, except for TimeVAE-generated signals, which had a positive effect. Most data augmentations improved the baseline accuracy of FCN model, although TimeVAE augmentation led to a decrease of 10.9%. For ResNet model, most augmentations either reduced the accuracy or resulted in only a marginal improvement of 1%.</p>
<fig id="j_infor578_fig_015">
<label>Fig. 15</label>
<caption>
<p>Comparison of change in classification accuracy of data augmentation methods for each model with baseline (no augmentation) accuracy. Accuracy changes of: a) CNN-LSTM; b) FCN; c) ResNet; d) InceptionTime.</p>
</caption>
<graphic xlink:href="infor578_g015.jpg"/>
</fig>
<p>The changes in classification accuracy of each model achieved by various data augmentation methods are presented in Fig. <xref rid="j_infor578_fig_015">15</xref>. Separate graphs show the relative change in accuracy for each model as compared to baseline (without any augmentation) accuracy. CNN-LSTM model generally experienced a negative impact from most data augmentations, except for TimeVAE-generated signals, which had a positive effect. Most data augmentations improved the baseline accuracy of FCN model, although TimeVAE augmentation led to a decrease of 10.9%. For ResNet model, most augmentations either reduced accuracy or resulted in only a marginal improvement of 1%. The baseline classification accuracy of InceptionTime model was increased by almost all the data augmentations.</p>
<p>The plot analysis revealed that for FCN and InceptionTime models, almost all the data augmentation methods increased the baseline classification accuracy. However, for ResNet and CNN-LSTM models, most augmentations had a negative impact on the classification accuracy. Likely, the architecture of FCN and InceptionTime models allowed them to adapt more effectively to specific augmentation methods, thereby enhancing their overall classification accuracy and robustness.</p>
<p>The results of the second stage of the study showed that FCN model trained with augmented data classified CB load and defect states with an accuracy of <inline-formula id="j_infor578_ineq_101"><alternatives><mml:math>
<mml:mn>92.6</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi>
<mml:mo>±</mml:mo>
<mml:mn>1.54</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$92.6\% \pm 1.54\% $]]></tex-math></alternatives></inline-formula>. This model was comparable to ResNet and InceptionTime models in terms of classification accuracy, but FCN model was trained much faster due to its small number of trainable parameters. ResNet model achieved an accuracy of <inline-formula id="j_infor578_ineq_102"><alternatives><mml:math>
<mml:mn>92.1</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi>
<mml:mo>±</mml:mo>
<mml:mn>1.47</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$92.1\% \pm 1.47\% $]]></tex-math></alternatives></inline-formula> with data augmented by drifted Gaussian noise, while InceptionTime model achieved its best result of <inline-formula id="j_infor578_ineq_103"><alternatives><mml:math>
<mml:mn>92.5</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi>
<mml:mo>±</mml:mo>
<mml:mn>0.71</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$92.5\% \pm 0.71\% $]]></tex-math></alternatives></inline-formula> with data augmented by random Laplace and uniform noise. CNN-LSTM model achieved its highest accuracy of <inline-formula id="j_infor578_ineq_104"><alternatives><mml:math>
<mml:mn>75.7</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi>
<mml:mo>±</mml:mo>
<mml:mn>0.88</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$75.7\% \pm 0.88\% $]]></tex-math></alternatives></inline-formula> when trained with signals generated by TimeVAE model. The best accuracy results of FCN, ResNet, and InceptionTime models were very similar and differed only slightly, but the accuracy of CNN-LSTM model was much worse.</p>
</sec>
</sec>
</sec>
<sec id="j_infor578_s_020">
<label>4</label>
<title>Conclusion</title>
<p>In this research, we examined existing DL algorithms and DNN architectures for classifying CB states. In addition, various time series data augmentation methods applicable to CB tension signals were examined.</p>
<p>The study successfully developed and evaluated several DNN models for the classification of CB load and defect states using tension signals, specifically based on FCN, ResNet, InceptionTime, and CNN-LSTM architectures. FCN model was able to classify CB states with an accuracy of <inline-formula id="j_infor578_ineq_105"><alternatives><mml:math>
<mml:mn>92.6</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi>
<mml:mo>±</mml:mo>
<mml:mn>1.54</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$92.6\% \pm 1.54\% $]]></tex-math></alternatives></inline-formula>, making it the most accurate of the studied models. ResNet and InceptionTime models also performed well, with accuracies of <inline-formula id="j_infor578_ineq_106"><alternatives><mml:math>
<mml:mn>92.1</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi>
<mml:mo>±</mml:mo>
<mml:mn>1.47</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$92.1\% \pm 1.47\% $]]></tex-math></alternatives></inline-formula> and <inline-formula id="j_infor578_ineq_107"><alternatives><mml:math>
<mml:mn>92.5</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi>
<mml:mo>±</mml:mo>
<mml:mn>0.71</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$92.5\% \pm 0.71\% $]]></tex-math></alternatives></inline-formula>, respectively. CNN-LSTM model demonstrated the worst results, with a maximum accuracy of <inline-formula id="j_infor578_ineq_108"><alternatives><mml:math>
<mml:mn>75.7</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi>
<mml:mo>±</mml:mo>
<mml:mn>0.88</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$75.7\% \pm 0.88\% $]]></tex-math></alternatives></inline-formula> only.</p>
<p>The impact of various data augmentation methods on classification accuracy was also analysed. The combined addition of Laplace and drifted Gaussian noise increased the baseline (without any augmentation) accuracy of FCN model by 4.5% to <inline-formula id="j_infor578_ineq_109"><alternatives><mml:math>
<mml:mn>92.6</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi>
<mml:mo>±</mml:mo>
<mml:mn>1.54</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$92.6\% \pm 1.54\% $]]></tex-math></alternatives></inline-formula>. Adding Laplace and uniform noise increased the accuracy of InceptionTime model by 3.4% to <inline-formula id="j_infor578_ineq_110"><alternatives><mml:math>
<mml:mn>92.5</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi>
<mml:mo>±</mml:mo>
<mml:mn>0.71</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$92.5\% \pm 0.71\% $]]></tex-math></alternatives></inline-formula>. The classification accuracy of CNN-LSTM model trained with signals generated by TimeVAE increased by 11.4% to <inline-formula id="j_infor578_ineq_111"><alternatives><mml:math>
<mml:mn>75.7</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi>
<mml:mo>±</mml:mo>
<mml:mn>0.88</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$75.7\% \pm 0.88\% $]]></tex-math></alternatives></inline-formula>, but it still remained much lower than that of other models. The baseline accuracy of ResNet model increased by 1% only to <inline-formula id="j_infor578_ineq_112"><alternatives><mml:math>
<mml:mn>92.1</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi>
<mml:mo>±</mml:mo>
<mml:mn>1.47</mml:mn>
<mml:mi mathvariant="normal">%</mml:mi></mml:math><tex-math><![CDATA[$92.1\% \pm 1.47\% $]]></tex-math></alternatives></inline-formula> after training with drifted Gaussian noise augmented data.</p>
<p>These results underline the effectiveness of applying data augmentations to small CB tension signal datasets, enhancing the classification accuracy of models based on FCN and InceptionTime architectures. In classifying CB states, FCN-based model showed higher accuracy and speed compared to other models, despite having the lowest amount of trainable parameters. Successful application of FCN model demonstrated the importance of selecting and optimizing the right architecture for specific data and classification tasks.</p>
<p>CB status classification under fixed loads and rotation speed could be considered as a limitation of this study. A set of fixed parameters does not reflect real world conditions and future investigations should be based on random CB status classification on unseen experiment parameters. Empirical model parameters selection method can be also considered a limitation of this study.</p>
<p>Further research on CB state classification could aim to improve accuracy in classifying weights of similar mass (1 kg, 2 kg, 3 kg). Additionally, future research could explore advanced generative data augmentation techniques, for example, those utilizing GANs or other VAE architectures to enhance the quality of CB tension signal data.</p>
</sec>
</body>
<back>
<ack id="j_infor578_ack_001">
<title>Acknowledgements</title>
<p>This paper has received funding under postdoctoral fellowship project from the Research Council of Lithuania (LMTLT), agreement No. [S-PD-22-81].</p></ack>
<ref-list id="j_infor578_reflist_001">
<title>References</title>
<ref id="j_infor578_ref_001">
<mixed-citation publication-type="journal"><string-name><surname>Andrejiova</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Grincova</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Marasova</surname>, <given-names>D.</given-names></string-name> (<year>2021</year>). <article-title>Identification with machine learning techniques of a classification model for the degree of damage to rubber-textile conveyor belts with the aim to achieve sustainability</article-title>. <source>Engineering Failure Analysis</source>, <volume>127</volume>, <fpage>105564</fpage>.</mixed-citation>
</ref>
<ref id="j_infor578_ref_002">
<mixed-citation publication-type="journal"><string-name><surname>Bortnowski</surname>, <given-names>P.</given-names></string-name>, <string-name><surname>Król</surname>, <given-names>R.</given-names></string-name>, <string-name><surname>Ozdoba</surname>, <given-names>M.</given-names></string-name> (<year>2022</year>a). <article-title>Roller damage detection method based on the measurement of transverse vibrations of the conveyor belt</article-title>. <source>Eksploatacja i Niezawodność</source>, <volume>24</volume>(<issue>3</issue>), <fpage>510</fpage>–<lpage>521</lpage>.</mixed-citation>
</ref>
<ref id="j_infor578_ref_003">
<mixed-citation publication-type="journal"><string-name><surname>Bortnowski</surname>, <given-names>P.</given-names></string-name>, <string-name><surname>Kawalec</surname>, <given-names>W.</given-names></string-name>, <string-name><surname>Król</surname>, <given-names>R.</given-names></string-name>, <string-name><surname>Ozdoba</surname>, <given-names>M.</given-names></string-name> (<year>2022</year>b). <article-title>Types and causes of damage to the conveyor belt–review, classification and mutual relations</article-title>. <source>Engineering Failure Analysis</source>, <volume>140</volume>, <fpage>106520</fpage>.</mixed-citation>
</ref>
<ref id="j_infor578_ref_004">
<mixed-citation publication-type="journal"><string-name><surname>Chlap</surname>, <given-names>P.</given-names></string-name>, <string-name><surname>Min</surname>, <given-names>H.</given-names></string-name>, <string-name><surname>Vandenberg</surname>, <given-names>N.</given-names></string-name>, <string-name><surname>Dowling</surname>, <given-names>J.A.</given-names></string-name>, <string-name><surname>Holloway</surname>, <given-names>L.</given-names></string-name>, <string-name><surname>Haworth</surname>, <given-names>A.</given-names></string-name> (<year>2021</year>). <article-title>A review of medical image data augmentation techniques for deep learning applications</article-title>. <source>Journal of Medical Imaging and Radiation Oncology</source>, <volume>65</volume>(<issue>5</issue>), <fpage>545</fpage>–<lpage>563</lpage>.</mixed-citation>
</ref>
<ref id="j_infor578_ref_005">
<mixed-citation publication-type="journal"><string-name><surname>Dąbek</surname>, <given-names>P.</given-names></string-name>, <string-name><surname>Wróblewski</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Wodecki</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Bortnowski</surname>, <given-names>P.</given-names></string-name>, <string-name><surname>Ozdoba</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Król</surname>, <given-names>R.</given-names></string-name>, <string-name><surname>Zimroz</surname>, <given-names>R.</given-names></string-name> (<year>2023</year>). <article-title>Application of the methods of monitoring and detecting the belt mistracking in laboratory conditions</article-title>. <source>Applied Sciences</source>, <volume>13</volume>(<issue>4</issue>), <fpage>2111</fpage>.</mixed-citation>
</ref>
<ref id="j_infor578_ref_006">
<mixed-citation publication-type="other"><string-name><surname>Desai</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Freeman</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Wang</surname>, <given-names>Z.</given-names></string-name>, <string-name><surname>Beaver</surname>, <given-names>I.</given-names></string-name> (2021). Timevae: a variational auto-encoder for multivariate time series generation. <italic>arXiv preprint</italic> arXiv:<ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/2111.08095">2111.08095</ext-link>.</mixed-citation>
</ref>
<ref id="j_infor578_ref_007">
<mixed-citation publication-type="other"><string-name><surname>Goodfellow</surname>, <given-names>I.</given-names></string-name>, <string-name><surname>Pouget-Abadie</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Mirza</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Xu</surname>, <given-names>B.</given-names></string-name>, <string-name><surname>Warde-Farley</surname>, <given-names>D.</given-names></string-name>, <string-name><surname>Ozair</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Courville</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Bengio</surname>, <given-names>Y.</given-names></string-name> (2014). Generative adversarial nets. <italic>Advances in Neural Information Processing Systems</italic>, <italic>27</italic>.</mixed-citation>
</ref>
<ref id="j_infor578_ref_008">
<mixed-citation publication-type="other"><string-name><surname>Goubeaud</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Joußen</surname>, <given-names>P.</given-names></string-name>, <string-name><surname>Gmyrek</surname>, <given-names>N.</given-names></string-name>, <string-name><surname>Ghorban</surname>, <given-names>F.</given-names></string-name>, <string-name><surname>Schelkes</surname>, <given-names>L.</given-names></string-name>, <string-name><surname>Kummert</surname>, <given-names>A.</given-names></string-name> (2021). Using Variational Autoencoder to augment Sparse Time series Datasets. In: <italic>2021 7th International Conference on Optimization and Applications (ICOA)</italic>, pp. 1–6.</mixed-citation>
</ref>
<ref id="j_infor578_ref_009">
<mixed-citation publication-type="other"><string-name><surname>Gregor Hartmann</surname>, <given-names>K.</given-names></string-name>, <string-name><surname>Tibor Schirrmeister</surname>, <given-names>R.</given-names></string-name>, <string-name><surname>Ball</surname>, <given-names>T.</given-names></string-name> (2018). EEG-GAN: Generative adversarial networks for electroencephalograhic (EEG) brain signals. <italic>arXiv e-prints</italic>, arXiv:<ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/1806.01875v1">1806.01875v1</ext-link>.</mixed-citation>
</ref>
<ref id="j_infor578_ref_010">
<mixed-citation publication-type="journal"><string-name><surname>Hsiao</surname>, <given-names>T.-Y.</given-names></string-name>, <string-name><surname>Chang</surname>, <given-names>Y.-C.</given-names></string-name>, <string-name><surname>Chou</surname>, <given-names>H.-H.</given-names></string-name>, <string-name><surname>Chiu</surname>, <given-names>C.-T.</given-names></string-name> (<year>2019</year>). <article-title>Filter-based deep-compression with global average pooling for convolutional networks</article-title>. <source>Journal of Systems Architecture</source>, <volume>95</volume>, <fpage>9</fpage>–<lpage>18</lpage>.</mixed-citation>
</ref>
<ref id="j_infor578_ref_011">
<mixed-citation publication-type="journal"><string-name><surname>Huang</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Chakraborty</surname>, <given-names>P.</given-names></string-name>, <string-name><surname>Sharma</surname>, <given-names>A.</given-names></string-name> (<year>2023</year>). <article-title>Deep convolutional generative adversarial networks for traffic data imputation encoding time series as images</article-title>. <source>International Journal of Transportation Science and Technology</source>, <volume>12</volume>(<issue>1</issue>), <fpage>1</fpage>–<lpage>18</lpage>.</mixed-citation>
</ref>
<ref id="j_infor578_ref_012">
<mixed-citation publication-type="journal"><string-name><surname>Iglesias</surname>, <given-names>G.</given-names></string-name>, <string-name><surname>Talavera</surname>, <given-names>E.</given-names></string-name>, <string-name><surname>González-Prieto</surname>, <given-names>Á.</given-names></string-name>, <string-name><surname>Mozo</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Gómez-Canaval</surname>, <given-names>S.</given-names></string-name> (<year>2023</year>). <article-title>Data augmentation techniques in time series domain: a survey and taxonomy</article-title>. <source>Neural Computing and Applications</source>, <volume>35</volume>(<issue>14</issue>), <fpage>10123</fpage>–<lpage>10145</lpage>.</mixed-citation>
</ref>
<ref id="j_infor578_ref_013">
<mixed-citation publication-type="chapter"><string-name><surname>Iwana</surname>, <given-names>B.K.</given-names></string-name>, <string-name><surname>Uchida</surname>, <given-names>S.</given-names></string-name> (<year>2021</year>). <chapter-title>Time series data augmentation for neural networks by time warping with a discriminative teacher</chapter-title>. In: <source>2020 25th International Conference on Pattern Recognition (ICPR)</source>. <publisher-name>IEEE</publisher-name>, pp. <fpage>3558</fpage>–<lpage>3565</lpage>.</mixed-citation>
</ref>
<ref id="j_infor578_ref_014">
<mixed-citation publication-type="journal"><string-name><surname>Kingma</surname>, <given-names>D.P.</given-names></string-name>, <string-name><surname>Welling</surname>, <given-names>M.</given-names></string-name> (<year>2019</year>). <article-title>An introduction to variational autoencoders</article-title>. <source>Foundations and Trends® in Machine Learning</source>, <volume>12</volume>(<issue>4</issue>), <fpage>307</fpage>–<lpage>392</lpage>.</mixed-citation>
</ref>
<ref id="j_infor578_ref_015">
<mixed-citation publication-type="journal"><string-name><surname>Klištincová</surname>, <given-names>N.</given-names></string-name>, <string-name><surname>Pin</surname>, <given-names>L.</given-names></string-name>, <string-name><surname>Puškárová</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Giannino</surname>, <given-names>D.</given-names></string-name>, <string-name><surname>Bučková</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Lambreva</surname>, <given-names>M.D.</given-names></string-name>, <string-name><surname>Manfredini</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Canfora</surname>, <given-names>L.</given-names></string-name>, <string-name><surname>Pangallo</surname>, <given-names>D.</given-names></string-name>, <string-name><surname>Pinzari</surname>, <given-names>F.</given-names></string-name> (<year>2024</year>). <article-title>From farm to fork: fungal and bacterial contaminants and their diagnostics in the production steps of ready-to-eat salads</article-title>. <source>Trends in Food Science &amp; Technology</source>, <volume>150</volume>, <elocation-id>104573</elocation-id>.</mixed-citation>
</ref>
<ref id="j_infor578_ref_016">
<mixed-citation publication-type="chapter"><string-name><surname>Li</surname>, <given-names>X.-G.</given-names></string-name>, <string-name><surname>Miao</surname>, <given-names>C.-Y.</given-names></string-name>, <string-name><surname>Wang</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Zhang</surname>, <given-names>Y.</given-names></string-name> (<year>2011</year>). <chapter-title>Automatic defect detection method for the steel cord conveyor belt based on its X-ray images</chapter-title>. In: <source>2011 International Conference on Control, Automation and Systems Engineering (CASE)</source>, pp. <fpage>1</fpage>–<lpage>4</lpage>.</mixed-citation>
</ref>
<ref id="j_infor578_ref_017">
<mixed-citation publication-type="chapter"><string-name><surname>Long</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Shelhamer</surname>, <given-names>E.</given-names></string-name>, <string-name><surname>Darrell</surname>, <given-names>T.</given-names></string-name> (<year>2015</year>). <chapter-title>Fully convolutional networks for semantic segmentation</chapter-title>. In: <source>Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition</source>, pp. <fpage>3431</fpage>–<lpage>3440</lpage>.</mixed-citation>
</ref>
<ref id="j_infor578_ref_018">
<mixed-citation publication-type="journal"><string-name><surname>Santos</surname>, <given-names>A.A.</given-names></string-name>, <string-name><surname>Rocha</surname>, <given-names>F.A.</given-names></string-name>, <string-name><surname>Reis</surname>, <given-names>A.J.d.R.</given-names></string-name>, <string-name><surname>Guimarães</surname>, <given-names>F.G.</given-names></string-name> (<year>2020</year>). <article-title>Automatic system for visual detection of dirt buildup on conveyor belts using convolutional neural networks</article-title>. <source>Sensors</source>, <volume>20</volume>(<issue>20</issue>), <fpage>5762</fpage>.</mixed-citation>
</ref>
<ref id="j_infor578_ref_019">
<mixed-citation publication-type="journal"><string-name><surname>Sarker</surname>, <given-names>I.H.</given-names></string-name> (<year>2021</year>). <article-title>Deep learning: a comprehensive overview on techniques, taxonomy, applications and research directions</article-title>. <source>SN Computer Science</source>, <volume>2</volume>, <elocation-id>420</elocation-id>.</mixed-citation>
</ref>
<ref id="j_infor578_ref_020">
<mixed-citation publication-type="other"><string-name><surname>Smith</surname>, <given-names>K.E.</given-names></string-name>, <string-name><surname>Smith</surname>, <given-names>A.O.</given-names></string-name> (2020). Conditional GAN for timeseries generation. <italic>arXiv preprint</italic> arXiv:<ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/2006.16477">2006.16477</ext-link>.</mixed-citation>
</ref>
<ref id="j_infor578_ref_021">
<mixed-citation publication-type="chapter"><string-name><surname>Um</surname>, <given-names>T.T.</given-names></string-name>, <string-name><surname>Pfister</surname>, <given-names>F.M.</given-names></string-name>, <string-name><surname>Pichler</surname>, <given-names>D.</given-names></string-name>, <string-name><surname>Endo</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Lang</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Hirche</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Fietzek</surname>, <given-names>U.</given-names></string-name>, <string-name><surname>Kulić</surname>, <given-names>D.</given-names></string-name> (<year>2017</year>). <chapter-title>Data augmentation of wearable sensor data for Parkinson’s disease monitoring using convolutional neural networks</chapter-title>. In: <source>Proceedings of the 19th ACM International Conference on Multimodal Interaction</source>, pp. <fpage>216</fpage>–<lpage>220</lpage>.</mixed-citation>
</ref>
<ref id="j_infor578_ref_022">
<mixed-citation publication-type="journal"><string-name><surname>Wang</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Perez</surname>, <given-names>L.</given-names></string-name> (<year>2017</year>a). <article-title>The effectiveness of data augmentation in image classification using deep learning</article-title>. <source>Convolutional Neural Networks Vis. Recognit</source>, <volume>11</volume>(<issue>2017</issue>), <fpage>1</fpage>–<lpage>8</lpage>.</mixed-citation>
</ref>
<ref id="j_infor578_ref_023">
<mixed-citation publication-type="journal"><string-name><surname>Wang</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Miao</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Miao</surname>, <given-names>D.</given-names></string-name>, <string-name><surname>Yang</surname>, <given-names>D.</given-names></string-name>, <string-name><surname>Zheng</surname>, <given-names>Y.</given-names></string-name> (<year>2023</year>). <article-title>Hazard source detection of longitudinal tearing of conveyor belt based on deep learning</article-title>. <source>PLoS ONE</source>, <volume>18</volume>(<issue>4</issue>), <fpage>0283878</fpage>.</mixed-citation>
</ref>
<ref id="j_infor578_ref_024">
<mixed-citation publication-type="chapter"><string-name><surname>Wang</surname>, <given-names>Z.</given-names></string-name>, <string-name><surname>Yan</surname>, <given-names>W.</given-names></string-name>, <string-name><surname>Oates</surname>, <given-names>T.</given-names></string-name> (<year>2017</year>b). <chapter-title>Time series classification from scratch with deep neural networks: a strong baseline</chapter-title>. In: <source>2017 International Joint Conference on Neural Networks (IJCNN)</source>. <publisher-name>IEEE</publisher-name>, pp. <fpage>1578</fpage>–<lpage>1585</lpage>.</mixed-citation>
</ref>
<ref id="j_infor578_ref_025">
<mixed-citation publication-type="other"><string-name><surname>Yoon</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Jarrett</surname>, <given-names>D.</given-names></string-name>, <string-name><surname>Van der Schaar</surname>, <given-names>M.</given-names></string-name> (2019). Time-series generative adversarial networks. <italic>Advances in Neural Information Processing Systems</italic>, <italic>32</italic>.</mixed-citation>
</ref>
<ref id="j_infor578_ref_026">
<mixed-citation publication-type="journal"><string-name><surname>Zhang</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Shi</surname>, <given-names>H.</given-names></string-name>, <string-name><surname>Zhang</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Yu</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Zhou</surname>, <given-names>M.</given-names></string-name> (<year>2021</year>). <article-title>Deep learning-based damage detection of mining conveyor belt</article-title>. <source>Measurement</source>, <volume>175</volume>, <elocation-id>109130</elocation-id>.</mixed-citation>
</ref>
<ref id="j_infor578_ref_027">
<mixed-citation publication-type="journal"><string-name><surname>Žvirblis</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Petkevičius</surname>, <given-names>L.</given-names></string-name>, <string-name><surname>Bzinkowski</surname>, <given-names>D.</given-names></string-name>, <string-name><surname>Vaitkus</surname>, <given-names>D.</given-names></string-name>, <string-name><surname>Vaitkus</surname>, <given-names>P.</given-names></string-name>, <string-name><surname>Rucki</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Kilikevičius</surname>, <given-names>A.</given-names></string-name> (<year>2022</year>). <article-title>Investigation of deep learning models on identification of minimum signal length for precise classification of conveyor rubber belt loads</article-title>. <source>Advances in Mechanical Engineering</source>, <volume>14</volume>(<issue>6</issue>), <fpage>1</fpage>–<lpage>13</lpage>.</mixed-citation>
</ref>
</ref-list>
</back>
</article>
