Warning: fopen(/home/virtual/kwjs/journal/upload/ip_log/ip_log_2024-10.txt): failed to open stream: Permission denied in /home/virtual/lib/view_data.php on line 100 Warning: fwrite() expects parameter 1 to be resource, boolean given in /home/virtual/lib/view_data.php on line 101 Review on the Recent Welding Research with Application of CNN-Based Deep Learning Part I: Models and Applications

J Weld Join > Volume 39(1); 2021 > Article
Lee, Yi, Hyun, and Kim: Review on the Recent Welding Research with Application of CNN-Based Deep Learning Part I: Models and Applications

Abstract

During machine learning algorithms, deep learning refers to a neural network containing multiple hidden layers. Welding research based upon deep learning has been increasing due to advances in algorithms and computer hardwares. Among the deep learning algorithms, the convolutional neural network (CNN) has recently received the spotlight for performing classification or regression based on image input. CNNs enables end-to-end learning without feature extraction and in-situ estimation of the process outputs. In this paper, 18 recent papers were reviewed to investigate how to apply CNN models to welding. The papers was classified into 5 groups: four for supervised learning models and one for unsupervised learning models. The classification of supervised learning groups was based on the application of transfer learning and data augmentation. For each paper, the structure and performance of its CNN model were described, and also its application in welding was explained.

1. Introduction

Deep learning is one of the machine learning techniques, a kind of Artificial Intelligence (AI), with increasing applications to a wide range of sectors. Machine learning is not based on explicit rules, it learns rules using datasets, and is categorized into supervised learning, unsupervised learning, and reinforcement learning according to the types of training datasets. Supervised learning uses inputs and correct outputs as training datasets and it is applied to classification and regression problems1).
Artificial Neural Network (ANN), a type of machine learning, was developed by mimicking human neural networks and consists of an input layer, hidden layers, and an output layer. ANN is further categorized into a shallow neural network (SNN) where there is one hidden layer and a deep neural network (DNN) where there are two or more hidden layers. When machine learning is performed using DNN, it is called deep learning.
As for research that applies ANN to welding, since the error back-propagation algorithm2) was introduced in the mid-1980s, ANN has been actively applied to a number of welding processes. The examples of application includes prediction of weld quality from control parameters3,4), suggestion of appropriate control welding parameters from the welding of the desired weld attributes5), profile extraction from laser vision sensing image6), and quality and seam tracking using vision sensors7). However, there were some problems in expanded application of neural network models at the time. First, the sigmoid function or RBF (Radial Basis Function) function was mainly used as an activation function of neural networks, and learning of weights far from the output layers was difficult in training of DNN through the back-propagation algorithm. In addition, when images or time series signals were used, learning was performed through feature extraction rather than raw signals without feature extraction, and models that predict outputs from scalar variables such as control parameters were mainly developed, and thus application of the trained model to another production system was intrinsically difficult. When only the control parameters are used as the input parameters of the model, several environmental disturbances (e.g. humidity, temperature, electric power instability) cannot be considered, and it was difficult to set the same system inputs even by setting the same control parameters in other systems.
In recent years, ReLu (Rectified Linear unit) has been proposed as an activation function to enable training of DNN. With the introduction of Convolutional Neural Networks (CNNs)8), also known as convnets, which are mainly used for image recognition, and Recurrent Neural Network (RNN)9,10) for use in models using time-series variables, innovative development in deep learning technologies has been achieved including end- to-end learning.
Machine learning has been applied to various welding processes and the examples include the use of deep learning and reinforcement learning in the laser welding control11), quality prediction through deep learning in arc welding12), and quality classification through SVM (Supported Vector Machine), a machine learning algorithim 13,14).
In this paper, 18 recently published papers on CNN- based welding research are reviewed. As shown in Fig. 1, these papers were divided into 5 groups based on supervised/unsupervised learning, use of transfer learning, and application of data augmentation in data preprocessing.
Fig. 1
Classification of papers reviewed
jwj-39-1-10gf1.jpg

2. CNN Structure and Learning Methods

CNN is an artificial neural network that uses convolution operations, and CNN models has shown excellent performance in image recognition because the feature patterns of the images can be extracted by training through convolutional filters. In addition, manual extraction of features is not necessary since features are automatically learned, and models can be constructed based on pre-trained networks. With these advantages, research of applying CNNs for deep learning using images has shown a rapid increase in recent years.

2.1 Basic structure

CNN is composed of convolutional layers, pooling layers, and fully connected layers. In the convolutional layer, inputs within a kernel are connected by specific weights, and a feature map is constructed by information representing each kernel area and passed to the next layer. The convolution operation is performed by sliding the kernel window over the input data at a fixed stride. The output of the convolutional layer can be passed to the next layer as it is, but in most cases, pooling is performed, one of the down sampling operation, to prevent over-fitting and create features that are highly noise-resistant. Typical pooling methods are max pooling and mean pooling, in which the maximum and mean values are output within the specified kernel window, respectively. The fully connected layer is a layer in which nodes of the previous layer and those of the next layer are all connected in the same way as a traditional ANN. In the CNN model, the successive repetition of the convolutional layer and the pooling layer enables inclusion of local features in the image and also represents the global features and these are passed as the input of the fully connected layer, and the fully connected layer produces the final output.

2.2 Key techniques used in the reviewed papers

The techniques adopted for modeling and training are briefly introduced as follows, and more details can be found in References15).

2.2.1 Activation function

The activation function converts the input signal of a node in an ANN to an output signal, and the identity function, sigmoid, Tanh, ReLu, and softmax function are used as the activation function. In general, ANN performs learning of complex nonlinear phenomena by adopting a nonlinear function as an activation function. Among them, the sigmoid function is useful for converting all values into probabilities, and generates output values between 0 or 1, which can be used for binary classification. The Tanh function has a smooth shape like the sigmoid function, and has the advantage of being able to generate negative values because the output values range between -1 and 1 depending on the input value. The sigmoid function and the tanh function are disadvantageous for use in the learning of DNN because it is possible that the gradient of the function vanishes by saturation. However, the ReLU function outputs 0 when the input value is negative, and outputs the input value as it is when the input is 0 or positive. The softmax function used in the classification problem converts the input values between 0 and 1 and the sum of the output values is always 1. The softmax function increases the differences in the portions of the input values, making the portion of the largest value even larger and the portions of other values even smaller, and is used as the activation function of the last node for the classification neural network.

2.2.2 Transfer learning

Transfer learning performs learning by using models that has been pre-trained with large-scale data such as VGG, GoogLeNet, and MobileNet, rather than learning a complex CNN structure from scratch in the design of an image classification model using CNN and it allows a accurate model in a relatively time-efficient manner. Transfer learning adopts model optimization through fine-tuning. The fine-tuning strategies can be divided into three methods: retraining the entire model, fine- turning only the parameters of the fully connected (FC) layers, and fine-tuning some convolutional layers and all fully connected layers.

2.2.3 Data augmentation

In deep learning, the more number of training datasets are available, improved accuracy can be achieved. Insufficient number of training datasets leads to over- fitting and ensuring a large number of training datasets is important in the performance. When CNN uses images as input data and the number of training datasets is insufficient, data augmentation can be used to address the problem of insufficient input. Data augmentation is a method of adding artificially processed images to the given training datasets through effects such as flipping, rotation, and changing brightness to the input images. In this paper, for convenience of presentation, the method of using original data without data augmentation is indicated as learning using the original data.

3. Supervised learning with training from scratch

3.1 Original data utilization

Z. Guo et al.16) proposed a model for classifying normal and defective welds by applying CNN to electric resistance welding in the line pipe manufacturing process, and achieved an accuracy of 99.01%. The cross-section images of the welds were used as input and the sigmoid function was used in the output layer to classify normal and defective welds. The number of hidden layers of the CNN was 5, 7, and 13, and the difference in accuracy according to the network depth was examined. With increasing network depth, the average time per epoch was 261.68 s, 266.24 s, and 267.26 s in case of 5, 7, and 13 layers, respectively, indicating no significant difference but the error rates were 7.89%, 4.93%, and 0.99%, showing improved accuracy. It was reported that superior accuracy results were obtained compared to the error rate of 2.3% at human inspection.
Table 1
Summary of research using original data and training from scratch
Process Input Output Network No. of dataset Ref. No
Electric Resistance Welding 256×256 (px) image from cross-sections Classification
- Weld quality (Good/Defect)
CNN-FCN 13 HL*** 11026 (TR*) 304 (TE**) 16
Arc Welding 320×240 (px) image from weld surfaces Classification
- Weld quality (Good, Porosity, Undercut, Spatter)
CNN-FCN
5 HL
120 17
Laser Welding 351 features of in-situ signals from photo diodes, image sensor, and spectrometer Classification
- Weld quality (Good, Blowout, Humping, Undercut)
CNN-FCN 7 HL 7500 (TR) 5000 (TE) from 25 welding runs 18
Gas Tungsten Arc Welding 400×487 (px) in-situ weld pool image Classification
- Weld quality (2, 4, 6 classes)
12 models (6 CNNs & 6 FCNs) 26666 (TR) 6588 (TE) 20
Resistance Spot Welding 128×128 (px) image from weld surfaces Regression
- Strength
- Nugget diameter
Classification
- Fracture mode, Expulsion
CNN-FCN 11 HL 90 21
Laser Welding 121×121 (px) in-situ image of weld pool including keyhole Classification
- Penetration (2, 3, 4 classes)
10 Models (6 CNNs, {HOG, LBP, BOF, or SAE} + SVM) 10,000 (TR) 1000 (TE) 22

* TR: For training;

** TE: For test;

*** HL: Hidden Layer

A. Kjumaidi et al.17) proposed a CNN-based defect detection model using photo images of arc welding bead surface as input, and using only 3 hidden CNN layers and 2 hidden FC layer, test accuracy of 95.83% was achieved for 4-class classification of defects (good, porosity, undercut, and spatter).
Y. Zhang et al.18) performed in-situ measurement of welding signals using multi-sensor technology, and a CNN-based defect detection model was proposed based on the input signals. The multi-sensor system comprised a photodiode, a spectrometer, and two high-speed cameras. The collected photodiode signals were decomposed for each frequency by the Wavelet Packet Decomposition (WPD) method to obtain 320 feature signals. In addition, the spectrometer signals were collected by dividing the wavelength range between 400 nm and 900 nm into 25 bands. The information on the size and location of the keyholes was collected with No. 1 camera with 976 nm laser light and the bandpass filter applied, and the spatter and plume related information was collected from No. 2 camera with a wavelength range of 320 nm-750 nm. After collecting a total of 351 signals from various sensors at a frequency of 500 Hz, the rest were filled with value 1 to form 400 input nodes. For verification on the accuracy of the CNN model, the results were compared with the fully connected network (FCN) model19) result with same input nodes. In the comparison, the mean error of the FCN model was 9.65% and the mean error of the CNN model was 4.7%, indicating that the CNN model accuracy was superior to that of FCN model. For generalization of the proposed CNN model, the verification of the model was performed by experimenting under four welding conditions. As a result of CNN model classification, error rates of 1.73%, 8.0%, 11.2%, and 8.75% were obtained in each defect class (good, blowout, humping, undercut). When the cases of classification error were analyzed, there were cases where the blowout defect and the undercut defect were confused in detection, or the humping defect and good welds were confused.
D. Bacioiu et al.20) constructed various DNN models for welding defects classification in GTA welding of aluminum 5083 alloy and compared the model per- formances. In a total of 60 welding experiments, the in-situ molten pool images were used as input data, and the 6 CNN models and 6 FCN models were developed with defect classifications of 2-class (good, defective), 4-class (good, melt-through, contamination, lack of fusion) and 6-class (good, melt-through, contamination, lack of fusion, misalignment, lack of penetration). Then, the performance was compared according to input image resolution, model depth, and learning rate. In the evaluation of the input image resolution, the comparison of original data images of 800x974 pixels with sampled images at 400x487 and 25x30 pixels confirmed that in the 6-class classification the CNN model showed higher sensitivity to the final classification accuracy than the FCN model. In all models, the accuracy improved with increasing model depth, and learning did not converge at a learning rate of 10-3 or below. In all evaluations, CNN models showed higher accuracy than FCN models. S. Choi et al.21) presented a CNN model that predicts weld quality (tensile shear strength, nugget diameter, fracture mode of welds, expulsion occurrence) through the surface images of resistance spot welds for automobile steel sheets with a tensile strength of 980 MPa. The surface indentation and heat trace correlate with the weld quality, and thus the tensile shear strength, nugget diameter, fracture mode of welds, and expulsion occurrence were predicted with the accuracy of 98.6%, 98.8%, 100%, and 100%, respectively, through the CNN model. The generality of the model was verified through a test using data not included in the training, and the verification result confirmed that the prediction accuracy of the tensile shear strength and nugget diameter was 97.8% and 97.4%, respectively
Z. Zhang et al.22) used a CNN model to determine the penetration state of tailor-rolled blank (TRB) in-situ laser welding. The applied penetration state classification was 2-class (incomplete penetration, complete penetration), 3-class (incomplete penetration, desirable-complete penetration, overpenetration), and 4-class (incomplete penetration, desirable penetration, complete penetration, overpenetration). Six CNN models with different number of kernels and convolutional layers were trained and evaluated in comparison. It was confirmed that the increased complexity of the model did not necessarily lead to improved accuracy, and the optimal CNN model was determined considering the model complexity in addition to accuracy. Moreover, apart from the model complexity, the image sampling frequency (100 fps, 300 fps, 3000 fps) of the input images that can affect the latency of model was varied to conduct experiments in comparison. Since there was no significant difference in accuracy among models with difference sampling frequencies, 100 fps was selected with the lowest latency. To verify the applicability of the CNN model, it was compared with other methods [(HOG, LBP, Bag of features (BOF), or SAE)+SVM]. As a result of the verification, the CNN model in 2-class classification showed the highest prediction accuracy, and in the case of 3-class classification, the prediction accuracy of the CNN model was 99%, 93.8%, and 96.5%, respectively, indicating higher performance than other models.
When the penetration state was classified into 4 classes, the accuracy of each model was relatively low, but the stability in the accuracy of the CNN model was high.

3.2 Data augmentation

W. Hou et al.23) investigated a defect classification model (good, incomplete penetration, pore, slag inclusion, crack) from X-ray images of various welds in an open database GDXray. First, 3503 of 32×32-pixel images were extracted from 88 X-ray images by subsampling. Since the distribution of each type of defect is unbalanced, resampling was performed using ROS (Random Over Sampling), RUS (Random UnderSampling), and SMOTE (Synthetic Minority Oversampling TEchnique) methods. The RUS method is the simplest approach that randomly deletes samples from the majority class to balance the class distribution and the ROS method duplicates samples selected from the minority class and adds to the original dataset to balance the class distribution. In addition, the SMOTE method is a resampling method in which Euclidean distance is calculated to randomly find k-nearest neighbors to create new artificial samples. For classification models, traditional methods of Haralick texture features and HOG (Histograms of Oriented Gradients) were used as well as the deep learning method SSAE (Stacked Sparse Auto-Encoders), and two CNN models with different depths were used. For visual comparison on the results of extracted features through each model, each part was visualized into a 2D map using the t-SNE method to compare the accuracy. From the comparison, it was confirmed that the t-SNE map of the original images was disorderly, and the t-SNE map of the models with Haralick technique and HOG method also showed poor data classification performance. Both SSAE and CNN models showed better performance of classification in the t-SNE map. Between the two CNN models, the one with larger depth showed a better classification performance. Among them, the classification of samples without defects and samples with pores showed the best performance, and crack defects were the most difficult for classification. In the case of the ROS datasets, the mean accuracy of the defect classification for Haralick, HOG, SSAE, CNN1, and CNN2 was 60.9%, 81.6%, 91.4%, 94.7%, and 96.3%, respectively, in the case of the RUS datasets, the accuracy values were 60.7%, 67.5%, and 87.3. %, 76.6%, 79.9%, and for SMOTE datasets, the values were 63.1%, 82.2%, 90.5%, 94.4%, and 97.2%, respectively. In conclusion, they reported that the deep learning method showed better feature extraction performance than the traditional model, and that with SMOTE method, the problem of accuracy reduction from data imbalance can be resolved.
J. Park et al.24) proposed two-step CNN models for defect detection in engine transmission welds. In the first CNN model, the representation of collected images of the circular welding area of the engine transmission was converted from Polar coordinates to Cartesian coordinates and the center point was predicted. Assuming that the weld width was fixed, the background except the welds was removed, thereby optimizing the input data to the second defect detection CNN model. Compared with Hough circle, a traditional method, the two-step models have showed superior performance under dark lighting environments and with blurred images. Since there are not many defect data images in the training datasets, image transformation (rotation, translation, scaling) and image distortion (brightness and contrast adjustment) were performed to improve performance. In the second defect detection CNN model, the mini-batch gradient descent method can lead to biased learning due to data imbalance, class-specific batch sampling method with a set ratio of OK and NG images of each batch was used. For the defect detection model, the performance was comparatively evaluated between HOG+SVM, LBP (Local Binary Pattern)+SVM and four CNN models, and all CNN based models showed outstanding performances.
Table 2
Summary of research using data augmentation and training from scratch
Process Input Output Network Data augmentation Ref. No
X-ray Inspection 32×32 (px) X-ray images from GDXray-Weld Classification
- Weld quality (2, 5 classes)
5 Models (Haralick, HOG, SSAE, 2 CNNs) Subsampling ROS**, RUS*** SMOTE*4 23
Transmission Welds 80×256 (px) for center detection 32x352 (px) for defect inspection from weld surfaces Regression
- Center point
Classification
- Weld quality (Good, NG)
6 Models (HOG+SVM, LBP+SVM, 4 CNNs) Rotation Translation Scaling Brightness Adj. Contrast Adj. 24
Flux Cored Arc Welding 150×125 (px) in-situ image of weld pool Regression
- 8 molten pool features
CNN-FCN 10 HL* Horizontal flip Translation Scaling 25
Gas Tungsten Arc Weldng 100×100 (px) in-situ images of weld pool in three-way Classification
- Weld quality (Under-penetration, Full penetration, Burn through)
CNN-FCN 10 HL Noise addition
Rotation
26

* HL: Hidden Layer;

** ROS: Random Over-Sampling;

*** RUS: Random Under-Sampling;

* 4 SMOTE: Synthetic Minority Oversampling TEchnique

For automation of Flux Cored Arc (FCA) Welding, T. Ashida et al.25) used a CNN model to extract features of molten pool and calculated the distance between the leading of the molten pool and arc center and the width of the pool to investigate a feedback control system to form back bead. Data augmentation was performed using horizontal flip, translation, and scaling to increase the number of datasets from 2,400 to 12,000. When comparing between the proposed CNN model and the conventional feature-extraction image processing algorithms, the mean prediction error of the leading end position of the molten pool was reduced by 0.3 mm. The seam tracking and back bead formation was verified by implementing the developed feedback control algorithm to the butt weld joint with a joint gap from 3 mm to 10 mm along the weld seam which is also deviated by 10 mm from the pre-taught path.
Z. Zhang et al.26) developed a CNN model for in-situ weld defects prediction in pulsed gas tungsten arc (GTA) welding of aluminum alloys. In order to increase the training speed and accuracy of the CNN model, a system was devised to collect 3-way information simultaneously from the top front, top back, and back seam. Welding was performed on rectangular, dumbbell-shaped, and ladder-shaped specimens to generate welding defects. Downsampling was performed on the original images of 1392x1040 pixels with the arithmetic mean and thus input images of 100x100 pixels were used. For data augmentation, the number of datasets was increased by adding noise (salt and pepper noise) and randomly applying 15 and 30 degrees rotation to the input images. After the data augmentation, the accuracy improved by 3.88% from 92.3% to 96.18%, verifying the effectiveness of data augmentation. In addition, for optimization of CNN model, the accuracy and the final training speed were compared by varying 4 convolutional layer structures and the number of kernels. The training time and efficiency increased with the number of kernels, and the accuracy increased with the increase in the number of feature images excluding 2D images with only 0 in the pooling layer.

4. Supervised learning with transfer learning

4.1 Research using original data

H. Zhu et al.27) applied a CNN model for defects (normal, overlap, spatter, porosity) classification on weld surfaces of Gas Metal Arc Welding (GMAW). Transfer learning was performed using LeNet-5, and since the softmax function, which is frequently used in the last layer of the final fully connected layer of the CNN model, has a disadvantage of performance degradation when the number of training datasets is insufficient, Random Forest and SVM classifier were used for comparison of the performance. Preprocessing was performed such as median filtering, image enhancement by gradation processing, and OTSU image thresholding for all collected images of defects on weld surfaces. From the results obtained from softmax, SVM and Random Forest classifiers, the accuracy obtained using the softmax classifier was the lowest and the accuracy of the proposed Random Forest Classifier showed the best performance.
Table 3
Summary of research using original data and transfer learning
Process Input Output Base network No. of dataset Ref. No
Gas Metal Arc Welding 80×120 (px) image from weld surfaces Classification
- Weld quality (Normal, Overlap, Spatter, Porosity)
LeNet-5 +
Classification (Softmax, Random Forest, SVM)
320 (TR*) 80 (TE**) 27
Laser Welding 150×150 (px) image from weld surfaces Classification
- Weld quality (Normal, Porosity, Level misalignment)
AlexNet, VGG-16, Resnet-50, Densenet-121, MobileNetV3-Large 7217 (TR) 814(TE) 28
Gas Tungsten Arc Spot Weldng 640×480 (px) in-situ image of weld pool Classification
- Weld penetration (Under, Desirable, Excessive penetration)
ResNet 18-layers model 25643 (TR) 2851 (TE) 29
Laser Welding 224×224 (px) in-situ image from a coaxial camera Classification
- Pore or No pore
AlexNet 1971 (TR) 1276 (TE) 30

* TR: For training;

** TE: For Test

Y. Yang et. al28) used photographs of laser welding area taken by CMOS digital camera and transfer learning of a CNN model was performed for classification of weld quality into 3 classes (normal, porosity, level misalignment) and 2 classes (qualified, defect). In the basic VGG-16 model, the weight of the convolutional layer was fixed and training was performed only on the weight of the FC layer. The result of this model was compared with the other eight models (pre-optimization AlexNet, Resnet- 50, VGG-16, Densenet-121, MobileNetV3-Large and post-optimization pre-AlexNet, pre-Resnet-50, and pre- Densenet-121). The 2-class binary classification achieved accuracy of 90% or higher in all cases. For multiclass classification of 3-class, Resnet-50 and MobileNetV3- Large were the most advanced deep learning algorithm and showed good performance, but 3 GPUs were needed for the training. The optimized VGG-16 model in the study showed excellent performance with training time of 1 hour without having to use GPU.
W. Jiao et al.29) developed a weld penetration prediction model using CNN in the GTA spot welding process. Images were collected using the camera positioned on the top-side of the specimen, and the penetration state (partial, full and excessive penetration) was predicted through correlation with the area of the bright parts from the camera on the back side. For transfer learning, ResNet model with 18 layers was used and compared with a 9-layer CNN model. The accuracy of the ResNet model with transfer learning was 96.35%, and the test accuracy of the compared CNN model was 92.7%, confirming the superiority in the results obtained with transfer learning.
B. Zhang et al.30) constructed a model for prediction of the in-situ internal porosity status of aluminum 6061 alloy in laser welding. Consisting of 532 nm light and high-speed camera, in-situ weld-pool images were collected, and the training was performed using porosity occurrence information confirmed through image processing of the longitudinal cross-section. AlexNet-based transfer learning was performed, and most of the pores were successfully detected, but during the verification process, the pores with size of 100 μm or less, and some deep pores were not detected.

4.2 Data augmentation

N. Yang et al.31) proposed a LeNet-5-based transfer learning CNN model for defect classification using X-ray weld images. From the images acquired through the experiment, images around the weld area were extracted and the contrast in the images was improved by adjusting the gray level. Since there were not enough samples in the dataset, the images were rotated, shifted and blurred and noise was also added to increase the size of the dataset.
For optimization of the CNN model, each size of convolution kernels was varied for evaluation. The larger the kernel size, there was a slight increase in accuracy and the convergence became slower. A function that synthesized LReLU and Softplus with reference to 0 was used for classification to prevent saturation and it was confirmed that this function had superior performance than sigmoid function. When compared with the accuracy values of LeNet-5, ANN and SVM (0.758, 0.897, 0.982), the developed model showed a superior accuracy at 0.993.
C. V. Dung et al.32) developed a CNN model for detection of fatigue cracks from photographic images of welded joints in a structure. From the images collected from the structure, sub-images of 64x64-pixel were extracted and datasets were constructed by classification according to the crack occurrence. In order to resolve the problem of imbalanced training dataset due to smaller number of images with cracks than images without cracks, data augmentation was performed through rotation, shift, shear, zoom, and flip of the crack images. CNNs of three different structures which are shallow CNN, VGG-16-based BN(Bottle Neck) training, and VGG-16-based fine-turning training were evaluated. In all cases, the results with data augmentation showed that the accuracy improved by 2% or higher.
Table 4
Summary of research using data augmentation and transfer learning
Process Input Output Base network Data augmentation Ref. No
X-ray Inspection 60×60 (px) X-ray images Classification
- Weld quality (2, 5 classes)
LeNet-5 Rotation, Translation
Blurr, Noise
31
Transmission Welds 64×64 (px) image from surfaces Classification
- Crack, No Crack
VGG-16 Rotation, Shift Shear, Zoom, Flip 32

5. Unsupervised learning with autoencoder

CNN models are applied to predict or classify specific values in welding process but there are also cases when CNN models are used for feature extraction from acquired images. Autoencoder is unsupervised learning using CNN and reconstructs the original images through encoding and decoding processes. In the process, constraints such as noise are applied so that key features can be applied when the original images are reconstructed. Autoencoder is used when extracting features of the original images in high dimension or when reduced images are needed with the features retained.
J. Guenther et al.11) applied autoencoder method with CNN for feature extraction in laser welding and extracted 16 features from the input image. The extracted 16 features were used for reinforcement learning, and the weld quality was achieved by controlling the output in laser welding.
A. Muniategui et al.33) developed a quality determination model of resistance spot welding using a fuzzy algorithm. However, the fuzzy algorithm that uses the original images with high resolution of 1000x800-pixel as input is not efficient in terms of computation time for application in real-time process and the data storage space problem may arise due to the high storage capacity required. To overcome these limitations, an autoencoder was used to reduce the image resolution to 15x15 pixels while retaining the features of the image and the reduced images were applied to the fuzzy algorithm. Finally, quality prediction was achieved at an accuracy of 88% or higher in resistance spot welding.

6. Summary and Outlook

As can be seen from the review in this study, among various types of deep learning techniques, CNN models with images have been actively applied in welding research. The field of Welding is typically dominated by tacit knowledge rather than rule-based explicit knowledge, indicating high applicability of data-based deep learning, and the applications of CNN in welding research are expected to increase even further. The in-situ measurement of waveform-based time series signals have been actively used in determination of weld quality in previous studies, and in the future, there will be increased investigation and adoption of multi-sensor-based deep learning techniques where continuous waveform sensors and image sensors are applied simultaneously. Furthermore, at present, the images at the time of measurement are used for quality classification or regression, but in the future, hybrid models of combining RNN and CNN will be applied, leading to more intelligent models in which the information extracted from images in the past will be transferred to the current state prediction.

Acknowledgement

This research was supported by the MOTIE (Ministry of Trade, Industry, and Energy) in Korea, under the Fostering Global Talents for Innovative Growth Program (P0008750) supervised by the Korea Institute for Advancement of Technology (KIAT)

References

1. P. Kim, MATLAB Deep Learning, Apress Seoul, Korea. (2017) https://doi.org/10.1007/978-1-4842-2845-6_5
[CROSSREF] 
2. D.E. Rumelhart, G. E. Hinton, and R. J. Williams, Learning Representations by Back-Propagating Errors, Nature. 323(6088) (1986) 533–536. https://doi.org/10.1038/323533a0
[CROSSREF] 
3. K. andersen, G. E. Cook, G. Karsai, and K. Ramaswamy, Artificial Neural Networks Applied to Arc Welding Process Modeling and Control, IEEE Trans. Ind. Appl. 26(5) (1990) 824–830. https://doi.org/10.1109/28.60056
[CROSSREF] 
4. G.E. Cook, R. J. Barnett, K. Andersen, and A. M. Strauss, Weld Modeling and Control Using Artificial Neural Networks, IEEE Trans. Ind. Appl. 31(6) (1995) 1484–1491. https://doi.org/10.1109/28.475745
[CROSSREF] 
5. H.S. Moon and S. J. Na, A Neuro-Fuzzy Approach to Select Welding Conditions for Welding Quality Impro- vement in Horizontal Fillet Welding, J. Manuf. Syst. 15(6) (1996) 392–403. https://doi.org/10.1016/S0278-6125(97)83053-1
[CROSSREF] 
6. K. Bae and S. J. Na, A Study of Vision-Based Measure- ment of Weld Joint Shape Incorporating the Neural Network, Proc. Inst. Mech. Eng., Part B:J. Eng. Manuf. 208(1) (1994) 61–69. https://doi.org/10.1243/PIME_PROC_1994_208_060_02
[CROSSREF] 
7. S. Lee, W. Chang, W. Yoo, and S. J. Na, A Study on a Vision Sensor Based Laser Welding System for Bellows, J. Manuf. Syst. 19(4) (2000) 249–255. https://doi.org/10.1016/S0278-6125(01)80004-2
[CROSSREF] 
8. A. I. Krizhevsky, G. E. Sutskever, and . Hinton, Imagenet Classification with Deep Convolutional Neural Networks, Advances in Neural Information Processing systems. (2012) Lake Taho, Navada, USA1097–1105. https://doi.org/10.1145/3065386
[CROSSREF] 
9. K. Cho, B. Van Merrieboer, C. Gulcehre, D. Bahdanau, F. Bougares, H. Schwenk, and Y. Bengio, Learning Phrase Representations Using RNN Encoder-Eecoder for Sta- tistical Machine Translation, arXiv preprint arXiv:1406.1078. (2014) https://doi.org/10.3115/v1/d14-1179
[CROSSREF] 
10. J. Weston, S. Chopra, and A. Bordes, Memory Networks, arXiv preprint arXiv. 1410 3916–2014.
11. J. Guenther, P. M. Pilarski, G. Helfrich, H. Shen, and K. Diepold, Intelligent Laser Welding Through Representa- tion, Prediction, and Control Learning:an Architecture with Deep Neural Networks and Reinforcement Learning, Mechatron. 34 (2016) 1–11. https://doi.org/10.1016/j.mechatronics.2015.09.004
[CROSSREF] 
12. M.S. Kim, S. M. Shin, S. Kim, and D. H. Rhee, A Study on the Algorithm for Determining Back Bead Gene- ration in GMA Welding Using Deep Learning, J. Weld. Join. 36(2) (2018) 74–81. https://doi.org/10.5781/JWJ.2018.36.2.11
[CROSSREF] 
13. D. Petkovic, Prediction of Laser Welding Quality by Com- putational Intelligence Approaches, Optik. 140 (2017) 597–600. https://doi.org/10.1016/j.ijleo.2017.04.088
[CROSSREF] 
14. J. Chen, T. Wang, X. Gao, and L. Wei, Real-Time Mo- nitoring of High-Power Disk Laser Welding Based on Support Vector Machine, Comput. Ind. 94 (2018) 75–81. https://doi.org/10.1016/j.compind.2017.10.003
[CROSSREF] 
15. F. Chollet, Deep Learning With Python, Manning Pu- blications Co. New York, NY, USA(2018)
16. Z. Guo, S. Ye, Y. Wang, and C. Lin, Resistance Wel- ding Spot Defect Detection with Convolutional Neural Networks, International Conference on Computer Vision Systems. (2017) Shenzhen, China169–174. https://doi.org/10.1007/978-3-319-68345-4_15
[CROSSREF] 
17. A. Khumaidi, E. M. Yuniarno, and M. H. Purnomo, Welding Defect Classification Based on Convolution Neural Network (CNN) and Gaussian Kernel, Inter- national Seminar on Intelligent Technology and Its Applications. (2017) Surabaya, Indonesia261–265. https://doi.org/10.1109/ISITIA.2017.8124091
[CROSSREF] 
18. Y. Zhang, D. You, X. Gao, N. Zhang, and P. P. Gao, Welding Defects Detection Based on Deep Learning with Multiple Optical Sensors During Disk Laser Welding of Thick Plates, J. Manuf. Syst. 51 (2019) 87–94. https://doi.org/10.1016/j.jmsy.2019.02.004
[CROSSREF] 
19. Y. Zhang, D. You, X. Gao, and S. Katayama, online Monitoring of Welding Status Based on a DBN Model During Laser Welding, Eng. 5(4) (2019) 671–678. https://doi.org/10.1016/j.eng.2019.01.016
[CROSSREF] 
20. D. Bacioiu, G. Melton, M. Papaelias, and R. Shaw, Automated Defect Classification of Aluminium 5083 TIG Welding Using HDR Camera and Neural Networks, J. Manuf. Process. 45 (2019) 603–613. https://doi.org/10.1016/j.jmapro.2019.07.020
[CROSSREF] 
21. S. Choi, I. Hwang, Y. Kim, B. Kang, and M. Kang, Prediction of the Weld Qualities Using Surface App- earance Image in Resistance Spot Welding, Met. 9(8) (2019) 831. https://doi.org/10.3390/met9080831
[CROSSREF] 
22. Z. Zhang, B. Li, W. Zhang, R. Lu, S. Wada, and Y. Zhang, Real-Time Penetration State Monitoring Using Convolutional Neural Network for Laser Welding of Tailor Rolled Blanks, J. Manuf. Syst. 54 (2020) 348–360. https://doi.org/10.1016/j.jmsy.2020.01.006
[CROSSREF] 
23. W. Hou, Y. Wei, Y. Jin, and C. Zhu, Deep Features Based on a DCNN Model for Classifying Imbalanced Weld Flaw Types, Measurement. 131 (2019) 482–489. https://doi.org/10.1016/j.measurement.2018.09.011
[CROSSREF] 
24. J.K. Park, W. H. An, and D.J. Kang, Convolutional Neural Network Based Surface Inspection System for Non-Patterned Welding Defects, Int. J. Precis. Eng. Manuf. 20(3) (2019) 363–374. https://doi.org/10.1007/s12541-019-00074-4
[CROSSREF] 
25. T. Ashida, A. Okamoto, K. Ozaki, M. Hida, and T. Yama- shita, Development of Image Sensing Technology for Automatic Welding (Image Recognition by Deep Learning), Kobelco Technol. Rev. 37 (April., 2019) 77–81.
26. Z. Zhang, G. Wen, and S. Chen, Weld Image Deep Learning-Based On-Line Defects Detection Using Convolutional Neural Networks for Al Alloy in Robotic Arc Welding, J. Manuf. Process. 45 (2019) 208–216. https://doi.org/10.1016/j.jmapro.2019.06.023
[CROSSREF] 
27. H. Zhu, W. Ge, and Z. Liu, Deep Learning-Based Classification of Weld Surface Defects, Appl. Sci. 9(16) (2019) 3312. https://doi.org/10.3390/app9163312
[CROSSREF] 
28. Y. Yang, L. Pan, J. Ma, R. Yang, Y. Zhu, Y. Yang, and L. Zhang, A High-Performance Deep Learning Algo- rithm for the Automated Optical Inspection of Laser Welding, Appl. Sci. 10(3) (2020) 933. https://doi.org/10.3390/app10030933
[CROSSREF] 
29. W. Jiao, Q. Wang, Y. Cheng, and Y. Zhang, End-To- End Prediction of Weld Penetration:A Deep Learning and Transfer Learning Based Method, J. Manuf. Process. (2020) https://doi.org/10.1016/J.Jmapro.2020.01.044
[CROSSREF]  [PUBMED]  [PMC] 
30. B. Zhang, K. M. Hong, and Y. C. Shin, Deep-Learning- Based Porosity Monitoring of Laser Welding Process, Manuf. Lett. 23 (2020) 62–66. https://doi.org/10.1016/j.mfglet.2020.01.001
[CROSSREF] 
31. N. Yang, L. Niu, and G. Mi. Chen, X-Ray Weld Image Classification Using Improved Convolutional Neural Network, AIP Conference Proceedings. (1995) (2018) 020035. https://doi.org/10.1063/1.504∾
[CROSSREF] 
32. C.V. Dung, H. Sekiya, S. Hirano, T. Okatani, and C. Miki, A Vision-Based Method for Crack Detection in Gusset Plate Welded Joints of Steel Bridges Using Deep Convolutional Neural Networks, Autom. Constr. 102 (2019) 217–229. https://doi.org/10.1016/j.autcon.2019.02.013
[CROSSREF] 
33. AB. Muniategui, L. Heriz, M. Eciolaza, A. Ayuso, I. Iturrioz, . Quintana, and P. Lvarez, Spot Welding Monitoring System Based on Fuzzy Classification and Deep Learning, IEEE International Conference on Fuzzy Systems, Naples, Italy. (2017) 1–6. https://doi.org/10.1109/fuzz-ieee.2017.8015618
[CROSSREF] 


ABOUT
BROWSE ARTICLES
ARTICLE CATEGORY 
FOR CONTRIBUTORS
Editorial Office
#304, San-Jeong Building, 23, Gukhoe-daero 66-gil, Yeongdeungpo-gu, Seoul 07237, Korea
Tel: +82-2-538-6511    Fax: +82-2-538-6510    E-mail: koweld@kwjs.or.kr                

Copyright © 2024 by The Korean Welding and Joining Society.

Developed in M2PI