In Situ Nondestructive Fatigue‐Life Prediction of Additive Manufactured Parts by Establishing a Process–Defect–Property Relationship

The presence of process‐induced internal defects (i.e., pores, microcracks, and lack‐of‐fusions) significantly deteriorates the structural durability of parts fabricated by additive manufacturing. However, traditional defects characterization techniques, such as X‐ray CT and ultrasonic scanning, are costly and time‐consuming. There is a research gap in the nondestructive evaluation of fatigue performance directly from the process signature of laser‐based additive manufacturing processes. Herein, a novel two‐phase modeling methodology is proposed for fatigue life prediction based on in situ monitoring of thermal history. Phase (I) includes a convolutional neural network designed to detect the relative size of the defects (i.e., small gas pores and large lack‐of‐fusions) by leveraging processed thermal images. Subsequently, a fatigue‐life prediction model is trained in Phase (II) by incorporating the defect characteristics extracted from Phase (I) to evaluate the fatigue performance. Estimating defect characteristics from the in situ thermal history facilitates the fatigue predicting process.


Introduction
The additive manufacturing (AM) domain contains novel fabrication techniques, which are defined as the process of joining the material together in a layer-by-layer manner, to make 3D objects. [1] AM enables fabricating complex geometries where most of them are not feasible in the domain of conventional manufacturing techniques. The significant design flexibility offered by AM techniques can save a noticeable amount of money and time if applied correctly. However, the broader adoption of AM processes has faced challenging issues, especially at satisfying quality standards and process repeatability. [2] The compromised structural performance resulted from processinduced defects is the main challenge against the continued adoption of AM in different industries. [3,4] Among different modes of mechanical failures, fatigue failure, that is, failure under cyclic loading, is the dominant failure mode in mission-critical applications. [5] This is due to the fact that fatigue is a local phenomenon; thus, it is more directly affected by the microstructural features. [5] 62% of aircraft structures have had failures due to fatigue, where only 14% of them were because of mechanical overload. [6] Meeting fatigue and durability requirements has proven to be a challenging task for AM parts. [3,4] Process and design parameters have shown a significant impact on the microstructure and defect properties of AM parts, and thus largely determine their fatigue life behavior. [3,5] In the absence of voids and inclusions, which typically serve as crack initiation sites, slip bands usually drive the crack initiation in material. [7] Slip length and slip planarity are the two most important factors in determining fatigue properties of titanium alloys. [8] Slip length can be controlled by microstructural features including α laths, phase boundaries and colony boundaries, while slip planarity is governed by oxygen and aluminum content as well as secondary precipitation (Ti3Al). [7] Since fatigue cracks tend to initiate at the longest existing crystallographic slip bands in the microstructure, reducing the maximum dislocation slip length is critical for enhancing the material resistance against fatigue crack initiation (i.e., HCF strength). Crystallographic orientations of the adjacent grains as well as the grain size and geometry act as a barrier for DOI: 10.1002/aisy.202000268 The presence of process-induced internal defects (i.e., pores, microcracks, and lack-of-fusions) significantly deteriorates the structural durability of parts fabricated by additive manufacturing. However, traditional defects characterization techniques, such as X-ray CT and ultrasonic scanning, are costly and timeconsuming. There is a research gap in the nondestructive evaluation of fatigue performance directly from the process signature of laser-based additive manufacturing processes. Herein, a novel two-phase modeling methodology is proposed for fatigue life prediction based on in situ monitoring of thermal history. Phase (I) includes a convolutional neural network designed to detect the relative size of the defects (i.e., small gas pores and large lack-of-fusions) by leveraging processed thermal images. Subsequently, a fatigue-life prediction model is trained in Phase (II) by incorporating the defect characteristics extracted from Phase (I) to evaluate the fatigue performance. Estimating defect characteristics from the in situ thermal history facilitates the fatigue predicting process. microcrack growth. [9,10] The crack path deflection across a grain boundary is strongly influenced by the orientation of neighbored grains. High-angle colony boundaries are an effective barrier to microcrack growth and the crack will be arrested/retarded when none of the available slip planes are oriented closely. [10][11][12] Fatigue behavior of AM metallic materials is currently dominated by process-induced defects, such as gas-containing pores and lack of fusion (LOF) defects. When the microstructure is nearly homogeneous and the residual stresses in an AM part are limited-for instance, by post-fabrication heat treatments or preheating the build plate-variations in size, location, and spacing of defects are the main reasons for the uncertainty and scatter in the AM fatigue data. [13][14][15] Therefore, defect characteristics (i.e., size, shape, location, and distribution) are leading influential parameters affecting fatigue life performance. [16] In this article, all the specimens underwent a homogenizing and stress relieving heat treatment to minimize the effects of residual stresses and microstructural variations, yet microstructural features, including the grain size, grain orientation, and misorientation angles, dictated by the manufacturing and postmanufacturing heat treatment processes, are taken into account for fatigue-life prediction. Several studies have demonstrated that one of the main issues in the fatigue-life description of AM parts is due to scattering in the high-cycle fatigue (HCF) regime, mostly resulting from the presence of process-induced defects. [3,16,17] Thus, being able to predict the fatigue behavior of AM materials based on the defect characteristics would be an important step towards improving the reliability of AM parts. [14] Fatigue performance of an AM part may not necessarily be dominated by the most massive defect, preferably a combination of microstructural features, including grain size and defect characteristics (i.e., size, shape, location, and distribution) typically compete together for initiating cracks. [18] It has been shown that the crack initiating defect for AM Ti-6Al-4V is rarely simply the largest defect; rather, various aspects affect the tendency for a defect to initiate a crack, including proximity to the surface and to other pores. [19] This indicates that the fatigue life prediction based on the largest of the defects found in multiple specimens can be conservative in some cases. As a result, without considering the defect location and spacing, it is not possible to discriminate critical defects from harmless ones. Thus, a successful in situ qualification and certification algorithm requires the ability to identify the formation of defects during the build process and link the characteristics of detected anomalies to the performance and durability of the fabricated part.
Developing proper applications for fracture mechanics-based damage tolerance principles provide a mechanism for effective risk mitigation in dealing with inherent material flaws, such as manufacturing-induced defects. [20] In this concept, that is, crack-growth-based and defect-tolerant modeling of fatigue, each anomaly is treated as a sharp crack from the first mission cycle. [15,21] This methodology is an appropriate tool for expressing the relationship between fatigue resistance and material quality and is widely used in aerospace industries. [20] , [22][23][24] One of the practical methods to predict the "fatigue" behavior of materials has been proven to be the use of the "crack propagation" concept since the mid-1980s by the Advisory Group for Aerospace Research and Development (AGARD) Structures and Materials Panel technical reports (R-732, R-767) on studying "small-crack" behavior. [25][26][27] Evidently, AM process-induced defects have a different formation mechanism and a disparate impact on fracturecritical properties. [20] In the presence of large lack of fusion (LOF) defects, crack-growth-based modeling of fatigue-that is, approximating LOFs as the preexisting sharp cracks-is a promising method for analyzing the fatigue-life. [14] However, for gas porosity, the assumption of preexisting sharp cracks may produce too conservative results. [20,28] In addition, for AM parts, due to the formation of a large number of defects with various sizes and locations, several defects may have the potential to serve as the crack initiation site under cyclic loading. [3] Using crackgrowth-based concepts to predict the fatigue life for multiple initiations of cracks and their coalesce, which is prevalent in AM parts, is challenging. [14] The microstructure-based multistage fatigue (MSF) model, developed initially for cast AL alloy, directly involves the effects of microstructural features and defect characteristics, that is, defect size, nearest neighbor distance (NND), and void volume fraction in the fatigue-life prediction. [29] MSF model characterizes fatigue life in three stages; incubation, microstructurally small crack growth, and long crack growth; based on microstructural features. MSF model used the local constrained cyclic microplasticity phenomenon to address the inclusions in fatigue crack incubation as well as microstructurally small crack growth, which also includes the impact of crystallographic orientation on crack tip displacement as the driving force.
In literature, casual relationships such as process-structureproperty linkages have been established by relying mainly on observations made from cleverly designed control experiments and physics-based models. According to vision setting documents, it has come clear to the experts in material science and engineering that data science and analytics have provided a third set of distinct, powerful tools for establishing the mentioned casual relationships. [30][31][32] Therefore, a data-driven methodology is proposed in this article to characterize the fatigue life from thermal history by creating the process-defect-structure relationship.
The traditional fatigue-life prediction models are using X-ray CT scans or similar post-manufacturing technologies to derive defect features such as defect size or nearest neighbor distance. The issue of nondestructive evaluation methods (X-ray CT scans) is the expense and unavailability of these machines for industrial sections. This study bridges this gap by proposing a model to predict the defect features from the process signature (i.e., thermal history). Many studies have aimed at building defect models based on morphological characteristics of a melt pool. A melt pool is an expression given to a superheated area of molten metal in proximity to the laser-material interface. [1] Melt pools are captured as high-resolution thermal images during the build process using a powerful co-axial thermal camera. These images carry the most critical information of the AM process. Morphological characteristics of a melt pool can refer to its depth, size, temperature distribution, and other geometrical features. For example, Tan et al. used melt pool dimensions extracted from a coaxially aligned high-speed camera in a welding process. [33] Song et al. proposed a predictive model to control the melt pool temperature for a high-power diode laser cladding process. [34] Clijsters et al. developed an anomaly detection model based on in situ melt pool dimensions for the selective laser melting (SLM) process. [35] Moreover, Lane et al. proposed a model based on high-speed thermographic measurements to monitor the melt pool region (characterized by contours and melt pool depth) in a laser powder bed fusion (LPBF) process. [36] Heigel et al. computed the melt pool length using a high-speed thermal camera in an LPBF process. [37] Finally, Fisher et al. studied the correlation between melt pool dimensions with temperature information gathered using a visible-light camera in an LPBF process. [38] Additionally, some studies have worked on qualitative and quantitative observation of melt pool dynamics where none of them directly worked on detecting defect formation. [39][40][41] Many studies, similar to the current one, have incorporated powerful techniques such as statistical analysis methods, machine learning (ML), or computer vision (CV) techniques to extract higher-order and yet prominent features of melt pools instead of morphological ones. For example, a neural network model based on data collected from high-speed cameras is developed by Lou et al. to identify the relationship between the process parameters and keyhole defect formation. [42] Off-nominal defects are detected by statistically comparing the pixel-wise trans-layer emitted light intensity profiles by Grasso et al. [43] Khanzadeh et al. proposed a clusteringbased methodology to analyze the melt pool morphologies extracted from a radial base function using the melt pool boundary captured by a thermal camera. [44,45] Dongsen et al. proposed a model based on both simple statistical and convolutional neural network descriptors to discriminate between melt pools generated under three different process parameters. [46] Finally, Scime and Beuth incorporated computer vision and unsupervised machine learning based on a scale-variant description of the melt pool to detect key-holing porosity and balling instabilities. [47] The aforementioned studies have mainly focused on methodologies to predict the occurrence of defects within the parts, and their limitation is that they have not characterized the magnitude of predicted defects. Having a method to predict the size of defects paves the way toward extracting defect features that play a critical role in describing the fatigue life. Therefore, we propose a methodology to predict the relative size of defects (small or large in terms of significant axis length) based on observed thermal history, that is, melt pool images.
In this article, a multiphase data-driven methodology is proposed to predict the final fatigue life based on in situ process signatures, that is, real-time thermal history. [48] The proposed method is a two-phase process. The objective of Phase (I) methodology is to study and extract the size and location of defects by establishing the thermal-defect relationship within the additive manufacturing processes (Figure 1a,b). Traditional defect characterization techniques, such as X-ray computed tomography (CT) and ultrasonic scanning, are time-consuming and expensive. Moreover, X-ray CT scans provide inferior results for bulk specimens and cannot serve as the main tool for real-time modeling or monitoring for AM processes. Defect prediction is one of the significant areas of study in the AM community. So far, the main objective of the previous studies was predicting the defect occurrence rather than defect size characterization. [47] , [49][50][51] To move the current state-of-the-art approaches one step forward, the relative size of the defect is predicted based on processed thermal images. After data preprocessing steps (data transformation), a convolutional neural network (CNN) is designed to discriminate the size of anomalies effectively.
The main objective of Phase (II) is to establish the defectfatigue relationship by modeling the fatigue behavior based on defect properties and microstructural characteristics ( Figure 1c); therefore, completing the process-defect-property relationship. The proposed methodology considers four parameters (i.e., defect size, nearest neighbor distance, grain size, and grain misorientation) to predict the fatigue life of Ti-6AL-4V parts. Furthermore, grain size and grain misorientation are added to the model due to their noticeable influence on fatigue life and the fact that they cannot be predicted from thermal history. The main purpose of this study is the identification of defect features from thermal history and their impact on the fatigue behavior. Therefore, the feature extraction of the microstructure  is not discussed in this article. Grain features can be learned from historical experiments on Ti-6AL-4V parts. The proposed methodology provides a parametric representation of fatigue life where each fatigue curve is characterized using a set of parametric coefficients. Two regression models, stepwise linear regression (SLR) and Gaussian process regression (GPR), are trained to establish the defect-fatigue relationship by predicting the parametric fatigue coefficients from defect and grain features.
The remainder of the article is organized as follows. Section 2 reviews the background information alongside the overview of the methodology. Sections 3 and 4 provide Phase (I) characterization of establishing the thermal-defect relationship and Phase (II) methodology for determining the defect-fatigue relationship, respectively. Section 5 includes the uncertainty analysis. Section 6 discusses the modeling results and validations, and Section 7 states the concluding remarks. Fatigue life is determined through a process where a material undergoes cycle-by-cycle damage of fluctuating stresses or strains. In fatigue experiments, the applied stress is not large enough to break the specimen instantly (immediate failure); instead, the material stays under a certain number of load fluctuations until a failure, that is, when the accumulated damage reaches a certain threshold.
In fatigue life prediction, the main idea is to predict the number of cycles for specific stress or strain range. In this article, the proposed fatigue model is aimed at predicting the number of cycles for a particular strain range. In the following paragraphs, microstructure-based fatigue modeling (MSF) is discussed thoroughly since the proposed model is developed based on MSF modeling.
In MSF, fatigue life is decomposed into three successive stages based on the microstructural properties of crack growth, as follows [52] where N TOTAL stands for the total number of cycles to failure, N INC stands for the number of cycles to fatigue crack incubation, N MSC stands for the number of cycles for microstructurally small crack propagation, and N LC stands for the number of cycles for long crack propagation. The basis for the MSF was initially founded upon multiscale finite element simulations, and only microstructural data was applied toward the end of the development. The incubation stage of the MSF model is based upon a modified Coffin-Manson law. The MSC stage of the MSF model is based upon crack tip displacement (CTD). The long crack (LC) stage is based upon a modified Paris law. The governing equations in the multistage fatigue (MSF) model are listed here. [52] Generalized incubation model Microstructurally small crack growth and long crack growth The total fatigue life (Equation (1)) includes three parts, namely incubation, microstructurally small crack (MSC) growth, and long crack (LC) growth. Equations (2) and (3) describe the ratio of the plastic zone to the particle size or defect (l=D). The ratio of the plastic zone to the particle size is described as a function of strain amplitude using ε th , ε per , η lim , and r parameters. The incubation life is modeled in Equation (4) based on a modified Coffin-Manson law using the cyclic plastic deformation (microplasticity). As in the Coffin-Manson law, C INC and a are the coefficient and exponent fit to experimental data, respectively. The β term is a nonlocal maximum plastic shear strain amplitude located around the particle or defect. To model multiaxial fatigue, a variation of the nonlocal maximum plastic shear strain amplitude expression is presented in Equation (5) where σ y is the yield stress and k are determined from experiments. The parameters τ 0 f , γ 0 f , b 0 , c 0 are torsion fatigue parameters for both strain-life and stress-life, and μ is the shear modulus. Equation (6) describes the microstructural small crack growth regime, where ΔCTD is the crack tip displacement range, χ is a material constant, and ΔCTD th is the threshold for crack tip displacement. Equation (7) defines the crack tip displacement as a function of remote loading, where C I , C II , ξ, ϖ, ξ 0 , ϖ 0 , ζ constants are determined through small crack growth experiments. The GS parameter captures the effect of grain size on the microstructure small crack behavior. Long crack growth behavior is presented in Equation (8). The effective stress intensity factor is defined as ΔK eff ¼ K max À K op , where K max is the maximum stress intensity factor and K op is the opening stress intensity factor. Finally, the faster of the small crack and long crack calculated growth rates is employed for constant amplitude fatigue life prediction (9). According to Bu and Stephens, physically-short cracks have a size less than 1.0 mm, where physically long cracks are in the range of several millimeters. [53] Therefore, it should be noted that for the purpose of this study, according to the fact that experiments are conducted on small laboratory specimens, the N LC is not considered in calculating the final fatigue life, and predicting fatigue life at the very high cycle (VHCF) is not within this work's scope.
As it is evident, MSF uses a thorough and complex combination of mathematical and mechanical calculations to predict the fatigue life of a particular part. Additionally, the number of parameters involved in these calculations is significantly large (over 40), which includes mechanical properties, microstructure parameters, material incubation, small crack, and long crack constants. Although these parameters cover a majority of fatigue modeling aspects, it complicates the simple fatigue predicting task where it is required to input all the parameters correctly to acquire a better estimation. This study provides a specificpurpose fatigue prediction of Ti-6AL-4V material from handpicked parameters of the MSF model, such as defect size, nearest neighbor distance, grain size, and grain misorientation. Moreover, the proposed methodology is nondestructive and directly predicts the total fatigue life N total for a given range of strain amplitude. It is essential to mention that a data-driven surrogate modeling procedure enables the reverse optimization of the process to derive an understanding of the thermal history and design inputs based on fatigue life.

Proposed Framework Overview
The objective of this article is to present a two-phase modeling approach to characterize the thermal-defect-fatigue relationship for nondestructive fatigue life prediction. In Phase (I), the thermal-defect relationship is established for identifying the location of defect occurrence and size characterization. The major challenge in Phase (I) is the high dimensionality and complex spatial-temporal correlation in the thermal image stream. To account for the dimensionality issue, prior to the CNN, a spherical transformation technique is used to reduce the noise and dimensionality of the thermal images. Predicting the relative size of the defect enables us to characterize the defect features to get on step closer in estimating the fatigue life, while the existing methods only predict the defect occurrence without further characterizing its size.
In Phase (II), the defect-fatigue relationship is described by leveraging the MSF model and functional regression analysis. The main challenge concerned with this phase is the lack of experimental samples. Therefore, the MSF model is leveraged to estimate a range of fatigue lives for Ti-6AL-4V. To directly predict the fatigue curve, a parametric representation of the curve is introduced using a polynomial function. The polynomial coefficients can be estimated using the defect and grain features (as shown in Figure 2). Two methods, namely stepwise linear regression (SLR) and Gaussian process regression (GPR) are used to correlate the defect and grain features to the fatigue polynomial coefficients. The fatigue curve can be reconstructed using the predicted coefficients and the known strain amplitude range. Further details of the methodologies are discussed in the following sections.

Phase (I) Modeling: Thermal-Defect Relationship
Phase (I) builds a model to predict the relative size of defects from in situ thermal history. The size of a defect is a continuous variable that naturally brings out the idea of a regression model. Unfortunately, the resolution of X-ray data is limited, and therefore, regression learning is a poor choice in this case. For the issue of fatigue life, the size of the defects plays a critical role in reducing the number of cycles to failure, and therefore, detecting large defects is more critical than the smaller ones. Thus, defects are divided into two groups based on a predefined threshold, creating a multiclass classification problem. This procedure is divided into three steps: 1) data transformation, 2) CNN design, and 3) defect prediction (model I), which are discussed in detail within the following sections.

Data Transformation
One way to represent in situ thermal history is by using Spatiotemporal melt pool images captured by a high-resolution thermal www.advancedsciencenews.com www.advintellsyst.com camera during the build process. The melt pool images have different shapes and sizes. One practical idea in literature is to transform the images into the spherical domain where all images have identical domains. Moreover, this transformation helps to decrease the size of images slightly. Each pixel in the melt pool has three attributes, x-location (x index ), y-location (y index ), and temperatures (z temp ). The spherical transformation is transforming cartesian coordinate (x index , y index , z temp ) to spherical coordinate (r, θ, φ) where r is the radial distance, θ is the polar angle, and φ is the azimuthal angle, using the following set of formulations Assuming that each image has a size I 1 Â I 2 , the x-axis and y-axis of each image is in the range of ½0, I 1 and ½0, I 2 , respectively. The method used by Seifi et al. does not centralize the ranges before the transformation. [54] Using their methodology causes the melt pool images to lose valuable information at the peak of melt pools. The reason behind this fact is the different thermal distribution of cylinder thermal images compared to the thin wall. One example of the thermal image transformed by both methods is illustrated in Figure 3, where there is an evident data loss at the peak of the melt pool in Figure 3a. In the current study, the range of each melt-pool is changed to ½À I 1 2 , I 1 2 and ½À I 2 2 , I 2 2 for the x-axis and y-axis, respectively. After data transformation, a nonparametric surface interpolation is used to create continuous data out of discrete data points. The interpolation function helps to extract a smaller grid of information and smooth the data to remove the thermal noise, especially at lower temperatures.

Convolutional Neural Network Design
Melt pool images are a set of high dimensional captures carrying the essential information about the AM process. This information is not visible to the naked eye or the simple analytical tools. Therefore, a higher-order technique is required to extract the quality relevant information from melt pool images and deal with the curse of dimensionality. The primary information that a melt pool image carries is the thermal information in the form of temperatures. Thus, melt pool images are considered single-channel images, meaning that they are similar to grayscale images. A grayscale image contains the magnitude of each datapoint with a value between 0 (black) and 1 (white). For CNN, melt pool images are converted into grayscale images using max-min global standardization.
A convolutional neural network (CNN) is a type of deep neural network with the frequent use of analyzing visual imagery that is mainly used for image and video recognition, image  [54] www.advancedsciencenews.com www.advintellsyst.com classification, etc. CNN uses convolution, a specialized kind of linear operation, instead of the general matrix multiplication in at least one of the layers of the network. [55] The CNN consists of several layers to learn the output from input images and prevent overfitting. The main layers are the input layer, output layer, and hidden layers. The hidden layers consist of the convolution layer, pooling layer, activation function layer, normalization layer, fully connected layer, softmax layer, and a classification layer. These concepts are briefly explained in the following sections.

Convolutional Layer
A convolutional layer abstracts the input image tensor to a feature map by applying convolutional filters to the input images. The output shape depends on several convolutional layer parameters, such as filter size, stride (step size), and padding. [56] The output width and height of the convolutional layer is The output of the convolutional layer is passed to the next layer.

Pooling Layer
A pooling layer aims at reducing the spatial size of the feature map and removing unnecessary spatial information by the down-sampling operation. Ordinary pooling operation is class max-pooling that returns the maximum of the rectangular region of input based on pooling size. [57] In the case of overlapping regions, the output of the pooling layer is

Activation Function Layer
The activation function is a linear or nonlinear function that maps the set of inputs to a node to an output. The most common activation function is called rectified linear unit (ReLU) layer that performs a threshold operation, as follows [58] f ðxÞ ¼ This threshold operation does not change the size of the input.

Batch Normalization Layer
A batch normalization layer is used to normalize the input based on the mean and variance of a mini-batch. [59] The normalized values arê where x i is the input variable, μ B and σ 2 B are the mean and the variance of the mini-batch, and ε is the property epsilon that improves the numerical stability in case that the variance is near zero.

Fully Connected Layer
A fully connected layer connects all of its neurons to the neurons in the preceding layer, similar to a fully connected neural network.

Softmax Layer
A softmax layer applies a softmax function to the input to get a number between 0 and 1, wherein the case of classification can be interpreted as the probability of class 1 for the corresponding input. The standard (unit) softmax function is where K is the number classes, 0 ≤ S r ðxÞ ≤ 1 and P k j¼1 S j ¼ 1. The softmax function is usually considered as the multiclass generalization of the sigmoid function. [60]

Classification Layer
A classification layer calculates the cross-entropy loss function for a multiclass classification problem where classes are mutually exclusive. The classification layer uses the output of the softmax layer to assign each input to one of the K mutually exclusive categories. The cross-entropy function is [60] loss ¼ À (18) where N is the total number of samples, K is the total number of classes, t ij is an indicator variable that shows the true value if ith sample belongs to the jth class, and S ij is the output of the softmax layer for sample i and class j.

Model (I): Defect Prediction
There are many ways to design a CNN, and it is the designer's choice on how to design the layers and assign the parameters. In this article, a simple design of CNN is used that provides reasonable accuracy. The objective of CNN is to learn the defective behavior from processed melt pool images. In this section, at first, the design of the CNN used in this article is provided, and afterward, the possible training options are mentioned. By using the different layers defined in the previous section, the design mentioned in Table 1 is used in this article. The details of the CNN design are mentioned in the modeling results section.
The next step is training the network based on the available samples. Before the training, several options are required to be defined, such as the solver for training the network, initial www.advancedsciencenews.com www.advintellsyst.com learning rate, the maximum number of epochs, and validation procedure. The final output of the CNN is the label for each image, whether the image is healthy, small anomaly, or large anomaly. This classification provides the information required to build the fatigue prediction model.

Phase (II) Modeling: Defect-Fatigue Relationship
Phase I helped to identify the defect distribution within an AM part using thermal history. In Phase II, the objective is to extract the defect features from predicted defect distribution and use it to predict the fatigue life of the part. This phase is divided into four steps: 1) extracting defect characteristics, 2) fatigue curve representation, and 3) fatigue life prediction (models II and III). These topics are elaborated upon in the following sections.

Extracting Defect Related Features
Defects are voids within a manufactured part that seriously affect its mechanical behavior. To present defects' intensity and spatial distribution, two features, defect size, and nearest neighbor distance, are utilized. [29] These features are further used to predict fatigue life.

Defect Size
The defect size is the first feature demonstrating the size of the largest defect in the material at the scale of a micrometer. Several works have shown that the main life-limiting failure mechanism of AM parts is due to cracks initiating from the defects located near the surface. [13] , [61][62][63][64][65] Therefore, the most dominating defect (the largest/closest to the surface) determines the HCF life, where a single defect leads to a crack that propagates until failure. Thus, the information of the largest defect (mostly lack-of-fusion defects) is crucial in characterizing the long-life regimes (i.e., HCF), and the effect of smaller effects can be negligible. This variable is mentioned by the name p size throughout the article. It is highly probable for the crack to initiate at the location of the largest defect; that is the reason one of the features is allocated to the size of the largest defect.
Calculation: The output of the defect prediction model (model (I)) is a discrete value, such as small or large defects. Therefore, the exact magnitude of the largest defect is unavailable. The approximate approach to determine the p size is to choose the average size of the large defects seen in the current samples as the maximum size of the defect in the material (400 μm). In case that there is no large defect, the mean value of the size of the small defects is sufficient (100 μm). This approach is identical to choosing an upper bound for p size which leads to a lower bound for fatigue life. Future research can focus on how to accurately predict the defect size instead of small or large classification.

Nearest Neighbor Distance
For short-life regimes (LCF), large defects may not dominate the fatigue life of the parts. Instead, a combination of defect features (i.e., size, shape, location, and distribution of all types of defects), and microstructural characteristics may compete for initiating cracks. Nearest neighbor distance (NND) is a feature to characterize the spatial distribution of the defects. This feature is mentioned by the name p nnd throughout the article. The more defects are located near each other, the less value of the NND, leading to poor fatigue life. Additionally, the distance of a defect from the surface plays a critical role in measuring NND. A near-surface defect is more likely to reduce fatigue life; hence, decreasing the value of NND. The approximate approach to calculate NND involves defining a neighboring criterion. In this article, the defects, within the 4D i radius of defect i, are assumed to be neighbors of defect i. Having neighboring criteria in hand, neighbors of each defect can be determined (NN i ). By measuring the distance between each defect and its neighbors (DD ij ∶j ∈ NN i ), the average neighboring distance between defects DD i can be calculated. Similarly, the distance to the nearest surface of each defect (DS i ) is calculated as well. The minimum value for both min i DS i and min jNN i j≥5 DD i is calculated, and their minimum is minðdd min , ds min Þ is set to be the approximate value of p nnd . Calculation: As mentioned for p size , the output of the model I defines each defect as small or large. Other than defect features, two microstructural characteristics are used in this article to account for the grain effect of fatigue life prediction. Grain size (g size ) and grain misorientation (g mo ) are considered as design parameters and can be determined from historical experiments. Having these four features in hand, the fatigue life prediction model can be developed.

Fatigue Curve Representation
Predicting a curve from multiple parameters is a challenging task. One effective approach is to represent a curve with a parametric function. Predicting coefficients of a parametric function is a much more effective approach where the number of parameters is much smaller than the number of data points within the fatigue curve. Additionally, the number of cycles (N total Þ for different strains in a fatigue curve are scattered between 1 and 10 8 . Fitting a parametric function to this data is extremely inefficient and has a large fitting error. To solve this issue, the natural logarithm of the N total is used, which makes it suitable for parametric modeling.
In this article, a family of polynomial functions is used to fit the fatigue curves as follows where ε a is the given strain range, ϵ is the error term, a q is the coefficient of the polynomial term ε q a , Q is the total degree of polynomial function, and f ðε a Þ is the function of predicting fatigue life, which is equal to the logarithm of N Total . The degree of the polynomial vastly affects the output of parametric modeling. Choosing a small Q creates a function lacking the power to represent the fatigue curve truly, and for a large Q, the function becomes unnecessarily complex and may overfit the curve. Thus, the best value for Q is the minimum value that satisfies a high adjusted R-squared and aptly correlated the functional fatigue parameters to the input features. After representing the fatigue curve with functional parameters, the next step is to build a model to predict them from defect features and grain characteristics.

Fatigue Life Prediction Models
In this section, two models are represented to predict the fatigue curve parameters, introduced in the previous section, from defect features and grain characteristics. The input to both models is from feature space, and the output is from the polynomial coefficient space (fatigue curve parameters). The first model is a simple stepwise linear regression (SLR) model where efficient (the criterion is discussed later in the article) features from feature space are added to the model in a step-by-step manner. The second model is a probabilistic model named Gaussian process regression (GPR), where all the features from feature space are considered for training the model. First, the feature space, both input, and output, is defined, and later, both models are discussed in detail.
The main variables in feature space are defect size (p size ), nearest neighbor distance (p nnd ), grain size (g size ), and grain misorientation (g mo ). The input feature space (X ) contains the solo, interactive, and quadratic effect of features, as follows: 1) Main effect: p size , p nnd , g size , g nnd ; 2) Interactive effect: p size Â p nnd , p size Â g size , : : : , g size Â g nnd ; 3) Quadratic effect: p 2 size , p 2 nnd , g 2 size , g 2 nnd . The output space contains the polynomial coefficients fa 1 , : : : , a Q g extracted from Section 4.2. The general formulation of this methodology is represented as the following equations a q ¼ G q ðX Þ þ ε, ∀q ∈ 0, 1, : : : , Q where a q is the qth polynomial coefficient, X is the variables from feature space (main, interactive, and quadratic effect), ε is the error, and G q is the function that models the relationship between the feature space and the qth order polynomial coefficient. Overall, it is required Q functions to finalize the model. G q can be either SLR or GPR. The following two sections will discuss the details of these two models to predict the polynomial coefficients from input space features.

Model (II.1): Stepwise Linear Regression (SLR)
Stepwise linear regression (SLR) is a model that automatically adds or removes the explanatory variables based on a predefined criterion. [66] The criterion can be chosen from multiple choices such as a change in the sum of squared error, an increase in the value of R-squared, or an increase in the value of adjusted R-squared. The criterion chosen in this article is adjusted R-squared with a determined threshold. If the increase in adjusted R-squared is greater than the threshold (1%), the candidate variable is added to the model. The step-by-step algorithm is shown in Figure 4. For each polynomial coefficient fa 1 , : : : , a Q g, one SLR is trained based on steps defined in Figure 4. Final regression models are examined in the experimental results section.

Modell (II.2): Gaussian Process Regression (GPR)
Linear regression used in the previous section offers models with virtues like the simplicity of implementation and interpretation. The main drawback is that it has limited flexibility where the relationship between input and output variables cannot be approximated as a linear function. Gaussian process regression (GPR) is a powerful nonparametric probabilistic model that offers flexibility to model nonlinear relationships. Assume that the training set contains the data in the form of fðx i , a ij Þ; i ¼ 1, 2, : : : , ng where x i ∈ ℛ d is from feature space and a ij ∈ ℛ∶j ¼ 1, : : : , Q is one of the polynomial coefficients. A GPR explains the response using latent variables lðx i Þ; i ¼ 1, : : : , n, from a Gaussian process (GP), and basis functions h. The covariance function between the latent variables captures the smoothness of the response variable, and the basis function h projects the input variables x i into a p-dimensional space. A Gaussian process (GP) is a set of random variables where every finite set of them have joint Gaussian distribution. If lðxÞ is a GP, for finite observation set fx 1 , x 2 , : : : , x n ; x i ∈ ℛ d g, www.advancedsciencenews.com www.advintellsyst.com random variables lðx 1 Þ, lðx 2 Þ, : : : , lðx n Þ have joint Gaussian distribution. Every GP is determined by its mean mðxÞ and covariance function kðx, x 0 Þ. Therefore, if lðxÞ is a GP, then E½lðxÞ ¼ mðxÞ and Cov½lðxÞ, lðx 0 Þ ¼ E½ðlðxÞ À mðxÞÞðlðx 0 ÞÀ mðx 0 ÞÞ ¼ kðx, x 0 Þ. Now the GPR model can be represented by hðxÞ T β þ lðxÞ where lðxÞ are from a GP with mean zero and covariance kðx, x 0 Þ, hðxÞ are the set of basis functions that transform every observation x from original space ℛ d into a new feature vector space ℛ p and β is a p Â 1 vector of coefficients for basis function. This model represents a GPR, and one instance of response can be modeled as which demonstrates the probabilistic nature of GPR. More details of GPR can be found in a book written by Williams and Rasmussen. [67] All the hyperparameters of the GPR model are optimized to minimize the cross-validation loss. These hyperparameters are basis function, Kernel function, sigma (noise standard deviation), and standardization. A limitedmemory Broyden-Fletcher-Goldfarb-Shanno (L-BFGS) algorithm is used to optimize the aforementioned model parameters.
More details for L-BFGS can be found in Chapter 7 of the numerical optimization book. [68] 5. Uncertainty Analysis Assume that the number of actual large and small defects are shown with N L and N S , respectively. P L i is the probability that large defect i is classified as large. Similarly, P S i is the probability that small defect i is classified as small. If there exists a large defect inside the part, the presence of small defects does not affect the p size . The probability that neither of the large defects is classified as large is Q N L i ð1 À P L i Þ. For the proposed defect prediction model, if this probability is near zero, it means that the model is classifying at least one large defect correctly. In case that there is no large defect within the material, a similar approach can be applied with small defect probabilities P S i . Moreover, at the presence of the condition Q N L i ð1 À P L i Þ % 0, the calculation of p size is robust to misclassifications. The values assigned for p size along with their probabilities are represented in Table 2.
A similar approach is applied for p nnd using the same notation. Two tables are required to calculate p nnd , the first one is for distance to the surface of the defects, and the second one is for the clustered average distance between neighboring defects. The probabilities are calculated separately for both tables. Assume that P i ¼ P L i þ P S i is the probability of predicting defect i. Comparing the value of P L i with P S i defines the diameter of the defect for neighboring criterion (P L i > P S i ! D i ¼ 400). Table 3 is created for distance to the surface of every predicted defect where defects are sorted based on their DS i in an ascending order.
If defect i is not detected from thermal analysis, the algorithm moves to the second defect with the smallest distance to the surface. In this case, the p nnd increases by DS iþ1 À DS i , without considering the table for the average neighboring distance between defects. Table 4 is built for the average neighboring distance between defects where they are sorted by DD i in an ascending order. It is important to mention that only defects with at least   www.advancedsciencenews.com www.advintellsyst.com five neighbors are considered. Set NN k,l i is the lth combination set of choosing k neighbors from neighbor set NN i .
According to Table 4, defect i is detected with a probability P i Q iÀ1 j¼1 ð1 À P j Þ. To be considered as an effective defect in the algorithm of NND, at least 5 of its neighbors should be identified with probability Thus, the overall probability of assigning the value DD i to In case that defect i goes undetected, the next smallest value in the table is chosen. It is important to mention that these two tables for DS i and DD i should be considered together while choosing a value for p nnd . Assume that the change in p size and p nnd are shown with Δ size and Δ nnd . Using the G q ðX Þ function, the new polynomial fatigue coefficients can be derived using the following equation whereX contains the new defect features asp size ¼ p size þ Δ size andp nnd ¼ p nnd þ Δ nnd . Incorporating the new fatigue polynomial coefficients into the fatigue function, the new fatigue curve is generated. The difference between these two curves for a giving strain value is the change in fatigue life due to misclassification.
Further in the article, the effect of change in p size and p nnd due to misclassification is numerically examined.

Modeling Results and Validation
The objective of this section is to examine the performance of the proposed multiphase methodology to see if the defect prediction based on thermal history can be used to predict fatigue life accurately. The validation procedure is demonstrated in Figure 5, where rounded blue rectangles are the proposed methodology, and the red boxes are the traditional procedure. According to Figure 5, two separate validation procedures are used to examine the performance of the proposed method. Following are the description of the datasets used for the validation

Thermal Images
Thermal images are collected from a direct laser deposition (DLD) process where an OPTOMEC LENS 750 machine (with a 1 kW Nd: YAG laser) was used to fabricate Ti-6Al-4V cylinders, and the machine is equipped with a thermal imaging camera, one co-axial pyrometer camera to capture melt pools (Stratonics, Inc.), and one infrared camera to capture the global heat flow (Sierra-Olympic Technologies, Inc. Viento320). Seifi et al. have provided the LENS machine and camera setup details. [54] This dataset is a collection of thermal images captured by a co-axial pyrometer camera, carrying the main information about the build process. The laser has used a unidirectional scan pattern with 15 tracks per layer and a finishing contour path. There is a hatch rotation of 180 between consecutive layers. The experimental setup of the laser machine is as follows: 1) Laser power: 300 (W). 2) Laser travel speed: 40 (ipm). 3) Powder flow rate: 3 (rpm). 4) Layer thickness: 0.015 (in). 5) Hatch distance: 0.02 (in).
The captured thermal images are projected into a 752 Â 480 size frame and further cropped to a smaller frame with more concentration on the peak of melt pools. The cylinder is finished by depositing 69 layers on top of each other, where the thermal camera collected 2827 none-noise images of the fabrication process.   www.advancedsciencenews.com www.advintellsyst.com

Defect Data
The fabricated cylinder is scanned using a Nikon X-ray CT XT H225 (https://www.nikonmetrology.com/en-us/x-ray-ct/x-ray-ctsystems-225kv-and-320kv-ct-inspection-and-metrology) machine with a maximum voltage of 225 kV. This machine is suitable for inspecting small to medium parts. The voltage and the current of the X-ray machine are set to 180 kV and 98 μA to scan the cylinder. The scanned X-ray file is processed using the myVGL application, and defect information is extracted in an excel sheet. The primary columns are: 1) Probability: a software assigned value to emphasize the accuracy of a detected void to be pore. The probability column is used to remove the less significant defects. The diameter represents the size of the defects, and center x-y-z data demonstrates the location of the defects and helps to match them with thermal images. The average and standard deviation of all the defects are 150 and 120 μm. The histogram of the size of defects is shown in Figure 6. This figure contains all the defects, regardless of their probabilities. In this article, only high probability defects are considered in the modeling. While most defects are small and large defects rarely happen during the build, the threshold is set to 200 μm to distinguish between frequent small defects and rare large defects.

Fatigue Experiments
Fatigue experiments are highly costly and time-consuming; therefore, the MSF model is calibrated based on experimental data on Ti-6Al-4V. [61,69] Afterward, the MSF model is used to estimate a range of possible fatigue lives based on different combinations of defect size, NND, grain size, and grain misorientation. These experiments are used to validate the defect-fatigue model (Phase (II)).

Dataset Description
Thermal history carries the primary information about the process signature, and here, it is used to predict the relative size of defects. Thermal history is a set of high-resolution images of melt pools captured by a thermal camera, called the pyrometer camera. These melt pool images have different shapes and sizes, which makes the comparison of melt pool images extra hard and less meaningful. To amend this issue, melt pool images are transformed from a cartesian to a spherical coordinate system. The pyrometer camera captures the images in a frame with size 752 Â 480, which can be cropped to smaller images with more concentration on the peak of melt pools. The current cylinder dataset contains 69 layers of 2827 melt pool images, each with a size of 80 Â 80 pixels. The location of melt pools is known with three parameters, x-location, y-location, and layer number. Out of the 2827 melt pools, 2634 are flagged normal, 110 are flagged with a small defect, and 83 are flagged with a large defect. The threshold to divide defects into small and large is 200 μm, comparing the major axis length. All the transformed melt pools are converted into single-channel (grayscale) images.

Model (I): Defect Prediction by CNN
The specific architecture of CNN is mentioned in Table 5.
The model is using a solver named stochastic gradient descent with momentum (SGDM) to train the network. The initial learning rate is 0.01, the maximum number of epochs is ten, and the validation frequency is every ten iterations. To prevent CNN from overfitting, 20% of the training dataset is used to validate the network based on predefined validation frequency. Due to the unbalanced number of samples in classes, a stratified data split approach is incorporated to preserve the percentage of samples for each category in sub-sets. Sixfold cross-validation is used to examine the performance of the proposed model by using metrics such as precision, recall, and f-score that are defined as follows True or false terms illustrated the correct and incorrect prediction of the classification algorithm, respectively; additionally, positive and negative terms represent the objective class of the algorithm at the time (healthy, small defect, or large defect) and the opposing classes, respectively. The averaged confusion matrix alongside averaged accuracy metrics is reported in Table 6.
Regarding the performance of the CNN in the test set, out of 13 large defects in the fabricated part, 2 (15%) are misclassified as healthy (nonexistent), and 2 (15%) are misclassified as a small defect. Additionally, out of 18 small defects in the test set, 2 (11%) are misclassified as large, and 3 (16%) are misclassified as healthy (nonexistent). These numbers are trivial for healthy samples, but the sample size is larger as well. The averaged f-score for healthy, small defect, and large defect classes are 99%, 74%, and 72%, respectively. Overall, considering the small number of defect samples, the accuracy of the model is acceptable. The uncertainty analysis is used to account for the fact that the model can misclassify the images, especially the large ones.
To further investigate the accuracy of the model (I), the receiver operating characteristic (ROC) curve of the classification model is plotted. The ROC curve is a graphical measure that demonstrates the diagnostic ability of a binary classifier by varying the discrimination threshold. The ROC curve can be created by drawing the true positive rate (TPR) against the false positive rate (FPR) for different threshold values. TPR is known as the recall factor, introduced earlier, and FPR is known as the probability of false alarm (type I error). [70] The ROC graph depicts the tradeoffs between TPR (benefits) and FPR (costs). The best possible classification model would result in a ROC curve with a point in the upper left corner (100 % recall).
Additionally, a random guess classifier would yield a diagonal line from the bottom left to the top right corner, which is called the "no-discrimination" or "chance" line. The ROC curve of a fair classification model should lie between the random and the best possible curves. Moreover, the area under the curve (AUC) is a measure that calculates the area under the ROC curve. AUC demonstrates the probability of ranking a randomly selected positive instance higher than a randomly selected negative example by the classifier. Greater AUC means better average performance. [70] The ROC response of sixfold cross-validation is shown in Figure 7. The AUC of different folds for all of the classes is near one.

Phase (II) Validation: Fatigue Life Prediction
In this section, the fatigue prediction models defined earlier are implemented using generated experiments. Prediction accuracy, as well as model performances, are analyzed. In the following section, experiment generation is discussed, and both proposed models are implemented along with their results.

Experiment Generation
Even though fatigue experiments are extremely expensive and time-consuming, it is highly unlikely to fabricate specimens with similar defects and microstructure properties regarding the AM process's uncertain nature. Additionally, there are not enough fatigue specimens in hand from an additive manufacturing process to train a reliable prediction model. This uncertain nature of AM processes and lack of sufficient samples are the main reasons behind incorporating the MSF model as the backbone of this study. The MSF model is calibrated based on experimental data on Ti-6Al-4V, and different combinations of defect size, NND, grain size, and grain misorientation are used to estimate a range for the fatigue life of Ti-6Al-4V material. For this purpose, the acceptable range of these features is defined based on the defect statistics determined through fractography and microstructural examination of the fatigue specimens. These ranges are shown in Table 7. Figure 8 illustrates the maximum change in fatigue life for each feature while keeping the other three features at their mean value. After defining the features range, a full factorial design of   [29] The remaining parameters of the MSF model are kept at a constant level, specifically for the Ti-6AL-4V parts.
In addition to the training samples aforementioned, a testing sample set is required to examine the performance of the proposed methodologies. While a full factorial design is used to generate training samples, a near-random method, named Latin hypercube sampling (LHS), is used to generate the testing samples. The LHS was described by researchers from Los Alamos National Laboratory in 1979. [71] LHS uses the idea behind the Latin square and generalizes it to an arbitrary number of dimensions. A Latin square is a n Â n matrix containing n distinct values where each value occurs exactly once in each row and each column.
Assume there exist N variables, and the range of each is divided into M equally probable intervals. Then, M samples points while satisfying the Latin hypercube requirement, where each specific row and column only contains one sample. This requirement forces each variable to have the same number of divisions, M. One of the main advantages of the LHS is that

Fatigue Models
Each fatigue experiment is represented with a polynomial function. The degree of the polynomial function is a hyper-parameter and should be chosen based on the training set. Therefore, a 10-fold cross-validation method is incorporated using the training set to select the degree of the polynomial for fatigue curve modeling comparing three measures. The first measure is the root mean square error (RMSE) for predicting the polynomial coefficients.
The second measure is called the error percentage. Error percentage is derived by dividing the magnitude of error in fatigue life prediction to the maximum range of fatigue life within the training dataset. Maximum fatigue life is achieved when p size , g size and g mo are at their minimum, and p nnd is at its maximum. Minimum fatigue life is achieved oppositely. Each fatigue curve is created from 101 data points where for each strain value, fatigue life (number of cycles) is recorded. At each data point, the error percentage is calculated, and afterward, the average error percentage for each sample is recorded.
The third measure is called the dissimilarity score. The dissimilarity score calculates the resemblance of the predicted fatigue curve to the actual one, and it is defined as the standard deviation of error percentage at 101 data points of a fatigue curve. Both error percentages and dissimilarity scores should be near zero. According to Figure 9, Q ¼ 4 is the optimum degree of the polynomial function to represent each fatigue curve. Subsequently, the accuracy of the model for predicting the polynomial coefficient and fatigue curves is examined using the testing set. First, the root-mean-square error (RMSE) of both regression models (SLR and GPR) for predicting the polynomial coefficients are represented. Second, the predicted coefficients are used to construct the fatigue curves based on the known strain range.
Before using SLR or GPR, both input features and output coefficients are normalized to compare performance measures. The RMSE resulting from predicting output coefficients is illustrated in Table 8, where SLR performs slightly better.
After predicting the coefficients, the fatigue curves are constructed and compared with the original curves. One sample output, along with the boxplot of the error-percentage for all 40 testing samples, is shown in Figure 10. According to Figure 9, 2 out of 40 for SLR and 3 out of 40 samples for GPR have an error percentage near 20%, and the remaining samples have less than 10% error in prediction. Overall, GPR has a slightly better performance compared to the SLR due is low error variance.

Numerical Analysis of the Uncertainty
In this section, a numerical analysis of the uncertainty in defect prediction and its effect on fatigue life is provided. For this purpose, the dataset is randomly divided into train, validation, and test sets, and the prediction probabilities of the samples in the test set are calculated. The test set has N S ¼ 16 small defects and N L ¼ 13 large defects. The probability table for p size is shown  Figure 9. Illustration of the results of three measures with different degrees of the polynomial function. in Table 9. The probability of classifying at least one defect is 1, meaning that the model (I) can classify at least one large defect and therefore, the estimatedp size is 400 μm. Additionally, during the condition Q N L i ð1 À P L i Þ % 0 holds, the model is robust to misclassification of the small defects.
The probability table for p nnd based defects' distance to the surface is provided in Table 10. The probability table for defects' distance to distance is not delivered due to the scatteredness of the defects. Therefore, DD i is insignificant compared to DS i , and Table 10 can be used for an approximation of p nnd . According to Table 10, while the probability of the first defect is less than 0.5, the second defect is used for NND calculation. Thus, the estimatedp nnd is equal to 26 μm.
The actual values of p size and p nnd are 477 and 21 μm, respectively, for the test set. The changes in these two variables are Δ size ¼p size À p size ¼ 400À477 ¼ À77μm and Δ nnd ¼p nnd À p nnd ¼ 26À21 ¼ 5 μm. The fatigue life resulted from these two sets of defect parameters is calculated where grain features are kept constant at their mean values (g size ¼ 100, g mo ¼ 15). The results are illustrated in Figure 11, where the average fatigue life error is 298 and 276 cycles for GPR and SLR, respectively.

Conclusion
One of the main challenges that AM processes are facing is the uncertainty in the structural performance of AM parts. This issue has been a significant barrier in front of the broader adoption of AM by industries. This uncertainty can be seen in different modes of failure, especially fatigue failure, the dominant failure mode in various sectors. The amount and distribution of defects alongside microstructural features play a crucial role in characterizing the fatigue life of AM parts. Thus, it is necessary to develop a successful in situ qualification and certification methodology not only to identify the formation of defects but also to characterize the fatigue life based on detected defects.
In this article, a novel data-driven multiphase nondestructive in situ fatigue prediction methodology is developed to predict the fatigue resistance based on defect characteristics. The proposed method is developed in two phases. The objective of Phase (I) is The model fails to identify 1.82E-47 Figure 10. a) One sample output where the solid green curve is the actual fatigue life, SLR prediction is shown with the blue dotted curve, and the GPR prediction is shown with the red dash-dotted curve. b) The boxplot representation of error percentage for 40 testing samples.
www.advancedsciencenews.com www.advintellsyst.com to characterize the defect features based on in situ thermal history. Thermal history, that is, melt pool images, carries the most critical information about the build process; therefore, these images are processed and utilized to predict the relative defect size. The model (I) demonstrates this procedure and wraps it up by training a classification model that predicts the relative size of the defects (small or large). The objective of Phase (II) is to build the relationship between the defect properties extracted from Phase I and the fatigue behavior based on the MSF model. MSF model was used to identify the defect-related input features. Each fatigue curve is represented with a parametric polynomial function, and polynomial parameters are predicted from defect properties and microstructural features. Stepwise linear regression (SLR) and Gaussian process regression (GPR) approaches are used to predict the final fatigue parameters. The performance of Phase (I) methodology is validated using a cylinder fabricated by a direct laser deposition process, and Phase (II) model is validated based on MSF fatigue life simulations. Considering the small number of samples, the proposed methodologies have noticeable accuracies. Although MSF is a potent physics-based prediction model for fatigue lives with useful utilities, it depends on a handful of parameters to provide a prediction. Additionally, the AM parts are prone to defects, and identifying the defect-related parameters using the postmanufacturing methods is expensive and time-consuming; therefore, the proposed methodology significantly accelerates the qualification and certification of additive manufacturing processes.
Some exciting topics remain open for future research. First, the proposed methodology can be extended to add a new phase to account for the design-thermal relationship. Second, an inverse robust design exploration can be developed to generate satisfying design solutions. www.advancedsciencenews.com www.advintellsyst.com