Abstract
Additive manufacturing (AM) has revolutionized the way we design, prototype, and produce complex parts with unprecedented geometries. However, the lack of understanding of the functional properties of 3D-printed parts has hindered their adoption in critical applications where reliability and durability are paramount. This paper proposes a novel approach to the functional qualification of 3D-printed parts via physical and digital twins. Physical twins are parts that are printed under the same process conditions as the functional parts and undergo a wide range of (destructive) tests to determine their mechanical, thermal, and chemical properties. Digital twins are virtual replicas of the physical twins that are generated using finite element analysis (FEA) simulations based on the 3D shape of the part of interest. We propose a novel approach to transfer learning, specifically designed for the fusion of diverse, unstructured 3D shape data and process inputs from multiple sources. The proposed approach has demonstrated remarkable results in predicting the functional properties of 3D-printed lattice structures. From an engineering standpoint, this paper introduces a comprehensive and innovative methodology for the functional qualification of 3D-printed parts. By combining the strengths of physical and digital twins with transfer learning, our approach opens up possibilities for the widespread adoption of 3D printing in safety-critical applications. Methodologically, this work presents a significant advancement in transfer learning techniques, specifically addressing the challenges of multi-source (e.g., digital and physical twins) and multi-input (e.g., 3D shapes and process variables) transfer learning.
1 Introduction
Additive manufacturing (AM), commonly known as 3D printing, has become a popular manufacturing technique due to its ability to produce complex parts with unique geometries. The versatility of 3D printing has made it a popular choice in various industries, such as aerospace and healthcare. However, the lack of understanding of the functional properties of 3D-printed parts has hindered their adoption in critical applications where reliability and durability are essential. Destructive testing is a common method for assessing the functional properties of a part, but in many cases, it may not be possible or practical to perform destructive testing on the part to be used. Destructive testing may damage the part, rendering it unusable. In addition, destructive testing is often time-consuming and expensive, which may not be feasible for large-scale production. Furthermore, the material properties of 3D-printed parts can vary significantly depending on the printing process, material type, and post-processing techniques used.
Existing research on the functional qualification of 3D-printed parts has mainly focused on the characterization of material properties or the development of predictive models. However, these approaches are often limited in their ability to capture the complex interactions between the as-printed 3D shape, material properties, and the printing process variability.
We hypothesize that the only way to verify the functional properties of a part without destructive testing is through the use of physical and digital twins. Physical twins are printed under the same process conditions as the functional parts and undergo a wide range of tests to determine their mechanical, thermal, and chemical properties. Digital twins are virtual replicas of physical counterparts, created through finite element analysis (FEA) simulations. More formally, according to NIST [1], “A digital twin is the electronic representation—the digital representation—of a real-world entity, concept, or notion, either physical or perceived.” In our case, these replicas are generated from 3D scans of the object and enable precise assessment of the impact of shape inaccuracies on functional characteristics. Together, physical and digital twins enable accurate predictions of the functional properties of the parts, without the need for destructive testing on the part to be used. Figure 1 illustrates the concept behind the AUDIT (e.g., Functional Qualification in Additive Manufacturing via Physical and Digital Twins) framework, which combines physical and digital twins for functional qualification.
To accurately predict functional characteristics, this paper proposes a novel transfer learning technique for data fusion between heterogeneous process data and unstructured 3D shape data. By utilizing the knowledge from digital and physical twins, as well as the process conditions, this approach enables the accurate prediction of functional characteristics.
Overall, this paper presents a comprehensive approach to the functional qualification of 3D-printed parts. This approach has the potential to significantly improve the adaptation of 3D-printed parts in critical applications, where their functional properties must be accurately verified.
The remainder of the article is organized as follows. Section 2 provides a brief literature review. Then the proposed AUDIT framework for functional qualification in additive manufacturing via physical and digital twins is introduced in Sec. 3. Section 4 validates the proposed methodology by using a real-world case study of 3D-printed polylactic acid (PLA) lattice structures. Furthermore, the performance of the proposed method is compared with existing benchmark methods in terms of estimation accuracy. Finally, we conclude the article with a short discussion and an outline of future research topics in Sec. 5.
2 Literature Review
In recent years, 3D printing has emerged as a transformative manufacturing technology with a wide range of applications across industries. As the demand for functional qualification of 3D-printed parts in safety-critical applications continues to grow, researchers and practitioners have explored various approaches to address this critical aspect. This literature review examines the existing methods and their limitations, highlighting the need for further advancements in the field.
2.1 Destructive Testing for Functional Analysis in 3D Printing.
Destructive testing has long been a common practice for assessing the mechanical properties and performance of manufactured products. Mishra and Senthil [2] investigated the relationship between applied force and breaking strain of 3D-printed PLA parts using destructive testing with a universal testing machine (UTM). Zeng et al. [3] used destructive compression testing to study the behavior of honeycomb structures, revealing the correlation between material bonding and fracture location. Li et al. [4] quantitatively measured the post-yield crushing stress of honeycombs through destructive testing, enabling the evaluation of hierarchical honeycombs. Han et al. [5] conducted destructive testing on 3D-printed concrete walls to derive a calculation formula for predicting failure loads.
However, the applicability of the aforementioned methods to 3D-printed parts is limited due to the irreversible nature of the destructive testing process. Once a product is subjected to destructive testing, it becomes unusable, rendering it impractical for functional qualification. This limitation raises the need for alternative methods that can provide functional assessment without compromising the integrity of the part.
2.2 Finite Element Analysis for Functional Analysis in 3D Printing.
FEA has been widely utilized to simulate and predict the behavior of structures, including 3D-printed components. FEA offers valuable insights into the mechanical response of parts under different loading conditions. Cao et al. [6] compared experimental and simulation results for different lattice structures, investigating the effect of lattice shape parameters on stress using numerical simulations. Lesueur et al. [7] explored the effect of internal structure geometry on the yield of a structure using an FEA model, validated with experimental results.
However, FEA has its limitations, particularly when it comes to capturing the variability introduced by different process conditions in additive manufacturing [8]. Factors such as temperature, layer thickness, printing speed, and material properties can significantly influence the functional properties of 3D-printed parts. Several studies have investigated the accuracy of computer-aided design (CAD)-based compression simulations. Belhabib and Guessasma [9] found that filament-based computations closely matched the experimental deformation trends in the compression of hollow structures, but they overestimated the performance of hollow structures by an average of 43%. Abbot et al. [10] also observed significant discrepancies between simulation and physical compression test results. FEA models, with their inherent assumptions and simplifications, may not fully account for the variations of process conditions during 3D printing, limiting their accuracy in functional qualification.
2.3 Transfer Learning for Multi-Input, Multi-Source 3D Transfer Learning.
Transfer learning has emerged as a powerful technique in machine learning, enabling knowledge transfer from one domain to another. While transfer learning has shown promising results in various applications, its extension to the multi-input, multi-source setting in additive manufacturing is still limited. The lack of large-scale and diverse datasets, as well as the challenges associated with obtaining samples representative of the entire design space, pose obstacles to effectively applying transfer learning to functional qualification in 3D printing. Unlike structured data types, 3D point cloud data introduce complexities stemming from its inherent spatial and geometric attributes. The irregular nature of point clouds, their varying densities, and the incorporation of both structural and textural information necessitate specialized methodologies that can effectively capture these features. The focus on multi-source domain adaptation often revolves around structured tabular or image data [11–13]. These challenges demand a tailored approach that considers the unique characteristics of 3D data. Consequently, current transfer learning methods are not readily available to tackle the specific challenges posed by the functional qualification of 3D-printed parts.
2.4 Digital Twins and Transfer Learning in Additive Manufacturing.
In the realm of additive manufacturing, both digital twins and transfer learning have garnered attention. The literature spans diverse dimensions, ranging from initial conceptual visions endorsing the integration of digital twins in metal additive manufacturing for improved process models [14], to the exploration of optimal process conditions [15], and their subsequent adaptation to novel shapes using transfer learning [16]. However, it is important to note, that these methods do not predict shape variation or the functional attributes of individual products. Instead, they focus on batch shape optimization by suggesting optimal settings for specific shapes. Generative design has been explored to generate intricate geometries via numerical simulations [17], but it fails to account for quality and structural concerns arising from process variation during printing. Additionally, a range of work has tackled defect classification [18]. However, a classification of defects may not offer the nuanced understanding needed for certain situations, such as evaluating the significance of a keyhole pore and whether it necessitates product rejection. Several comprehensive review articles [19,20] underscore the need for functional qualification of 3D-printed parts that integrates both physical and digital twins, capturing the information from complex and diverse data sources.
The limitations discussed above highlight the challenges in functional qualification for 3D-printed parts. Destructive testing is not universally applicable due to the irreversible nature of the process. FEA may not capture process variability, and transfer learning methods lack applicability in the multi-input, multi-source setting.
To overcome these limitations and address the research gaps identified in the literature review, the AUDIT framework makes several contributions. In particular, AUDIT includes:
Integration of Physical and Digital Twins: introduces methodology for functional qualification by combining physical and digital twin concepts, bridging the gap between real-world physical processes and their virtual representations through a cohesive twin system.
Consideration of Process Variability: Proposed a novel methodology to model and account for inherent variations in physical processes, enhancing the accuracy of predictions for functional characteristics, and considering real-world manufacturing conditions and associated uncertainties.
Incorporation of 3D Shape Data: Provided comprehensive representation of physical objects by integrating 3D shape data into the modeling process, enhancing the understanding of complex relationships between 3D shape data, process settings, and functional properties in additive manufacturing
Multi-Source, Multi-Input Transfer Learning: Proposed a transfer learning framework that can effectively leverage information from diverse sources and inputs, including different data types, for unsupervised transfer learning.
In conclusion, the AUDIT framework provides a comprehensive solution that addresses the limitations of functional qualification in the current literature. Its contributions pave the way for further research and development, offering a pathway to overcome these challenges and establish robust approaches for functional qualification in the field of 3D printing.
3 AUDIT Methodology
This section presents the AUDIT framework as an approach to functional qualification enabled by multi-source, multi-input transfer learning via contrastive learning with augmentations. We consider a specific data scenario, where we assume 3D measurements of the 3D-printed part are available. However, in the context of products and additive manufacturing processes, where obtaining 3D measurements can be challenging, it is possible to replace the 3D point clouds by utilizing 2D imaging data from each print layer, which can be represented as a 3D tensor or stack of 2D images. We would like to emphasize recent techniques enabling the reconstruction of 3D point cloud data from 2D image stacks through complementary data fusion with process features [21]. This promising approach addresses the discretization issue of 2D image stacks and holds potential for application in our functional qualification work. Additionally, in the literature, treating 3D point cloud data as a stack of 2D images is widely adopted and effective [22]. Leveraging the well-established image processing capabilities of convolutional neural networks (CNNs) optimized for 2D data proves advantageous in various applications involving 3D point cloud data [23]. While this representation introduces further discretization, it offers valuable benefits, such as compatibility with established data acquisition methodologies in additive manufacturing and the effectiveness of CNN methodologies, leading to promising results in practical scenarios [24]. Importantly, in certain applications, obtaining 3D scans may not be feasible, especially for extremely complex shapes. However, the shape features remain crucial for determining functional properties. In such cases, utilizing 2D image stacks might be the best possible approach to extract valuable insights and predict functional behavior effectively. Therefore, without sacrificing generality, we assume that the 3D point cloud measurements can be substituted with layer-wise 2D imaging data.
From the object of interest, which is the functional part intended for field use, a set of 3D point cloud measurements is available, where S is the subscript for a 3D shape, i is the sample index, NOoI is the total number of samples, and the sample consists of a set of unstructured, varying-sized 3D measurement points (i.e., ). The object of interest represents the target domain, for which we aim to enhance prediction accuracy. Additionally, a set of process variables is available, which are the same for the object of interest and the physical twin, since they are printed under the same process conditions. Note that we do not assume the availability of the functional property output variables (i.e., unlabeled dataset) in the AUDIT framework, which enhances the practicality of this approach by eliminating the need for an extensive dataset obtained through destructive testing. The destructive testing labels of the object of interest are solely utilized to verify the model's performance.
For the digital twin, a set of 3D measurement point clouds of the 3D-printed part is available, which represents the shape of the part. In addition, a set of material properties of the part is available too, where the subscript M denotes the material. Furthermore, functional property output variables are obtained via computer simulation (e.g., finite element analysis), where , and denotes the dimension of the (multivariate) functional property, or dimension of the output variables. For the digital twin, the process variables are not available because it did not undergo a physical printing process.
The physical twin, which is manufactured under identical process conditions as the object of interest, exhibits similar design features but utilizes less material to conserve resources. However, it still allows for destructive testing to assess the impact of the process conditions. For the physical twin, a set of 3D measurement point clouds and a set of process variables for the 3D-printed part are available. Note that the object of interest and the physical twin are printed at the same time on one print bed, so they are printed under the same process conditions. Additionally, functional property output variables are obtained from physical testing procedures, where . For the physical twin, we did not measure specific material properties for each sample, and hence no material properties are available.
Based on this dataset, we study the problem of unsupervised domain adaptation for 3D point cloud models by adapting a 3D model parametrized by θ, where θ is obtained from multiple labeled sources, multi-input domains (i.e.,, ) to an unlabeled target domain (i.e., object of interest). The main objective of the AUDIT framework is to improve the performance of the model on the unseen test set of the target domain, which requires careful consideration of the architecture and loss function to enable the learning of transferable features.
3.1 Architecture Design and Big Picture.
To accomplish this objective, we employ multi-input encoders for both the target (object of interest) and source domains (physical and digital twins), allowing us to learn domain-specific features. These features are then used in the discriminative head of the model, which benefits from a shared encoder structure that learns transferable features. The proposed framework is presented in Fig. 2, offers a high-level overview.
Our architectural design hypothesis is based on the notion that 3D networks progressively process domain-specific nontransferable features while acquiring domain-invariant features. To capture these distinctions, we utilize domain-specific encoders to learn low-level features that are unique to each data domain. Subsequently, we concatenate the features from the low-dimensional feature space and perform contrastive alignment to achieve instance-level feature alignment. Finally, we incorporate a discriminative head that provides a supervision signal for both the physical and digital twin domains. A discriminative head refers to the final layers of a deep learning model that is responsible for making regression predictions or classifying inputs based on the learned features extracted by the preceding layers. This head also predicts pseudo-labels for the target domain, which represents the object of interest. These pseudo-labels are continuously updated during the joint optimization process.
Transfer learning for 3D objects is challenging due to significant geometry shifts, such as variations in density and occlusion ratios caused by diverse physical environments and sensor configurations. Unlike 2D domain models trained on the backbone ImageNet, 3D point cloud modeling lacks a well-trained, transferable backbone. One of the reasons is the difficulty in reducing domain shifts for low-level geometric representations in the 3D model architecture.
Our architecture addresses this challenge by leveraging domain-specific 3D encoders that learn distinct mapping functions to convert unstructured 3D point clouds into a low-dimensional feature space. In our method, “domain-specific encoders” refer to separate neural network encoders designed to extract distinctive features from different source domains (i.e., digital twin, physical twin, and object of interest domain). The encoders are domain-specific since they do not share model parameters with other domains to capture unique features present within each domain. The use of domain-specific encoders offers several advantages:
– Separation of Domain-Specific Features: Domain-specific encoders facilitate the disentanglement of domain-specific features from shared features. By forcing the encoders to focus on capturing domain-specific characteristics, we enable the model to differentiate between features that are intrinsic to each domain and those that are shared across domains.
– Enhanced Feature Discriminability: When the encoders are tailored to their respective domains, the learned features become more discriminative. This discriminability improves the model's ability to capture subtle differences and adapt to variations. Domain-specific encoders prevent the model from merging domain-specific attributes into a single, less informative representation.
– Robustness to Domain Shift: Domain-specific encoders contribute to the model's robustness against domain shift. As each encoder specializes in learning domain-specific representations, the model becomes more adaptable to variations between sources and target domains. This adaptability is crucial for effective alignment in scenarios where domains exhibit dissimilarities due to changes in data collection conditions or sensing mechanisms.
In summary, domain-specific encoders in our method play a crucial role in enhancing the effectiveness of the contrastive instance alignment. This enables domain adaptation on the target domain while maintaining performance on the source domain, facilitating bidirectional knowledge sharing. Shared 3D encoders co-train with data samples from both domains, compressing the outputs of domain-specific encoders.
3.2 Selection of Physical Twin.
The design of the physical twin should be tailored to the specific attributes of the object of interest and the goals of the functional qualification. For example, in our case study, we were particularly concerned with the compressive force behavior of lattice structures. Hence, through preliminary experiments and FEA simulations, we identified high-stress regions that occur for certain printing parameters/settings under specific forces and orientations. Based on those results, we can choose a design that preserves the high-stress regions while conserving materials and reducing printing time. A practical approach to constructing the physical twins is to utilize parametric models accessible through CAD tools. These models allow us to extract and incorporate significant design characteristics, which have been identified through prior experimentation and FEA analysis. The connection between the physical twin and the object of interest stems from our machine learning model. This is in contrast to the current industrial practice of printing two identical parts, subsequently subjecting one to destructive testing. This existing approach lacks a definitive link or model to establish the equivalence and relationship between the two parts and leads to high material usage and scrap rates. Our method effectively remedies this shortcoming. Simultaneously printing a physical twin alongside the object of interest enables us to evaluate how variations in the printing parameters influence the functionality and specifications of the object of interest. In the case study, while we have selected a quarter section of the lattice as the physical twin, the impact of the excluded links is contained within our training dataset. Through the integration of digital and physical twins into a comprehensive machine learning model, we can significantly enhance the precision of predictions for functional characteristics. This enhancement is achieved by considering process variability through the physical twin and process variables, alongside factoring in the effects of 3D shapes and materials via the digital twin. The efficacy of this approach has been demonstrated in our case study. While we acknowledge that further exploration is needed to apply these concepts to more intricate component shapes and diverse functional qualification objectives, our intention is to introduce a novel concept through our paper. We aim to address a critical problem—functional qualification in AM—which will undoubtedly require further investigation for the development of appropriate physical twin designs tailored to new objectives and components. A promising direction for the design of physical twins is to draw inspiration from accelerated testing techniques in reliability theory. For instance, one potential strategy involves proportionally reducing the wall thickness of the part and simplifying its structures in the physical twin. However, it is crucial to incorporate uncertainty quantification methodologies to establish confidence levels for the relationships established by our model. The robustness and generalizability of our approach depend on the careful selection of physical twin (source domain) shapes that share large similarities with the object of interest. By choosing a large overlap of geometric attributes and design complexities between physical twin and object of interest, it is possible to improve the method's ability to effectively adapt to the target domain.
3.3 Contrastive Instance Alignment.
– Scale Invariance for Varying Point Density: 3D point clouds are inherently sparse and exhibit varying point densities across domains due to different sensor configurations. By choosing the cosine similarity, we leverage its scale-invariant nature to ensure that our instance alignment method is not affected by the overall density or magnitude of points in each point cloud.
– Directional Information for Spatial Relationships: In the realm of 3D point clouds, capturing spatial relationships is very important. Cosine similarity considers the direction of the vectors in the high-dimensional space, allowing us to capture the alignment based on the orientations of the vectors rather than just their magnitudes. This feature becomes valuable when aligning instances to preserve spatial structures and geometric arrangements, crucial for 3D point cloud tasks like shape matching and object recognition across domains.
– Sparse Data Handling in High Dimensions: Cosine similarity is effective when dealing with high-dimensional and sparse point cloud data. In some point cloud representations like voxels, the majority of elements in point cloud vectors might be zero, rendering traditional distance metrics less effective.
3.4 Hard Sample Mining.
In the context of 3D measurements in manufacturing, variability arises from factors such as variations in point cloud density and the presence of occlusions. These factors have a significant impact on the effectiveness of contrastive instance alignment. Point cloud density can vary between the object of interest and the physical twin, with some measurement techniques producing sparse point clouds while others generate denser ones. This discrepancy poses a challenge when aligning sample distributions, as the contrastive instance alignment approach may favor densely populated areas, potentially neglecting patterns with sparse point clouds. Moreover, additive manufacturing often involves complex geometries and occlusions, making it challenging to capture complete object geometry. Consequently, pseudo-labels used for alignment may not accurately represent patterns with severe occlusions. Acknowledging and addressing these factors is essential for accurate and comprehensive 3D transfer learning in additive manufacturing and other manufacturing domains.
To efficiently obtain hard samples, Ref. [25] proposed a novel algorithm that optimizes without directly conducting gradient-based optimization. This algorithm combines two components: simulating object occlusions by altering the geometry of easy samples and discarding critical points along the gradient direction from existing dense point clouds. The attribution score, indicating the contribution of each point to the discriminative loss, plays a crucial role in hard sample mining. Aggregating highly scored points identifies important segments/subsets in a point cloud. By discarding points with high attribution scores, a “hard sample” is created for the model to predict. The transformed point clouds achieved through hard sample mining contribute to effective contrastive instance alignment by reducing the distribution mismatch induced by pseudo-labels in the target domain. Figure 3 illustrates the hard sample mining algorithm, involving random viewpoint selection, calculation of point attributions, and deletion of points with large attribution scores until the termination criterion is met .
The transformed point clouds focus on effective contrastive instance alignment by reducing the distribution mismatch of the target (object of interest) domain induced by pseudo-labels.
Figure 4 illustrates the AUDIT procedure specifically designed for multi-source constructive alignment. In the context of functional qualification in additive manufacturing, the conventional use of contrastive loss tends to effectively align easily recognizable 3D objects, such as parts with consistent point density and minimal occlusions. However, this approach often neglects the challenging samples encountered in additive manufacturing, where different parts may exhibit variations in point density and severe occlusions due to complex geometries or intricate designs. Consequently, using the conventional contrastive loss may lead to a mismatch in point density and occlusion ratio between the sample distribution of pseudo-labels and the ground truths in the object of interest (target) domain. To overcome this issue, we leverage the hard sample mining algorithm [25] in the additive manufacturing context. This algorithm transforms point clouds by considering specific geometry mismatches across the object of interest and physical twin. By addressing these challenges, the hard sample mining algorithm significantly improves the domain alignment, enabling more effective transfer learning and alignment among the objects of interest, the physical and digital twins.
3.5 Unified AUDIT Framework.
We propose a stepwise training procedure with a warm-up process to train the AUDIT framework as shown in Algorithm 1. Specifically, we first pre-train the source models of the physical and digital twins on the labeled source domain and use them to generate pseudo-labels on the target set of the object of interest. We then conduct hard sample mining [25] and augment the target set. Next, we warm up the model following Eq. (6), which allows a more stable convergence in the early stages of training. For the remaining epochs, we update the pseudo-labels using stepwise co-training. During this process gradually adapts to the object of interest (target) domain while maintaining the in-domain performance.
The model architecture details are available from the source code of the AUDIT method, which will be open source upon paper publication. The hyperparameters are tuned using Bayesian optimization.
AUDIT algorithm for functional qualification via multi-source, multi-input transfer learning between digital and physical twins and the object of interest
Inputs:
Digital twin (Source 1): Labeled point cloud dataset from source domain
Physical twin (Source 2): Labeled point cloud dataset from source domain
Material properties for digital twin simulation:
Process measurements (identical for object of interest and physical twin):
Object of interest (target): Unlabeled input point cloud from target domain
Algorithm parameters: Network architecture and termination tolerance
Output:
Learned network weights θ of model f(·)
Algorithm:
Pre-train base model
finit = fit(DDT, DPT)
Generate pseudo-labels for target domain samples
Mine hard samples to augment the target set
Initialize the model with 3D base model
Warm start of AUDIT model
AUDIT iteration: Iteration index k
While not converged:
6.1 Update pseudo-labels:
- 6.2 Add new hard samples to the target dataset [25]
6.3 Model update
Termination check:
3.6 Discussion of Properties and Limitations.
The effectiveness of transfer learning is not always guaranteed, unless its basic assumptions are satisfied: (1) the learning tasks of the domains are related/similar; (2) the source domain and target domain data distributions are not too different; and (3) a suitable model can be applied to both domains. Violations of these assumptions may lead to negative transfer (NT), i.e., introducing source domain data/knowledge undesirably decreases the learning performance in the target domain. We would like to highlight two key properties of our approach in handling this issue: contrastive instance alignment using cosine similarity and the incorporation of hard sample mining:
– Contrastive Instance Alignment Using Cosine Similarity: Our method employs contrastive instance alignment as a core component. By utilizing cosine similarity, we focus on aligning instances while considering their relative positions in the feature space. This approach helps mitigate negative transfer by promoting the alignment of instances that share semantic similarities from both the source and target domains. This alignment encourages these instances to group together in the feature space, effectively preserving domain-specific characteristics. The use of cosine similarity as a distance metric promotes the alignment of semantically similar instances, even in scenarios where other distance measures might not be as effective.
– Hard Sample Mining: Negative transfer can stem from including irrelevant or conflicting source domain instances. To mitigate this, our method employs hard sample mining during the instance alignment process. Hard sample mining involves identifying challenging samples from the source domain that are difficult to align with the target domain. By focusing on these challenging instances, our approach reduces the likelihood of introducing undesirable knowledge from the source domain into the target domain. This strategy enhances the model's adaptability by prioritizing instances that contribute positively to the alignment process.
Through the synergistic application of contrastive instance alignment with cosine similarity and the incorporation of hard sample mining, our approach actively addresses the negative transfer challenge.
Acknowledging the potential scarcity of physical twins in comparison to digital twins, which may lead to imbalanced data, we emphasize the adaptability of our method with slight adjustments. These adaptations are outlined as follows:
– Parallel Feature Learning With Imbalanced Data: Within the framework of co-training, the imbalanced domain can be considered as one view while the balanced (or artificially balanced) domain serves as the other. This preserves the original co-training mechanism while mitigating the imbalance challenge.
– Data-Specific Sampling: When imbalances are present in source and target domains, distinct sampling strategies can be employed for each dataset. Strategies such as oversampling, undersampling, or adaptive sampling can be applied to address the imbalance in the imbalanced domain, while the balanced (or artificially balanced) domain adheres to regular co-training principles without extensive alterations.
However, we acknowledge that data imbalances can introduce biases in the alignment process. Our method is designed with a certain level of flexibility to accommodate such concerns via more advanced strategies as follows:
– Loss Re-Weighting: Implementing loss re-weighting mechanisms assigns higher weights to instances from the minority domain. This corrective measure counteracts the impact of imbalanced data by directing the model's focus toward underrepresented instances, thus fostering a more balanced alignment.
– Transfer Learning Techniques: The use of transfer learning techniques enables the utilization of pre-trained models or features from the imbalanced domain to initiate the alignment process. This leverages knowledge transfer from both the digital and physical twins, even when the latter's data are limited.
In conclusion, our model offers a robust approach for functional qualification, employing both physical and digital twins within the framework of multi-source, multi-input unsupervised transfer learning. Nonetheless, we recognize certain limitations that present exciting opportunities for future research endeavors.
4 AUDIT Case Study—3D-Printed Lattice Structures
We conducted a real-world case study to demonstrate the potential of the AUDIT framework for functional qualification in additive manufacturing via physical and digital twins. Our experiments use fused filament fabrication (FFF) to print PLA specimens. The printed object of interest is a body-centered cubic (BCC) lattice situated within a cubic-primitive (CP) (Fig. 5(a)). The dimensions of the lattice unit cube are 5 cm × 5 cm × 5 cm. To create a physical twin, we extracted a one-fourth portion of the BCC-CP lattice structure. This approach allowed us to produce a physical twin that required less material compared to the object of interest, while still retaining similar design features (Fig. 5(b)). The physical twin and the object of interest were printed using identical process conditions. This enables us to understand how the printing conditions of both parts impact their functional properties.
4.1 Experimental Setup.
The specimens for the experiments have been printed using a Prusa MK3S FFF printer developed by Prusa Research, Prague, Czech Republic. The measurement setup is complemented by a FLIR T360 thermal imaging infrared camera with 1.3 MP resolution and a FARO Quantum ScanArm with laser line probe. A microcomputer is used to log the nozzle and print bed temperature. A noise detector is installed to collect acoustic emission signals of the process. The experimental setup is visualized in Fig. 5(c).
To capture the influence of process parameters on the FFF printing process, a space-filling Latin hypercube design with N = 60 samples is utilized. The corresponding process parameter ranges are reported in Table 1.
Process setting | Range |
---|---|
Printing speed | 35–100 mm/s |
Fan speed | 0–100% |
Nozzle temperature | 190–240 °C |
Print bed temperature | 40–75 °C |
Extrusion width | 0.35–0.55 mm |
Process setting | Range |
---|---|
Printing speed | 35–100 mm/s |
Fan speed | 0–100% |
Nozzle temperature | 190–240 °C |
Print bed temperature | 40–75 °C |
Extrusion width | 0.35–0.55 mm |
When conducting the experiments, 12 experiments failed due to improper process parameter combinations, resulting in 48 samples in total. For each of those experiments, 3D measurement point cloud data from the FARO Scanner are available for both the object of interest as well as the physical twin. Additionally, we recorded in situ sensing data from six heterogeneous data sources ranging from three data types tabular, functional curve, and image data. The process data sources along with their typical dimensions are listed in Table 2, where L = 250 denotes the number of printing layers.
Process data | Data type | Data dimension |
---|---|---|
Process settings (no in situ monitoring):
| Tabular | ℝ3 |
Nozzle temperature: | Functional curve | ℝ1000×L |
Print bed temperature: | Functional curve | ℝ1000×L |
Infrared image: | Image | 320 × 240 × L |
Process data | Data type | Data dimension |
---|---|---|
Process settings (no in situ monitoring):
| Tabular | ℝ3 |
Nozzle temperature: | Functional curve | ℝ1000×L |
Print bed temperature: | Functional curve | ℝ1000×L |
Infrared image: | Image | 320 × 240 × L |
In terms of data preprocessing, the functional curves of the nozzle and bed temperature are fixed to a length of 1000 using dynamic time warping. The point clouds of the object of interest and the physical twin are up- or down-sampled to a fixed-point number of Np = 60,000 resulting in a data dimension ℝ60,000×3 for each sample. Note that these measurement points are unstructured and can exhibit irregular spatial arrangements and varying densities. In contrast to structured point clouds, these measurement locations are not consistent across different samples.
To model the heterogeneous input data, we use the following data-type-specific feature extractors: for the tabular data, employ a fully-connect multi-layer perceptron (MLP) to extract features from , , and . For the two functional curves (i.e., , ), we utilize the deep CNN architecture proposed by Yang et al. [26] in an autoencoder setting. As the feature extractor for the infrared images , we utilize a convolutional autoencoder structure proposed by Ref. [27]. Here we elaborate further on our rationale for choosing specific neural network architectures tailored to distinct dataset types:
– Fully-Connected MLP: For tabular datasets, we opted for a fully-connected MLP due to its effectiveness in handling structured data. Tabular data typically consist of features organized in rows and columns, making them well-suited for MLPs. The architecture's ability to learn complex relationships among features enables accurate predictions in such scenarios.
– Deep CNN Architecture: A deep CNN architecture was selected for functional curve datasets as it excels in processing sequential data. Functional curves represent time-series or sequential data, where the order of the data points is crucial. The hierarchical nature of CNNs allows them to capture both local and global patterns in the functional curves, making them a powerful choice for this dataset.
– Convolutional Autoencoder Structure: For infrared image datasets, we employed a convolutional autoencoder structure. Autoencoders are particularly suitable for learning efficient representations from high-dimensional data like images. The convolutional autoencoder's ability to encode essential features and reconstruct the images with high fidelity is essential for achieving accurate predictions with infrared images.
Each chosen architecture was carefully tailored to the specific attributes and intricacies of its respective dataset. This approach aimed to maximize performance and ensure robust predictions. The neural network designs were aligned with the inherent nature of the data types, striving for optimal outcomes in each case. It is worth noting that these selections are widely acknowledged in literature. However, we acknowledge the potential necessity of fine-tuning architectures for particular applications to further elevate performance and address unique challenges inherent to different scenarios.
Furthermore, we conducted individual quasi-static compression tests on the object of interest and physical twin samples using a Shimadzu AG-IC 20 kN UTM. To carry out the compression test, we positioned the lattice specimens on a rigid plate, with an upper rigid plate descending to apply compression at an engineering strain rate of 0.001 s−1. Since the layer-wise fabrication process introduces anisotropy in material properties, all lattices in this study were compressed along the rise (printing) direction for consistency. The displacement recorded by the UTM and the contact force measured by the load cell attached to the upper plate were converted into engineering stress–strain curves. Additionally, we utilized a digital camera to capture optical images of the entire crushing process, enabling future analysis of the deformation mechanism. Figure 6 presents a visual representation of the destructive testing performed on one object of interest for different contact forces. The experiment began with Fig. 6(a) where no displacement was initially applied. As the experiment progressed, a gradual displacement was exerted on the top surface in a top-down direction to compress the part. In Fig. 6(b), the force reached a magnitude of 103.1 N. Continuing the experiment, both the displacement and force continued to increase. However, as the structure started to crack, the two left front struts of the CP became detached and flew away, and the front right strut also developed a crack. Consequently, the force decreased to 86.2 N, as depicted in Fig. 6(c). Subsequently, the force was absorbed by the inner lattice structure (BCC). As the back right top strut of the BCC structure also cracked, the force further dropped to 43.7 N, as illustrated in Fig. 6(d).
The primary aim of this case study is to accurately predict the maximum compressive strength of the object of interest , as it serves as a crucial indicator for assessing the functional qualification of lattice structures. The histograms and the fitted PDF shown in Fig. 7 provide evidence of substantial variation in functional performance due to different printing process conditions. Each of the 48 maximum compressive force results in Fig. 7 corresponds to a distinct set of process settings, as a space-filling design of experiments was employed. Additionally, a moderate correlation of 53.43% (Pearson) is observed between the maximum compressive force of the object of interest and its physical twin , as they were printed under identical process conditions. This correlation further shows the connection between the functional properties of the object of interest and its physical twin, which is induced by the identical printing process conditions.
In our case study, our focus was on predicting the maximum compressive strength of the object of interest—a pivotal parameter for numerous applications. However, we recognize the significance of extending our predictions to encompass other functional properties relevant in diverse applications. To enable predictions for these additional functional properties, a series of carefully designed experiments is essential. These experiments should encompass various materials, geometries, and process conditions to establish comprehensive datasets. These datasets will then serve as the foundation for training our model to make precise predictions across a range of functional properties. The accuracy of these predictions depends on the correlation among the 3D shape, digital twin, and measured process conditions with functional properties.
4.2 Finite Element Analysis Simulations for the Digital Twin.
To create a digital twin of the object, a 3D scan was performed of the object of interest. Finite element simulations were then conducted using the 3D scan data to analyze the compressive behavior. The ansysTM software, specifically the static structural module, was utilized. The boundary condition involved fixing the bottom surfaces of the lattice structure while applying uniform stress on the top surfaces. Table 3 presents the input parameters used in the simulations, which were obtained from the literature on bulk PLA filament [28].
Density (kg/m3) | Elastic modulus | Poisson's ratio |
---|---|---|
ρ = 1240 | E = 3500 MPa | ν = 0.35 |
Density (kg/m3) | Elastic modulus | Poisson's ratio |
---|---|---|
ρ = 1240 | E = 3500 MPa | ν = 0.35 |
The output of the simulation is the breaking displacement value at the ultimate tensile strength of PLA (42 MPa) with a precision of two decimal microns. Figure 8 depicts the lattice structure of the digital twin at various times throughout the simulation.
We find that the small shape discrepancies from the ideal design have only a moderate impact on the functional properties, as indicated by a Pearson correlation of 22.31% between the breaking displacement and the maximum compressive force of the object of interest .
A far more crucial factor influencing the functional characteristics is the 3D printing process conditions. Optimizing printing process conditions enables enhanced control over the functional properties of printed objects. By emphasizing the relationship between process conditions and functional properties, it paves the way for continuous improvement and design optimization in future research.
4.3 Benchmark Methods.
We evaluated the AUDIT framework against various benchmarks such as linear regression, multi-layer perceptron, supervised transfer learning (pre-training), and a data augmentation scheme called PointAugment. In the following, we give a brief overview of those benchmark methods.
Two supervised regression models, namely linear regression and MLP, are used as benchmarks. Linear regression is chosen for its simplicity, while MLP can capture nonlinear relationships. These models use a combination of features related to the 3D printing process conditions and functional property outcome variables from digital and physical twins to predict the maximum compressive force of the object of interest. Various combinations of input features are explored for both models to identify the optimal setup. Table 4 displays the different input feature settings for each model configuration.
Setting 1 | Setting 2 | Setting 3 | Setting 4 | |
---|---|---|---|---|
Input features | Process setting values:
| Setting 1 + : destructive testing result of physical twin | Setting 1 + : simulated result of digital twin | Setting 1 + + |
Output |
Setting 1 | Setting 2 | Setting 3 | Setting 4 | |
---|---|---|---|---|
Input features | Process setting values:
| Setting 1 + : destructive testing result of physical twin | Setting 1 + : simulated result of digital twin | Setting 1 + + |
Output |
Furthermore, we utilize a widely used supervised transfer learning technique called pre-training [29]. Initially, the model is pre-trained on the source dataset (digital and physical twins) and then fine-tuned on the target dataset of interest. This transfer learning approach mitigates overfitting caused by PointNet's neural network structure when dealing with smaller datasets. During pre-training, a certain percentage of the initial layers are frozen, allowing the model to leverage generalized information. The optimal number of frozen layers is determined through iterative exploration. Subsequently, the model is fitted to the object of interest (target) dataset.
Finally, we compare AUDIT with PointAugment, an advanced augmentation algorithm for 3D point cloud data [30]. PointAugment generates new samples by augmenting existing ones, enriching data diversity. Unlike fixed strategies, PointAugment trains an augmentor alongside the model, using sample-aware augmentation based on geometric structure. It optimizes performance through adversarial learning and automates the augmentation process, improving dataset enrichment.
Although these benchmarks have the advantage of accessing target labels, we include them to evaluate AUDIT's unsupervised domain adaptation performance.
4.4 Case Study Prediction Results.
In this section, we compare the AUDIT framework with the benchmarks using normalized root mean squared error (NRMSE) for future dataset comparisons. Table 5 presents the average NRMSE from ten-fold cross-validation (CV). In particular, we used a nested cross-validation setup, where the outer loop performs ten-fold cross-validation for model evaluation (43 samples (90%) for training and 5 samples (10%) for testing), while the inner loop splits the training data further to tune hyperparameters on a smaller training set (34 samples, 80%) and a validation set (9 samples, 20%) to tune the hyperparameters. It is important to note that the object of interest labels used for performance evaluation in AUDIT are only employed during training in the benchmark methods. Additionally, we provide the un-normalized root mean squared error (RMSE) to scale back to the dataset being analyzed .
Method | NRMSE | RMSE |
---|---|---|
Linear regression—Setting 1 | 6.066 (3.214) | 1319.420 (699.081) |
Linear regression—Setting 2 | 5.940 (2.668) | 1291.975 (580.329) |
Linear regression—Setting 3 | 6.065 (2.745) | 1319.205 (597.204) |
Linear regression—Setting 4 | 5.952 (2.242) | 1294.718 (487.785) |
MLP—Setting 1 | 7.407 (4.363) | 1611.130 (949.229) |
MLP—Setting 2 | 6.961 (3.018) | 1514.087 (654.766) |
MLP—Setting 3 | 7.446 (2.542) | 1619.608 (552.954) |
MLP—Setting 4 | 2.781 (1.303) | 604.947 (283.484) |
Pre-training (PointNet) | 20.973 (13.269) | 4561.662 (2884.613) |
PointAugment | 0.280 (0.076) | 61.067 (16.628) |
AUDIT (ours) | 0.014 (0.001) | 3.002 (0.337) |
Method | NRMSE | RMSE |
---|---|---|
Linear regression—Setting 1 | 6.066 (3.214) | 1319.420 (699.081) |
Linear regression—Setting 2 | 5.940 (2.668) | 1291.975 (580.329) |
Linear regression—Setting 3 | 6.065 (2.745) | 1319.205 (597.204) |
Linear regression—Setting 4 | 5.952 (2.242) | 1294.718 (487.785) |
MLP—Setting 1 | 7.407 (4.363) | 1611.130 (949.229) |
MLP—Setting 2 | 6.961 (3.018) | 1514.087 (654.766) |
MLP—Setting 3 | 7.446 (2.542) | 1619.608 (552.954) |
MLP—Setting 4 | 2.781 (1.303) | 604.947 (283.484) |
Pre-training (PointNet) | 20.973 (13.269) | 4561.662 (2884.613) |
PointAugment | 0.280 (0.076) | 61.067 (16.628) |
AUDIT (ours) | 0.014 (0.001) | 3.002 (0.337) |
AUDIT demonstrates superior performance compared to all other models in Table 5, showcasing its potential for functional qualification in additive manufacturing. AUDIT achieves a small RMSE of the maximum compressive force of approximately 3 N, demonstrating the accurate prediction of functional properties.
Although the benchmark methods (e.g., linear regression, MLP, pre-training, and PointAugment) have access to target labels, they still fall short in achieving satisfactory performance due to limited sample size, high-dimensional data complexity, and the inability to consider all relevant data sources. In contrast, AUDIT surpasses them by effectively co-training with labeled source data and augmented hard samples, while also incorporating heterogeneous process conditions and material properties with its multi-input architecture.
Our framework incorporates hard sample mining as an augmentation strategy. During each training epoch, this technique generates a set of challenging “hard” samples for augmentation, enhancing the model's generalizability and preventing overfitting. Furthermore, to assess the model's ability to generalize to unseen data, we employed a nested ten-fold cross-validation methodology. We closely monitored the model's performance on a validation set during training and applied early stopping. If we detected signs of overfitting, such as a decrease in validation performance while training performance continued to improve, we stopped training to mitigate overfitting.
However, it is essential to acknowledge the limitations of this study. The current application of this novel machine learning approach is limited to a small dataset and simple geometries. More R&D efforts are needed for further exploration to assess its generalizability to larger datasets and more complex geometries.
5 Conclusion
In conclusion, this paper introduces the AUDIT framework—a novel approach to the functional qualification of 3D-printed parts using physical and digital twins. By combining the strengths of physical and digital twins with transfer learning techniques, the AUDIT framework enables accurate predictions of the functional properties of 3D-printed parts without the need for destructive testing. The case studies on 3D-printed lattice structures highlight the potential of this approach in enhancing the functional qualification of critical 3D-printed parts. By considering real-world manufacturing process conditions and incorporating the FEA analysis of the 3D shape (digital twin), AUDIT provides a more holistic evaluation of 3D-printed functional properties. Additionally, the framework introduces transfer learning techniques for additive manufacturing processes, enabling the fusion of heterogeneous 3D shape data from multiple sources to enhance understanding of the relationships between 3D shapes, process conditions, and functional properties.
Although the framework has undergone evaluation using a dataset of 3D-printed lattice structures, there is a need for future work to expand and validate its applicability across a broader range of 3D-printed parts. This verification should encompass different printing processes and materials to ensure the framework's effectiveness in diverse scenarios. Additionally, future research should focus on improving the efficiency and sampling procedures of 3D data acquisition with intricate designs or internal structures (e.g., computed tomography scanning) to generate digital twins in those challenging applications.
Furthermore, this work has the potential to enable the development of control and compensation schemes based on the functional properties of the products. The AUDIT model establishes a link between heterogeneous process variables and functional properties, enabling inverse optimization and control of 3D printing parameters.
Overall, the AUDIT framework offers a comprehensive solution for functional qualification in 3D printing. Progress in this field holds the potential to facilitate the widespread adoption of 3D printing in safety-critical applications.
Acknowledgment
The authors would like to extend their sincere appreciation to Dr. Chuck Zhang, the Harold E. Smalley Professor, and Yifeng Wang, a Ph.D. Student, both affiliated with the H. Milton Stewart School of Industrial and Systems Engineering at Georgia Institute of Technology. Their invaluable assistance with the quasi-static compression tests conducted on the Universal Testing Machine (UTM) is deeply acknowledged.
Funding Data
This research is funded by National Science Foundation Award ID 2019378.
Conflict of Interest
There are no conflicts of interest.
Data Availability Statement
The datasets generated and supporting the findings of this article are obtainable from the corresponding author upon reasonable request.
Nomenclature
- i =
sample index
- E =
elastic modulus
- L =
number of printing layers
- S =
shape
- T =
target domain (object of interest)
- =
neural network model
- NOoI =
number of samples from the object of interest
- NPT =
number of samples from the physical twin
- =
dimension of the (multivariate) functional property output variables
- =
number of measurement points in object of interest point cloud of sample i
- =
feature instance from the source domain
- =
feature instance from the target domain
- =
hard sample point cloud of the object of interest
- =
maximal value of the functional property output variable of the object of interest in the dataset
- =
minimal value of the functional property output variable of the object of interest in the dataset
- =
discriminative loss
- =
inter-class loss
- =
intra-class loss
- =
contrastive alignment loss
- =
hard sample loss
- =
process variables
- =
functional property output variables of the object of interest
- =
functional property output variables of the digital twin
- =
functional property output variables of the physical twin
- =
3D measurement point cloud of the object of interest
- =
material properties for the digital twin
- =
3D measurement point cloud of the physical twin
- δ =
dynamic upper bounding parameter for hard sample mining
- θ =
parameters of the neural network model
- λ =
tuning parameter balance domain adaptation and learning of the discriminative task
- ρ =
density
- τ =
tuning parameter for the strength of domain adaptation
- ν =
Poisson's ratio
- 3D =
three-dimensional
- AM =
additive manufacturing
- AUDIT =
functional qualification in additive manufacturing via physical and digital twins
- BCC =
body-centered cubic
- CAD =
computer-aided design
- CP =
cubic-primitive
- CV =
cross-validation
- DT =
digital twin
- FEA =
finite element analysis
- FFF =
fused filament fabrication
- MLP =
multi-layer perceptron
- NRMSE =
normalized root mean squared error
- OoI =
object of interest
- PDF =
probability distribution function
- PLA =
polylactic acid
- PT =
physical twin
- RMSE =
root mean squared error
- UTM =
universal testing machine