A Data-Driven Approach to Complex Voxel Predictions in Grayscale Digital Light Processing Additive Manufacturing Using U-Nets and Generative Adversarial Networks

Small. 2023 Dec;19(50):e2301987. doi: 10.1002/smll.202301987. Epub 2023 Jul 6.

Abstract

Data-driven U-net machine learning (ML) models, including the pix2pix conditional generative adversarial network (cGAN), are shown to predict 3D printed voxel geometry in digital light processing (DLP) additive manufacturing. A confocal microscopy-based workflow allows for the high-throughput acquisition of data on thousands of voxel interactions arising from randomly gray-scaled digital photomasks. Validation between prints and predictions shows accurate predictions with sub-pixel scale resolution. The trained cGAN performs virtual DLP experiments such as feature size-dependent cure depth, anti-aliasing, and sub-pixel geometry control. The pix2pix model is also applicable to larger masks than it is trained on. To this end, the model can qualitatively inform layer-scale and voxel-scale print failures in real 3D-printed parts. Overall, machine learning models and the data-driven methodology, exemplified by U-nets and cGANs, show considerable promise for predicting and correcting photomasks to achieve increased precision in DLP additive manufacturing.

Keywords: 3D printing; digital light processing; machine learning; neural networks; pix2pix; stereolithography; vat photopolymerization.