<p>Applications of X-ray computed tomography (CT) for porosity characterization of engineering materials often involve an extended data analysis workflow that includes CT reconstruction of raw projection data, binarization, labeling and mesh extraction. It is often desirable to map the porosity in larger samples but the computational challenge of reducing gigabytes of raw data to porosity information poses a critical bottleneck. In this work, we describe algorithms and implementation of an end-to-end porosity mapping code that processes raw projection data from a synchrotron CT instrument into a porosity map and visualization in the form of triangular face mesh. Towards this objective, we report the development of a novel subset reconstruction scheme for X-ray CT using filtered backprojection and a convolutional neural network that allows us to reconstruct arbitrarily-shaped subsets of a tomography object. We build upon this scheme to implement the complete code for porosity mapping. The code first detects possible voids by performing a coarse reconstruction on down-sampled projections and then improves the shape of those voids with higher detail offered by reconstructing selected subsets from the original raw data. We report measurements of the time taken by this code to perform complete processing from raw data to a triangular face mesh for several visualization scenarios on a single highperformance workstation equipped with GPU. We show that we can now visualize local porosity within a 8 gigavoxel CT volume (12 gigabytes raw data) within 1 to 2 minutes and a 64 gigavoxel CT volume (100 gigabytes of raw data) within 3 to 7 minutes.</p>
<p>Applications of X-ray computed tomography (CT) for porosity characterization of engineering materials often involve an extended data analysis workflow that includes CT reconstruction of raw projection data, binarization, labeling and mesh extraction. It is often desirable to map the porosity in larger samples but the computational challenge of reducing gigabytes of raw data to porosity information poses a critical bottleneck. In this work, we describe algorithms and implementation of an end-to-end porosity mapping code that processes raw projection data from a synchrotron CT instrument into a porosity map and visualization in the form of triangular face mesh. Towards this objective, we report the development of a novel subset reconstruction scheme for X-ray CT using filtered backprojection and a convolutional neural network that allows us to reconstruct arbitrarily-shaped subsets of a tomography object. We build upon this scheme to implement the complete code for porosity mapping. The code first detects possible voids by performing a coarse reconstruction on down-sampled projections and then improves the shape of those voids with higher detail offered by reconstructing selected subsets from the original raw data. We report measurements of the time taken by this code to perform complete processing from raw data to a triangular face mesh for several visualization scenarios on a single highperformance workstation equipped with GPU. We show that we can now visualize local porosity within a 8 gigavoxel CT volume (12 gigabytes raw data) within 1 to 2 minutes and a 64 gigavoxel CT volume (100 gigabytes of raw data) within 3 to 7 minutes.</p>
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
customersupport@researchsolutions.com
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
Copyright © 2025 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.