|
|
|
Physics-informed neural network for material identification via distortion-robust polychromatic x-ray attenuation correction in photon-counting detectors |
| Xin Yan(闫欣)1,2, Jie Zhang(张杰)2,†, Kai He(何凯)2,‡, Yiheng Liu(刘毅恒)2, Yuetong Zhao(赵悦彤)2, Gang Wang(王刚)2, Xinlong Chang(常新龙)1, and Youwei Zhang(张有为)3 |
1 Rocket Force University of Engineering, Xi'an 710025, China; 2 State Key Laboratory of Ultrafast Optical Science and Technology, Xi'an Institute of Optics and Precision Mechanics, Chinese Academy of Sciences, Xi'an 710119, China; 3 National Gravitation Laboratory, MOE Key Laboratory of Fundamental Physical Quantities Measurement, and School of Physics, Huazhong University of Science and Technology, Wuhan 430074, China |
|
|
|
|
Abstract Spectral distortions in photon-counting detectors (PCDs) fundamentally limit the quantitative accuracy of material identification. While machine learning is used for compensation, current data-driven methods often lack physical constraints, limiting their interpretability and reliability across varying conditions. To address this issue, we propose a physics-informed neural network (PINN) framework that explicitly embeds the Beer-Lambert law into the learning architecture. By integrating an explicit differential layer to extract high-order curvature features from distorted spectra, the model enables direct inference of the effective atomic number and areal density. This approach effectively leverages the $Z$-dependent non-linear profile of the photoelectric effect, even when explicit absorption edges are outside the primary detection window. Simulation results establish a high-precision benchmark for $Z_{\rm{eff}}$ estimation in the target low-$Z$ range (613), with an RMSE of 0.2111. Experimental validation on a CdZnTe-PCD further demonstrates that this accuracy improvement is preserved under realistic pulse pile-up and noise conditions, achieving an RMSE of 0.2457 and an $R^{{2}}$ of 0.9670. Compared with conventional physical correction methods (typically $\pm 0.5$ error margin), the proposed framework provides improved precision, with 92.86 % of $Z_{\rm{eff}}$ estimation errors falling within $\pm 0.4$, corresponding to an approximately 20 % tighter error bound. These results confirm that the proposed framework effectively mitigates spectral distortion, providing a robust, calibration-free solution for precise material identification of low-$Z$ materials in industrial non-destructive testing.
|
Received: 12 February 2026
Revised: 24 March 2026
Accepted manuscript online: 01 April 2026
|
|
PACS:
|
07.85.Fv
|
(X- and γ-ray sources, mirrors, gratings, and detectors)
|
| |
07.05.Mh
|
(Neural networks, fuzzy logic, artificial intelligence)
|
| |
81.70.-q
|
(Methods of materials testing and analysis)
|
| |
87.59.-e
|
(X-ray imaging)
|
|
| Fund: Project supported by the Natural Science Basic Research Program — General Program (Grant No. 2025JC-YBMS-712). |
Corresponding Authors:
Jie Zhang,E-mail:zhangjie@opt.ac.cn;Kai He,E-mail:hekai@opt.ac.cn
E-mail: zhangjie@opt.ac.cn;hekai@opt.ac.cn
|
Cite this article:
Xin Yan(闫欣), Jie Zhang(张杰), Kai He(何凯), Yiheng Liu(刘毅恒), Yuetong Zhao(赵悦彤), Gang Wang(王刚), Xinlong Chang(常新龙), and Youwei Zhang(张有为) Physics-informed neural network for material identification via distortion-robust polychromatic x-ray attenuation correction in photon-counting detectors 2026 Chin. Phys. B 35 050702
|
[1] Wen X, Chen S, Song K, Yu H, Li X and Zhong L 2025 Comput. Mater. Contin. 854173 [2] Morigi M P and Albertin F 2022 J. Imaging 8119 [3] Schwartz F R, Sodickson A D, Pickhardt P J, Sahani D V, Lev M H and Gupta R 2025 Radiology 314 e240662 [4] Johns H E and Cunningham J R 1983 The Physics of Radiology, 4th edn. (Springfield: Charles C Thomas) p. 134 [5] Alvarez R E and Macovski A 1976 Phys. Med. Biol. 21733 [6] Taguchi K and Iwanczyk J S 2013 Med. Phys. 40100901 [7] Kreisler B 2022 Eur. J. Radiol. 149110229 [8] Hayashi H, Kimoto N, Asahara T, Asakawa T, Lee C and Katsumata A 2021 Photon Counting Detectors for x-ray Imaging: Physics and Applications (Cham: Springer) p. 59 [9] Schlomka J P, Roessl E, Dorscheid R, Dill S, Martens G, Istel T, Baumer C, Herrmann C, Bonn R, Fink C, Nikisch E, Mattson J, Carmi R and Proksa R 2008 Phys. Med. Biol. 534031 [10] Cammin J, Xu J, Barber W C, Iwanczyk J S, Hartsough N E and Taguchi K 2014 Med. Phys. 41041905 [11] Ponchut C 2008 Nucl. Instrum. Methods Phys. Res. A 591311 [12] Hong W, Wen X L, Wei L, Zhu B, Wu Y C, Dong K G, Jiao C Y, Wu B and He Y L 2017 Chin. Phys. B 26025204 [13] Guo R R, Xu Y D, Zha G Q, Wang T and Jie W Q 2018 Chin. Phys. B 27127202 [14] Nan R H, Wang P F, Jian Z Y and Li X J 2017 Acta Phys. Sin. 66206101(in Chinese) [15] Yu H, Zhang M M, Du Y Y, Xi S Z, Zha G Q and Jie W Q 2021 J. Synth. Cryst. 501883 [16] Zhu W F, Wu J J, Tang X Q, Zhao R, Lu P Z, Qu B B, Li M Y and Qin F 2023 Nucl. Tech. 46050402 [17] Cao M, Deng J, He W, Zhou Y, Xie R, Jiang Y and Wang L 2023 Appl. Phys. A 129595 [18] Zakharchenko A A, Rybka A V, Davydov L N, Vierovkin A A, Kutny V E and Khazhmuradov M A 2014 East Eur. J. Phys. 168 [19] Nag R, Chaudhuri S K, Kleppinger J W, Karadavut O and Mandal K C 2021 J. Mater. Sci. Mater. Electron. 3226740 [20] Roy U N, Camarda G S, Cui Y, Yang G and James R B 2021 Sci. Rep. 1110338 [21] Hu K, Li F, Chen L, Liang F T and Jin G 2015 Chin. Phys. Lett. 32030701 [22] Kimoto N, Hayashi H, Nishigami R, Kobayashi D, Maeda T and Katsumata A 2024 Emerging Radiation Detection (Cham: Springer) p. 69 [23] Kimoto N, Nishigami R, Kobayashi D, Maeda T, Asahara T, Goto S, Kanazawa Y, Katsumata A, Yamamoto S and Hayashi H 2026 Radiat. Phys. Chem. 239113237 [24] Shikhaliev P M 2012 Phys. Med. Biol. 571595 [25] Kimoto N, Hayashi H, Asakawa T, Lee C, Asahara T, Maeda T, Goto S, Kanazawa Y, Katsumata A, Yamamoto S and Okada M 2021 Appl. Radiat. Isot. 170109617 [26] Murata K and Ogawa K 2021 Sci. Symp. Med. Imaging Conf. (Piscataway) p. 1217 [27] Sasaki T, Kuno H, Nomura K, Muramatsu Y, Aokage K, Samejima J, Taki T, Goto E, Wakabayashi M, Furuya H, Taguchi H and Kobayashi T 2025 Jpn. J. Radiol. 431132 [28] Lantz M and Ongie G 2023 Proc. IEEE 20th Int. Symp. Biomed. Imaging (Cartagena) p. 1 [29] Darui L, Jordan M M and Willie J P 2025 ACS Photonics 126071 [30] Chaudhuri S K, Kleppinger J W, Karadavut O, Nag R, Panta R, Agostinelli F, Sheth A, Roy U N, James R B and Mandal K C 2022 J. Mater. Sci. Mater. Electron. 331452 [31] Nadkarni R, Allphin A, Clark D P and Badea C T 2022 Phys. Med. Biol. 67155003 [32] Eguizabal A, Oktem O and Persson M U 2022 arXiv:2208.03360[physics.med-ph] [33] Weiss M, Brierley N, von Schmid M and Meisen T 2024 NDT 21 [34] Raissi M, Perdikaris P and Karniadakis G E 2019 J. Comput. Phys. 378686 [35] Chen Y, Lu L, Karniadakis G E and Dal Negro L 2020 Opt. Express 2811618 [36] Baldan M, Di Barba P and Lowther D A 2022 Proc. IEEE 20th Biennial Conf. Electromagn. Field Comput. (Denver) p. 1 [37] Zhang R, Li F and Liu J 2026 Chin. Phys. B 35030201 [38] Guo J, Wang H, Gu S and Hou C 2024 Chin. Phys. B 33050701 [39] Li W K, Zhao Z and Zhang Y T 2025 Acta Phys. Sin. 74235207(in Chinese) [40] Liu X M, Zhang Z Y and Liu W J 2023 Chin. Phys. Lett. 40070501 [41] Tian S F, Li B and Zhang Z 2024 Chin. Phys. Lett. 41030201 [42] Adler J and Oktem O 2018 IEEE Trans. Med. Imaging 371322 [43] Toscano J D, Oommen V, Varghese A J, Zou Z, Ahmadi Daryakenari N, Wu C and Karniadakis G E 2025 Mach. Learn. Comput. Sci. Eng. 115 [44] Yang S, Kim H, Hong Y, Yee K, Maulik R and Kang N 2024 Comput. Methods Appl. Mech. Eng. 428117075 |
| No Suggested Reading articles found! |
|
|
Viewed |
|
|
|
Full text
|
|
|
|
|
Abstract
|
|
|
|
|
Cited |
|
|
|
|
Altmetric
|
|
blogs
Facebook pages
Wikipedia page
Google+ users
|
Online attention
Altmetric calculates a score based on the online attention an article receives. Each coloured thread in the circle represents a different type of online attention. The number in the centre is the Altmetric score. Social media and mainstream news media are the main sources that calculate the score. Reference managers such as Mendeley are also tracked but do not contribute to the score. Older articles often score higher because they have had more time to get noticed. To account for this, Altmetric has included the context data for other articles of a similar age.
View more on Altmetrics
|
|
|