| ELECTROMAGNETISM, OPTICS, ACOUSTICS, HEAT TRANSFER, CLASSICAL MECHANICS, AND FLUID DYNAMICS |
Prev
Next
|
|
|
Image-free single-pixel semantic segmentation for complex scene based on multi-scale U-Net |
| Tengfei Liu(刘腾飞), Yanfeng Bai(白艳锋)†, Jianxia Chen(陈健霞), Jintao Zhai(翟锦涛), Siqing Xiang(向思卿), Xianwei Huang(黄贤伟), and Xiquan Fu(傅喜泉)‡ |
| College of Computer Science and Electronic Engineering, Hunan University, Changsha 410082, China |
|
|
|
|
Abstract Single-pixel imaging (SPI) receives widespread attention due to its superior anti-interference capabilities, and image segmentation technology can effectively facilitate its recognition and information extraction. However, the complexity of the target scene and plenty of imaging time in SPI make it challenging to achieve high-quality and concise segmentation. In this paper, we investigate the image-free intricate scene semantic segmentation in SPI. Using "learned" illumination patterns allows for the full extraction of the object's spatial information, thereby enabling pixel-level segmentation results through the decoding of the received measurements. Simulation and experimentation show that, in the absence of image reconstruction, the mean intersection over union (MIoU) of segmented image can reach higher than 85%, and the Dice coefficient (DICE) close to 90% even at the sampling ratio of 5%. Our approach may be favorable to applications in medical image segmentation and autonomous driving field.
|
Received: 29 March 2025
Revised: 12 May 2025
Accepted manuscript online: 16 June 2025
|
|
PACS:
|
42.30.-d
|
(Imaging and optical processing)
|
| |
42.30.Va
|
(Image forming and processing)
|
| |
42.30.Tz
|
(Computer vision; robotic vision)
|
|
| Fund: Project supported by the Fundamental Research Funds for the Central Universities of China (Grant No. 531118010757). |
Corresponding Authors:
Yanfeng Bai, Xiquan Fu
E-mail: yfbai@hnu.edu.cn;fuxq@hnu.edu.cn
|
Cite this article:
Tengfei Liu(刘腾飞), Yanfeng Bai(白艳锋), Jianxia Chen(陈健霞), Jintao Zhai(翟锦涛), Siqing Xiang(向思卿), Xianwei Huang(黄贤伟), and Xiquan Fu(傅喜泉) Image-free single-pixel semantic segmentation for complex scene based on multi-scale U-Net 2026 Chin. Phys. B 35 014202
|
[1] Saleh K, Szénási S and Vámossy Z 2021 IEEE 19th World Symposium on Applied Machine Intelligence and Informatics (SAMI) 000477 [2] Zaitoun N M and Aqel M J 2015 Procedia Comput. Sci. 65 797 [3] 1978 IEEE Trans. Syst. Man Cybern. 8 630 [4] Canny J 1986 IEEE Trans. Pattern Anal. Mach. Intell. PAMI- 8 679 [5] Adams R and Bischof L 1994 IEEE Trans. Pattern Anal. Mach. Intell. 16 641 [6] Ronneberger O, Fischer P and Brox T 2015 Medical image computing and computer-assisted intervention, MICCAI 2015: 18th international conference, Munich, Germany, October 5–9, 2015, proceedings, part III 18, pp. 234–241 [7] Badrinarayanan V, Kendall A and Cipolla R 2017 IEEE Trans. Pattern Anal. Mach. Intell. 39 2481 [8] He K, Zhang X, Ren S and Sun J 2016 Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770–778 [9] He K, Gkioxari G, Dollár P and Girshick R 2017 Proceedings of the IEEE international conference on computer vision, pp. 2961–2969 [10] Zhou Y, Guo S X, Zhong F and Zhang T 2019 Chin. Phys. B 28 84204 [11] Hou H Y, Zhao Y N, Han J C, Cao D Z, Zhang S H, Liu H C and Liang B L 2023 Chin. Phys. B 32 064201 [12] Bie S H, Wang C H, Lv R B, Bao Q Q, Fu Q, Meng S Y and Chen X H 2023 Chin. Phys. B 32 128702 [13] Hardy N D and Shapiro J H 2013 Phys. Rev. A 87 023820 [14] Hardy N D and Shapiro J H 2011 Phys. Rev. A 84 063824 [15] Liu J J, Bai Y F, Huang X W, Tan W, Nan S Q, Zou X P F and Fu X Q 2021 Appl. Sci.-Basel 11 4115 [16] Zou X, Huang X, Liu C, Tan W, Bai Y and Fu X 2023 Opt. Laser Technol. 167 109807 [17] Zhou L Y, Huang X W, Fu Q, Zou X P F, Bai Y F and Fu X Q 2021 Chin. Opt. Lett. 19 121101 [18] Pan L, Deng C J, Yu C P, Yue S, Gong W L and Han S S 2021 Chin. Opt. Lett. 19 041103 [19] Lin L X, Cao J, Zhou D, Cui H and Hao Q 2022 Opt. Express 30 11243 [20] Fan Y R, Bai Y F, Fu Q, Zhang R, Zhou L Y, Zhu X H, Zou X P F and Fu X Q 2024 Opt. Commun. 566 130684 [21] Cheng J and Han S S 2004 Phys. Rev. Lett. 92 093903 [22] Liu H C and Zhang S 2017 Appl. Phys. Lett. 111 031110 [23] Wang C L, Mei X D, Pan L, Wang P W, Li W, Gao X, Bo Z W, Chen M L, Gong W L and Han S S 2018 Remote Sens. 10 732 [24] Liu H, Bian L and Zhang J 2023 Opt. Laser Technol. 157 108600 [25] Liu X, Han T, Zhou C, Huang J, Ju M, Xu B and Song L 2023 Opt. Express 31 9945 [26] Gibson G M, Johnson S D and PadgettMJ 2020 Opt. Express 28 28190 [27] He Z Q and Dai S S 2024 AIP Adv. 14 045316 [28] Zhao H, Shi J, Qi X, Wang X and Jia J 2017 Proceedings of the IEEE conference on computer vision and pattern recognition 2881 [29] Cohen G, Afshar S, Tapson J and Van Schaik A 2017 International joint conference on neural networks (IJCNN) 2921 [30] Dong C, Loy C C and Tang X 2016 Computer Vision, ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 11–14, 2016, Proceedings, Part II 14 391 [31] Zhang Z B, Li X, Zheng S J, YaoMH, Zheng G A and Zhong J G 2020 Opt. Express 28 13269 [32] Hou H and Andrews H 1978 IEEE Trans. Acoust. Speech Signal Process. 26 508 [33] Floyd R W 1976 Proc. Soc. Inf. Disp. 17 75 |
| No Suggested Reading articles found! |
|
|
Viewed |
|
|
|
Full text
|
|
|
|
|
Abstract
|
|
|
|
|
Cited |
|
|
|
|
Altmetric
|
|
blogs
Facebook pages
Wikipedia page
Google+ users
|
Online attention
Altmetric calculates a score based on the online attention an article receives. Each coloured thread in the circle represents a different type of online attention. The number in the centre is the Altmetric score. Social media and mainstream news media are the main sources that calculate the score. Reference managers such as Mendeley are also tracked but do not contribute to the score. Older articles often score higher because they have had more time to get noticed. To account for this, Altmetric has included the context data for other articles of a similar age.
View more on Altmetrics
|
|
|