[ad_1]
Structured light can significantly enhance information capacity, due to its coupling of spatial dimensions and multiple degrees of freedom. In recent years, the combination of structured light patterns with image processing and machine intelligence has shown vigorous development potential in fields such as communication and detection.
One of the most notable features of a structured light field is the two and three-dimensional distribution of its amplitude information. This feature can effectively integrate with maturely developed image processing technology and can also achieve cross-medium information transmission by virtue of machine learning technology currently driving profound changes.
Complex structured light fields based on coherent superposition states can carry abundant spatial amplitude information. By further combining spatial nonlinear conversion, significant increases in information capacity can be realized.
Zilong Zhang from Beijing Institute of Technology and Yijie Shen from Nanyang Technological University, along with their teams’ members, proposed a new method for enhancing information capacity based on complex mode coherent superposition states and their spatial nonlinear conversion. By integrating machine vision and deep learning technologies, they achieved large-angle point-to-multipoint information transmission with a low bit error rate.
The study is published in the journal Laser & Photonics Reviews.
In this model, Gaussian beams are used to obtain spatial nonlinear conversion (SNC) of structured light through a spatial light modulator. Convolutional neural networks (CNN) are used to identify the intensity distribution of the beams.
By comparing the basic superposition mode and the SNC mode, it is observed that with the increase in order of constituent eigenmodes of the basic mode, the encoding capability of HG superposition mode is significantly better than LG mode, and the mode encoding capacity after spatial structured nonlinear conversion can be significantly improved.
To verify the encoding and decoding performances based on the above model, a 50×50-pixel color image was transmitted. The RGB dimensions of the image were divided into 5 chromaticity levels, comprising a total 125 kinds of chromaticity information, each encoded by 125 HG coherent superposition states. Additionally, different degrees of phase jitter caused by atmospheric turbulence were loaded onto these 125 modes through a DMD spatial light modulator and trained with deep learning technology to form a dataset.
Further, using nonlinear conversion, the analysis of higher capacity decoding effects was implemented, in which 530 SNC modes were selected for experimental measurement of the confusion matrix to these modes by convolutional neural networks. The experimental findings indicate that due to more distinct structural features, SNC modes can still ensure similarly low bit error rates while significantly increasing data capacity, with a data recognition accuracy up to 99.5%.
Additionally, the experiment also verified the machine vision pattern recognition capability under conditions of diffuse reflection, achieving simultaneous high-precision decoding by multiple receiving cameras with observation angles up to 70°.
More information:
Zilong Zhang et al, Spatial Nonlinear Conversion of Structured Light for Machine Learning Based Ultra‐Accurate Information Networks (Laser Photonics Rev. 18(6)/2024), Laser & Photonics Reviews (2024). DOI: 10.1002/lpor.202470039
Provided by
Chinese Academy of Sciences
Citation:
Machine learning method uses nonlinear optics and structured light to expand information network accuracy and capacity (2024, July 23)
retrieved 23 July 2024
from https://phys.org/news/2024-07-machine-method-nonlinear-optics-network.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no
part may be reproduced without the written permission. The content is provided for information purposes only.
[ad_2]
Source link