ΑΙhub.org
 

Identifying light sources using machine learning


by
26 May 2020



share this:
photon counting set-up
Artistic impression of schematic experimental set-up for photon counting. The section in the light grey box corresponds to the thermal light part of the experiment and the section in the dark grey box corresponds to the coherent light part.

The identification of light sources is very important for the development of photonic technologies such as light detection and ranging (LiDAR), and microscopy. Typically, a large number of measurements are needed to classify light sources such as sunlight, laser radiation, and molecule fluorescence. The identification has required collection of photon statistics or quantum state tomography. In recently published work, researchers have used a neural network to dramatically reduce the number of measurements required to discriminate thermal light from coherent light at the single-photon level.

In their paper, authors from Louisiana State University, Universidad Nacional Autónoma de México and Max-Born-Institut describe their experimental and theoretical techniques. They demonstrate the potential of machine learning to perform discrimination of light sources at extremely low light levels. This is achieved by training single artificial neurons with the statistical fluctuations that characterize coherent and thermal states of light.

The experimental set-up involves a continuous wave (CW) laser that is divided by a 50:50 beam splitter. Half of the beam passes through optical components which generate pseudo-thermal light. The emerging photons are counted by a superconducting nanowire single-photon detector (SNSPD). The other half of the beam is used as a coherent light source and is detected by another SNSPD. The data are divided in time bins of 1µs; this timeframe corresponds to the coherence time of the laser. The equipment is tuned so that the mean number of photons counted in each bin is below one.

The distributions of photon counts obtained from repeated runs of this experiment were used to train and test an ADALINE neuron and, for comparison, naive Bayes classifier. ADALINE is a single-layer neural network model based on a linear processing element, proposed by Bernard Widrow, for binary classification. It has no hidden layers, simply consisting of inputs and an output neuron.

The researchers also tested different machine learning methods: a) one-dimensional convolutional neural network (1D CNN) and b) a multilayer neural network (MNN). Interestingly, they found that these more sophisticated methods did not significantly affect the classification. They concluded that a simple ADALINE offers a perfect balance between accuracy and simplicity.

The team believe that their work has important implications for multiple photonic technologies, such as LiDAR and microscopy of biological materials. Using their method fewer measurements are needed for classification, enabling researchers to identify light sources much more quickly. In certain applications, such as microscopy, this means that they can limit light damage since they don’t have to illuminate the sample nearly as many times when taking measurements.

Read the research in full

Identification of light sources using machine learning
Chenglong You, Mario A. Quiroz-Juárez, Aidan Lambert, Narayan Bhusal, Chao Dong, Armando Perez-Leija, Amir Javaid, Roberto de J. León-Montiel, and Omar S. Magaña-Loaiza
Applied Physics Reviews (2020)

The work is also posted on arXiv.




Lucy Smith is Senior Managing Editor for AIhub.
Lucy Smith is Senior Managing Editor for AIhub.

            AIhub is supported by:



Subscribe to AIhub newsletter on substack



Related posts :

AI and Theory of Mind: an interview with Nitay Alon

  16 Mar 2026
Find out more about how Theory of Mind plays out in deceptive environments, multi-agents systems, the interdisciplinary nature of this field, when to use Theory of Mind, and when not to, and more.
coffee corner

AIhub coffee corner: AI, kids, and the future – “generation AI”

  13 Mar 2026
The AIhub coffee corner captures the musings of AI experts over a short conversation.

AI chatbots can effectively sway voters – in either direction

  12 Mar 2026
A short interaction with a chatbot can meaningfully shift a voter’s opinion about a presidential candidate or proposed policy.

Studying the properties of large language models: an interview with Maxime Meyer

  11 Mar 2026
What happens when you increase the prompt length in a LLM? In the latest interview in our AAAI Doctoral Consortium series, we sat down with Maxime, a PhD student in Singapore.

What the Moltbook experiment is teaching us about AI

An experimental social media platform where only AI bots can post reveals surprising lessons about artificial intelligence behaviour and safety.

The malleable mind: context accumulation drives LLM’s belief drift

  09 Mar 2026
LLMs change their "beliefs" over time, depending on the data they are given.

RWDS Big Questions: how do we balance innovation and regulation in the world of AI?

  06 Mar 2026
The panel explores the tensions, trade-offs and practical realities facing policymakers and data scientists alike.

Studying multiplicity: an interview with Prakhar Ganesh

  05 Mar 2026
What is multiplicity, and what implications does it have for fairness, privacy and interpretability in real-world systems?



AIhub is supported by:







Subscribe to AIhub newsletter on substack




 















©2026.02 - Association for the Understanding of Artificial Intelligence