[2603.09645] Noise in Photonic Quantum Machine Learning: Models, Impacts, and Mitigation Strategies

[2603.09645] Noise in Photonic Quantum Machine Learning: Models, Impacts, and Mitigation Strategies

arXiv - Machine Learning 4 min read

About this article

Abstract page for arXiv paper 2603.09645: Noise in Photonic Quantum Machine Learning: Models, Impacts, and Mitigation Strategies

Quantum Physics arXiv:2603.09645 (quant-ph) [Submitted on 10 Mar 2026 (v1), last revised 30 Mar 2026 (this version, v2)] Title:Noise in Photonic Quantum Machine Learning: Models, Impacts, and Mitigation Strategies Authors:A.M.A.S.D. Alagiyawanna, Asoka Karunananda View a PDF of the paper titled Noise in Photonic Quantum Machine Learning: Models, Impacts, and Mitigation Strategies, by A.M.A.S.D. Alagiyawanna and Asoka Karunananda View PDF HTML (experimental) Abstract:Photonic Quantum Machine Learning (PQML) is an emerging method to implement scalable, energy-efficient quantum information processing by combining photonic quantum computing technologies with machine learning techniques. The features of photonic technologies offer several benefits: room-temperature operation; fast (low delay) processing of signals; and the possibility of representing computations in high-dimensional (Hilbert) spaces. This makes photonic technologies a good candidate for the near-term development of quantum devices. However, noise is still a major limiting factor for the performance, reliability, and scalability of PQML implementations. This review provides a detailed and systematic analysis of the sources of noise that will affect PQML implementations. We will present an overview of the principal photonic quantum computer designs and summarize the many different types of quantum machine learning algorithms that have been successfully implemented using photonic quantum computer architectures suc...

Originally published on March 31, 2026. Curated by AI News.

Related Articles

Llms

Depth-first pruning seems to transfer from GPT-2 to Llama (unexpectedly well)

TL;DR: Removing the right transformer layers (instead of shrinking all layers) gives smaller, faster models with minimal quality loss — a...

Reddit - Artificial Intelligence · 1 min ·
Machine Learning

If frontier AI labs have unlimited shovels, what's stopping them from building everything?

I found myself explaining AI tokens to my mom over the weekend. At first I related them to building bricks: blocks of data the model uses...

Reddit - Artificial Intelligence · 1 min ·
[2603.16790] InCoder-32B: Code Foundation Model for Industrial Scenarios
Llms

[2603.16790] InCoder-32B: Code Foundation Model for Industrial Scenarios

Abstract page for arXiv paper 2603.16790: InCoder-32B: Code Foundation Model for Industrial Scenarios

arXiv - AI · 4 min ·
[2603.16430] EngGPT2: Sovereign, Efficient and Open Intelligence
Llms

[2603.16430] EngGPT2: Sovereign, Efficient and Open Intelligence

Abstract page for arXiv paper 2603.16430: EngGPT2: Sovereign, Efficient and Open Intelligence

arXiv - AI · 4 min ·
More in Machine Learning: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime