Foundations of Quantum Computing: I-Demystifying ‎‎Quantum Paradoxes

Abstract
Speedy developments in Quantum Technologies mandate that fundamentals of Quantum Computing are well explained and ‎understood. Meanwhile, paradigms of so-called quantum non-locality, wave function (WF) “collapse”, “Schrödinger cat” and ‎some other historically popular misconceptions continue to feed mysteries around quantum phenomena. Arguing that above ‎misinterpretations stem from classically minded and experimentally unverifiable perceptions, recasting Principle of ‎Superposition (PS) and key experimental details into classical notions. Revisiting main components of general quantum ‎measurement protocols (analyzers and detectors), and explaining paradoxes of WF collapse and Schrödinger cat. Reminding ‎that quantum measurements routinely reveal correlations dictated by conservation laws in each individual realization of the ‎quantum ensemble, manifesting “correlation-by-initial conditions” in contrast to traditional “correlation-by-interactions”. We ‎reiterate: Quantum Mechanics (QM) is not a dynamical theory in the same sense the Classical Mechanics (CM) is – it is a ‎statistical phenomenology, as established in 1926 by Born’s postulate. That is, while QM rests on conservation laws in each ‎individual outcome, it does not indicate how exactly a specific outcome is selected. This selection remains fundamentally ‎random and represents true randomness of QM, the latter being a statistical paradigm with a WF standing for a complex-‎valued distribution function. Finally, PS is the backbone of a quantum measurement process: PS can be conveniently viewed ‎as a composition of partial distributions into the total distribution – similar to classical probability mixtures – and is ‎effectuated experimentally by the analyzer part of a measuring device. ‎