Quantum Computing Basics and the Most Important Algorithms
Quantum mechanics provides the basis for a new computing paradigm. In the 1990s Shor's algorithm which exponentially speeds up classical integer factorization demonstrated the theoretical power of quantum computers. Recently, the development of quantum computers is making great progress and we need to learn how we can use the advantage of quantum computers.
In this course you will learn the basics of quantum computing and the most important algorithms of quantum computing.
Quantum Computing technology has advanced greatly in recent years. In the past 3 decades alone, new technologies have brought about a huge leap in Quantum Computation, from running the first ever quantum algorithm back in 1992 using Nuclear Magnetic Resonance(NMR), to running Quantum Neural Networks nowadays on a 64-Qbit Quantum Processing Unit (QPU) via Cloud. The technological advancement has accelerated greatly, more significantly over the past 2 years marked by the first announcement of a potential Quantum Supremacy. Moreover, Quantum Hardware developers have revealed their roadmap to reach a 1000 Qbit machine hinting that at an exponential rate, a million Qbit machine might be expected.
The emerging technology of Quantum Computation is bringing together computer scientists, physicists, chemists, statisticians and a broad range of engineers from the STEM (Science, Technology, Engineering and Mathematics) fields in a great effort to reach production level performance. Additionally, the high impact potential of this technology may bring revolutionary advancements over the whole spectrum of human activity.