Introduction to Quantum Computing: From a Layperson to a Programmer in 30 Steps

Publication Date


Document Type





This textbook introduces quantum computing to readers who do not have much background in linear algebra. The author targets undergraduate and master students, as well as non-CS and non-EE students who are willing to spend about 60 -90 hours seriously learning quantum computing. Readers will be able to write their program to simulate quantum computing algorithms and run on real quantum computers on IBM-Q. Moreover, unlike the books that only give superficial, “hand-waving” explanations, this book uses exact formalism so readers can continue to pursue more advanced topics based on what they learn from this book.

  • Encourages students to embrace uncertainty over the daily classical experience, when encountering quantum phenomena;
  • Uses narrative to start each section with analogies that help students to grasp the critical concept quickly;
  • Uses numerical substitutions, accompanied by Python programming and IBM-Q quantum computer programming, as examples in teaching all critical concepts.


Electrical Engineering