Quantum computing explained in terms of elementary linear algebra, emphasizing computation and algorithms and requiring no background in physics.
This introduction to quantum algorithms is concise but comprehensive, covering many key algorithms. It is mathematically rigorous but requires minimal background and assumes no knowledge of quantum theory or quantum mechanics. The book explains quantum computation in terms of elementary linear algebra; it assumes the reader will have some familiarity with vectors, matrices, and their basic properties, but offers a review of the relevant material from linear algebra. By emphasizing computation and algorithms rather than physics, it makes quantum algorithms accessible to students and researchers in computer science who have not taken courses in quantum physics or delved into fine details of quantum effects, apparatus, circuits, or theory.
by Kenneth Grahame
by Charles Darwin
by Henry David Thoreau
by Charles W. Chesnutt
by W. Somerset Maugham
by Richard J. Borden, Darron Collins
by Kenneth King
by J. David Henry
by John J. McKay
by Philippe J. Dubois, Elise Rousseau
by Zane Grey
by Charles Dickens
Sign up for our email newsletter