A new quantum computer can execute calculations in mere moments that would take the most advanced supercomputers 47 years to process.

  • DarkGamer@kbin.social
    link
    fedilink
    arrow-up
    4
    ·
    1 year ago

    Quantum computers aren’t faster because they “process” multiple “possibilities” at once.

    @hetscop I thought eigenstates/qbits were what made quantum computing faster and gave it its additional capabilities. For example, if you had an 8-bit integer represented by a bunch of qbits in a superposition of states, it would have every possible value from 0-256 and could be computed with as though it were every possible value at once until it is observed, the probability wave collapses, and a finite value emerges. Is this not the case?

    I read up a bit on Shor’s algorithm, and while I don’t fully understand it all it seems to take advantage of superposition of states destructively, which sounds a lot like “processing multiple possibilities at once.”

    When we input a superposition through our system and measure the remainder, we get a superposition of all possible powers that result in only that remainder. And this remaining superposition repeats with a period of the power [math didn’t paste nicely]

    Just as a classical Fourier Transform translates a wave as a function of time into a function of frequency, so too does a Quantum Fourier Transform (QFT), with a superposition as an output with a frequency of the input.

    So, if we input a single state into a QFT, the output will be a superposition of states with varying weights or probabilities that form a wave with the input state as the frequency. And if we input multiple states into a QFT, the output will be a superposition of superpositions - with destructive and constructive interference combining superpositions into one wave. And if our input to the QFT is a superposition with period [math didn’t paste nicely] source

    If my understanding is inaccurate, can you recommend a good source to understand this? Thanks!

    • ZickZack@kbin.social
      link
      fedilink
      arrow-up
      4
      ·
      1 year ago

      For example, if you had an 8-bit integer represented by a bunch of qbits in a superposition of states, it would have every possible value from 0-256 and could be computed with as though it were every possible value at once until it is observed, the probability wave collapses, and a finite value emerges. Is this not the case?

      Not really, or at least it’s not a good way of thinking about it. Imagine it more like rigging coin tosses: You don’t have every single configuration at the same time, but rather you have a joint probability over all bits which get altered to produce certain useful distributions.
      To get something out, you then make a measurement that returns the correct result with a certain probability (i.e. it’s a probabilistic turing machine rather than a nondeterministic one).

      This can be very useful since sampling from a distribution can sometimes be much nicer than actually solving a problem (e.g. you replace a solver with a simulator of the output).
      In traditional computing this can also be done but that gives you the fundamental problem of sampling from very complex probability distributions which involves approximating usually intractable integrals.

      However, there are also massive limitations to the type of things a quantum computer can model in this way since quantum theory is inherently linear (i.e. no climate modelling regardless of how often people claim they want to do it).
      There’s also the question of how many things exist where it is more efficient to build such a distribution and sample from it, rather than having a direct solver.
      If you look at the classic quantum algorithms (e.g. https://en.wikipedia.org/wiki/Quantum_algorithm), you can see that there aren’t really that many algorithms out there (this is of course not an exhaustive list but it gives a pretty good overview) where it makes sense to use quantum computing and pretty much all of them are asymptotically barely faster or the same speed as classical ones and most of them rely on the fact that the problem you are looking at is a black-box one.

      Remember that one of the largest useful problems that was ever solved on a quantum computer up until now was factoring the number 21 with a specialised version of Shor’s algorithm that only works for that number (since the full shor would need many orders of magnitude more qbits than exist on the entire planet).

      There’s also the problem of logical vs physical qbits: In computer science we like to work with “perfect” qbits that are mathematically ideal, i.e. are completely noise free. However, physical qbits are really fragile and attenuate to pretty much anything and everything, which adds a lot of noise into the system. This problem also gets worse the larger you scale your system.

      The latter is a fundamental problem: the entire clue of quantum computers is that you can combine random states to “virtually” build a complex distribution before you sample from it. This can be much faster since the virtual model can look dependencies that are intractable to work with on a classical system, but that dependency monster also means that any noise in the system is going to negatively affect everything else as you scale up to more qbits.
      That’s why people expect real quantum computers to have many orders of magnitude more qbits than you would theoretically need.

      It also means that you cannot trivially scale up a physical quantum algorithm: Physical grovers on a list with 10 entries might look very different than a physical grover with 11 entries.
      This makes quantum computing a nonstarter for many problems where you cannot pay the time it takes to engineer a custom solution.
      And even worse: you cannot even test whether your fancy new algorithm works in a simulator, since the stuff you are trying to simulate is specifically the intractable quantum noise (something which, ironically, a quantum computer is excellent at simulating).

      In general you should be really careful when looking at quantum computing articles, since it’s very easy to build some weird distribution that is basically impossible for a normal computer to work with, but that doesn’t mean it’s something practical e.g. just starting the quantum computer, “boop” one bit, then waiting for 3ns will give you a quantum noise distribution that is intractable to simulate with a computer (same thing is true if you don’t do anything with a computer: there’s literal research teams of top scientists whose job boils down to “what are quantum computers computing if we don’t give them instructions”).

      Meanwhile, the progress of classical or e.g. hybrid analog computing is much faster than that of quantum computing, which means that the only people really deeply invested into quantum computing are the ones that cannot afford to miss, just in case there is in fact something:

      • finance
      • defence
      • security
    • hetscop@kbin.social
      link
      fedilink
      arrow-up
      2
      ·
      1 year ago

      (typed this out yesterday before @ZickZack s excellent answer, but couldn’t post it at the time due to maintenance…)

      No, you’ve got it wrong. This is a fairly common missunderstanding which is perpetuated by a lot of coverage about the topic being sloppy.

      You could argue that there is a grain of truth to the idea of processing multiple possibilities at once, but it’s a bit more complicated than that and the way it’s usually presented leads to people building a bad intuition of how it works. If you do get in to the nitty-gritty of Shors algorithm it feels to me at least a bit like a weird hack that shouldn’t work at all or at least not be faster than the normal way to compute prime factors. It isn’t a general speedup, just in certain cases where you can exploit quantum mechanics in clever ways.

      Of the top of my head the SMBC comic about it is actually pretty good. This article makes basically the same points, but a bit more elaborated (note that it was written a while ago so the part about the current state of quantum computing is outdated). I noticed that Veritasum put out a YouTube video which I haven’t watched, but he is in my experience good at explaining physics and math so I think that there’s a good chance that it’ll hold up. I remember liking this Minute Physics video about Shor’s algorithm too, if you wanna get a better understanding of it.

      I should clarify that I’m not a quantum phycisist, I’ve just done a couple of internet deep dives on the topic but I can’t say that I fully understand quantum computing at all. I do think my understanding of it is better than the one in this article and others like it.