As we head toward the end of the calendar year, a few items:
- Google published a new result in Nature a few days ago. This made a big news splash, including this accompanying press piece from google themselves, this nice article in Quanta, and the always thoughtful blog post by Scott Aaronson. The short version: Physical qubits as made today in the superconducting platform favored by google don't have the low error rates that you'd really like if you want to run general quantum algorithms on a quantum computer, which could certainly require millions of steps. The hope of the community is to get around this using quantum error correction, where some number of physical qubits are used to function as one "logical" qubit. If physical qubit error rates are sufficiently low, and these errors can be corrected with enough efficacy, the logical qubits can function better than the physical qubits, ideally being able to undergo a sequential operations indefinitely without degradation of their information. One technique for this is called a surface code. Google have implemented this in their most recent chip 105 physical qubit chip ("Willow"), and they seem to have crossed a huge threshold: When they increase the size of their correction scheme (going from a 3 (physical qubit) \(\times\) 3 (physical qubit) to 5 \(\times\) 5 to 7 \(\times\) 7), the error rates of the resulting logical qubits fall as hoped. This is a big deal, as it implies that larger chips, if they could be implemented, should scale toward the desired performance. This does not mean that general purpose quantum computers are just around the corner, but it's very encouraging. There are many severe engineering challenges still in place. For example, the present superconducting qubits must be tweaked and tuned. The reason google only has 105 of them on the Willow chip is not that they can't fit more - it's that they have to have wires and control capacity to tune and run them. A few thousand really good logical qubits would be needed to break RSA encryption, and there is no practical way to put millions of wires down a dilution refrigerator. Rather, one will need cryogenic control electronics.
- On a closely related point, google's article talks about how it would take a classical computer ten septillion years to do what its Willow chip can do. This is based on a very particularly chosen problem (as I mentioned here five years ago) called random circuit sampling, looking at the statistical properties of the outcome of applying random gate sequences to a quantum computer. From what I can tell, this is very different than what most people mean when they think of a problem to benchmark a quantum computer's advantage over a classical computer. I suspect the typical tech-literate person considering quantum computing wants to know, if I ask a quantum computer and a classical computer to factor huge numbers or do some optimization problem, how much faster is the quantum computer for a given size of problem? Random circuit sampling feels instead much more to me like comparing an experiment to a classical theory calculation. For a purely classical analog, consider putting an airfoil in a windtunnel and measuring turbulent flow, and comparing with a computational fluids calculation. Yes, the windtunnel can get you an answer very quickly, but it's not "doing" a calculation, from my perspective. This doesn't mean random circuit sampling is a poor benchmark, just that people should understand it's rather different from the kind of quantum/classical comparison they may envision.
- On one unrelated note: Thanks to a timey inquiry from a reader, I have now added a search bar to the top of the blog. (Just in time to capture the final decline of science blogging?)
- On a second unrelated note: I'd be curious to hear from my academic readers on how they are approaching generative AI, both on the instructional side (e.g., should we abandon traditional assignments and take-home exams? How do we check to see if students are really learning vs. becoming dependent on tools that have dubious reliability?) and on the research side (e.g., what level of generative AI tool use is acceptable in paper or proposal writing? What aspects of these tools are proving genuinely useful to PIs? To students? Clearly generative AI's ability to help with coding is very nice indeed!)
3 comments:
On the use of GenAI tools in research:
The adoption has come quicker than everyone thought it would be, yet I think it is just not being acknowledged openly. However, it seems great tool to increase ones productivity once you've figured out limits of each tool. Tools like Perplexity, paper-qa and github co-pilot have improved my ability to get things done.
I agree completely. I think it is bit frustrating that AI tools like Perplexity, Cursor and ChatGPT are seen as the enemy of serious science, when their output is indistinguishable from human researchers whose papers routinely contain mistakes (hallucinations), misquote references, and sometimes indulge in data manipulation, fabrication and concealment. It is like working with slide rules, because using calculators will blunt our superior intellect. Galactica, paper-qa and others have made it clear that factories of incremental research masquerading as serious science can be automated, for whatever it is worth. If we really want LLMs not to come up with random ideas and pretend they are true, we should hold ourselves to the same standards and write papers with the humility we seem to demand of LLMs, so that LLMs and graduate students trained on these papers imbibe the ability to put error bars on their creative ideas (aka hallucinations) and the tools to fact-check them.
Anon@11:59, can you say a little more about how you use Perplexity or paper-qa in ways that you find helpful? I've heard of Perplexity but I don't know what distinguishes it from the other LLM-based systems. I've never played with paper-qa. Can tools like that handle math (like, if I wanted to glean different expressions for some quantity from the literature), or are they entirely text-sensitive?
Anon@4:56, I think it all comes down to what people are able to learn and do. Knowing how to read the literature and learn from it has historically been a valuable skill; how much of that is retained or should be retained vs. being replaced with new skills (e.g. how to craft prompts to get the most out of LLM-based tools) is an open question.
Post a Comment