Category Archives: Quantum Computing
3 Quantum Computing Stocks to Buy Be Millionaire-Makers: May – InvestorPlace
Source: Bartlomiej K. Wroblewski / Shutterstock.com
Dont miss out on this exceptional chance to invest in quantum computing stocks to buy that could be millionaire makers while their valuations remain low. These innovative tech companies are developing cutting-edge quantum computing systems with the potential to generate massive returns for investors who get in early.
The quantum computing stocks featured below are poised to commercialize their technology across multiple industries. Quantum computing promises to transform various sectors of our world, from financial services to medical research. Also, it may enable groundbreaking advances and discoveries that arent possible with traditional classical computing.
The three quantum computing stocks to buy outlined in this article represent the best opportunities investors have to compound their wealth to seven figures. Weve only just started to see the potential of this industry and understand the implications of this new tech.
So, here are three quantum computing stocks for investors who want to earn a potential seven-figure sum.
Source: zakiahza / Shutterstock.com
Hewlett Packard Enterprise (NYSE:HPE) focuses on IT and quantum computing through its Intelligent Edge segment. The company has demonstrated significant achievements in quantum computing research.
HPEs Intelligent Edge segment provides solutions that bring computation closer to the data source. Integrating quantum computing capabilities with Intelligent Edge technologies can offer unique advantages, such as real-time data processing and enhanced decision-making capabilities at the networks edge.
Most recently, the Intelligent Edge segment reported revenue of $902 million, an increase of 9% year-over-year. This segment continues to grow, driven by strong demand for edge computing solutions. The company also achieved an EPS of $0.48, which surpassed the consensus estimate of $0.45. This compares to an EPS of $0.63 in the same quarter of the previous year.
HPE is a well-known brand akin to a more modern version of IBM (NYSE:IBM). It could be a good pick for those who like to stay with the blue-chip options while also having the potential to mint new millionaires.
IonQ (NYSE:IONQ) is a leader in developing trapped-ion quantum computers and making significant strides in the field. The company collaborates with major cloud platforms.
IonQs primary technology involves trapped-ion quantum computers, which utilize ions trapped in electromagnetic fields as qubits. This technology is known for its high-fidelity operations and stability.
Recently, IonQ achieved a milestone of 35 algorithmic qubits with its IonQ Forte system, a year ahead of schedule. This achievement allows the system to handle more sophisticated and more extensive quantum circuits. IonQs growth and technological advancements have been recognized in various industry lists, such as Fast Companys 2023 Next Big Things in Tech List and Deloittes 2023 Technology Fast 500 List.
With a market cap of just 1.79 billion, it remains a small-cap quantum computing stock that could hold significant upside potential for investors. Its developments so far have been promising, and it could prove to be a company that will make early investors rich.
Pure-play quantum computing company Rigetti Computing (NASDAQ:RGTI) is known for its vertically integrated approach. This includes designing and manufacturing quantum processors.
Rigetti has achieved a significant milestone with its 128-qubit chip, which promises to advance quantum computing capabilities and enable new applications. This development is a key part of Rigettis roadmap to scale up quantum systems and improve performance metrics.
Also, in Q1 2024, Rigetti reported a 99.3% median 2-qubit gate fidelity on its 9-qubit Ankaa-class processor. This high level of fidelity is crucial for reliable quantum computations and positions Rigetti well against competitors.
The market cap of RGTI is a fraction of IONQs at just under 200 million at the time of writing. Its progress is similarly impressive, so it could hold significant upside and potentially mint a new generation of millionaires with a large enough investment.
On the date of publication, Matthew Farley did not have (either directly or indirectly) any positions in the securities mentioned in this article. The opinions expressed are those of the writer, subject to the InvestorPlace.com Publishing Guidelines.
Matthew started writing coverage of the financial markets during the crypto boom of 2017 and was also a team member of several fintech startups. He then started writing about Australian and U.S. equities for various publications. His work has appeared in MarketBeat, FXStreet, Cryptoslate, Seeking Alpha, and the New Scientist magazine, among others.
See the original post:
3 Quantum Computing Stocks to Buy Be Millionaire-Makers: May - InvestorPlace
Amazon taps Finland’s IQM for its first EU quantum computing service – TNW
IQM Garnet, a 20-qubit quantum processing unit (QPU) is now available via Amazon Web Services (AWS) the first quantum computer accessible via AWS cloud in the European Union.
Finnish quantum hardware startup IQM is based outside of Helsinki, Finland. AWS previously has collaborations in place with IonQ, Oxford Quantum Circuits, QuEra, and Rigetti for its quantum cloud service known as Braket, but this will be the first AWS quantum processor hosted within the EU.
This also means that it is the first time Amazons quantum services will be accessible to end users in its AWS Europe (Stockholm) Region. It is also the first time IQMs quantum computers will be available in an on-demand structure via the cloud, and with AWS pay-as-you-go pricing.
We are very honoured to be part of the Amazon network and work together with a global tech company, Jan Goetz, co-CEO and co-founder at IQM told TNW. For IQM, this is a great opportunity to scale our offering globally and collaborate with leading end-users around the world.
Goetz further added that the joint offering was a great step forward for cloud quantum computing, and would enable cloud users to test novel types of algorithms and use-cases to develop their business.
As most of our readers will probably know, todays noisy and error-prone quantum computers cannot really do all that much yet. However, the technology is currently advancing incredibly fast. Learning to work with it will not happen overnight.
As such a whole business model has sprung up around getting organisations and corporations quantum-ready, so that they wont be caught off guard when quantum utility arrives. Todays smaller qubit systems are also training grounds for software developers, many of whom are working on solving the issue of error correction. In the context of cloud, IQM Garnet is mostly used by quantum algorithm engineers to develop IP around quantum compilers, algorithms, and error correction schemes, Max Haeberlein, Head of Cloud at IQM told TNW. IQM Garnet offers a highly homogenous layout and has cutting-edge fidelities, allowing users to effectively expand algorithms to the full size of the chip.
At the same time, Haeberlein said, the company offers IQM Garnet at affordable rates, which is especially important for the growing quantum algorithm startup scene.
IQM, founded in 2018, is Europes leading quantum hardware developer in superconducting qubits. In the beginning of next year the company plans to add a high-fidelity IQM Radiance 54-qubit quantum computer to its portfolio.
This, according to Haeberlein, will enable users to extend quantum algorithms beyond the point where they can still be classically emulated by supercomputers. In 2026, we release IQM Radiance with 150 qubits, where we will see the first commercial algorithm applications of scale in the domain of finance, automotive, life sciences, and chemicals, he adds.
Read the original here:
Amazon taps Finland's IQM for its first EU quantum computing service - TNW
The 2025 Millionaire’s Club: 3 Quantum Computing Stocks to Buy Now – InvestorPlace
The current rage is about artificial intelligence, but the advancement of the AI field relies on a few key elements including quantum computing
Source: Faces Portrait / Shutterstock.com
If youre on the hunt for quantum computing stocks to buy, youre in the right place. For the past several years, artificial intelligence (AI) has taken the front stage. Notjust in the techfield,but also in the stock market.AI advancement has been tremendous, allowingbusinesses, bothlarge and small,to automate some of their processes.Some of the largest companies have ramped uptheirinvesting in AI teams and divisions, amounting to billions of dollars in additional capital justto keep up with others in the field.
However,there is a newcomer to the field that is independent of AI but will complement it in the future, which is quantum computing. But what is it exactly?Quantum computing uses specialized algorithms and hardware while using quantum mechanics to solve complex problems that typical computerswill takeeithertoo long to solveorcannot solve entirely.Although the world of quantum computing and AI is incredibly complex, we have simplified investing in the field by narrowing it down to 3 companiesthat areat the forefront of the industry, while still being diversified.
Source: Piotr Swat / Shutterstock.com
Nvidia (NASDAQ:NVDA) is an American-based and international leader in the designing and manufacturing of graphics processing units. Although the companys main focus currently is on the AI market, it also has a division that focuses on the quantum computing industry. The stock currently trades at about $924, with a price target of $1,100. This new price target is almost $200 more than the current trading price of a stock, signifying a significant upside potential for Nvidia.
The company accelerates quantum computing centers around the world with its proprietary CUDA-Q platform. The platform also ties quantum computing into AI, allowing the system to solve new and countless problems much faster than before.
The stock currently trades at 36.53x forward earnings. This is about 20% lower in comparison to the stocks own five-year average forward price to earnings (P/E) ratio of 46.14x. Thus, considering what the stock usually trades for, it might be relatively undervalued and at a great point to scoop up some shares.
The company that goes hand in hand with the internet is Alphabet (NASDAQ:GOOG, NASDAQ:GOOGL). The American company first started as a search engine in the late 90s with its main goal of creating the perfect search engine. Fast forward 25 years and you now have a multi-trillion-dollar international company with departments in tech, consumer electronics, data, AI, e-commerce and quantum computing. The companys stock currently trades at about $177 but is on track to rise to an average of $195, with a high of $225 in the next 12 months.
In recent years, it has set out to build the best quantum computing for otherwise impossible problems with the introduction of XPRIZE Quantum Applications and Quantum AI. The program is designed to advance the field of algorithms relating to quantum computing with real-world applications.
As such, the company is in a quickly growing phase, and EPS is forecast to soar from $5.80 last year to over $7.84 by 2025. This makes it a great pick for any investor.
Intel (NYSE:INTC) has specialized in semiconductors since its founding in Mountain View, California in 1968. The company is the worlds largest manufacturer of semiconductors and CPUs and has been since its founding. Intels stock is at about $32 and the average price target is $39.63, with a low of $17 and a high of $68. This would mean an upside potential of almost 24%, on average.
The company has invested heavily in quantum computing in the past several years and is currently putting its expertise to good use, creating hot silicon spin qubits. Qubits are essentially small computing devices that perform differently than typical transistors while also operating at high temperatures.
The company is working diligently on applying the qubits into quantum computing chips that can be used to advance countless fields, while also working with AI systems. All this work is not for nothing. The company is translating this into earnings growth with EPS expected to rise from $1.01 at the end of this year to $1.80 by the end of 2025. As such, this stock should be on any investors watch list.
On the date of publication, Ian Hartana and Vayun Chugh did not hold (either directly or indirectly) any positions in the securities mentioned in this article. The opinions expressed in this article are those of the writer, subject to the InvestorPlace.com Publishing Guidelines.
Chandler Capital is the work of Ian Hartana and Vayun Chugh. Ian Hartana and Vayun Chugh are both self-taught investors whose work has been featured in Seeking Alpha. Their research primarily revolves around GARP stocks with a long-term investment perspective encompassing diverse sectors such as technology, energy, and healthcare.
View original post here:
The 2025 Millionaire's Club: 3 Quantum Computing Stocks to Buy Now - InvestorPlace
The power of Quantum Computing – The Cryptonomist
One of the exponential technologies that is not yet getting its fair share of love from the general public and media is Quantum computing. In the past few years, I had the privilege of spending time discussing it with people from CERN and the Fermi Lab, but my conversation with Scott Crowder, Vice President IBM Quantum Adoption and Business Development, had the right mix of theory and real-life examples, which will make anyone understand the potential of this field of research and its business applications. AI will keep its hype for a good while, as we see from its pervasive presence in every corner of the internet. Quantum can be the next big thing. This is our dialogue.
Who are you and what do you do for a living?
My name is Scott Crowder, and I run IBMs Quantum efforts to boost its adoption, together with our partners and industry clients. Our goal is to build a useful Quantum computing infrastructure and to help the world make a Quantum-safe transition, in the next ten years or so. I am an engineer by training and had worked on semi-conductors in the past, before taking on the role of CTO for IBM Systems. With Quantum, its the first time where we have a use first attitude, where we try things with partners, we teach and learn with our clients, before we scale-up projects. Its interesting and its fun.
What are the three killer use cases for Quantum, for what we know now?
Firstly, simulating nature, like materials science new materials, or chemistry, for example better battery chemistry, to mention something that is very hot right now. We do physics simulations or try to understand how complex proteins would behave. These are operations that entail higher computing power than what we could do with todays computers.
Secondly, we try to find patterns out of complex data. For example, a classification of a piece of data as fraud or not. If there is some structure in the data before us, Quantum computing is much better than classical computers to give meaning to it and even pick up things like false positives. This is extremely useful, if we want to make sense of the world.
Lastly, I would say, portfolio optimization, finding efficiencies, and distribution optimization. There are direct and huge applications here, for multiple industries. Think of the mobility or logistics markets, for example. This third use case is slightly farther out from us, in terms of time to market, when compared to the first two.
Where are we really, when it comes to Quantum adoption in the real world?
To simplify it: Quantum is better at doing what it does best, namely simulations. For sure, to do it at scale, larger systems are needed. So, we are looking at 2030 and beyond. What we are doing now is, lets say, algorithmic explorations. We work with a mix of partners: heavy industry conglomerates, banking, pharma, transportation, and startups. And, obviously, universities and research institutions.
Big Tech is also into Quantum, even though the talk of the town is AI. Intel, Microsoft, Google, AWS: all have investments and programs in Quantum, with different approaches to it.
What is the future business model of Quantum? How are you going to sell it?
Its hard to say right now. We must make some assumptions. Its probably going to continue to be, in the medium term, a cloud service, where partners have access to the Quantum capabilities we have built, via API calls, and they can interact with our experts, who help with the prototyping and the training. Basically, its going to be the same as a standard cloud business model. There will be ad hoc projects for sure, where the stakes are high, and we can unlock tremendous economic value. In a way, the approach is more like how we weave CPUs and GPUs into a compute fabric, and not via a single application per se, like a Chat GPT for Quantum.
What would you say is the number one risk associated with Quantum?
Cybersecurity is for sure the number one risk. Future, more powerful Quantum computers will crack at some point the current asymmetric cryptography, which protects public and private information, for example (mobile data, payments, medical records, etc). The math for that already exists. There are Quantum-safe cryptography solutions, but a full ecosystem of security providers and coding will need to change, to account for the Quantum shift, and to make sure we have a Quantum safe era.
Where can we find you and learn more about Quantum?
A simple search for anything related to IBM Quantum will do. I am also active on social media, like LinkedIn. IBM writes a lot of articles on Quantum. We need to talk about it publicly, and have people understand this is real, and it has great potential to bring tremendous value to society and business, across all industries. You may think this is science fiction, as its going to hit us in our next decade, but it is a new way of approaching complex problems. It could help other applications and use cases, as well, like AI, and this is why its the right moment to talk Quantum.
Read this article:
The power of Quantum Computing - The Cryptonomist
Aramco signs agreement with Pasqal to deploy first quantum computer in the Kingdom of Saudi Arabia – Aramco
Aramco, one of the worlds leading integrated energy and chemicals companies, has signed an agreement with Pasqal, a global leader in neutral atom quantum computing, to install the first quantum computer in the Kingdom of Saudi Arabia.
The agreement will see Pasqal install, maintain, and operate a 200-qubit quantum computer, which is scheduled for deployment in the second half of 2025.
Ahmad Al-Khowaiter, Aramco EVP of Technology & Innovation, said: Aramco is delighted to partner with Pasqal to bring cutting-edge, high-performance quantum computing capabilities to the Kingdom. In a rapidly evolving digital landscape, we believe it is crucial to seize opportunities presented by new, impactful technologies and we aim to pioneer the use of quantum computing in the energy sector. Our agreement with Pasqal allows us to harness the expertise of a leading player in this field, as we continue to build state-of-the-art solutions into our business. It is also further evidence of our contribution to the growth of the digital economy in Saudi Arabia.
Georges-Olivier Reymond, Pasqal CEO & Co-founder, said: The era of quantum computing is here. No longer confined to theory, it's transitioning to real-world applications, empowering organisations to solve previously intractable problems at scale. Since launching Pasqal in 2019, we have directed our efforts towards concrete quantum computing algorithms immediately applicable to customer use cases. Through this agreement, we'll be at the forefront of accelerating commercial adoption of this transformative technology in Saudi Arabia. This isn't just any quantum computer; it will be the most powerful tool deployed for industrial usages, unlocking a new era of innovation for businesses and society.
The quantum computer will initially use an approach called analog mode. Within the following year, the system will be upgraded to a more advanced hybrid analog-digital mode, which is more powerful and able to solve even more complex problems.
Pasqal and Aramco intend to leverage the quantum computer to identify new use cases, and have an ambitious vision to establish a powerhouse for quantum research within Saudi Arabia. This would involve leading academic institutions with the aim of fostering breakthroughs in quantum algorithm development a crucial step for unlocking the true potential of quantum computing.
The agreement also accelerates Pasqal's activity in Saudi Arabia, having established an office in the Kingdom in 2023, and follows the signing of a Memorandum of Understanding between the companies in 2022 to collaborate on quantum computing capabilities and applications in the energy sector. In 2023, Aramco's Wa'ed Ventures also participated in Pasqal's Series B fundraising round.
View original post here:
Aramco signs agreement with Pasqal to deploy first quantum computer in the Kingdom of Saudi Arabia - Aramco
How Nvidia co-founder plans to turn Hudson Valley into a tech powerhouse greater than Silicon Valley – New York Post
A co-founder of chip maker Nvidia is bankrolling a futuristic quantum computer system at Rensselaer Polytechnic Institute and wants to turn New Yorks Hudson Valley into a tech powerhouse.
Curtis Priem, 64, donated more than $75 million so that the Albany-area college could obtain the IBM-made computer the first such device on a university campus anywhere in the world, the Wall Street Journal reported.
The former tech executive and RPI alum said his goal is to establish the area around the school, based in Troy, into a hub of talent and business as quantum computing becomes more mainstream in the years ahead.
Weve renamed Hudson Valley as Quantum Valley, Priem told the Journal. Its up to New York whether they want to become Silicon State not just a valley.
The burgeoning technology uses subatomic quantum bits, or qubits, to process data much faster than conventional binary computers. The devices are expected to play a key role in the development of advanced AI systems.
Priem will reportedly fund the whopping $15 million per year required to rent the computer, which is kept in a building that used to be a chapel on RPIs campus.
RPI PresidentMartin Schmidt told the newspaper that the school will begin integrating the device into its curriculum and ensure it is accessible to the student body.
Representatives for IBM and RPI did not immediately return The Posts request for comment.
An electrical engineer by trade, Priem co-founded Nvidia alongside its current CEO Jensen Huang and Chris Malachowsky in 1993. He served as the companys chief technology officer until retiring in 2003.
Priem sold most of his stock in retirement and used the money to start a charitable foundation.
He serves as vice chair of the board at RPI and has reportedly donated hundreds of millions of dollars to the university.
Nvidia has surged in value as various tech firms rely on its computer chips to fuel the race to develop artificial intelligence.
The companys stock has surged 95% to nearly $942 per share since January alone. Nvidias market cap exceeds $2.3 trillion, making it the worlds third-most valuable company behind Microsoft and Apple.
In November 2023, Forbes estimated that Priem would be one of the worlds richest people, with a personal fortune of $70 billion, if he hadnt sold off most of his Nvidia shares.
ISC 2024 A Few Quantum Gems and Slides from a Packed QC Agenda – HPCwire
If you were looking for quantum computing content, ISC 2024 was a good place to be last week there were around 20 quantum computing related sessions. QC even earned a slide in Kathy Yelicks opening keynote Beyond Exascale. Many of the quantum sessions (and, of course, others) were video-recorded and ISC has now made them freely accessble.
Not all were recorded. For example what sounded like a tantalizing BOF panel Toward Hardware Agnostic Standards in Hybrid HPC/Quantum Computing featuring Bill Gropp (NCSA, University of Illinois), Philippe Deniel (Commissariat Energie Atomique (CEA)), Mitsuhisa Sato (RIKEN), Travis Humble (ORNL), Venkatesh Kannan (Irelands High Performance Centre), and Kristel Michielsen (Julich Supercomputing Center). Was sorry to miss that.
Regardless, theres a wealth of material online and its worth looking through the ISC 2024 inventory for subjects, speakers, and companies of interest (registration may be required). Compiled below are a few QC soundbites from ISC.
Yelick, vice chancellor for research at the University of California, covered a lot of ground in her keynote examining the tension and opportunities emerging from the clash of traditional FP64 HPC and mixed-precision AI and how the commercial supply line of advanced chips is changing. Quantum computing earned a much smaller slice.
I really just have this one slide about quantum. Theres been some really exciting progress if you have been following this and things like error correction over the last year with really, significant improvements in terms of the ability to build error corrected quantum systems. On the other hand, I would say we dont yet have an integrated circuit kind of transistor model yet, right. Weve got a bunch of transistors, [i.e.] weve got a whole bunch of different kinds of qubits that you can build, [and] theres still some debate [over them].
In fact, the latest one of the latest big error correction results was actually not for the superconducting qubits, which is what a lot of the early startups were in, but for the AMO (atomic, molecular, optical) physics. So this is really looking at the fact that were not yet at a place where we can rely on this for the next generation of computing, which is not to say that we should be ignoring it. Im really interested to see how [quantum computing evolves and] also thinking about how much classical computing were going to need with quantum because thats also going to be a big challenge with quantum. [Its] very exciting, but its not replacing also general purpose kind of computing that we do for science and engineering.
Not sure if thats a glass half-full or half-empty perspective. Actually, many of the remaining sessions tackled the questions she posed, including the best way to implement hyrbid HPC-Quantum system, error correction and error mitigation, and the jostling among competing qubit types.
It was easy to sympathize (sort of) with speakers presenting at the Quantum Computing Status of Technologies session, moderated by Valeria Bartsch of Fraunhofer CFL. The speakers came from companies developing different qubit modalities and, naturally, at least a small portion of their brief talks touted their company technology.
She asked, Heres another [submitted question]. What is the most promising quantum computing technology that your company is not developing yourself? I love that one. And everybody has to answer it now. You can think for a few seconds.
Very broadly speaking neutral atom, trapped ion, and superconducting are perhaps the most advanced qubit modalities currently and each speaker presented a bit of background on their companies technology and progress. Trapped ions boast long coherence times but somewhat slower swicthing speeds. Superconducting qubits are fast, and perhaps easier to scale, but error prone. Neutral atoms also have long coherence times but have so far been mostly used for analog computing though efforts are moving quickly to implement gate-based computing. To Hayes point, Marjorana (topology) qubits would be inherently resistant to error.
Not officially part of the ISC program, Hyperion delivered its mid-year HPC market update online just before the conference. The full HPCwire coverage is here and Hyperion said it planned to put its recorded presentation and slides available on its website. Chief Quantum Analyst Bob Sorensen provided a brief QC snapshot during the update predicting the WW QC market will surpass $1 billion in 2025.
Sorensen noted, So this is a quick chart (above) that just shows the combination of the last four estimates that we made, you can see starting in 2019, all the way up to this 2023 estimate that reaches that $1.5 billion in 2026 I talked about earlier. Now my concern here is always its dangerous to project out too far. So we do tend to limit the forecast to these kinds of short ranges, simply because a nascent sector like quantum, which has so much potential, but at the same time has some significant technical hurdles to overcome [which] means that there can be an inflection point most likely though in the upward direction.
He also pointed out that a new use case, a new breakthrough in modality or algorithms, any kind of significant driver that brings more interest in and performance to quantum kick can significantly change the trajectory here on the upside.
Sorensen said, Just to give you a sense of how these vendors that we spoke to looked at algorithms, we see the big three are still the big three in mod-sim, optimization, and AI with with some interest in cybersecurity aspects, post quantum encryption kinds of research and such as well as Monte Carlo processes taking advantage of quantum stability to generate random number generator, provable random numbers to support the Monte Carlo processing.
Interesting here is that were seeing a lot more other (17%). This is the first time weve seen that. We think it is [not so much] about new algorithms, but perhaps hybrid mod-sim optimized or machine learning that feeds into the optimization process. So we think were seeing more hybrid applications emerging as people take a look at the algorithms and decide what solves the use case that they have in hand, he said.
Satoshi Matsuoka, director of RIKEN Center for Computational Science, provided a quick overview of Fugaku plans for incorporating quantum computing as well as touching on the status of the ABCI-Q project. He, of course, has been instrumental with both systems. Both efforts emphasize creating a hybrid HPC-AI-Quantum infrastructure.
The ABCI-Q infrastructure (slide below) will be a variety of quantum-inspired and actual quantum hardware. Fujitsu will supply the former systems. Currently, quantum computers based on neutral atoms, superconducting qubits, and photonics are planned. Matsuoka noted this is well-funded a few $100 million with much of the work done geared toward industry.
Rollout of the integrated quantum-HPC hybrid infrastructure at Fugaku is aimed at the 2024/25 timeframe. Its also an ambitious effort.
About the Fugaku effort, Matsuoka said, [This] project is funded by a different ministry, in which we have several real quantum computers, IBMs Heron (superconducting QPU), a Quantinuum (trapped ion qubits), and quantum simulators. So real quantum computers and simulators to be coupled with Fugaku.
The objective of the project [is to] come up with a comprehensive software stack, such that when the real quantum computers that are more useful come online, then we can move the entire infrastructure along with any of those with quantum computers along with their successors to be deployed to solve real problems. This will be one of the largest hybrid supercomputers.
The aggressive quantum-HPC integration sounds a lot like what going on in Europe. (See HPCwire coverage, Europes Race towards Quantum-HPC Integration and Quantum Advantage)
The topic of benchmarking also came up during Q&A at one session. A single metric such as the Top500 is generally not preferred. But what then, even now during the so-called NISQ (noisy intermediate-scale quantum) computing era?
One questioner said, Lets say interesting algorithms and problems. Is there anything like, and Im not talking about a top 500 list for quantum computers, like an algorithm where we can compare systems? For example, Shors algorithm. So who did it and what is the best performance or the largest numbers you were able to factorize?
Hayes (Quantinuum) said, So we havent attempted to run Shors algorithm, and interesting implementations of Shors algorithm are going to require fault tolerance to factor a number that a classical computer cant. But you know, that doesnt mean it cant be a nice benchmark to see which company can factor the largest one. I did show some data on the quantum Fourier transform. Thats a primitive in Shors algorithm. I would say that thatd be a great candidate for benchmarking the progress and fault tolerance.
More interesting benchmarks for the NISC era are things like quantum volume, and theres some other ones that can be standardized, and you can make fair comparisons. So we try to do that. You know, theyre not widely or universally adopted, but there are organizations out there trying to standardize them. Its difficult getting everybody marching in the same direction.
Corcoles (IBM) added, I think benchmarking in quantum has an entire community around it, and they have been working on it for more than a decade. I read your question as focusing on application-oriented benchmarks versus system-oriented benchmarks. There are layers of subtlety there as well. If we think about Shors algorithm, for example, there were recent works last year suggesting theres more than one way to run Shors. Depending on the architecture, you might choose one or another way.
An architecture that is faster might choose to run many circuits in parallel that can capture Shors algorithm and then do a couple of processing or architecture that that might might take more time they just want to run one single circuit with high probability measure the right action. You could compare run times, but theres probably going to be differences that add to the uncertainty of what what technology you will use, meaning that there might be a regime of factoring, where you might want to choose one aspect or another, but then your particular physical implement, he said.
Macri (QuEra) said, My point is were not yet at the point where we can really [compare systems]. You know we dont want to compete directly with our technologies. I would say that especially in for what concerns applications we need to adopt a collaborative approach. So for example, there are certain areas where these benchmarks that you mentioned are not really applicable. One of them is a quantum simulation and we have seen really a lot of fantastic results from our technology, as well as from ion traps and superconducting qubits.
It doesnt really make sense really to compare the basic features of the technologies so that, you know, we can a priori, identify what is the specific application the result that you want to achieve. I would say lets focus on advancing the technology we see. We already know that there are certain types of devices that outperform others for specific applications. And then we will, we will decide these perhaps at a later stage. But I agreed for for very complex tasks, such as quantum Fourier transform, or perhaps the Shors algorithm, but I think, to be honest, its still too preliminary [for effective system comparisons].
As noted this was a break-out year for quantum at ISC which has long had quantum sessions but not as many. Europes aggressive funding, procurements, and HPC-quantum integration efforts make it clear it does not intend to be left behind in the quantum computing land rush, with, hopefully, a gold rush to follow.
Stay tuned.
Go here to see the original:
ISC 2024 A Few Quantum Gems and Slides from a Packed QC Agenda - HPCwire
Alice & Bob’s Cat Qubit Research Published in Nature – HPCwire
PARIS and BOSTON, May 23, 2024 Alice & Bob, a global leader in the race for fault-tolerant quantum computing, today announced the publication of its foundational research in Nature, showcasing significant advancements in cat qubit technology.
The study, Quantum control of a cat-qubit with bit-flip times exceeding ten seconds, realized in collaboration with the QUANTIC Team (Mines Paris PSL, Ecole Normale Suprieure and INRIA), demonstrates an unprecedented improvement in the stability of superconducting qubits, marking a critical milestone towards useful fault-tolerant quantum computing.
The researchers have significantly extended the bit-flip times from milliseconds to tens of secondsthousands of times better than any other superconducting qubit type.
Quantum computers face two types of errors: bit-flips and phase-flips. Cat qubits exponentially reduce bit-flips, which are analogous to classical bit flips in digital computing. As a result, the remaining phase-flips can be addressed more efficiently with simpler error correcting codes.
The researchers used Alice & Bobs Boson 3 chipset for this record-breaking result, which features a cat qubit design named TomCat. TomCat employs an efficient quantum tomography (measurement) protocol that allows for the control of quantum states without the use of a transmon, a common circuit used by many quantum companies, but one of the major sources of bit-flips for cat qubits. This design also minimizes the footprint of the qubit on the chip, removing drivelines, cables, instruments, making this stable qubit scalable. Recently, Alice & Bob made publicly available their new Boson 4 chipset that reaches over 7 minutes of bit-flip lifetime. The results from this Nature Publication can therefore be reproduced by users on Boson 4 over Google Cloud.
Although Alice & Bobs latest Boson chips are getting closer to the company bit-flip protection targets, Alice & Bob plans to further advance their technology. The next iterations will focus on boosting the cat qubit phase-flip time and readout fidelity to reach the requirements of their latest architecture to deliver a 100 logical qubit quantum computer.
Key advances highlighted in the research include:
About Alice & Bob
Alice & Bob is a quantum computing company based in Paris and Boston whose goal is to create the first universal, fault-tolerant quantum computer. Founded in 2020, Alice & Bob has already raised 30 million in funding, hired over 95 employees and demonstrated experimental results surpassing those of technology giants such as Google or IBM. Alice & Bob specializes in cat qubits, a pioneering technology developed by the companys founders and later adopted by Amazon. Demonstrating the power of its cat architecture, Alice & Bob recently showed that it could reduce the hardware requirements for building a useful large-scale quantum computer by up to 200 times compared with competing approaches. Alice & Bob cat qubit is available for anyone to test through cloud access.
Source: Alice & Bob
View original post here:
Alice & Bob's Cat Qubit Research Published in Nature - HPCwire
NIST quantum-resistant algorithms to be published within weeks, top White House advisor says – The Record from Recorded Future News
Update, May 24: Includes correction from NIST about the number of algorithms to be released.
The U.S. National Institute of Standards and Technology (NIST) will release post-quantum cryptographic algorithms in the next few weeks, a senior White House official said on Monday.
Anne Neuberger, the White Houses top cyber advisor, told an audience at the Royal United Services Institute (RUSI) in London that the release of the algorithms was a momentous moment, as they marked a major step in the transition to the next generation of cryptography.
The transition is being made in apprehension of what is called a cryptographically relevant quantum computer (CRQC), a device theoretically capable of breaking the encryption thats at the root of protecting both corporate and national security secrets, said Neuberger. NIST made a preliminary announcement of the algorithms in 2022.
Following publication, a spokesperson for NIST told Recorded Future News it was planning to release three finalized algorithms this summer and not four, as Neuberger had said in London.
Conrad Prince, a former official at GCHQ and now a distinguished fellow at RUSI, told Neuberger that during his previous career there had consistently been a concern about hostile states having the capability to decrypt the plaintext of secure messages, although this capability was consistently estimated at being roughly a decade away and had been for the last 20 years.
Neuberger said the U.S. intelligence communitys estimate is similar, the early 2030s, for when a CRQC would be operational. But the time-frame is relevant, said the White House advisor, because there is national security data that is collected today and even if decrypted eight years from now, can still be damaging.
Britains NCSC has warned that contemporary threat actors could be collecting and storing intelligence data today for decryption at some point in the future.
Given the cost of storing vast amounts of old data for decades, such an attack is only likely to be worthwhile for very high-value information, stated the NCSC. As such, the possibility of a CRQC existing at some point in the next decade is a very relevant threat right now.
Neuberger added: Certainly theres some data thats time sensitive, you know, a ship that looks to be transporting weapons to a sanctioned country, probably in eight years we dont care about that anymore.
Publishing the new NIST algorithms is a protection against adversaries collecting the most sensitive kinds of data today, Neuberger added.
A spokesperson for NIST told Recorded Future News: The plan is to release the algorithms this summer. We dont have anything more specific to offer at this time.
But publishing the algorithms is not the last step in moving to a quantum-resistant computing world. The NCSC has warned it is actually just the second step in what will be a very complicated undertaking.
Even if any one of the algorithms proposed by NIST achieves universal acceptance as something that is unbreakable by a quantum computer, it would not be a simple matter of just swapping those algorithms in for the old-fashioned ones.
Part of the challenge is that most systems that currently depend on public-key cryptography for their security are not necessarily capable of running the resource-heavy software used in post-quantum cryptography.
Ultimately, the security of public key cryptographic systems relies on the mathematical difficulty of factoring very large prime numbers something that traditional computers find exhaustingly difficult.
However, research by American mathematician Peter Shor, published in 1994, proposed an algorithm that could be run on a quantum computer for finding these prime factors with far more ease; potentially undermining some of the key assumptions about what makes public-key cryptography secure.
The good news, according to NCSC, is that while advances in quantum computing are continuing to be made, the machines that exist today are still limited, and suffer from relatively high error rates in each operation they perform, stated the agency.
But the NCSC warned that in the future, it is possible that error rates can be lowered such that a large, general-purpose quantum computer could exist, but it is impossible to predict when this may happen.
Recorded Future
Intelligence Cloud.
No previous article
No new articles
Alexander Martin
is the UK Editor for Recorded Future News. He was previously a technology reporter for Sky News and is also a fellow at the European Cyber Conflict Research Initiative.
Go here to see the original:
NIST quantum-resistant algorithms to be published within weeks, top White House advisor says - The Record from Recorded Future News
Glimpse of next-generation internet – Harvard Office of Technology Development
May 20th, 2024
By Anne Manning, Harvard Staff Writer Published in the Harvard Gazette
An up close photo of the diamond silicon vacancy center.
Its one thing to dream up a next-generation quantum internet capable of sending highly complex, hacker-proof information around the world at ultra-fast speeds. Its quite another to physically show its possible.
Thats exactly what Harvard physicists have done, using existing Boston-area telecommunication fiber, in a demonstration of the worlds longest fiber distance between two quantum memory nodes. Think of it as a simple, closed internet carrying a signal encoded not by classical bits like the existing internet, but by perfectly secure, individual particles of light.
The groundbreaking work, published in Nature, was led by Mikhail Lukin, the Joshua and Beth Friedman University Professor in the Department of Physics, in collaboration with Harvard professors Marko Lonar and Hongkun Park, who are all members of the Harvard Quantum Initiative. The Nature work was carried out with researchers at Amazon Web Services.
The Harvard team established the practical makings of the first quantum internet by entangling two quantum memory nodes separated by optical fiber link deployed over a roughly 22-mile loop through Cambridge, Somerville, Watertown, and Boston. The two nodes were located a floor apart in Harvards Laboratory for Integrated Science and Engineering.
Showing that quantum network nodes can be entangled in the real-world environment of a very busy urban area is an important step toward practical networking between quantum computers.
Mikhail Lukin, the Joshua and Beth Friedman University Professor in the Department of Physics
Quantum memory, analogous to classical computer memory, is an important component of a quantum computing future because it allows for complex network operations and information storage and retrieval. While other quantum networks have been created in the past, the Harvard teams is the longest fiber network between devices that can store, process, and move information.
Each node is a very small quantum computer, made out of a sliver of diamond that has a defect in its atomic structure called a silicon-vacancy center. Inside the diamond, carved structures smaller than a hundredth the width of a human hair enhance the interaction between the silicon-vacancy center and light.
The silicon-vacancy center contains two qubits, or bits of quantum information: one in the form of an electron spin used for communication, and the other in a longer-lived nuclear spin used as a memory qubit to store entanglement, the quantum-mechanical property that allows information to be perfectly correlated across any distance.
(In classical computing, information is stored and transmitted as a series of discrete binary signals, say on/off, that form a kind of decision tree. Quantum computing is more fluid, as information can exist in stages between on and off, and is stored and transferred as shifting patterns of particle movement across two entangled points.)
Map showing path of two-node quantum network through Boston and Cambridge. Credit: Can Knaut via OpenStreetMap
Using silicon-vacancy centers as quantum memory devices for single photons has been a multiyear research program at Harvard. The technology solves a major problem in the theorized quantum internet: signal loss that cant be boosted in traditional ways.
A quantum network cannot use standard optical-fiber signal repeaters because simple copying of quantum information as discrete bits is impossible making the information secure, but also very hard to transport over long distances.
Silicon-vacancy-center-based network nodes can catch, store, and entangle bits of quantum information while correcting for signal loss. After cooling the nodes to close to absolute zero, light is sent through the first node and, by nature of the silicon vacancy centers atomic structure, becomes entangled with it, so able to carry the information.
Since the light is already entangled with the first node, it can transfer this entanglement to the second node, explained first author Can Knaut, a Kenneth C. Griffin Graduate School of Arts and Sciences student in Lukins lab. We call this photon-mediated entanglement.
Over the last several years, the researchers have leased optical fiber from a company in Boston to run their experiments, fitting their demonstration network on top of the existing fiber to indicate that creating a quantum internet with similar network lines would be possible.
Showing that quantum network nodes can be entangled in the real-world environment of a very busy urban area is an important step toward practical networking between quantum computers, Lukin said.
A two-node quantum network is only the beginning. The researchers are working diligently to extend the performance of their network by adding nodes and experimenting with more networking protocols.
The paper is titled Entanglement of Nanophotonic Quantum Memory Nodes in a Telecom Network. The work was supported by the AWS Center for Quantum Networkings research alliance with the Harvard Quantum Initiative, the National Science Foundation, the Center for Ultracold Atoms (an NSF Physics Frontiers Center), the Center for Quantum Networks (an NSF Engineering Research Center), the Air Force Office of Scientific Research, and other sources.
Harvard Office of Technology Development enabled the strategic alliance between Harvard University and Amazon Web Services (AWS) to advance fundamental research and innovation in quantum networking.
Tags: Alliances, Collaborations, Quantum Physics, Internet, Publication
Press Contact: Kirsten Mabry | (617) 495-4157
Read more:
Glimpse of next-generation internet - Harvard Office of Technology Development