It’s no secret: quantum computing has been poised to be “the next big thing” for years. But recent developments in the quantum ecosystem, including major investments by companies such as IBM, Google, Microsoft and others, are the best indicators that now is the time to begin preparing for potentially viable quantum applications—and to identify where […]
Learning quantum theory requires dedication and a willingness to challenge classical assumptions. Quantum interference, particularly for massive particles, is a pivotal example in this journey. The Schrödinger equation, inspired by de Broglie’s hypothesis, revolutionized our understanding by revealing the wavelike nature of even massive particles. This phenomenon not only deepens our grasp of nature but also fuels innovations in quantum applications, from quantum sensing to quantum computing. Yet many students don't have the opportunity to run experiments that require sophisticated hardware. Not anymore!
Mads Bahrami, et al.
Announcements & Events
Today we celebrate a new waypoint on our journey of nearly four decades with the release of Version 14.0 of Wolfram Language and Mathematica. Over the two years since we released Version 13.0 we’ve been steadily delivering the fruits of our research and development in .1 releases every six months. Today we’re aggregating these—and more—into Version 14.0.
Today we’re launching Version 13.3 of Wolfram Language and Mathematica—both available immediately on desktop and cloud. It’s only been 196 days since we released Version 13.2, but there’s a lot that’s new, not least a whole subsystem around LLMs.
We originally invented the concept of “Notebooks” back in 1987, for Version 1.0 of Mathematica. And over the past 36 years, Notebooks have proved to be an incredibly convenient medium in which to do—and publish—work (and indeed, I, for example, have created hundreds of thousands of them). And, yes, eventually the basic concepts of Notebooks […]
Prompts are how one channels an LLM to do something. LLMs in a sense always have lots of “latent capability” (e.g. from their training on billions of webpages). But prompts—in a way that’s still scientifically mysterious—are what let one “engineer” what part of that capability to bring out.
So far, we mostly think of LLMs as things we interact directly with, say through chat interfaces. But what if we could take LLM functionality and “package it up” so that we can routinely use it as a component inside anything we’re doing? Well, that’s what our new LLMFunction is about.
A few weeks ago, in collaboration with OpenAI, we released the Wolfram plugin for ChatGPT, which lets ChatGPT use Wolfram Language and Wolfram|Alpha as tools, automatically called from within ChatGPT. One can think of this as adding broad “computational superpowers” to ChatGPT, giving access to all the general computational capabilities and computational knowledge in Wolfram […]