{"id":168137,"date":"2024-01-04T02:36:07","date_gmt":"2024-01-04T07:36:07","guid":{"rendered":"https:\/\/www.immortalitymedicine.tv\/were-on-the-brink-of-the-biggest-changes-to-computings-dna-and-its-not-just-quantum-thats-coming-pc-gamer\/"},"modified":"2024-08-18T11:32:37","modified_gmt":"2024-08-18T15:32:37","slug":"were-on-the-brink-of-the-biggest-changes-to-computings-dna-and-its-not-just-quantum-thats-coming-pc-gamer","status":"publish","type":"post","link":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/quantum-computing\/were-on-the-brink-of-the-biggest-changes-to-computings-dna-and-its-not-just-quantum-thats-coming-pc-gamer.php","title":{"rendered":"We&#8217;re on the brink of the biggest changes to computing&#8217;s DNA and it&#8217;s not just quantum that&#8217;s coming &#8211; PC Gamer"},"content":{"rendered":"<p><p>        This article was originally published on 30th June this        year and we are republishing it today as part of a series        celebrating some of our favourite pieces from the past 12        months.      <\/p>\n<p>      Read more: the future of CPUs    <\/p>\n<p>    Computers are built around logic: performing mathematical    operations using circuits. Logic is built around things such as    Addersnot the snake; the basic circuit that adds together two    numbers. This is as true of today's microprocessors as all    those going back to the very beginning of computing history.    You could go back to an abacus and find that, at some    fundamental level, it does the same thing as your shiny gaming    PC. It's just much, much less capable.  <\/p>\n<p>    Nowadays, processors can do a lot of mathematical calculations    using any number of complex circuits in a single clock. And a    lot more than just add two numbers together, too. But to get to    your shiny new gaming CPU, there has been a process of    iterating on the classical computers that came before, going    back centuries.  <\/p>\n<p>    As you might imagine, building something entirely different to    that is a little, uh, tricky, but that's what some are striving    to do, with technologies like quantum and neuromorphic    computingtwo distinct concepts that could change computing for    good.  <\/p>\n<p>    \"Quantum computing is a technology that, at least by name, we    have become very accustomed to hearing about and is always    mentioned as 'the future of computing',\" says Carlos Andrs    Trasvia Moreno, software engineering coordinator at    CETYS    Ensenada.  <\/p>\n<p>    Quantum computers utilise qubits, or quantum bits. Unlike a    classical bit, which can only exist in one of two states, these    qubits can exist in two states and a superposition of those two    states. It's zero, one, or both zero and one at the same time.    And if that sounds awfully confusing, that's because it is, but    it also has immense potential.  <\/p>\n<p>    Quantum computers are expected to be powerful enough to break    modern-day 'unbreakable' encryption, accelerate medicine    discover, re-shape how the global economy transports goods,    explore the stars, and pretty much revolutionise anything    involving massive number crunching.  <\/p>\n<p>    The problem is, quantum computers are immensely difficult to    make, and maybe even more difficult to run.  <\/p>\n<p>              Social Links Navigation            <\/p>\n<\/p>\n<p>    \"One of the main drawbacks of quantum computing is its    high-power consumption, since it works with algorithms of far    greater complexity than that of any current CPU,\" Moreno    continues. \"Also, it requires an environment of near absolute    zero temperatures, which worsens the power requirements of the    system. Lastly, they are extremely sensitive to environmental    disturbances such as heat, light and vibrations.  <\/p>\n<p>        We're scratching the surface there with quantum computing.      <\/p>\n<p>    \"Any of these can alter the current quantum states and produce    unexpected outcomes.\"  <\/p>\n<p>    And while you can sort of copy the function of classical logic    with qubitswe're not starting entirely at zero in developing    these machinesto exploit a quantum computer's power requires    new and complex quantum algorithms that we're only just getting    to grips with.  <\/p>\n<p>    IBM is one company investing heavily in quantum computing,    aiming to create a quantum computer with 4,158 or more qubits    by 2025. Google also has its fingers in    quantum.  <\/p>\n<\/p>\n<p>    Admittedly, we're still a long way off ubiquitous 'quantum    supremacy', which is the moment when a quantum computer is    better than today's top classical supercomputers. Google did    claim it did just    that back in 2019, though that may have turned out to be    something of a     niche achievement, but nonetheless an impressive one.    Either way, in practical terms, we're just not there yet.  <\/p>\n<\/p>\n<p>    They're a real pain to figure out, to put it scientifically.    But that's never stopped a good engineer yet.  <\/p>\n<p>    \"I do think that we're scratching the surface there with    quantum computing. And again, just like we broke the laws of    physics with silicon over and over and over again, I think we    break the laws of physics here, too,\" Marcus Kennedy, general    manager of gaming at Intel, tells me.  <\/p>\n<p>          Marcus Kennedy        <\/p>\n<p>    There's more immediate potential for the future of computing in    artificial intelligence, your favourite 2023 buzzword. But it    really is a massive and life-changing development for many, and    I'm not just talking about that clever-sounding,    slightly-too-argumentative chatbot in your browser. We're only    scratching the surface of AI's uses today, and to unlock those    deeper, more impactful uses there's a whole new type of chip in    the works.  <\/p>\n<p>    \"Neuromorphic computing is, in my mind, the most viable    alternative [to classical computing],\" Moreno says.  <\/p>\n<p>    \"In a sense, we could say that neuromorphic computers are    biological neural networks implemented on hardware. One would    think it's simply translating a perceptron to voltages and    gates, but it's actually a closer imitation on how brains work,    on how actual neurons communicate amongst each other through    synapsis.\"  <\/p>\n<p>    What is neuromorphic computing? The answers in the name, neuro,    meaning related to the nervous system. A neuromorphic computer    aims to imitate the greatest computer, and most complex    creation, ever known to man: the brain.  <\/p>\n<p>    \"I think we'll get to a place where the processing capability    of those neuromorphic chips far outstrips the processing    capability of a monolithic die based on an x86 architecture, a    traditional kind of architecture. Because the way the brain    operates, we know it has the capacity and the capability that    far outstrips anything else,\" Kennedy says.  <\/p>\n<p>    \"The most effective kind of systems tend to look very much like    things that you see in nature.\"  <\/p>\n<p>    Neuromorphic chips are yet to reach their breakthrough moment,    but they're coming. Intel has a couple of neuromorphic chips in    development today,     Loihi and Loihi 2.  <\/p>\n<\/p>\n<p>    And what is a neuromorphic chip, really? Well, it's a brain,    with neurons and synapses. But since they're still crafted from    silicon, think of them as a sort of hybrid of a classical    computer chip and the biology of the brain.  <\/p>\n<p>    And not necessarily a big brainLoihi 2 has 1 million neurons    and 120 million synapses, which is many orders of magnitude    smaller than a human brain with roughly 86 billion neurons and    trillions of synapses. It's hard to count them all, as    you might imagine, so we don't really know precisely, but we    have big ol' brains. You can brag about that all you want to    your smaller-brained animal companions.  <\/p>\n<p>    A cockroach is estimated to have as many synapses as Loihi 2,    for a better understanding of the grey matter scale we're    talking about here.  <\/p>\n<p>    \"We claim you don't need to be that complex that the brain has    its function, but if you're going to do computing, you just    need some of the basic functions of a neuron and synapse to    actually make it work,\"     Dr. Mark Dean told me in 2021.  <\/p>\n<p>          Dr. Mark Dean        <\/p>\n<p>    Neuromorphic computing has a lot of room to grow, and with a    rapidly growing interest in AI, this nascent technology may    prove to be the key to powering those ever-more-impressive AI    models you keep reading about.  <\/p>\n<p>        The amount of processing power would surpass any of the        existing products with just a fraction of the energy.      <\/p>\n<p>    You might think that AI models are running just fine today,    which is primarily thanks to Nvidia's graphics cards running    the show. But the reason neuromorphic computing is so    tantalising to some is \"that it can heavily reduce the power    consumption of a processor, whilst still managing the same    computational capabilities of modern chips,\" Moreno says.  <\/p>\n<p>    \"In comparison, the human brain is capable of hundreds of    teraflops of processing power with only 20 watts of energy    consumption, whilst a modest graphics card can output 40-50    teraflops of power with an energy consumption of 450 watts.\"  <\/p>\n<p>    Basically, \"If a neuromorphic processor were to be developed    and implemented in a GPU, the amount of processing power would    surpass any of the existing products with just a fraction of    the energy.\"  <\/p>\n<p>    Sound appealing? Yeah, of course it does. Lower energy    consumption isn't only massive for the potential computing    power it could bring about, it's massive for using less energy,    which has knock-on effects for cooling, too.   <\/p>\n<p>    \"Changing the architecture of computing would also require a    different programming paradigm to be implemented, which in its    own will also be an impressive feat,\" Moreno continues.  <\/p>\n<\/p>\n<p>    Building a neuromorphic chip is one thing, programming for it    is something else. That's one reason why Intel's neuromorphic    computing framework is open-source, you need a lot of hands on    deck to get this sort of project off the ground.  <\/p>\n<p>    \"The thing that we haven't cracked yet is the software behind    how to leverage the structure,\" Kennedy says. \"And so you can    create a chip that looks very much like a brain, the software    is really what makes it function like a brain. And to date, we    haven't cracked that nut.\"  <\/p>\n<p>    It'll take some time before we entirely replace AI accelerators    with something that resembles a brain. Or Adders and binary    functions, that are as old as computing itself, with quantum    computers. Yet experiential attempts have already begun to    replace classical computing as we know it.  <\/p>\n<p>    A recent breakthrough claimed by Microsoft sees the company    very bullish on quantum's future, and there's also recently    been IBM predicting quantum computers will    outperform classical ones in important tasks within two    years.  <\/p>\n<p>    In the words of Intel's Kennedy, \"I think we're getting there.\"  <\/p>\n<p><!-- Auto Generated --><\/p>\n<p>Continued here:<br \/>\n<a target=\"_blank\" href=\"https:\/\/www.pcgamer.com\/quantum-classical-neuromorphic-computing\/\" title=\"We're on the brink of the biggest changes to computing's DNA and it's not just quantum that's coming - PC Gamer\" rel=\"noopener\">We're on the brink of the biggest changes to computing's DNA and it's not just quantum that's coming - PC Gamer<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p> This article was originally published on 30th June this year and we are republishing it today as part of a series celebrating some of our favourite pieces from the past 12 months. Read more: the future of CPUs Computers are built around logic: performing mathematical operations using circuits. Logic is built around things such as Addersnot the snake; the basic circuit that adds together two numbers <a href=\"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/quantum-computing\/were-on-the-brink-of-the-biggest-changes-to-computings-dna-and-its-not-just-quantum-thats-coming-pc-gamer.php\">Continue reading <span class=\"meta-nav\">&rarr;<\/span><\/a><\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"limit_modified_date":"","last_modified_date":"","_lmt_disableupdate":"","_lmt_disable":"","footnotes":""},"categories":[494694],"tags":[],"class_list":["post-168137","post","type-post","status-publish","format-standard","hentry","category-quantum-computing"],"modified_by":null,"_links":{"self":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts\/168137"}],"collection":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/comments?post=168137"}],"version-history":[{"count":0,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts\/168137\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/media?parent=168137"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/categories?post=168137"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/tags?post=168137"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}