{"id":1075685,"date":"2024-05-06T02:47:03","date_gmt":"2024-05-06T06:47:03","guid":{"rendered":"https:\/\/www.immortalitymedicine.tv\/apples-ai-research-suggests-features-are-coming-for-siri-artists-and-more-the-verge\/"},"modified":"2024-08-18T12:53:40","modified_gmt":"2024-08-18T16:53:40","slug":"apples-ai-research-suggests-features-are-coming-for-siri-artists-and-more-the-verge","status":"publish","type":"post","link":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/ai\/apples-ai-research-suggests-features-are-coming-for-siri-artists-and-more-the-verge.php","title":{"rendered":"Apples AI research suggests features are coming for Siri, artists, and more. &#8211; The Verge"},"content":{"rendered":"<p><p>      It would be easy to think that Apple is late to the game on      AI. Since late 2022, when ChatGPT took the world by storm,      most of Apples competitors have fallen over themselves to      catch up. While Apple has certainly talked about AI and even      released some products with AI in mind, it seemed to be      dipping a toe in rather than diving in headfirst.    <\/p>\n<p>      But over the last few months, rumors and reports have      suggested that Apple has, in fact, just been biding its time,      waiting to make its move.There have been reports in      recent weeks that Apple is talking to both       OpenAI and Google about powering some of its AI features,      and the company has also been       working on its own model, called Ajax.    <\/p>\n<p>      If you look through Apples published AI research, a picture      starts to develop of how Apples approach to AI might come to      life. Now, obviously, making product assumptions based on      research papers is a deeply inexact science  the line from      research to store shelves is windy and full of potholes. But      you can at least get a sense of what the company is      thinking about and how its AI features might      work when Apple starts to talk about them at its annual      developer conference, WWDC, in June.    <\/p>\n<p>      I suspect you and I are hoping for the same thing here:      Better Siri. And it looks very much like Better Siri is      coming! Theres an assumption in a lot of Apples research      (and in a lot of the tech industry, the world, and      everywhere)that large language models will immediately      make virtual assistants better and smarter. For Apple,      getting to Better Siri means making those models as fast as      possible and making sure theyre everywhere.    <\/p>\n<p>      In iOS 18, Apple plans to have all its AI features running on      an on-device, fully offline model,       Bloomberg recently reported. Its tough to build      a good multipurpose model even when you have a network of      data centers and thousands of state-of-the-art GPUs  its      drastically harder to do it with only the guts inside your      smartphone. So Apples having to get creative.    <\/p>\n<p>      In a paper called LLM in a flash:      Efficient Large Language Model Inference with Limited      Memory (all these papers have really boring titles but      are really interesting, I promise!), researchers devised a      system for storing a models data, which is usually stored on      your devices RAM, on the SSD instead. We have demonstrated      the ability to run LLMs up to twice the size of available      DRAM [on the SSD], the researchers wrote, achieving an      acceleration in inference speed by 4-5x compared to      traditional loading methods in CPU, and 20-25x in GPU. By      taking advantage of the most inexpensive and available      storage on your device, they found, the models can run faster      and more efficiently.    <\/p>\n<p>      Apples researchers also created a system called EELBERT that can      essentially compress an LLM into a much smaller size without      making it meaningfully worse. Their compressed take on      Googles Bert model was 15 times smaller  only 1.2 megabytes      and saw only a 4 percent reduction in quality. It did      come with some latency tradeoffs, though.    <\/p>\n<p>      In general, Apple is pushing to solve a core tension in the      model world: the bigger a model gets, the better and more      useful it can be, but also the more unwieldy, power-hungry,      and slow it can become. Like so many others, the company is      trying to find the right balance between all those things      while also looking for a way to have it all.    <\/p>\n<p>      A lot of what we talk about when we talk about AI products is      virtual assistants  assistants that know things, that can      remind us of things, that can answer questions, and get stuff      done on our behalf. So its not exactly shocking that a lot      of Apples AI research boils down to a single question: what      if Siri was really, really, really good?    <\/p>\n<p>      A group of Apple researchers has been working on a way to use Siri      without needing to use a wake word at all; instead of      listening for Hey Siri or Siri, the device might be able      to simply intuit whether youre talking to it. This problem      is significantly more challenging than voice trigger      detection, the researchers did acknowledge, since there      might not be a leading trigger phrase that marks the      beginning of a voice command. That might be why another      group of researchers developed a system to more accurately detect      wake words. Another paper      trained a model to better understand rare words, which are      often not well understood by assistants.    <\/p>\n<p>      In both cases, the appeal of an LLM is that it can, in      theory, process much more information much more quickly. In      the wake-word paper, for instance, the researchers found that      by not trying to discard all unnecessary sound but,      instead, feeding it all to the model and letting it process      what does and doesnt matter, the wake word worked far more      reliably.    <\/p>\n<p>      Once Siri hears you, Apples doing a bunch of work to make      sure it understands and communicates better. In one paper, it      developed a      system called STEER (which stands for Semantic Turn      Extension-Expansion Recognition, so well go with STEER) that      aims to improve your back-and-forth communication with an      assistant by trying to figure out when youre asking a      follow-up question and when youre asking a new one. In      another, it uses LLMs to better understand ambiguous      queries to figure out what you mean no matter how you say      it. In uncertain circumstances, they wrote, intelligent      conversational agents may need to take the initiative to      reduce their uncertainty by asking good questions      proactively, thereby solving problems more effectively.            Another paper aims to help with that, too: researchers      used LLMs to make assistants less verbose and more      understandable when theyre generating answers.    <\/p>\n<p>      Whenever Apple does talk publicly about AI, it tends to focus      less on raw technological might and more on the day-to-day      stuff AI can actually do for you. So, while theres a lot of      focus on Siri  especially as Apple looks to compete with      devices like the Humane AI Pin, the Rabbit R1, and Googles      ongoing smashing of Gemini into all of Android there      are plenty of other ways Apple seems to see AI being useful.    <\/p>\n<p>      One obvious place for Apple to focus is on health: LLMs      could, in theory, help wade through the oceans of biometric      data collected by your various devices and help you make      sense of it all. So, Apple has been researching how to      collect and collate all of your motion data, how to use gait      recognition and your headphones to identify you, and how to      track and understand your heart rate data. Apple also created      and released the largest multi-device multi-location      sensor-based human activity dataset available after      collecting data from 50 participants with multiple on-body      sensors.    <\/p>\n<p>      Apple also seems to imagine AI as a creative tool. For one      paper, researchers interviewed a bunch of animators,      designers, and engineers and built a system called      Keyframer that enable[s] users to iteratively construct      and refine generated designs. Instead of typing in a prompt      and getting an image, then typing another prompt to get      another image, you start with a prompt but then get a toolkit      to tweak and refine parts of the image to your liking. You      could imagine this kind of back-and-forth artistic process      showing up anywhere from the Memoji creator to some of      Apples more professional artistic tools.    <\/p>\n<p>      In another      paper, Apple describes a tool called MGIE that lets you      edit an image just by describing the edits you want to make.      (Make the sky more blue, make my face less weird, add      some rocks, that sort of thing.) Instead of brief but      ambiguous guidance, MGIE derives explicit visual-aware      intention and leads to reasonable image editing, the      researchers wrote. Its initial experiments werent perfect,      but they were impressive.    <\/p>\n<p>      We might even get some AI in Apple Music: for a paper called      Resource-constrained      Stereo Singing Voice Cancellation, researchers explored      ways to separate voices from instruments in songs      which could come in handy if Apple wants to give      people tools to, say, remix songs the way you can on TikTok      or Instagram.    <\/p>\n<p>      Over time, Id bet this is the kind of stuff youll see Apple      lean into, especially on iOS. Some of it Apple will build      into its own apps; some it will offer to third-party      developers as APIs. (The recent Journaling Suggestions      feature is probably a good guide to how that might work.)      Apple has always trumpeted its hardware capabilities,      particularly compared to your average Android device; pairing      all that horsepower with on-device, privacy-focused AI could      be a big differentiator.    <\/p>\n<p>      But if you want to see the biggest, most ambitious AI thing      going at Apple, you need to know about Ferret. Ferret      is a multi-modal large language model that can take      instructions, focus on something specific youve circled or      otherwise selected, and understand the world around it. Its      designed for the now-normal AI use case of asking a device      about the world around you, but it might also be able to      understand whats on your screen. In the Ferret paper,      researchers show that it could help you navigate apps, answer      questions about App Store ratings, describe what youre      looking at, and more. This has really exciting implications      for accessibility but could also completely change the way      you use your phone  and your Vision Pro and \/ or smart      glasses someday.    <\/p>\n<p>      Were getting way ahead of ourselves here, but you can      imagine how this would work with some of the other stuff      Apple is working on. A Siri that can understand what you      want, paired with a device that can see and understand      everything thats happening on your display, is a phone that      can literally use itself. Apple wouldnt need deep      integrations with everything; it could simply run the apps      and tap the right buttons automatically.    <\/p>\n<p>      Again, all this is just research, and for all of it to work      well starting this spring would be a legitimately unheard-of      technical achievement. (I mean, youve tried chatbots  you      know theyre not great.) But Id bet you anything were going      to get some big AI announcements at WWDC. Apple CEO Tim Cook      even teased as much in February, and basically      promised it on this weeks earnings call. And two things      are very clear: Apple is very much in the AI race, and it      might amount to a total overhaul of the iPhone. Heck, you      might even start willingly using Siri! And that would be      quite the accomplishment.    <\/p>\n<p><!-- Auto Generated --><\/p>\n<p>Original post:<\/p>\n<p><a target=\"_blank\" rel=\"nofollow noopener\" href=\"https:\/\/www.theverge.com\/2024\/5\/5\/24147995\/apple-siri-ai-research-chatbot-creativity\" title=\"Apples AI research suggests features are coming for Siri, artists, and more. - The Verge\">Apples AI research suggests features are coming for Siri, artists, and more. - The Verge<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p> It would be easy to think that Apple is late to the game on AI.  <a href=\"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/ai\/apples-ai-research-suggests-features-are-coming-for-siri-artists-and-more-the-verge.php\">Continue reading <span class=\"meta-nav\">&rarr;<\/span><\/a><\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"limit_modified_date":"","last_modified_date":"","_lmt_disableupdate":"","_lmt_disable":"","footnotes":""},"categories":[1234935],"tags":[],"class_list":["post-1075685","post","type-post","status-publish","format-standard","hentry","category-ai"],"modified_by":null,"_links":{"self":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts\/1075685"}],"collection":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/comments?post=1075685"}],"version-history":[{"count":0,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts\/1075685\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/media?parent=1075685"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/categories?post=1075685"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/tags?post=1075685"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}