{"id":1075651,"date":"2024-02-22T02:36:55","date_gmt":"2024-02-22T07:36:55","guid":{"rendered":"https:\/\/www.immortalitymedicine.tv\/ai-and-you-openais-sora-previews-text-to-video-future-first-ivy-league-ai-degree-cnet\/"},"modified":"2024-08-18T12:53:16","modified_gmt":"2024-08-18T16:53:16","slug":"ai-and-you-openais-sora-previews-text-to-video-future-first-ivy-league-ai-degree-cnet","status":"publish","type":"post","link":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/ai\/ai-and-you-openais-sora-previews-text-to-video-future-first-ivy-league-ai-degree-cnet.php","title":{"rendered":"AI and You: OpenAI&#8217;s Sora Previews Text-to-Video Future, First Ivy League AI Degree &#8211; CNET"},"content":{"rendered":"<p><p>    AI developments are happening pretty fast. If you don't stop    and look around once in a while, you could miss them.  <\/p>\n<p>    Fortunately, I'm looking around for you and what I saw this    week is that competition between OpenAI, maker of ChatGPT and Dall-E, and Google continues to    heat up in a way that's worth paying attention to.  <\/p>\n<p>    A week after updating its Bard chatbot and changing the name to    Gemini, Google's DeepMind AI subsidiary previewed the next    version of its generative AI chatbot. DeepMind told CNET's Lisa    Lacy that Gemini 1.5 will be rolled out \"slowly\" to regular    people who sign up for a wait list and will be available now    only to developers and enterprise customers.  <\/p>\n<p>    Gemini 1.5 Pro, Lacy reports, is \"as capable    as\" the Gemini 1.0 Ultra model, which Google announced on Feb.    8. The 1.5 Pro model has a win rate -- a measurement of how    many benchmarks it can outperform -- of 87% compared with the    1.0 Pro and 55% against the 1.0 Ultra. So the 1.5 Pro is    essentially an upgraded version of the best available model    now.  <\/p>\n<p>    Gemini 1.5 Pro can ingest video, images, audio and text to    answer questions, added Lacy. Oriol Vinyals, vice president of    research at Google DeepMind and co-lead of Gemini, described    1.5 as a \"research release\" and said the model is \"very    efficient\" thanks to a unique architecture that can answer    questions by zeroing in on expert sources in that particular    subject rather than seeking the answer from all possible    sources.  <\/p>\n<p>    Meanwhile, OpenAI announced a new    text-to-video model called Sora that's capturing a lot of    attention because of the photorealistic videos it's able to    generate. Sora can \"create videos of up to 60 seconds featuring    highly detailed scenes, complex camera motion, and multiple    characters with vibrant emotions.\" Following up on a promise it    made, along with Google    and Meta last week, to watermark AI-generated images    and video, OpenAI says it's also creating tools to detect    videos created with Sora so they can be identified as being AI    generated.  <\/p>\n<p>    Google and Meta have    also announced their own gen AI text-to-video creators.  <\/p>\n<p>    Sora, which means \"sky\" in Japanese, is also being called    experimental, with OpenAI limiting access for now to so-called    \"red teamers,\" security experts and researchers who will assess    the tool's potential harms or risks. That follows through on    promises made as part of President Joe Biden's AI executive order last    year, asking developers to submit the results of safety    checks on their gen AI chatbots before releasing them publicly.    OpenAI said it's also looking to get feedback on Sora from some    visual artists, designers and filmmakers.  <\/p>\n<p>    How do the photorealistic videos look? Pretty realistic. I    agree with the The New    York Times, which described the short demo videos -- \"of    wooly mammoths trotting through a snowy meadow, a monster    gazing at a melting candle and a Tokyo street scene seemingly    shot by a camera swooping across the city\" -- as \"eye    popping.\"  <\/p>\n<p>    The MIT    Review, which also got a preview of Sora, said the \"tech    has pushed the envelope of what's possible with text-to-video    generation.\" Meanwhile, The    Washington Post noted Sora could exacerbate an already    growing problem with video deepfakes, which have been used to    \"deceive voters\" and scam consumers.  <\/p>\n<p>    One X commentator summarized it this way: \"Oh boy    here we go what is real anymore.\" And OpenAI CEO Sam Altman    called the news about its video generation model a \"remarkable    moment.\"  <\/p>\n<p>    You can see the four examples of what Sora can produce on OpenAI's intro site,    which notes that the tool is \"able to generate complex scenes    with multiple characters, specific types of motion, and    accurate details of the subject and background. The model    understands not only what the user has asked for in the prompt,    but also how those things exist in the physical world. The    model has a deep understanding of language, enabling it to    accurately interpret prompts and generate compelling characters    that express vibrant emotions.\"  <\/p>\n<p>    But Sora has its weaknesses, which is why OpenAI hasn't yet    said whether it will actually be incorporated into its    chatbots. Sora \"may struggle with accurately simulating the    physics of a complex scene and may not understand specific    instances of cause and effect. For example, a person might take    a bite out of a cookie, but afterward, the cookie may not have    a bite mark. The model may also confuse spatial details of a    prompt, for example, mixing up left and right.\"  <\/p>\n<p>    All of this is to remind us that tech is a tool -- and that    it's up to us humans to decide how, when, where and why to use    that technology. In case you didn't see it, the trailer for the    new Minions movie (Despicable Me 4: Minion    Intelligence) makes this point cleverly, with its sendup of    gen AI deepfakes and Jon Hamm's voiceover of how \"artificial    intelligence is changing how we see the worldtransforming the    way we do business.\"  <\/p>\n<p>    \"With artificial intelligence,\" Hamm adds over the minions'    laughter, \"the future is in good hands.\"  <\/p>\n<p>    Here are the other doings in AI worth your attention.  <\/p>\n<p>    Twenty tech companies, including Adobe, Amazon, Anthropic,    ElevenLabs, Google, IBM, Meta, Microsoft, OpenAI, Snap, TikTok    and X, agreed at a security conference in    Munich that they will voluntarily adopt \"reasonable    precautions\" to guard against AI tools being used to mislead or    deceive voters ahead of elections.  <\/p>\n<p>    \"The intentional and undisclosed generation and distribution of    Deceptive AI Election content can deceive the public in ways    that jeopardize the integrity of electoral processes,\" the text    of the accord says, according    to NPR. \"We affirm that the protection of electoral    integrity and public trust is a shared responsibility and a    common good that transcends partisan interests and national    borders.\"  <\/p>\n<p>    But the agreement is \"largely symbolic,\" the    Associated Press reported, noting that \"reasonable    precautions\" is a little vague.  <\/p>\n<p>    \"The companies aren't committing to ban or remove deepfakes,\"    the AP said. \"Instead, the accord outlines methods they will    use to try to detect and label deceptive AI content when it is    created or distributed on their platforms. It notes the    companies will share best practices with each other and provide    'swift and proportionate responses' when that content starts to    spread.\"  <\/p>\n<p>    AI has already been used to try to trick voters in the US and    abroad. Days before the New Hampshire presidential primary,    fraudsters sent an AI robocall that mimicked President Biden's voice, asking them not to    vote in the primary. That prompted the Federal Communications    Commission this month to make AI-generated robocalls illegal.    The AP said that \"Just days before Slovakia's    elections in November, AI-generated audio recordings    impersonated a candidate discussing plans to raise beer prices    and rig the election. Fact-checkers scrambled to identify them    as false as they spread across social media.\"  <\/p>\n<p>    \"Everybody recognizes that no one tech company, no one    government, no one civil society organization is able to deal    with the advent of this technology and its possible nefarious    use on their own,\" Nick Clegg, president of global affairs for    Meta, told the    Associated Press in an interview before the summit.  <\/p>\n<p>    Over 4 billion people are set to vote in key elections this    year in more than 40 countries,. including the US, The Hill    reported.  <\/p>\n<p>    If you're concerned about how deepfakes may be used to scam you    or your family members -- someone calls your grandfather and    asks them for money by pretending to be you -- Bloomberg    reporter Rachel Metz has a good idea. She suggests it may be    time for us all to create a \"family    password\" or safe word or phrase to share with our family    or personal network that we can ask for to make sure we're    talking to who we think we're talking to.  <\/p>\n<p>    \"Extortion has never been easier,\" Metz reports. \"The kind of    fakery that used to take time, money and technical know-how can    now be accomplished quickly    and cheaply by nearly anyone.\"  <\/p>\n<p>    That's where family passwords come in, since they're \"simple    and free,\" Metz said. \"Pick a word that you and your family (or    another trusted group) can easily remember. Then, if one of    those people reaches out in a way that seems a bit odd -- say,    they're suddenly asking you to deliver 5,000 gold bars to a    P.O. Box in Alaska -- first ask them what the password is.\"  <\/p>\n<p>    How do you pick a good password? She offers a few suggestions,    including using a word you don't say frequently and that's not    likely to come up in casual conversations. Also, \"avoid making    the password the name of a pet, as those are easily guessable.\"  <\/p>\n<p>    Hiring experts have told me it's going to take years to build an AI-educated workforce,    considering that gen AI tools like ChatGPT weren't released    until late 2022. So it makes sense that learning platforms like    Coursera, Udemy,    Udacity, Khan Academy and many universities    are offering online courses and certificates to upskill today's    workers. Now the University of Pennsylvania's School of    Engineering and Applied Science said it's the first Ivy League school to offer an    undergraduate major in AI.  <\/p>\n<p>    \"The rapid rise of generative AI is transforming virtually    every aspect of life: health, energy, transportation, robotics,    computer vision, commerce, learning and even national    security,\" Penn said    in a Feb. 13 press release. \"This produces an urgent need    for innovative, leading-edge AI engineers who understand the    principles of AI and how to apply them in a responsible and    ethical way.\"  <\/p>\n<p>    The bachelor of science in AI offers coursework in machine    learning, computing algorithms, data analytics and advanced    robotics and will have students address questions about \"how to    align AI with our social values and how to build trustworthy AI    systems,\" Penn professor Zachary Ives said.  <\/p>\n<p>    \"We are training students for jobs that don't yet exist in    fields that may be completely new or revolutionized by the time    they graduate,\" added Robert    Ghrist, associate dean of undergraduate education in Penn    Engineering.  <\/p>\n<p>    FYI, the cost of an undergraduate education at Penn, which    typically spans four years, is over $88,000    per year (including housing and food).  <\/p>\n<p>    For those not heading to college or who haven't signed up for    any of those online AI certificates, their AI upskilling may    come courtesy of their current employee. The Boston Consulting    Group, for its Feb. 9 report, What    GenAI's Top Performer Do Differently, surveyed over 150    senior executives across 10 sectors. Generally:  <\/p>\n<p>    Bottom line: companies are starting to look at existing job    descriptions and career trajectories, and the gaps they're    seeing in the workforce when they consider how gen AI will    affect their businesses. They've also started offering gen AI    training programs. But these efforts don't lessen the need for    today's workers to get up to speed on gen AI and how it may    change the way they work -- and the work they do.  <\/p>\n<p>    In related news, software    maker SAP looked at Google search data to see which states    in the US were most interested in \"AI jobs and AI business    adoption.\"  <\/p>\n<p>    Unsurprisingly, California ranked first in searches for \"open    AI jobs\" and \"machine learning jobs.\" Washington state came in    second place, Vermont in third, Massachusetts in fourth and    Maryland in fifth.  <\/p>\n<p>    California, \"home to Silicon Valley and renowned as a global    tech hub, shows a significant interest in AI and related    fields, with 6.3% of California's businesses saying that they    currently utilize AI technologies to produce goods and services    and a further 8.4% planning to implement AI in the next six    months, a figure that is 85% higher than the national average,\"    the study found.  <\/p>\n<p>    Virginia, New York, Delaware, Colorado and New Jersey, in that    order, rounded out the top 10.  <\/p>\n<p>    Over the past few months, I've highlighted terms you should    know if you want to be knowledgeable about what's happening as    it relates to gen AI. So I'm going to take a step back this    week and provide this vocabulary review for you, with a link to    the source of the definition.  <\/p>\n<p>    It's worth a few minutes of your time to know these seven    terms.  <\/p>\n<p>    Anthropomorphism: The tendency for people    to attribute humanlike qualities or characteristics to an AI    chatbot. For example, you may assume it's kind or cruel based    on its answers, even though it isn't capable of having    emotions, or you may believe the AI is sentient because it's    very good at mimicking human language.  <\/p>\n<p>    Artificial    general intelligence (AGI): A description of programs that    are as capable as -- or even more capable than -- than a human.    While full general intelligence is still off in the    future, models are growing in sophistication. Some have demonstrated skills    across multiple domains ranging from chemistry to    psychology, with task performance paralleling human    benchmarks.  <\/p>\n<p>    Generative    artificial intelligence (gen AI): Technology that creates    content -- including text, images, video and computer code --    by identifying patterns in large quantities of training data    and then creating original material that has similar    characteristics.  <\/p>\n<p>    Hallucination: Hallucinations are    unexpected and incorrect responses from AI programs that can    arise for reasons that aren't yet fully known. A language model    might suddenly bring up fruit salad recipes when you were    asking about planting fruit trees. It might also make up    scholarly citations, lie about data you ask it to analyze or    make up facts about events that aren't in its training data.    It's not fully understood why this happens but can arise from    sparse data, information gaps and misclassification.  <\/p>\n<p>    Large language model (LLM):    A type of AI model that can generate human-like text and is    trained on a broad dataset.  <\/p>\n<p>    Prompt    engineering: This is the act of giving AI an instruction so    it has the context it needs to achieve your goal. Prompt    engineering is best associated with OpenAI's ChatGPT,    describing the tasks users feed into the algorithm. (e.g. \"Give    me five popular baby names.\")  <\/p>\n<p>    Temperature: In simple terms, model    temperature is a parameter that controls how random a language    model's output is. A higher temperature means the model takes    more risks, giving you a diverse mix of words. On the other    hand, a lower temperature makes the model play it safe,    sticking to more focused and predictable responses.  <\/p>\n<p>    Model temperature has a big impact on the quality of the text    generated in a bunch of [natural language processing] tasks,    like text generation, summarization and translation.  <\/p>\n<p>    The tricky part is finding the perfect model temperature for a    specific task. It's kind of like Goldilocks trying to find the    perfect bowl of porridge -- not too hot, not too cold, but just    right. The optimal temperature depends on things like how    complex the task is and how much creativity you're looking for    in the output.  <\/p>\n<p>    Editors' note: CNET is using an AI engine to help create    some stories. For more, seethis post.  <\/p>\n<p><!-- Auto Generated --><\/p>\n<p>Read the original post: <\/p>\n<p><a target=\"_blank\" rel=\"nofollow noopener\" href=\"https:\/\/www.cnet.com\/tech\/computing\/ai-and-you-openais-sora-previews-text-to-video-future-first-ivy-league-ai-degree\/\" title=\"AI and You: OpenAI's Sora Previews Text-to-Video Future, First Ivy League AI Degree - CNET\">AI and You: OpenAI's Sora Previews Text-to-Video Future, First Ivy League AI Degree - CNET<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p> AI developments are happening pretty fast. If you don't stop and look around once in a while, you could miss them <a href=\"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/ai\/ai-and-you-openais-sora-previews-text-to-video-future-first-ivy-league-ai-degree-cnet.php\">Continue reading <span class=\"meta-nav\">&rarr;<\/span><\/a><\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"limit_modified_date":"","last_modified_date":"","_lmt_disableupdate":"","_lmt_disable":"","footnotes":""},"categories":[1234935],"tags":[],"class_list":["post-1075651","post","type-post","status-publish","format-standard","hentry","category-ai"],"modified_by":null,"_links":{"self":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts\/1075651"}],"collection":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/comments?post=1075651"}],"version-history":[{"count":0,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts\/1075651\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/media?parent=1075651"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/categories?post=1075651"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/tags?post=1075651"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}