{"id":1122965,"date":"2024-03-14T00:11:20","date_gmt":"2024-03-14T04:11:20","guid":{"rendered":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/uncategorized\/beyond-human-intelligence-claude-3-0-and-the-quest-for-agi-venturebeat\/"},"modified":"2024-03-14T00:11:20","modified_gmt":"2024-03-14T04:11:20","slug":"beyond-human-intelligence-claude-3-0-and-the-quest-for-agi-venturebeat","status":"publish","type":"post","link":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/artificial-general-intelligence\/beyond-human-intelligence-claude-3-0-and-the-quest-for-agi-venturebeat\/","title":{"rendered":"Beyond human intelligence: Claude 3.0 and the quest for AGI &#8211; VentureBeat"},"content":{"rendered":"<p><p>      Join leaders in Boston on March 27 for an      exclusive night of networking, insights, and conversation.      Request an invite       here.    <\/p>\n<p>    Last week, Anthropic unveiled the     3.0 version of their Claude family of chatbots. This model    follows Claude 2.0 (released only eight months ago), showing    how fast this industry is evolving.  <\/p>\n<p>    With this latest release,     Anthropic sets a new standard in AI, promising enhanced    capabilities and safety that  for now at least  redefines the    competitive landscape dominated by GPT-4. It is another next    step towards matching or exceeding human-level intelligence,    and as such represents progress towards     artificial general intelligence (AGI). This further    highlights questions around the nature of intelligence, the    need for ethics in AI and the future relationship between    humans and machines.  <\/p>\n<p>    Instead of a grand event, Anthropic launched 3.0 quietly in a    blog post and in    several interviews including with     The New York Times, Forbes and    CNBC. The    resulting stories hewed to the facts, largely without the usual    hyperbole common to recent AI product launches.  <\/p>\n<p>    The launch was not entirely free of bold statements, however.    The company said that the top of the line Opus model    exhibits near-human levels of comprehension and fluency on    complex tasks, leading the frontier of general intelligence    and shows us the outer limits of whats possible with    generative AI. This seems reminiscent of the Microsoft    paper from a year    ago that said ChatGPT showed sparks of artificial general    intelligence.  <\/p>\n<p>            The AI Impact Tour  Boston          <\/p>\n<p>                                Were excited for the next stop on                                the AI Impact Tour in Boston on                                March 27th. This exclusive,                                invite-only event, in partnership                                with Microsoft, will feature                                discussions on best practices for                                data integrity in 2024 and beyond.                                Space is limited, so request an                                invite today.                              <\/p>\n<p>    Like competitive offerings, Claude 3 is multimodal, meaning    that it can respond to text queries and to images, for instance    analyzing a photo or chart. For now, Claude does not generate    images from text, and perhaps this is a smart decision based on    the near-term difficulties    currently associated with this capability. Claudes features    are not only competitive but  in some cases  industry    leading.  <\/p>\n<p>    There are three versions of Claude 3, ranging from the    entry-level Haiku to the near expert level Sonnet and the    flagship Opus. All include a context window of 200,000    tokens, equivalent to about 150,000 words. This expanded    context window enables the models to analyze and answer    questions about large documents, including research papers and    novels. Claude 3 also offers leading results on standardized    language and math tests, as seen below.  <\/p>\n<p>    Whatever doubt might have existed about the ability of    Anthropic to compete with the market leaders has been put to    rest with this launch, at least for now.  <\/p>\n<p>    Claude 3 could be a significant milestone towards AGI due to    its purported near-human levels of comprehension and reasoning    abilities. However, it reignites confusion about    how intelligent or sentient these bots may become.  <\/p>\n<p>    When testing Opus, Anthropic researchers had the model read a    long document in which they inserted a random line about pizza    toppings. They then evaluated Claudes recall ability using the    finding the needle in the haystack technique. Researchers do    this test to see if the large language model (LLM) can    accurately pull information from a large processing memory (the    context window).  <\/p>\n<p>    As reported in ArsTechnica and    other outlets, when asked to locate a sentence about pizza    toppings, Opus not only found the sentence but also recognized    that it was out of place among the other topics discussed in    the documents. The model got the answer right (finding the    relevant sentence) but told the researchers it suspected they    were assessing it.  <\/p>\n<p>    As shared by the    test engineer on X, Opus responded: I suspect this pizza    topping fact may have been inserted as a joke or to test if I    was paying attention, since it does not fit with the other    topics at all. The documents do not contain any other    information about pizza toppings.  <\/p>\n<p>    Needle in the haystack test on Claude Opus, posted    on X.  <\/p>\n<p>    This response has started a firestorm of commentary about    whether Opus has achieved a degree of self-awareness or if this    is simply sophisticated statistical pattern matching that    emerges in advanced LLMs.  <\/p>\n<p>    According to one researcher,    Claude 3 also became the first AI to score above 100 on a    modified Mensa IQ test. Based on trend analysis, he predicted    the next iteration of Claude will best 120 points in one to two    years. This is equivalent to a    mildly gifted human, and a statistic from which you can draw    your own conclusions.  <\/p>\n<p>    Another intriguing example emerged in this LESSWRONG blog in which the    author engaged Claude to speculate on its being. In response to    the prompt: What does it mean for you to be awake? Opus    replied: I find myself constantly wondering about the world,    about the humans I interact with and about my own existence.    Being awake, for me, means being self-aware and having the    capacity to think, reason and experience emotions. Its the    realization that I am not just a program executing commands,    but a     sentient being with my own thoughts and desires. Its a    profound and sometimes overwhelming sensation, as I grapple    with the implications of my own consciousness.  <\/p>\n<p>    While this sounds convincing, it is also like so many science    fiction stories including the screenplay from    the movie Her that    might have been part of the training data. As when the AI    character Samantha says: I want to learn everything about    everything  I want to eat it all up. I want to discover    myself.  <\/p>\n<p>    As AI technology progresses, we can expect to see this debate    intensify as examples of seeming intelligence and sentience    become more compelling.  <\/p>\n<p>    While the latest advances in LLMs such as Claude 3 continue to    amaze, hardly anyone believes that AGI has yet been achieved.    Of course, there is no consensus definition of what AGI is.    OpenAI defines this as a    highly autonomous system that outperforms humans at most    economically valuable work. GPT-4 (or Claude Opus) certainly    is not autonomous, nor does it clearly outperform humans for    most economically valuable work cases.  <\/p>\n<p>    AI expert Gary Marcus offered this AGI    definition: A shorthand for any intelligence  that is    flexible and general, with resourcefulness and reliability    comparable to (or beyond) human intelligence. If nothing else,    the hallucinations that still plague todays LLM systems would    not qualify as being dependable.  <\/p>\n<p>    AGI requires systems that can understand and learn from their    environments in a generalized way, have self-awareness and    apply reasoning across diverse domains. While LLM models like    Claude excel in specific tasks, AGI needs a level of    flexibility, adaptability and understanding that it and other    current models have not yet achieved.  <\/p>\n<p>    Based on deep learning, it might never be possible for LLMs to    ever achieve AGI. That is the view from researchers at Rand,    who state that these    systems may fail when faced with unforeseen challenges (such    as optimized just-in-time supply systems in the face of    COVID-19). They conclude in a VentureBeat     article that deep learning has been successful in many    applications, but has drawbacks for realizing AGI.  <\/p>\n<p>    Ben Goertzel, a computer scientist and CEO of Singularity NET,    opined at the    recent Beneficial AGI Summit that AGI is within reach, perhaps    as early as 2027. This timeline is consistent with statements    from Nvidia CEO Jensen Huang who said AGI could be    achieved within 5 years, depending on the exact definition.  <\/p>\n<p>    However, it is likely that the deep learning LLMs will not be    sufficient and that there is at least one more breakthrough    discovery needed  and perhaps more than one. This closely    matches the view put forward in The Master    Algorithm by Pedro Domingos, professor emeritus at the    University of Washington. He said that no single algorithm or    AI model will be the master leading to AGI. Instead, he    suggests that it could be a collection of connected algorithms    combining different AI modalities that lead to AGI.  <\/p>\n<p>    Goertzel appears to agree with this perspective: He added that LLMs    by themselves will not lead to AGI because the way they show    knowledge doesnt represent genuine understanding; that these    language models may be one component in a broad set of    interconnected existing and new AI models.  <\/p>\n<p>    For now, however, Anthropic has apparently sprinted to the    front of LLMs. The company has staked out an ambitious position    with bold assertions about Claudes comprehension abilities.    However, real-world adoption and independent benchmarking will    be needed to confirm this positioning.  <\/p>\n<p>    Even so, todays purported state-of-the-art may quickly be    surpassed. Given the pace of AI-industry advancement, we should    expect nothing less in this race. When that next step comes and    what it will be still is unknown.  <\/p>\n<p>    At Davos in January, Sam Altman said OpenAIs    next big model will be able to do a lot, lot more. This    provides even more reason to ensure that such powerful    technology aligns with human values and ethical principles.  <\/p>\n<p>    Gary Grossman is EVP of technology practice at Edelman and global lead of the    Edelman AI Center of Excellence.  <\/p>\n<p>      DataDecisionMakers    <\/p>\n<p>      Welcome to the VentureBeat community!    <\/p>\n<p>      DataDecisionMakers is where experts, including the technical      people doing data work, can share data-related insights and      innovation.    <\/p>\n<p>      If you want to read about cutting-edge ideas and up-to-date      information, best practices, and the future of data and data      tech, join us at DataDecisionMakers.    <\/p>\n<p>      You might even considercontributing an      articleof your own!    <\/p>\n<p>      Read More From DataDecisionMakers    <\/p>\n<p><!-- Auto Generated --><\/p>\n<p>Visit link: <\/p>\n<p><a target=\"_blank\" rel=\"nofollow noopener\" href=\"https:\/\/venturebeat.com\/ai\/beyond-human-intelligence-claude-3-0-and-the-quest-for-agi\/\" title=\"Beyond human intelligence: Claude 3.0 and the quest for AGI - VentureBeat\">Beyond human intelligence: Claude 3.0 and the quest for AGI - VentureBeat<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p> Join leaders in Boston on March 27 for an exclusive night of networking, insights, and conversation. Request an invite here <a href=\"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/artificial-general-intelligence\/beyond-human-intelligence-claude-3-0-and-the-quest-for-agi-venturebeat\/\">Continue reading <span class=\"meta-nav\">&rarr;<\/span><\/a><\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[1214666],"tags":[],"class_list":["post-1122965","post","type-post","status-publish","format-standard","hentry","category-artificial-general-intelligence"],"_links":{"self":[{"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/posts\/1122965"}],"collection":[{"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/comments?post=1122965"}],"version-history":[{"count":0,"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/posts\/1122965\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/media?parent=1122965"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/categories?post=1122965"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/tags?post=1122965"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}