{"id":1118692,"date":"2023-10-18T02:23:30","date_gmt":"2023-10-18T06:23:30","guid":{"rendered":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/uncategorized\/nvidias-banking-on-tensorrt-to-expand-its-generative-ai-dominance-the-verge\/"},"modified":"2023-10-18T02:23:30","modified_gmt":"2023-10-18T06:23:30","slug":"nvidias-banking-on-tensorrt-to-expand-its-generative-ai-dominance-the-verge","status":"publish","type":"post","link":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/ai\/nvidias-banking-on-tensorrt-to-expand-its-generative-ai-dominance-the-verge\/","title":{"rendered":"Nvidia&#8217;s banking on TensorRT to expand its generative AI dominance &#8211; The Verge"},"content":{"rendered":"<p><p>    Nvidia     announced that its adding support for its TensorRT-LLM SDK    to Windows and models like Stable Diffusion as it aims to make    large language models (LLMs) and related tools run faster.    TensorRT speeds up inference, the process of going through    pretrained information and calculating probabilities to come up    with a result  like a newly generated Stable Diffusion image.    With this software, Nvidia wants to play a bigger part on that    side of generative AI.  <\/p>\n<p>    TensorRT-LLM breaks down LLMs like Metas    Llama 2 and other AI models like     Stability AIs Stable Diffusion to let them run faster on    Nvidias H100 GPUs. The company said that by running LLMs    through TensorRT-LLM, this acceleration significantly improves    the experience for more sophisticated LLM use  like writing    and coding assistants.  <\/p>\n<p>    This way Nvidia can not only provide the GPUs that train and    run LLMs but also provide the software that allows models to    run and work faster so users dont seek other ways to make    generative AI cost-efficient. The company said TensorRT-LLM    will be available publicly to anyone who wants to use or    integrate it and can access the SDK on its site.  <\/p>\n<p>    Nvidia already has a near monopoly on the powerful chips that    train LLMs like GPT-4 and to train and run one, you    typically need a lot of GPUs. Demand has skyrocketed for its    H100 GPUs; estimated prices have reached $40,000 per chip. The    company announced    a newer version of its GPU, the GH200, coming next year. No    wonder Nvidias revenues    increased to $13.5 billion in the second quarter.  <\/p>\n<p>    But the world of generative AI moves fast, and new methods to    run LLMs without needing a lot of expensive GPUs have come out.    Companies like    Microsoft and     AMD announced theyll make their own chips to lessen the    reliance on Nvidia.  <\/p>\n<p>    And companies have set their sights on the inference side of AI    development. AMD plans to buy software     company Nod.ai to help LLMs specifically run on AMD chips,    while companies like SambaNova    already offer services that make it easier to run models as    well.  <\/p>\n<p>    Nvidia, for now, remains the hardware leader in generative AI,    but it already looks like its angling for a future where    people dont have to depend on buying huge numbers of its    GPUs.  <\/p>\n<p><!-- Auto Generated --><\/p>\n<p>Continued here:<\/p>\n<p><a target=\"_blank\" rel=\"nofollow noopener\" href=\"https:\/\/www.theverge.com\/2023\/10\/17\/23920945\/nvidia-gpus-tensor-llms-ai\" title=\"Nvidia's banking on TensorRT to expand its generative AI dominance - The Verge\">Nvidia's banking on TensorRT to expand its generative AI dominance - The Verge<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p> Nvidia announced that its adding support for its TensorRT-LLM SDK to Windows and models like Stable Diffusion as it aims to make large language models (LLMs) and related tools run faster. TensorRT speeds up inference, the process of going through pretrained information and calculating probabilities to come up with a result like a newly generated Stable Diffusion image.  <a href=\"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/ai\/nvidias-banking-on-tensorrt-to-expand-its-generative-ai-dominance-the-verge\/\">Continue reading <span class=\"meta-nav\">&rarr;<\/span><\/a><\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[187743],"tags":[],"class_list":["post-1118692","post","type-post","status-publish","format-standard","hentry","category-ai"],"_links":{"self":[{"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/posts\/1118692"}],"collection":[{"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/comments?post=1118692"}],"version-history":[{"count":0,"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/posts\/1118692\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/media?parent=1118692"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/categories?post=1118692"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/tags?post=1118692"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}