{"id":169668,"date":"2024-06-28T02:39:40","date_gmt":"2024-06-28T06:39:40","guid":{"rendered":"https:\/\/www.immortalitymedicine.tv\/iii-artificial-intelligence-and-the-economy-implications-for-central-banks-bis-org\/"},"modified":"2024-08-18T12:47:44","modified_gmt":"2024-08-18T16:47:44","slug":"iii-artificial-intelligence-and-the-economy-implications-for-central-banks-bis-org","status":"publish","type":"post","link":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/artificial-intelligence\/iii-artificial-intelligence-and-the-economy-implications-for-central-banks-bis-org.php","title":{"rendered":"III. Artificial intelligence and the economy: implications for central banks &#8211; bis.org"},"content":{"rendered":"<p><p>Key takeaways        <\/p>\n<p>    The advent of large language models (LLMs) has catapulted    generative artificial intelligence (gen AI) into popular    discourse. LLMs have transformed the way people interact with    computers  away from code and programming interfaces to    ordinary text and speech. This ability to converse through    ordinary language as well as gen AI's human-like capabilities    in creating content have captured our collective imagination.  <\/p>\n<p>    Below the surface, the underlying mathematics of the latest AI    models follow basic principles that would be familiar to    earlier generations of computer scientists. Words or sentences    are converted into arrays of numbers, making them amenable to    arithmetic operations and geometric manipulations that    computers excel at.  <\/p>\n<p>    What is new is the ability to bring mathematical order at scale    to everyday unstructured data, whether they be text, images,    videos or music. Recent AI developments have been enabled by    two factors. First is the accumulation of vast reservoirs of    data. The latest LLMs draw on the totality of textual and    audiovisual information available on the internet. Second is    the massive computing power of the latest generation of    hardware. These elements turn AI models into highly refined    prediction machines, possessing a remarkable ability to detect    patterns in data and fill in gaps.  <\/p>\n<p>    There is an active debate on whether enhanced pattern    recognition is sufficient to approximate \"artificial general    intelligence\" (AGI), rendering AI with full human-like    cognitive capabilities. Irrespective of whether AGI can be    attained, the ability to impose structure on unstructured data    has already unlocked new capabilities in many tasks that eluded    earlier generations of AI tools.1 The new generation of AI models could be a    game changer for many activities and have a profound impact on    the broader economy and the financial system. Not least, these    same capabilities can be harnessed by central banks in pursuit    of their policy objectives, potentially transforming key areas    of their operations.  <\/p>\n<p>    The economic potential of AI has set off a gold rush across the    economy. The adoption of LLMs and gen AI tools is proceeding at    such breathtaking speed that it easily outpaces previous waves    of technology adoption (Graph 1.A). For example, ChatGPT alone    reached one million users in less than a week and nearly half    of US households have used gen AI tools in the past 12 months.    Mirroring rapid adoption by users, firms are already    integrating AI in their daily operations: global survey    evidence suggests firms in all industries use gen AI tools    (Graph 1.B). To    do so, they are investing heavily in AI technology to tailor it    to their specific needs and have embarked on a hiring spree of    workers with AI-related skills (Graph 1.C). Most firms expect these trends to    only accelerate.2  <\/p>\n<p>    This chapter lays out the implications of these developments    for central banks, which impinge on them in two important ways.  <\/p>\n<p>    First, AI will influence central banks' core activities as    stewards of the economy. Central bank mandates revolve around    price and financial stability. AI will affect financial systems    as well as productivity, consumption, investment and labour    markets, which themselves have direct effects on price and    financial stability. Widespread adoption of AI could also    enhance firms' ability to quickly adjust prices in response to    macroeconomic changes, with repercussions for inflation    dynamics. These developments are therefore of paramount concern    to central banks.  <\/p>\n<p>    Second, the use of AI will have a direct bearing on the    operations of central banks through its impact on the financial    system. For one, financial institutions such as commercial    banks increasingly employ AI tools, which will change how they    interact with and are supervised by central banks. Moreover,    central banks and other authorities are likely to increasingly    use AI in pursuing their missions in monetary policy,    supervision and financial stability.  <\/p>\n<p>    Overall, the rapid and widespread adoption of AI implies that    there is an urgent need for central banks to raise their game.    To address the new challenges, central banks need to upgrade    their capabilities both as informed observers of the    effects of technological advancements as well as users    of the technology itself. As observers, central banks need to    stay ahead of the impact of AI on economic activity through its    effects on aggregate supply and demand. As users, they need to    build expertise in incorporating AI and non-traditional data in    their own analytical tools. Central banks will face important    trade-offs in using external vs internal AI models, as well as    in collecting and providing in-house data vs purchasing them    from external providers. Together with the centrality of data,    the rise of AI will require a rethink of central banks'    traditional roles as compilers, users and providers of data. To    harness the benefits of AI, collaboration and the sharing of    experiences emerge as key avenues for central banks to mitigate    these trade-offs, in particular by reducing the demands on    information technology (IT) infrastructure and human capital.    Central banks need to come together to form a \"community of    practice\" to share knowledge, data, best practices and AI    tools.  <\/p>\n<p>    The chapter starts with an overview of developments in AI,    providing a deep dive into the underlying technology. It then    examines the implications of the rise of AI for the financial    sector. The discussion includes current use cases of AI by    financial institutions and implications for financial    stability. It also outlines the emerging opportunities and    challenges and the implications for central banks, including    how they can harness AI to fulfil their policy objectives. The    chapter then discusses how AI affects firms' productive    capacity and investment, as well as labour markets and    household consumption, and how these changes in aggregate    demand and supply affect inflation dynamics. The chapter    concludes by examining the trade-offs arising from the use of    AI and the centrality of data for central banks and regulatory    authorities. In doing so, it highlights the urgent need for    central banks to cooperate.  <\/p>\n<p>    Artificial intelligence is a broad term, referring to computer    systems performing tasks that require human-like intelligence.    While the roots of AI can be traced back to the late 1950s, the    advances in the field of machine learning in    the 1990s laid the foundations of the current generation of AI    models. Machine learning is a collective term referring to    techniques designed to detect patterns in the data and use them    in prediction or to aid decision-making.3  <\/p>\n<p>    The development of deep learning in the 2010s    constituted the next big leap. Deep learning uses neural    networks, perhaps the most important technique in machine    learning, underpinning everyday applications such as facial    recognition or voice assistants. The main building block of    neural networks is artificial neurons, which take    multiple input values and transform them to output as a set of    numbers that can be readily analysed. The artificial neurons    are organised to form a sequence of layers that can be    stacked: the neurons of the first layer take the input data and    output an activation value. Subsequent layers then take the    output of the previous layer as input, transform it and output    another value, and so forth. A network's depth refers    to the number of layers. More layers allow neural networks to    capture increasingly complex relationships in the data. The    weights determining the strength of connections between    different neurons and layers are collectively called    parameters, which are improved (known as    learning) iteratively during training. Deeper networks    with more parameters require more training data but predict    more accurately.  <\/p>\n<p>    A key advantage of deep learning models is their ability to    work with unstructured data. They achieve this by    \"embedding\" qualitative, categorical or visual    data, such as words, sentences, proteins or images, into arrays    of numbers  an approach pioneered at scale by the Word2Vec    model (see Box A). These arrays of numbers    (ie vectors) are interpreted as points in a vector space. The    distance between vectors conveys some dimension of similarity,    enabling algebraic manipulations on what is originally    qualitative data. For example, the vector linking the    embeddings of the words \"big\" and \"biggest\" is very similar to    that between \"small\" and \"smallest\". Word2Vec predicts a word    based on the surrounding words in a sentence. The body of text    used for the embedding exercise is drawn from the open internet    through the \"common crawl\" database. The concept of embedding    can be taken further into mapping the space of economic ideas,    uncovering latent viewpoints or methodological approaches of    individual economists or institutions    (\"personas\"). The space of ideas can be linked    to concrete policy actions, including monetary policy    decisions.4  <\/p>\n<p>    The advent of LLMs allows neural networks to    access the whole context of a word rather than just its    neighbour in the sentence. Unlike Word2Vec, LLMs can now    capture the nuances of translating uncommon languages, answer    ambiguous questions or analyse the sentiment of texts. LLMs are    based on the transformer model (see Box B). Transformers rely on \"multi-headed    attention\" and \"positional encoding\" mechanisms to efficiently    evaluate the context of any word in the document. The context    influences how words with multiple meanings map into arrays of    numbers. For example, \"bond\" could refer to a fixed income    security, a connection or link, or a famous espionage    character. Depending on the context, the \"bond\" embedding    vector lies geometrically closer to words such as \"treasury\",    \"unconventional\" and \"policy\"; to \"family\" and \"cultural\"; or    to \"spy\" and \"martini\". These developments have enabled AI to    move from narrow systems that solve one specific task to more    general systems that deal with a wide range of tasks.  <\/p>\n<p>    LLMs are a leading example of gen AI    applications because of their capacity to understand    and generate accurate responses with minimal or even no prior    examples (so-called few-shot or zero-shot learning abilities).    Gen AI refers to AIs capable of generating content, including    text, images or music, from a natural language prompt. The    prompts contain instructions in plain language or examples of    what users want from the model. Before LLMs, machine learning    models were trained to solve one task (eg image classification,    sentiment analysis or translating from French to English). It    required the user to code, train and roll out the model into    production after acquiring sufficient training data. This    procedure was possible for only selected companies with    researchers and engineers with specific skills. An LLM has    few-shot learning abilities in that it can be given a task in    plain language. There is no need for coding, training or    acquiring training data. Moreover, it displays considerable    versatility in the range of tasks it can take on. It can be    used to first classify an image, then analyse the sentiment of    a paragraph and finally translate it into any language.    Therefore, LLMs and gen AI have enabled people using ordinary    language to automate tasks that were previously performed by    highly specialised models.  <\/p>\n<p>    The capabilities of the most recent crop of AI models are    underpinned by advances in data and    computing power. The increasing availability    of data plays a key role in training and improving models. The    more data a model is trained on, the more capable it usually    becomes. Furthermore, machine learning models with more    parameters improve predictions when trained with sufficient    data. In contrast to the previous conventional wisdom that    \"over-parameterisation\" degrades the forecasting ability of    models, more recent evidence points to a remarkable resilience    of machine learning models to over-parameterisation. As a    consequence, LLMs with well designed learning mechanisms can    provide more accurate predictions than traditional parametric    models in diverse scenarios such as computer vision, signal    processing and natural language processing (NLP).5  <\/p>\n<p>    An implication is that more capable models tend to be larger    models that need more data. Bigger models and larger data sets    therefore go together and increase computational demands. The    use of advanced techniques on vast troves of data would not    have been possible without substantial increases in computing    power  in particular, the computational resources employed by    AI systems  which has been doubling every six    months.6The    interplay between large amounts of data and computational    resources implies that just a handful of companies provide    cutting-edge LLMs, an issue revisited later in the chapter.  <\/p>\n<p>    Some commentators have argued that AI has the potential to    become the next general-purpose technology,    profoundly impacting the economy and society. General-purpose    technologies, like electricity or the internet, eventually    achieve widespread usage, give rise to versatile applications    and generate spillover effects that can improve other    technologies. The adoption pattern of general-purpose    technologies typically follows a J-curve: it is slow at first,    but eventually accelerates. Over time, the pace of technology    adoption has been speeding up. While it took electricity or the    telephone decades to reach widespread adoption, smartphones    accomplished the same in less than a decade. AI features two    distinct characteristics that suggest an even steeper J-curve.    First is its remarkable speed of adoption, reflecting ease of    use and negligible cost for users. Second is its widespread use    at an early stage by households as well as firms in all    industries.  <\/p>\n<p>    Of course, there is substantial uncertainty about the long-term    capabilities of gen AI. Current LLMs can fail    elementary logical reasoning tasks and struggle with    counterfactual reasoning, as illustrated in recent BIS    work.7 For example,    when posed with a logical puzzle that demands reasoning about    the knowledge of others and about counterfactuals, LLMs display    a distinctive pattern of failure. They perform flawlessly when    presented with the original wording of a puzzle, which they    have likely seen during their training. They falter when the    same problem is presented with small changes of innocuous    details such as names and dates, suggesting a lack of true    understanding of the underlying logic of statements.    Ultimately, current LLMs do not know what they do not know.    LLMs also suffer from the    hallucination problem: they    can present a factually incorrect answer as if it were correct,    and even invent secondary sources to back up their fake claims.    Unfortunately, hallucinations are a feature rather than a bug    in these models. LLMs hallucinate because they are trained to    predict the statistically plausible word based on some input.    But they cannot distinguish what is linguistically probable    from what is factually correct.  <\/p>\n<p>    Do these problems merely reflect the limits posed by the size    of the training data set and the number of model parameters? Or    do they reflect more fundamental limits to knowledge that is    acquired through language alone? Optimists acknowledge current    limitations but emphasise the potential of LLMs to exceed human    performance in certain domains. In particular, they argue that    terms such as \"reason\", \"knowledge\" and \"learning\" rightly    apply to such models. Sceptics point out the limitations of    LLMs in reasoning and planning. They argue that the main    limitation of LLMs derives from their exclusive reliance on    language as the medium of knowledge. As LLMs are confined to    interacting with the world purely through language, they lack    the tacit non-linguistic, shared understanding that can be    acquired only through active engagement with the real    world.8  <\/p>\n<p>    Whether AI will eventually be able to perform tasks that    require deep logical reasoning has implications for its    long-run economic impact. Assessing which tasks will be    impacted by AI depends on the specific cognitive abilities    required in those tasks. The discussion above suggests that, at    least in the near term, AI faces challenges in reaching    human-like performance. While it may be able to perform tasks    that require moderate cognitive abilities and even develop    \"emergent\" capabilities, it is not yet able to perform tasks    that require logical reasoning and judgment.  <\/p>\n<p>    The financial sector is among those facing the greatest    opportunities and risks from the rise of AI, due to its high    share of cognitively demanding tasks and data-intensive    nature.9 Table 1 illustrates the    impact of AI in four key areas: payments, lending, insurance    and asset management.  <\/p>\n<p>    Across all four areas, AI can substantially enhance    efficiency and lower    costs in back-end processing, regulatory    compliance, fraud detection and customer service. These    activities give full play to the ability of AI models to    identify patterns of interest in seemingly unstructured data.    Indeed, \"finding a needle in the haystack\" is an activity that    plays to the greatest strength of machine learning models. A    striking example is the improvement of know-your-customer (KYC)    processes through quicker data processing and the enhanced    ability to detect fraud, allowing financial institutions to    ensure better compliance with regulations while lowering    costs.10 LLMs are    also increasingly being deployed for customer service    operations through AI chatbots and co-pilots.  <\/p>\n<p>    In payments, the abundance of    transaction-level data enables AI models to overcome    long-standing pain points. A prime example comes from    correspondent banking, which has become a high-risk, low-margin    activity. Correspondent banks played a key role in the    expansion of cross-border payment activity by enabling    transaction settlement, cheque clearance and foreign exchange    operations. Facing heightened customer verification and    anti-money laundering (AML) requirements, banks have    systematically retreated from the business (Graphs 2.A and 2.B). Such    retreat fragments the global payment system by leaving some    regions less connected (Graph 2.C), handicapping their connectivity    with the rest of the financial system. The decline in    correspondent banking is part of a general de-risking trend,    with returns from processing transactions being small compared    with the risks of penalties from breaching AML, KYC and    countering the financing of terrorism (CFT)    requirements.11  <\/p>\n<p>    A key use case of AI models is to improve KYC and AML processes    by enhancing (i) the ability to understand the compliance and    reputational risks that clients might carry, (ii) due diligence    on the counterparties of a transaction and (iii) the analysis    of payment patterns and anomaly detection. By bringing down    costs and reducing risks through greater speed and automation,    AI holds the promise to reverse the decline in correspondent    banking.  <\/p>\n<p>    The ability of AI models to detect patterns in the data is    helping financial institutions address many of these    challenges. For example, financial institutions are using AI    tools to enhance fraud detection and to identify security    vulnerabilities. At the global level, surveys indicate that    around 70% of all financial services firms are using AI to    enhance cash flow predictions and improve liquidity management,    fine-tune credit scores and improve fraud detection.12  <\/p>\n<p>    In credit assessment and lending, banks have    used machine learning for many years, but AI can bring further    capabilities. For one, AI could greatly enhance credit scoring    by making use of unstructured data. In deciding whether to    grant a loan, lenders traditionally rely on standardised credit    scores, at times combined with easily accessible variables such    as loan-to-value or debt-to-income ratios. AI-based tools    enable lenders to assess individuals' creditworthiness with    alternative data. These can include consumers' bank account    transactions or their rental, utility and telecommunications    payments data. But they can also be of a non-financial nature,    for example applicants' educational history or online shopping    habits. The use of non-traditional data can significantly    improve default prediction, especially among underserved groups    for whom traditional credit scores provide an imprecise signal    about default probability. By being better able to spot    patterns in unstructured data and detect \"invisible primes\", ie    borrowers that are of high quality even if their credit scores    indicate low quality, AI can enhance financial    inclusion.13  <\/p>\n<p>    AI has numerous applications in insurance,    particularly in risk assessment and pricing. For example,    companies use AI to automatically analyse images and videos to    assess property damage due to natural disasters or, in the    context of compliance, whether claims of damages correspond to    actual damages. Underwriters, actuaries or claims adjusters    further stand to benefit from AI summarising and synthesising    data gathered during a claim's life cycle, such as call    transcripts and notes, as well as legal and medical paperwork.    More generally, AI is bound to play an increasingly important    role in assessing different types of risks. For example, some    insurance companies are experimenting with AI methods to assess    climate risks by identifying and quantifying emissions based on    aerial images of pollution. However, to the extent that AI is    better at analysing or inferring individual-level    characteristics in risk assessments, including those whose use    is prohibited by regulation, existing inequalities could be    exacerbated  an issue revisited in the discussion on the    macroeconomic impact of AI.  <\/p>\n<p>    In asset management, AI models are used to    predict returns, evaluate risk-return trade-offs and optimise    portfolio allocation. Just as LLMs assign different    characteristics to each word they process, they can be used to    elicit unobservable features of financial data (so-called asset    embeddings). This allows market participants to extract    information (such as firm quality or investor preferences) that    is difficult to discern from existing data. In this way, AI    models can provide a better understanding of the risk-return    properties of portfolios. Models that use asset embeddings can    outperform traditional models that rely only on observable    characteristics of financial data. Separately, AI models are    useful in algorithmic trading, owing to their ability to    analyse large volumes of data quickly. As a result, investors    benefit from quicker and more precise information as well as    lower management fees.14  <\/p>\n<p>    The widespread use of AI applications in the financial sector,    however, brings new challenges. These pertain to cyber security    and operational resilience as well as financial stability.  <\/p>\n<p>    The reliance on AI heightens concerns about cyber    attacks, which regularly feature among the top worries    in the financial industry. Traditionally, phishing emails have    been used to trick a user to run a malicious code (malware) to    take over the user's device. Credential phishing is the    practice of stealing a user's login and password combination by    masquerading as a reputable or known entity in an email,    instant message or another communication channel. Attackers    then use the victim's credentials to carry out attacks on    additional targets and gain further access.15 Gen AI could vastly expand hackers'    ability to write credible phishing emails or to write malware    and use it to steal valuable information or encrypt a company's    files for ransom. Moreover, gen AI allows hackers to imitate    the writing style or voice of individuals, or even create fake    avatars, which could lead to a dramatic rise in phishing    attacks. These developments expose financial institutions and    their customers to a greater risk of fraud.  <\/p>\n<p>    But AI also introduces altogether new sources of cyber    risk. Prompt injection attacks, one of the    most widely reported weaknesses in LLMs, refer to an attacker    creating an input to make the model behave in an unintended    way. For example, LLMs are usually instructed not to provide    dangerous information, such as how to manufacture napalm.    However, in the infamous grandma jailbreak, where the    prompter asked ChatGPT to pretend to be their deceased    grandmother telling a bedtime story about the steps to produce    napalm, the chatbot did reveal this information. While this    vulnerability has been fixed, others remain. Data poisoning    attacks refer to malicious tampering with the data an AI    model is trained on. For example, an attacker could adjust    input data so that the AI model fails to detect phishing    emails. Model poisoning attacks deliberately introduce    malware, manipulating the training process of an AI system to    compromise its integrity or functionality. This attack aims to    alter the model behaviour to serve the attacker's    purposes.16 As more    applications use data created by LLMs themselves, such attacks    could have increasingly severe consequences, leading to    heightened operational risks among financial institutions.  <\/p>\n<p>    Greater use of AI raises issues of bias and    discrimination. Two examples stand out. The first    relates to consumer protection and fair lending practices. As    with traditional models, AI models can reflect biases and    inaccuracies in the data they are trained on, posing risks of    unjust decisions, excluding some groups from socially desirable    insurance markets and perpetuating disparities in access to    credit through algorithmic discrimination.17 Consumers care about these risks:    recent evidence from a representative survey of US households    suggests a lower level of trust in gen AI than in    human-operated services, especially in high-stakes areas such    as banking and public policy (Graph 3.A) and when AI tools are provided by    big techs (Graph    3.B).18 The    second example relates to the challenge of ensuring data    privacy and confidentiality when dealing with growing volumes    of data, another key concern for users (Graph 3.C). In the light of the high    privacy standards that financial institutions need to adhere    to, this heightens legal risks. The lack of    explainability of AI models (ie their black box nature) as well    as their tendency to hallucinate amplify these risks.  <\/p>\n<p>    Another operational risk arises from relying on just a few    providers of AI models, which increases third-party    dependency risks. Market concentration arises from the    centrality of data and the vast costs of developing and    implementing data-hungry models. Heavy up-front investment is    required to build data storage facilities, hire and train    staff, gather and clean data and develop or refine algorithms.    However, once the infrastructure is in place, the cost of    adding each extra unit of data is negligible. This centrality    leads to so-called data gravity: companies that already have an    edge in collecting, storing and analysing data can provide    better-trained AI tools, whose use creates ever more data over    time. The consequence of data gravity is that only a few    companies provide cutting-edge LLMs. Any failure among or cyber    attack on these providers, or their models, poses risks to    financial institutions relying on them.  <\/p>\n<p>    The reliance of market participants on the same handful of    algorithms could lead to financial stability    risks. These could arise from AI's ubiquitous adoption    throughout the financial system and its growing capability to    make decisions independently and without human intervention    (\"automaticity\") at a speed far beyond human capacity. The    behaviour of financial institutions using the same algorithms    could amplify procyclicality and market volatility by    exacerbating herding, liquidity hoarding, runs and fire sales.    Using similar algorithms trained on the same data can also lead    to coordinated recommendations or outright collusive outcomes    that run afoul of regulations against market manipulation, even    if algorithms are not trained or instructed to    collude.19 In    addition, AI may hasten the development and introduction of new    products, potentially leading to new and little understood    risks.  <\/p>\n<p>    Central banks stand at the intersection of the monetary and    financial systems. As stewards of the economy through their    monetary policy mandate, they play a pivotal role in    maintaining economic stability, with a primary objective of    ensuring price stability. Another essential role is to    safeguard financial stability and the payment system. Many    central banks also have a role in supervising and regulating    commercial banks and other participants of the financial    system.  <\/p>\n<p>    Central banks are not simply passive observers in monitoring    the impact of AI on the economy and the financial system. They    can harness AI tools themselves in pursuit of their policy    objectives and in addressing emerging challenges. In    particular, the use of LLMs and AI can support central banks'    key tasks of information collection and statistical    compilation, macroeconomic and financial analysis to support    monetary policy, supervision, oversight of payment systems and    ensuring financial stability. As early adopters of machine    learning methods, central banks are well positioned to reap the    benefits of AI tools.20  <\/p>\n<p>    Data are the major resource that stand to become more valuable    due to the advent of AI. A particularly rich source of data is    the payment system. Such data present an enormous amount of    information on economic transactions, which naturally lends    itself to the powers of AI to detect patterns.21 Dealing with such data    necessitates adequate privacy-preserving techniques and the    appropriate data governance frameworks.  <\/p>\n<p>    The BIS Innovation Hub's Project Aurora explores some of these    issues. Using a synthetic data set emulating money laundering    activities, it compares various machine learning models, taking    into account payment relationships as input. The comparison    occurs under three scenarios: transaction data that are siloed    at the bank level, national-level pooling of data and    cross-border pooling. The models undergo training with known    simulated money laundering transactions and subsequently    predict the likelihood of money laundering in unseen synthetic    data.  <\/p>\n<p>    The project offers two key insights. First, machine learning    models outperform the traditional rule-based methods prevalent    in most jurisdictions. Graph neural networks, in particular,    demonstrate superior performance, effectively leveraging    comprehensive payment relationships available in pooled data to    more accurately identify suspect transaction networks. And    second, machine learning models are particularly effective when    data from different institutions in one or multiple    jurisdictions are pooled, underscoring a premium on    cross-border coordination in AML efforts (Graph 4).  <\/p>\n<p>    The benefits of coordination are further illustrated by Project    Agor. This project gathers seven central banks and private    sector participants to bring tokenised central bank money and    tokenised deposits together on the same programmable platform.  <\/p>\n<p>    The tokenisation built into Agor would allow the platform to    harness three capabilities: (i) combining messaging and account    updates as a single operation; (ii) executing payments    atomically rather than as a series of sequential updates; and    (iii) drawing on privacy-preserving platform resources for    KYC\/AML compliance. In traditional correspondent banking,    information checks and account updates are made sequentially    and independently, with significant duplication of effort    (Graph 5.A). In    contrast, in Agor the contingent performance of actions    enabled by tokenisation allows for the combination of assets,    information, messaging and clearing into a single atomic    operation, eliminating the risk of reversals (Graph 5.B). In turn,    privacy-enhancing data-sharing techniques can significantly    simplify compliance checks, while all existing rules and    regulations are adhered to as part of the pre-screening    process.22  <\/p>\n<p>    In the development of a new payment infrastructure like Agor,    great care must be taken to ensure potential gains are not lost    due to fragmentation. This can be done via access policies to    the infrastructure or via interoperability, as advocated in the    idea of the Finternet. This refers to multiple interconnected    financial ecosystems, much like the internet, designed to    empower individuals and businesses by placing them at the    centre of their financial lives. The Finternet leverages    innovative technologies such as tokenisation and unified    ledgers, underpinned by a robust economic and regulatory    framework, to expand the range and quality of savings and    financial services. Starting with assets that can be easily    tokenised holds the greatest promise in the near    term.23  <\/p>\n<p>    Central banks also see great benefits in using gen AI to    improve cyber security. In a recent BIS survey    of central bank cyber experts, a majority deem gen AI to offer    more benefits than risks (Graph 6.A) and think it can outperform    traditional methods in enhancing cyber security    management.24    Benefits are largely expected in areas such as the automation    of routine tasks, which can reduce the costs of time-consuming    activities traditionally performed by humans (Graph 6.B). But human    expertise will remain important. In particular, data scientists    and cyber security experts are expected to play an increasingly    important role. Additional cyber-related benefits from AI    include the enhancement of threat detection, faster response    times to cyber attacks and the learning of new trends,    anomalies or correlations that might not be obvious to human    analysts. In addition, by leveraging AI, central banks can now    craft and deploy highly convincing phishing attacks as part of    their cyber security training. Project Raven of the BIS    Innovation Hub is one example of the use of AI to enhance cyber    resilience (see Box C).  <\/p>\n<p>    The challenge for central banks in using AI tools comes in two    parts. The first is the availability of timely data, which is a    necessary condition for any machine learning application.    Assuming this issue is solved, the second challenge is to    structure the data in a way that yields insights. This second    challenge is where machine learning tools, and in particular    LLMs, excel. They can transform unstructured data from a    variety of sources into structured form in real time. Moreover,    by converting time series data into tokens resembling textual    sequences, LLMs can be applied to a wide array of time series    forecasting tasks. Just as LLMs are trained to guess the next    word in a sentence using a vast database of textual    information, LLM-based forecasting models use similar    techniques to estimate the next numerical observation in a    statistical series.  <\/p>\n<p>    These capabilities are particularly promising for    nowcasting. Nowcasting is a technique that    uses real-time data to provide timely insights. This method can    significantly improve the accuracy and timeliness of economic    predictions, particularly during periods of heightened market    volatility. However, it currently faces two important    challenges, namely the limited usability of timely data and the    necessity to pre-specify and train models for concrete    tasks.25 LLMs and gen    AI hold promise to overcome both bottlenecks (see Box D). For example, an LLM fine-tuned with    financial news can readily extract information from social    media posts or non-financial firms' and banks' financial    statements or transcripts of earning reports and create a    sentiment index. The index can then be used to nowcast    financial conditions, monitor the build-up of risks or predict    the probability of recessions.26 Moreover, by categorising texts into    specific economic topics (eg consumer demand and credit    conditions), the model can pinpoint the source of changes in    sentiment (eg consumer sentiment or credit risk). Such data are    particularly relevant early in the forecasting process when    traditional hard data are scarce.  <\/p>\n<p>    Beyond financial applications, AI-based nowcasting can also be    useful to understand real-economy developments. For example,    transaction-level data on household-to-firm or firm-to-firm    payments, together with machine learning models, can improve    nowcasting of consumption and investment. Another use case is    measuring supply chain bottlenecks with NLP, eg based on text    in the so-called Beige Book. After classifying sentences    related to supply chains, a deep learning algorithm classifies    the sentiment of each sentence and provides an index that    offers a real-time view of supply chain bottlenecks. Such an    index can be used to predict inflationary pressures. Many more    examples exist, ranging from nowcasting world trade to climate    risks.27  <\/p>\n<p>    Access to granular data can also enhance central banks' ability    to track developments across different industries and    regions. For example, with the help of AI, data from    job postings or online retailers can be used to track wage    developments and employment dynamics across occupations, tasks    and industries. Such a real-time and detailed view of labour    market developments can help central banks understand the    extent of technology-induced job displacements, how quickly    workers find new jobs and attendant wage dynamics. Similarly,    satellite data on aerial pollution or nighttime lights can be    used to predict short-term economic activity, while data on    electricity consumption can shed light on industrial production    in different regions and industries.28 Central banks can thereby obtain a more    nuanced picture of firms' capital expenditure and production,    and how the supply of and demand for goods and services are    changing.  <\/p>\n<p>    Central banks can also use AI, together with human expertise,    to better understand factors that contribute to    inflation. Neural networks can handle more input    variables compared with traditional econometric models, making    it possible to work with detailed data sets rather than relying    solely on aggregated data. They can further reflect intricate    non-linear relationships, offering valuable insights during    periods of rapidly changing inflation dynamics. If AI's impact    varies by industry but materialises rapidly, such advantages    are particularly beneficial for assessing inflationary    dynamics.  <\/p>\n<p>    Recent work in this area decomposes aggregate inflation into    various sub-components.29 In a first step, economic theory is used    to pre-specify four factors shaping aggregate inflation: past    inflation patterns, inflation expectations, the output gap and    international prices. A neural network then uses aggregate    series (eg the unemployment rate or total services inflation)    and disaggregate series (eg two-digit industry output) to    estimate the contribution of each of the four subcomponents to    overall inflation, accounting for possible non-linearities.  <\/p>\n<p>    The use of AI could play an important role in supporting    financial stability analysis. The strongest    suit of machine learning and AI methodologies is identifying    patterns in a cross-section. As such, they can be particularly    useful to identify and enhance the understanding of risks in a    large sample of observations, helping identify the    cross-section of risk across financial and non-financial firms.    Again, availability of timely data is key. For example, during    increasingly frequent periods of low liquidity and market    dysfunction, AI could help prediction through better monitoring    of anomalies across markets.30  <\/p>\n<p>    Finally, pairing AI-based insights with human judgment could    help support macroprudential regulation.    Systemic risks often result from the slow build-up of    imbalances and vulnerabilities, materialising in infrequent but    very costly stress events. The scarcity of data on such events    and the uniqueness of financial crises limit the stand-alone    use of data-intensive AI models in macroprudential    regulation.31    However, together with human expertise and informed economic    reasoning to see through the cycle, gen AI tools could yield    large benefits to regulators and supervisors. When combined    with rich data sets that provide sufficient scope to find    patterns in the data, AI could help in building early warning    indicators that alert supervisors to emerging pressure points    known to be associated with system-wide risks.  <\/p>\n<p>    In sum, with sufficient data, AI tools offer central banks an    opportunity to get a much better understanding of economic    developments. They enable central banks to draw on a richer set    of structured and unstructured data, and complementarily, speed    up data collection and analysis. In this way, the use of AI    enables the analysis of economic activity in real time at a    granular level. Such enhanced capabilities are all the more    important in the light of AI's potential impact on employment,    output and inflation, as discussed in the next section.  <\/p>\n<p>    AI is poised to increase productivity growth.    For workers, recent evidence suggests that AI directly raises    productivity in tasks that require cognitive skills (Graph 7.A). The use of    generative AI-based tools has had a sizeable and rapid positive    effect on the productivity of customer support agents and of    college-educated professionals solving writing tasks. Software    developers that used LLMs through the GitHub Copilot AI could    code more than twice as many projects per week. A recent    collaborative study by the BIS with Ant Group shows that    productivity gains are immediate and largest among less    experienced and junior staff (Box    E).32  <\/p>\n<p>    Early studies also suggest positive effects of AI on firm    performance. Patenting activity related to AI and the use of AI    are associated with faster employment and output growth as well    as higher revenue growth relative to comparable firms. Firms    that adopt AI also experience higher growth in sales,    employment and market valuations, which is primarily driven by    increased product innovation. These effects have materialised    over a horizon of one to two years. In a global sample, AI    patent applications generate a positive effect on the labour    productivity of small and medium-sized enterprises, especially    in services industries.33  <\/p>\n<p>    The macroeconomic impact of AI on productivity    growth could be sizeable. Beyond directly enhancing    productivity growth by raising workers' and firms' efficiency,    AI can spur innovation and thereby future productivity growth    indirectly. Most innovation is generated in occupations that    require high cognitive abilities. Improving the efficiency of    cognitive work therefore holds great potential to generate    further innovation. The estimates provided by the literature    for AI's impact on annual labour productivity growth (ie output    per employee) are thus substantive, although their range    varies.34 Through    faster productivity growth, AI will expand the economy's    productive capacity and thus raise aggregate supply.  <\/p>\n<p>    Higher productivity growth will also affect aggregate demand    through changes in firms' investment. While    gen AI is a relatively new technology, firms are already    investing heavily in the necessary IT infrastructure and    integrating AI models into their operations  on top of what    they already spend on IT in general. In 2023 alone, spending on    AI exceeded $150 billion worldwide, and a survey of US    companies' technology officers across all sectors suggests    almost 50% rank AI as their top budget item over the next    years.35  <\/p>\n<p>    An additional boost to investment could come from    improved prediction. AI adoption will lead to    more accurate predictions at a lower cost, which reduces    uncertainty and enables better decision-making.36 Of course, AI could also    introduce new sources of uncertainty that counteract some of    its positive impact on firm investment, eg by changing market    and price dynamics.  <\/p>\n<p>    Another substantial part of aggregate demand is household    consumption. AI could spur consumption by    reducing search frictions and improving matching, making    markets more competitive. For example, the use of AI agents    could improve consumers' ability to search for products and    services they want or need and help firms in advertising and    targeting services and products to consumers.37  <\/p>\n<p>    AI's impact on household consumption will also depend on how it    affects labour markets, notably labour demand and wages. The    overall impact depends on the relative strength of three forces    (Graph 8): by how    much AI raises productivity, how many new tasks it creates and    how many workers it displaces by making existing tasks    obsolete.  <\/p>\n<p>    If AI is a true general-purpose technology that raises total    factor productivity in all industries to a similar extent, the    demand for labour is set to increase across    the board (Graph    8, blue boxes). Like previous general-purpose technologies,    AI could also create altogether new tasks,    further increasing the demand for labour and spurring wage    growth (green boxes). If so, AI would increase aggregate    demand.  <\/p>\n<p>    However, the effects of AI might differ across tasks    and occupations. AI might benefit only some workers,    eg those whose tasks require logical reasoning. Think of nurses    who, with the assistance of AI, can more accurately interpret    x-ray pictures. At the same time, gen AI could make other tasks    obsolete, for example summarising documents, processing claims    or answering standardised emails, which lend themselves to    automation by LLMs. If so, increased AI adoption would lead to    displacement of some workers (Graph 8, red boxes). This could lead to    declines in employment and lower wage growth, with    distributional consequences. Indeed, results from a recent    survey of US households by economists in the BIS Monetary and    Economic Department in collaboration with the Federal Reserve    Bank of New York indicate that men, better-educated individuals    or those with higher incomes think that they will benefit more    from the use of gen AI than women and those with lower    educational attainment or incomes (Graph 7.B).38  <\/p>\n<p>    These considerations suggest that AI could have implications    for economic inequality. Displacement might    eliminate jobs faster than the economy can create new ones,    potentially exacerbating income inequality. A differential    impact of benefits across job categories would strengthen this    effect. The \"digital divide\" could widen, with individuals    lacking access to technology or with low digital literacy being    further marginalised. The elderly are particularly at risk of    exclusion.39  <\/p>\n<p>    Through the effects on productivity, investment and consumption    the deployment of AI has implications for output and inflation.    A BIS study illustrates the key mechanisms at work.40 As the source of a permanent    increase in productivity, AI will raise aggregate supply. An    increase in consumption and investment raises aggregate demand.    Through higher aggregate demand and supply,    output increases (Graph 9.A). In the short term, if households    and firms fully anticipate that they will be richer in the    future, they will increase consumption at the expense of    investment, slowing down output growth.  <\/p>\n<p>    The response of inflation will also depend on    households' and businesses' anticipation of future gains from    AI. If the average household does not fully anticipate gains,    it will increase today's consumption only modestly. AI will act    as a disinflationary force in the short run (blue line in    Graph 9.B), as    the impact on aggregate supply dominates. In contrast, if    households anticipate future gains, they will consume more,    making AI's initial impact inflationary (red line in Graph 9.B). Since past    general-purpose technologies have had an initial    disinflationary impact, the former scenario appears more    likely. But in either scenario, as economic capacity expands    and wages rise, the demand for capital and labour will steadily    increase. If these demand effects dominate the initial positive    shock to output capacity over time, higher inflation would    eventually materialise. How quickly demand forces increase    output and prices will depend not only on households'    expectations but also on the mismatch in    skills required in obsolete and newly created tasks.    The greater the skill mismatch (other things being equal), the    lower employment growth will be, as it takes displaced workers    longer to find new work. It might also be the case that some    segments of the population will remain permanently unemployable    without retraining. This, in turn, implies lower consumption    and aggregate demand, and a longer disinflationary impact of    AI.  <\/p>\n<p>    Another aspect that warrants further investigation is the    effect of AI adoption on price formation.    Large retail companies that predominately sell online use AI    extensively in their price-setting processes. Algorithmic    pricing by these retailers has been shown to increase both the    uniformity of prices across locations and the frequency of    price changes.41 For    example, when gas prices or exchange rates move, these    companies quickly adjust the prices in their online stores. As    the use of AI becomes more widespread, also among smaller    companies, these effects could become stronger. Increased    uniformity and flexibility in pricing can mean greater and    quicker pass-through of aggregate shocks to local prices, and    hence inflation, than in the past. This can ultimately change    inflation dynamics. An important aspect to consider is how    these effects could differ depending on the degree of    competition in the AI model and data market, which could    influence the variety of models used.  <\/p>\n<p>    Finally, the impact of AI on fiscal    sustainability remains an open question. All things    equal, an AI-induced boost to productivity and growth would    lead to a reduced debt burden. However, to the extent that    faster growth is associated with higher interest rates,    combined with the potential need for fiscal programmes to    manage AI-induced labour relocation or sustained spells of    higher unemployment rates, the impact of AI on the fiscal    outlook might be modest. More generally, the AI growth dividend    is unlikely to fully offset the spending needs that may arise    from the green transition or population ageing over the next    decades.  <\/p>\n<p>    The use of AI models opens up new opportunities for central    banks in pursuit of their policy objectives. A consistent theme    running through the chapter has been the availability of data    as a critical precondition for successful applications of    machine learning and AI. Data governance frameworks will be    part and parcel of any successful application of AI. Central    banks' policy challenges thus encompass both    models and data.  <\/p>\n<p>    An important trade-off arises between using    \"off-the-shelf\" models versus developing in-house fine-tuned    ones. Using external models may be more    cost-effective, at least in the short run, and leverages the    comparative advantage of private sector companies. Yet reliance    on external models comes with reduced transparency and exposes    central banks to concerns about dependence on a few external    providers. Beyond the general risks that market concentration    poses to innovation and economic dynamism, the high    concentration of resources could create significant operational    risks for central banks, potentially affecting their ability to    fulfil their mandates.  <\/p>\n<p>    Another important aspect relates to central banks' role as    users, compilers and disseminators of data.    Central banks use data as a crucial ingredient in their    decision-making and communication with the public. And they    have always been extensive compilers of data, either collecting    them on their own or drawing on other official agencies and    commercial sources. Finally, central banks are also providers    of data, to inform other parts of government as well as the    general public. This role helps them fulfil their obligations    as key stakeholders in national statistical systems.  <\/p>\n<p>    The rise of machine learning and AI, together with advances in    computing and storage capacity, have cast these aspects in an    urgent new light. For one, central banks now need to make sense    of and use increasingly large and diverse sets of structured    and unstructured data. And these data often reside in the hands    of the private sector. While LLMs can help process such data,    hallucinations or prompt injection attacks can lead to biased    or inaccurate analyses. In addition, commercial data vendors    have become increasingly important, and central banks make    extensive use of them. But in recent years, the cost of    commercial data has increased markedly, and vendors have    imposed tighter use conditions.  <\/p>\n<p>    The decision on whether to use external or internal models and    data has far-reaching implications for central banks'    investments and human capital. A key challenge is    setting up the necessary IT infrastructure,    which is greater if central banks pursue the road of developing    internal models and collecting or producing their own data.    Providing adequate computing power and software, as well as    training existing or hiring new staff, involves high up-front    costs. The same holds for creating a data lake, ie pooling    different curated data sets. Yet a reliable and safe IT    infrastructure is a prerequisite not only for big data analyses    but also to prevent cyber attacks.  <\/p>\n<p>    Hiring new or retaining existing staff with    the right mix of economic understanding and programming skills    can be challenging. As AI applications increase the    sophistication of the financial system over time, the premium    on having the right mix of skills will only grow. Survey-based    evidence suggests this is a top concern for central banks    (Graph 10).    There is high demand for data scientists and other AI-related    roles, but public institutions often cannot match private    sector salaries for top AI talent. The need for staff with the    right skills also arises from the fact that the use of AI    models to aid financial stability monitoring faces limitations,    as discussed above. Indeed, AI is not a substitute for human    judgment. It requires supervision by experts with a solid    understanding of macroeconomic and financial processes.  <\/p>\n<p>    How can central banks address these challenges and mitigate    trade-offs? The answer lies, in large part, in cooperation    paired with sound data governance practices.  <\/p>\n<p>    Collaboration can yield significant benefits and relax    constraints on human capital and IT. For one, the    pooling of resources and knowledge can lower    demands among central banks and could ease the resource    constraints on collecting, storing and analysing big data as    well as developing algorithms and training models. For example,    central banks could address rising costs of commercial data,    especially for smaller institutions, by sharing more granular    data themselves or by acquiring data from vendors through joint    procurement. Cooperation could also facilitate training staff    through workshops in the use of AI or the sharing of    experiences in conferences. This would particularly benefit    central banks with fewer staff and resources and with limited    economies of scale. Cooperation, for example by re-using    trained models, could also mitigate the environmental costs    associated with training algorithms and storing large amounts    of data, which consume enormous amounts of energy.  <\/p>\n<p>    Central bank collaboration and the sharing of experiences could    also help identify areas in which AI adds the most value and    how to leverage synergies. Common data    standards could facilitate access to publicly available data    and facilitate the automated collection of relevant data from    various official sources, thereby enhancing the training and    performance of machine learning models. Additionally, dedicated    repositories could be set up to share the open source code of    data tools, either with the broader public or, at least    initially, only with other central banks. An example is a    platform such as BIS Open Tech, which supports international    cooperation and coordination in sharing statistical and    financial software. More generally, central banks could    consider sharing domain-adapted or fine-tuned models in the    central banking community, which could significantly lower the    hurdles for adoption.42 Joint work on AI models is possible    without sharing data, so they can be applied even where there    are concerns about confidentiality.  <\/p>\n<p>    An example of how collaboration supports data collection and    dissemination is the jurisdiction-level statistics on    international banking, debt securities and over-the-counter    derivatives by the BIS. These data sets have a long history     the international banking statistics started in the 1970s. They    are a critical element for monitoring developments and risks in    the global financial system. They are compiled from submissions    by participating authorities under clear governance rules and    using well established statistical processes. At a more    granular level, arrangements for the sharing of confidential    bank-level data include the quantitative impact study data    collected by the Basel Committee on Banking Supervision and the    data on large global banks collected by the International Data    Hub. Other avenues to explore include sharing synthetic or    anonymised data that protect confidential information.  <\/p>\n<p>    The rising importance of data and emergence of new sources and    tools call for sound data governance    practices. Central banks must establish robust    governance frameworks that include guidelines for selecting,    implementing and monitoring both data and algorithms. These    frameworks should comprise adequate quality control and cover    data management and auditing practices. The importance of    metadata, in particular, increases as the range and variety of    data expand. Sometimes referred to as \"the data about the    data\", metadata include the definitions, source, frequency,    units and other information that define a given data set. This    metadata is crucial when privacy-preserving methods are used to    draw lessons from several data sets overseen by different    central banks. Machine readability is greatly enhanced when    metadata are standardised so that the machines know what they    are looking for. For example, the \"Findable, Accessible,    Interoperable and Reusable\" (FAIR) principles provide guidance    in organising data and metadata to ease the burden of sharing    data and algorithms.43  <\/p>\n<p>    More generally, metadata frameworks are crucial for a better    understanding of the comparability and limits of data series.    Central banks can also cooperate in this domain. For example,    the Statistical Data and Metadata Exchange (SDMX) standard    provides a common language and structure for metadata. Such    standards are crucial to foster data-sharing, lower the    reporting burden and facilitate interoperability. Similarly,    the Generic Statistical Business Process Model lays out    business processes for official statistics with a unified    framework and consistent terminology. Sound data governance    practices would also facilitate the sharing of confidential    data.  <\/p>\n<p>    In sum, there is an urgent need for central banks to    collaborate in fostering the development of a community    of practice to share knowledge, data, best practices    and AI tools. In the light of rapid technological change, the    exchange of information on policy issues arising from the role    of central banks as data producers, users and disseminators is    crucial. Collaboration lowers costs, and such a community would    foster the development of common standards. Central banks have    a history of successful collaboration to overcome new    challenges. The emergence of AI has hastened the need for    cooperation in the field of data and data governance.  <\/p>\n<p>    Graph 1.A: The adoption of ChatGPT is proxied by the ratio of    the maximum number of website visits worldwide for the period    November 2022April 2023 and the worldwide population with    internet connectivity. For more details on computer see US    Census Bureau; for electric power, internet and social media    see Comin and Hobijn (2004) and Our World in Data; for    smartphones, see Statista.  <\/p>\n<p>    Graph 1.B: Based on an April 2023 global survey with 1,684    participants.  <\/p>\n<p>    Graph 1.C: Data for capital invested in AI companies for 2024    are annualised based on data up to mid-May. Data on the    percentage of AI job postings for AU, CA, GB, NZ and US are    available for the period 201423; for AT, BE, CH, DE, ES, FR,    IT, NL and SE, data are available for the period 201823.  <\/p>\n<p>    Graph 2.A: Three-month moving averages.  <\/p>\n<p>    Graphs 2.B and 2.C: Correspondent banks that are active in    several corridors are counted several times. Averages across    countries in each region. Markers in panel C represent    subregions within each region. Grouping of countries by region    according to the United Nations Statistics Division; for    further details see unstats.un.org\/unsd\/methodology\/m49\/.  <\/p>\n<p>    Graph 3.A: Average scores in answers to the following question:    \"In the following areas, would you trust artificial    intelligence (AI) tools less or more than traditional    human-operated services? For each item, please indicate your    level of trust on a scale from 1 (much less trust than in a    human) to 7 (much more trust).\"  <\/p>\n<p>    Graph 3.B: Average scores and 95% confidence intervals in    answers to the following question: \"How much do you trust the    following entities to safely store your personal data when they    use artificial intelligence tools? For each of them, please    indicate your level of trust on a scale from 1 (no trust at all    in the ability to safely store personal data) to 7 (complete    trust).\"  <\/p>\n<p>    Graph 3.C: Average scores (with scores ranging from 1 (lowest)    to 7 (highest)) in answers to the following questions: (1) \"Do    you think that sharing your personal information with    artificial intelligence tools will decrease or increase the    risk of data breaches (that is, your data becoming publicly    available without your consent)?\"; (2) \"Are you concerned that    sharing your personal information with artificial intelligence    tools could lead to the abuse of your data for unintended    purposes (such as for targeted ads)?\"  <\/p>\n<p>    Graph 6.A: The bars show the share of respondents to the    question, \"Do you agree that the use of AI can provide more    benefits than risks to your organisation?\".  <\/p>\n<p>    Graph 6.B: The bars show the average score that respondents    gave to each option when asked to \"Rate the level of    significance of the following benefits of AI in cyber    security\"; the score scale of each option is from 1 (lowest) to    5 (highest).  <\/p>\n<p>    Graph 7.A: The bars correspond to estimates of the increase in    productivity of users that rely on generative AI tools relative    to a control group that did not.  <\/p>\n<p>    Acemoglu, D (2024): \"The simple macroeconomics of AI\",    Economic Policy, forthcoming.  <\/p>\n<p>    Agrawal, A, J Gans and A Goldfarb (2019): \"Exploring the impact    of artificial intelligence: prediction versus judgment\",    Information Economics and Policy, vol 47, pp16.  <\/p>\n<p>    ----- (2022): Prediction machines, updated and expanded:    the simple economics of artificial intelligence, Harvard    Business Review Press, 15 November.  <\/p>\n<p>    Ahir, H, N Bloom and D Furceri (2022): \"The world uncertainty    index\", NBER Working Papers, no 29763, February.  <\/p>\n<p>    Aldasoro, I, O Armantier, S Doerr, L Gambacorta and T Oliviero    (2024a): \"Survey evidence on gen    AI and households: job prospects amid trust concerns\",    BIS Bulletin, no 86, April.  <\/p>\n<p>    ----- (2024b): \"The gen AI gender gap\", BIS Working    Papers, forthcoming.  <\/p>\n<p>    Aldasoro, I, S Doerr, L Gambacorta, G Gelos and D Rees (2024):    \"Artificial intelligence, labour markets and inflation\",    mimeo.  <\/p>\n<p>    Aldasoro, I, S Doerr, L Gambacorta, S Notra, T Oliviero and D    Whyte (2024): \"Generative    artificial intelligence and cybersecurity in central    banking\", BIS Papers, no145, May.  <\/p>\n<p><!-- Auto Generated --><\/p>\n<p>See the rest here:<\/p>\n<p><a target=\"_blank\" rel=\"nofollow noopener\" href=\"https:\/\/www.bis.org\/publ\/arpdf\/ar2024e3.htm\" title=\"III. Artificial intelligence and the economy: implications for central banks - bis.org\">III. Artificial intelligence and the economy: implications for central banks - bis.org<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p> Key takeaways The advent of large language models (LLMs) has catapulted generative artificial intelligence (gen AI) into popular discourse. LLMs have transformed the way people interact with computers away from code and programming interfaces to ordinary text and speech. This ability to converse through ordinary language as well as gen AI's human-like capabilities in creating content have captured our collective imagination <a href=\"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/artificial-intelligence\/iii-artificial-intelligence-and-the-economy-implications-for-central-banks-bis-org.php\">Continue reading <span class=\"meta-nav\">&rarr;<\/span><\/a><\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"limit_modified_date":"","last_modified_date":"","_lmt_disableupdate":"","_lmt_disable":"","footnotes":""},"categories":[13],"tags":[],"class_list":["post-169668","post","type-post","status-publish","format-standard","hentry","category-artificial-intelligence"],"modified_by":null,"_links":{"self":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts\/169668"}],"collection":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/comments?post=169668"}],"version-history":[{"count":0,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts\/169668\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/media?parent=169668"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/categories?post=169668"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/tags?post=169668"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}