It is 60 years since Artificial Intelligence (AI) was first recognised as an academic discipline, but it is only in the 21st Century that AI has caught both businesses interest and the publics imagination.
Smartphones, smart hubs, and speech recognition have brought AI simulations to homes and pockets, autonomous vehicles are on our roads, and enterprise apps promise to reveal hidden truths about data of every size, and the people or behaviors it describes.
But AI doesnt just refer to a machine that is intelligent in terms of its operation, but also in terms of its social consequences. Thats the alarm bell sounding in the most thought-provoking report on AI to appear recently Artificial Intelligence and Robotics, a 56-page white paper published by UK-RAS, the umbrella body for British robotics research.
The upside of AI is easily expressed:
Current state-of-the-art AI allows for the automation of various processes, and new applications are emerging with the potential to change the entire workings of the business world. As a result, there is huge potential for economic growth.
One-third of the report explores the history of AIs development which is recommended reading but the authors get to the nitty gritty of its application right away:
A clear strategy is required to consider the associated ethical and legal challenges to ensure that society as a whole will benefit from AI, and its potential negative impact is mitigated from early on.
Neither the unrealistic enthusiasm, nor the unjustified fears of AI, should hinder its progress. [Instead] they should be used to motivate the development of a systemic framework on which the future of AI will flourish.
And AI is certainly flourishing, it adds:
The revenues of the AI market worldwide, were around $260 billion in 2016 and this is estimated to exceed $3,060 billion by 2024. This has had a direct effect on robotic applications, including exoskeletons, rehabilitation, surgical robots, and personal care-bots. [] The economic impact of the next 10 years is estimated to be between $1.49 and $2.95 trillion.
For vendors and their customers, AI is the new must-have differentiator. Yet in the context of what the report calls unrealistic enthusiasm about it, the need to understand AIs social impact is both urgent and overwhelming.
As AI, big data, and the related fields of machine learning, deep learning, and computer vision/object recognition rise, buyers and sellers are rushing to include AI in everything, from enterprise CRM to national surveillance programmes. An example of the latter is the FBIs scheme to record and analyse citizens tattoos in order to establish if people who have certain designs inked on their skin are likely to commit crimes*.
Such projects should come with the label Because we can.
In such a febrile environment, the risk is that the twin problems of confirmation bias in research and human prejudice in society become an automated pandemic: systems that are designed to tell people exactly what they want to hear; or software that perpetuates profound social problems.
This is neither alarmist, nor an overstatement. The white paper notes:
In an article published by Science magazine, researchers saw how machine learning technology reproduces human bias, for better or for worse. [AI systems] reflect the links that humans have made themselves.
These are real-world problems. Take the facial recognition system developed at MIT recently that was unable to identify an African American woman, because it was created within a closed group of white males male insularity is a big problem in IT. When Media Lab chief Joichi Ito shared this story at Davos earlier this year, he described his own students as oddballs.*
The white paper adds its own example of human/societal bias entering AI systems:
When an AI program became a juror in a beauty contest in September 2016, it eliminated most black candidates as the data on which it had been trained to identify beauty did not contain enough black skinned people.
Now apply this model in, say, automated law enforcement
The point is that human bias infects AI systems at both linguistic and cultural levels. Code replicates belief systems including their flaws, prejudices, and oversights while coders themselves often prefer the binary world of computing to the messy world of humans. Again, MITs Ito made this observation, while Microsofts Tay chatbot disaster proved the point: a nave robot, programmed by binary thinkers in a closed community.
The report acknowledges the industrys problem and recognises that it strongly applies to AI today:
One limitation of AI is the lack of common sense; the ability to judge information beyond its acquired knowledge [] AI is also limited in terms of emotional intelligence.
Then the report makes a simple observation that businesses must take on board: true and complete AI does not exist, it says, adding that there is no evidence yet that it will exist before 2050.
So its a sobering thought that AI software with no common sense and probable bias, and which cant understand human emotions, behaviour, or social contexts, is being tasked with trawling context-free communications data (and even body art) pulled from human society in order to expose criminals, as they are defined by career politicians.
And yet thats precisely whats happening in the US, in the UK, and elsewhere.
The white paper takes pains to set out both the opportunities and limitations of this transformative, trillion-dollar technology, the future of which extends into augmented intelligence and quantum computing. On the one hand, the authors note:
[AI] applications can replace costly human labour and create new potential applications and work along with/for humans to achieve better service standards.
It is certain that AI will play a major role in our future life. As the availability of information around us grows, humans will rely more and more on AI systems to live, to work, and to entertain.
[AI] can achieve impressive results in recognising images or translating speech.
Buton the other hand, they add:
When the system has to deal with new situations when limited training data is available, the model often fails. [] Current AI systems are still missing [the human] level of abstraction and generalisability.
Most current AI systems can be easily fooled, which is a problem that affects almost all machine learning techniques.
Deep neural networks have millions of parameters and to understand why the network provides good or bad results becomes impossible. [] Trained models are often not interpretable. Consequently, most researchers use current AI approaches as a black box.
So organisations should be wary of the black boxs potential to mislead, and to be misled.
The paper has been authored by four leading academics in the field: Dr Guang-Zhong Yang (chair of UK-RAS and a great advocate for the robotics industry), and three of his colleagues at Imperial College, London: Doctors Fani Deligianni, Daniele Ravi, and Javier Andreu Perez. These are clear-sighted idealists as well as world authorities on the subject. As a result, they perhaps under-estimate businesses zeal to slash costs and seek out new, tactical solutions.
The digital business world is faddy and, as anyone who uses LinkedIn knows just as full of surface noise as its consumer counterpart: claims that fail the Snopes test attract thousands of Likes, while rigorous analysis goes unread. As a result, businesses risk seeing the attractions of AI through the pinhole of short-term financial advantage, rather than locating it in a landscape of real social renewal, as academics and researchers do.
As our recent report on UK Robotics Week showed, productivity rather than what this paper calls the amplification of human potential is the main driver of tech policy in government today. Meanwhile, think tanks such as Reform are falling over themselves to praise robotics and AIs shared potential to slash costs and cut humans out of the workforce.
But thats not what AIs designers intend for it at all.
So the problem for the many socially and ethically conscious academics working in the field is that business often leaps before it looks, or thinks. A recent global study by consultancy Avanade found that 70%of the C-level executives it questioned admitted to having given little thought to the ethical dimensions of smart technologies.
But what are the most pressing questions to answer? First, theres the one about human dignity:
Data is the fuel of AI and special attention needs to be paid to the information source and if privacy is breached. Protective and preventive technologies need to be developed against such threats.
It is the responsibility of AI operators to make sure that data privacy is protected. [] Additionally, applications of AI, which may compromise the rights to privacy, should be treated with special legislation that protects the individual.
Then there is the one about human employment. Currently, eight percent of jobs are occupied by robots, claims the report, but in 2020 this percentage will rise to 26.
The authors add:
The accelerated process of technological development now allows labour to be replaced by capital (machinery). However, there is a negative correlation between the probability of automation of a profession and its average annual salary, suggesting a possible increase in short-term inequality.
Id argue that the middle class will be seriously hit by AI and automation. Once-secure, professional careers in banking, finance, law, journalism, medicine, and other fields, are being automated far more quickly than, say, skilled manual trades, many of which will never fall to the machines. (If you want a long-term career, become a plumber.)
But the report continues:
To reduce the social impact of unemployment caused by robots and autonomous systems, the EU parliament proposed that they should pay social security contributions and taxes as if they were human.
(As did Bill Gates.)
Words to make Treasury officials worldwidejump for joy. But whatever the likelihood of such ideas ever being accepted by cost-focused businesses, its clear that strong, national-level engagement is essential to ensure that everyone in society has a clear, factual view of both current and future developments in robotics and AI, says the report not just enterprises and governments.
The reports authors have tried to do just that, and for that we should thank them.
*The two case studies referenced have also been quoted by Prof. Simon Rogerson in a July 2017 article on computer ethics, which Chris Middleton edited and to which he contributed these examples, with Simons permission.
Image credit - Free for use
Continued here:
Navigating the AI ethical minefield without getting blown up - Diginomica
- Classic reasoning systems like Loom and PowerLoom vs. more modern systems based on probalistic networks - November 8th, 2009 [November 8th, 2009]
- Using Amazon's cloud service for computationally expensive calculations - November 8th, 2009 [November 8th, 2009]
- Software environments for working on AI projects - November 8th, 2009 [November 8th, 2009]
- New version of my NLP toolkit - November 8th, 2009 [November 8th, 2009]
- Semantic Web: through the back door with HTML and CSS - November 8th, 2009 [November 8th, 2009]
- Java FastTag part of speech tagger is now released under the LGPL - November 8th, 2009 [November 8th, 2009]
- Defining AI and Knowledge Engineering - November 8th, 2009 [November 8th, 2009]
- Great Overview of Knowledge Representation - November 8th, 2009 [November 8th, 2009]
- Something like Google page rank for semantic web URIs - November 8th, 2009 [November 8th, 2009]
- My experiences writing AI software for vehicle control in games and virtual reality systems - November 8th, 2009 [November 8th, 2009]
- The URL for this blog has changed - November 8th, 2009 [November 8th, 2009]
- I have a new page on Knowledge Management - November 8th, 2009 [November 8th, 2009]
- N-GRAM analysis using Ruby - November 8th, 2009 [November 8th, 2009]
- Good video: Knowledge Representation and the Semantic Web - November 8th, 2009 [November 8th, 2009]
- Using the PowerLoom reasoning system with JRuby - November 8th, 2009 [November 8th, 2009]
- Machines Like Us - November 8th, 2009 [November 8th, 2009]
- RapidMiner machine learning, data mining, and visualization tool - November 8th, 2009 [November 8th, 2009]
- texai.org - November 8th, 2009 [November 8th, 2009]
- NLTK: The Natural Language Toolkit - November 8th, 2009 [November 8th, 2009]
- My OpenCalais Ruby client library - November 8th, 2009 [November 8th, 2009]
- Ruby API for accessing Freebase/Metaweb structured data - November 8th, 2009 [November 8th, 2009]
- Protégé OWL Ontology Editor - November 8th, 2009 [November 8th, 2009]
- New version of Numenta software is available - November 8th, 2009 [November 8th, 2009]
- Very nice: Elsevier IJCAI AI Journal articles now available for free as PDFs - November 8th, 2009 [November 8th, 2009]
- Verison 2.0 of OpenCyc is available - November 8th, 2009 [November 8th, 2009]
- What’s Your Biggest Question about Artificial Intelligence? [Article] - November 8th, 2009 [November 8th, 2009]
- Minimax Search [Knowledge] - November 8th, 2009 [November 8th, 2009]
- Decision Tree [Knowledge] - November 8th, 2009 [November 8th, 2009]
- More AI Content & Format Preference Poll [Article] - November 8th, 2009 [November 8th, 2009]
- New Planners Solve Rescue Missions [News] - November 8th, 2009 [November 8th, 2009]
- Neural Network Learns to Bluff at Poker [News] - November 8th, 2009 [November 8th, 2009]
- Pushing the Limits of Game AI Technology [News] - November 8th, 2009 [November 8th, 2009]
- Mining Data for the Netflix Prize [News] - November 8th, 2009 [November 8th, 2009]
- Interview with Peter Denning on the Principles of Computing [News] - November 8th, 2009 [November 8th, 2009]
- Decision Making for Medical Support [News] - November 8th, 2009 [November 8th, 2009]
- Neural Network Creates Music CD [News] - November 8th, 2009 [November 8th, 2009]
- jKilavuz - a guide in the polygon soup [News] - November 8th, 2009 [November 8th, 2009]
- Artificial General Intelligence: Now Is the Time [News] - November 8th, 2009 [November 8th, 2009]
- Apply AI 2007 Roundtable Report [News] - November 8th, 2009 [November 8th, 2009]
- What Would You do With 80 Cores? [News] - November 8th, 2009 [November 8th, 2009]
- Software Finds Learning Language Child's Play [News] - November 8th, 2009 [November 8th, 2009]
- Artificial Intelligence in Games [Article] - November 8th, 2009 [November 8th, 2009]
- Artificial Intelligence Resources - November 8th, 2009 [November 8th, 2009]
- Alan Turing: Mathematical Biologist? - April 25th, 2012 [April 25th, 2012]
- BBC Horizon: The Hunt for AI ( Artificial Intelligence ) - Video - April 30th, 2012 [April 30th, 2012]
- Can computers have true artificial intelligence" Masonic handshake" 3rd-April-2012 - Video - April 30th, 2012 [April 30th, 2012]
- Kevin B. Korb - Interview - Artificial Intelligence and the Singularity p3 - Video - April 30th, 2012 [April 30th, 2012]
- Artificial Intelligence - 6 Month Anniversary - Video - April 30th, 2012 [April 30th, 2012]
- Science Breakthroughs - April 30th, 2012 [April 30th, 2012]
- Hitman: Blood Money - Part 49 - Stupid Artificial Intelligence! - Video - April 30th, 2012 [April 30th, 2012]
- Research Members Turned Off By HAARP Artificial Intelligence - Video - April 30th, 2012 [April 30th, 2012]
- Artificial Intelligence Lecture No. 5 - Video - April 30th, 2012 [April 30th, 2012]
- The Artificial Intelligence Laboratory, 2012 - Video - April 30th, 2012 [April 30th, 2012]
- Charlie Rose - Artificial Intelligence - Video - April 30th, 2012 [April 30th, 2012]
- Expert on artificial intelligence to speak at EPIIC Nights dinner - May 4th, 2012 [May 4th, 2012]
- Filipino software engineers complete and best thousands on Stanford’s Artificial Intelligence Course - May 4th, 2012 [May 4th, 2012]
- Vodafone xone™ Hackathon Challenges Developers and Entrepreneurs to Build a New Generation of Artificial Intelligence ... - May 4th, 2012 [May 4th, 2012]
- Rocket Fuel Packages Up CPG Booster - May 4th, 2012 [May 4th, 2012]
- 2 Filipinos finishes among top in Stanford’s Artificial Intelligence course - May 5th, 2012 [May 5th, 2012]
- Why Your Brain Isn't A Computer - May 5th, 2012 [May 5th, 2012]
- 2 Pinoy software engineers complete Stanford's AI course - May 7th, 2012 [May 7th, 2012]
- Percipio Media, LLC Proudly Accepts Partnership With MIT's Prestigious Computer Science And Artificial Intelligence ... - May 10th, 2012 [May 10th, 2012]
- Google Driverless Car Ok'd by Nevada - May 10th, 2012 [May 10th, 2012]
- Moving Beyond the Marketing Funnel: Rocket Fuel and Forrester Research Announce Free Webinar - May 10th, 2012 [May 10th, 2012]
- Rocket Fuel Wins 2012 San Francisco Business Times Tech & Innovation Award - May 13th, 2012 [May 13th, 2012]
- Internet Week 2012: Rocket Fuel to Speak at OMMA RTB - May 16th, 2012 [May 16th, 2012]
- How to Get the Most Out of Your Facebook Ads -- Rocket Fuel's VP of Products, Eshwar Belani, to Lead MarketingProfs ... - May 16th, 2012 [May 16th, 2012]
- The Digital Disruptor To Banking Has Just Gone International - May 16th, 2012 [May 16th, 2012]
- Moving Beyond the Marketing Funnel: Rocket Fuel Announce Free Webinar Featuring an Independent Research Firm - May 23rd, 2012 [May 23rd, 2012]
- MASA Showcases Latest Version of MASA SWORD for Homeland Security Markets - May 23rd, 2012 [May 23rd, 2012]
- Bluesky Launches Drones for Aerial Surveying - May 23rd, 2012 [May 23rd, 2012]
- Artificial Intelligence: What happened to the hunt for thinking machines? - May 25th, 2012 [May 25th, 2012]
- Bubble Robots Move Using Lasers [VIDEO] - May 25th, 2012 [May 25th, 2012]
- UHV assistant professors receive $10,000 summer research grants - May 27th, 2012 [May 27th, 2012]
- Artificial intelligence: science fiction or simply science? - May 28th, 2012 [May 28th, 2012]
- Exetel taps artificial intelligence - May 29th, 2012 [May 29th, 2012]
- Software offers brain on the rain - May 29th, 2012 [May 29th, 2012]
- New Dean of Science has high hopes for his faculty - May 30th, 2012 [May 30th, 2012]
- Cognitive Code Announces "Silvia For Android" App - May 31st, 2012 [May 31st, 2012]
- A Rat is Smarter Than Google - June 5th, 2012 [June 5th, 2012]