AI risks repeating social media era’s mistakes: Microsoft head

NEW DELHI: Breakneck development of artificial intelligence risked repeating mistakes made by the tech industry at the start of the social media era, Microsoft president Brad Smith told a business forum on Friday. Rapid advancements in AI have stoked global alarm over the technology’s potential for disinformation, misuse and upheaval of the labour market. But Smith suggested these misgivings were not reflected by the developers of the potentially revolutionary technology, whose optimism reminded him of the early years of social media platforms.

Back then, the tech industry “became a little too euphoric about all the good things that social media would bring to the world—and there have been many—without thinking about the risks as well,” he said. “We need to be clear-eyed, we need to be excited about the opportunities, but thoughtful, perhaps even concerned, about the downside. And we need to construct the guardrails from the outset,” he added. The rise of AI has raised both excitement and concerns about its potential to improve or replace tasks done by humans.

AI tools have shown in recent months the ability to generate essays, create realistic images, mimic voices of famous singers and even pass medical exams, among a slew of uses. But there are also worries that chatbots could flood the internet with disinformation, that biased algorithms will churn out racist material or that AI-powered automation could lay waste to entire industries. A United Nations report this week said AI was more likely to augment jobs than to destroy them, adding nonetheless that the tech would alter work intensity and the autonomy of workers.

It also said the effects of technology would vary greatly between professions and regions, with clerical workers facing the most exposure to changes and women more likely than men to see their jobs affected. Smith said it was clear the public “want to be confident that this new technology will remain under human control”. Mastercard chief executive Michael Miebach said companies needed to build trust over use of the tech and take action to address issues such as AI bias.

But he also said he believed that the risks around AI were “not terribly new” and should not impede further development of the technology. “Naturally regulation will be behind,” he said. “But that shouldn’t slow us down.” Both men were speaking in New Delhi alongside other world industry leaders at a meeting that is serving as a prelude to next month’s G20 summit in the Indian capital. – AFP

more recommended stories

Terms of Use:

  • This website Arabian Daily is an individual’s property, not used for any commercial or sales purposes. What you see here are one’s random thoughts in action. I, by no means, endorse any product or party through this, unless stated explicitly.
  • All work you will find here is copyrighted unless stated otherwise. No part of this work can be reproduced in any way with the exception of a) if you share our work, it should link back to this website; b) if you quote any part of our work, it should be properly credited to us with a link to this website.
  • All images used on this website have been taken from open source image websites on the Internet. If any of them are copyrighted to you and you want us to take them down or add credits, please feel free to contact us here, or by using the contact form on this page.
  • The views expressed on Arabian Daily are solely ours. They do not represent any party or any particular school of thought. This website does not promote racism in any form.
  • Privacy Policy:
    This website will respect the readers’ and the writer’s privacy. We do not sell any of their personal or contact information to another company. We do not put your information on spam lists. Also, and more importantly, we are not responsible for the privacy practices of any of our advertisers or website commenters.
  • Reserve Rights: We reserve the right to change the focus on this website, to shut it down, sell it or to change the terms of use at our own discretion.