<tt id="6hsgl"><pre id="6hsgl"><pre id="6hsgl"></pre></pre></tt>
          <nav id="6hsgl"><th id="6hsgl"></th></nav>
          国产免费网站看v片元遮挡,一亚洲一区二区中文字幕,波多野结衣一区二区免费视频,天天色综网,久久综合给合久久狠狠狠,男人的天堂av一二三区,午夜福利看片在线观看,亚洲中文字幕在线无码一区二区
          Global EditionASIA 中文雙語Fran?ais
          Opinion
          Home / Opinion / Op-Ed Contributors

          We need a precautionary approach to AI

          By Maciej Kuziemski | China Daily | Updated: 2018-05-08 07:06
          Share
          Share - WeChat
          Song Chen/China Daily

          For policymakers in any country, the best way to make decisions is to base them on evidence, however imperfect the available data may be. But what should leaders do when facts are scarce or nonexistent? That is the quandary facing those who have to grapple with the fallout of "advanced predictive algorithms"-the binary building blocks of machine learning and artificial intelligence (AI).

          In academic circles, AI-minded scholars are either "singularitarians" or "presentists". Singularitarians generally argue that while AI technologies pose an existential threat to humanity, the benefits outweigh the costs. But although this group includes many tech luminaries and attracts significant funding, its academic output has so far failed to prove their calculus convincingly.

          On the other side, presentists tend to focus on the fairness, accountability, and transparency of new technologies. They are concerned, for example, with how automation will affect the labor market. But here, too, the research has been unpersuasive. For example, MIT Technology Review recently compared the findings of 19 major studies examining predicted job losses, and found that forecasts for the number of globally "destroyed" jobs vary from 1.8 million to 2 billion.

          Simply put, there is no "serviceable truth" to either side of this debate. When predictions of AI's impact range from minor job-market disruptions to human extinction, clearly the world needs a new framework to analyze and manage the coming technological disruption.

          But every so often, a "post-normal" scientific puzzle emerges, something philosophers Silvio Funtowicz and Jerome Ravetz first defined in 1993 as a problem "where facts are uncertain, values in dispute, stakes high, and decisions urgent". For these challenges, of which AI is one, policy cannot afford to wait for science to catch up.

          At the moment, most AI policymaking occurs in the "Global North", which de-emphasizes the concerns of less-developed countries and makes it harder to govern dual-use technologies. Worse, policymakers often fail to consider the potential environmental impact, and focus almost exclusively on the anthropogenic effects of automation, robotics and machines.

          The precautionary principle is not without its detractors, though. While its merits have been debated for years, we need to accept that the lack of evidence of harm is not the same thing as evidence of lack of harm.

          For starters, applying the precautionary principle to the context of AI would help rebalance the global policy discussion, giving weaker voices more influence in debates that are currently monopolized by corporate interests. Decision-making would also be more inclusive and deliberative, and produce solutions that more closely reflected societal needs. The Institute of Electrical and Electronics Engineers, and The Future Society at Harvard's Kennedy School of Government are already spearheading work in this participatory spirit. Additional professional organizations and research centers should follow suit.

          Moreover, by applying the precautionary principle, governance bodies could shift the burden of responsibility to the creators of algorithms. A requirement of explainability of algorithmic decision-making can change incentives, prevent "blackboxing", help make business decisions more transparent, and allow the public sector to catch up with the private sector in technology development. And, by forcing tech companies and governments to identify and consider multiple options, the precautionary principle would bring to the fore neglected issues, like environmental impact.

          Rarely is science in a position to help manage an innovation long before the consequences of that innovation are available for study. But, in the context of algorithms, machine learning, and AI, humanity cannot afford to wait. The beauty of the precautionary principle lies not only in its grounding in international public law, but also in its track record as a framework for managing innovation in myriad scientific contexts. We should embrace it before the benefits of progress are unevenly distributed, or, worse, irreversible harm has been done.

          The author is a policy fellow at the School of Transnational Governance at the European University Institute.
          Project Syndicate

          Most Viewed in 24 Hours
          Top
          BACK TO THE TOP
          English
          Copyright 1994 - . All rights reserved. The content (including but not limited to text, photo, multimedia information, etc) published in this site belongs to China Daily Information Co (CDIC). Without written authorization from CDIC, such content shall not be republished or used in any form. Note: Browsers with 1024*768 or higher resolution are suggested for this site.
          License for publishing multimedia online 0108263

          Registration Number: 130349
          FOLLOW US
          主站蜘蛛池模板: 人妻另类 专区 欧美 制服| 日韩在线视频线观看一区| 国产99久久亚洲综合精品西瓜tv| 最近中文字幕完整版hd| 四虎影视国产精品永久在线| 妇女自拍偷自拍亚洲精品| 激情内射亚洲一区二区三区| 一本色道婷婷久久欧美| 国产欧美精品一区aⅴ影院| 欧美成人aaa片一区国产精品| 中文字幕va一区二区三区 | 亚洲A综合一区二区三区| 97成人碰碰久久人人超级碰oo| 欧美日韩精品免费一区二区三区| 丰满少妇特黄一区二区三区| 日本中文字幕在线播放| 亚洲精品久久久久国色天香| 久9视频这里只有精品试看| 国产又粗又猛又黄又爽无遮挡| 少妇粗大进出白浆嘿嘿视频| caoporen国产91在线| 日韩无矿砖一线二线卡乱| 国产亚洲人成网站在线观看 | 少妇又爽又刺激视频| 精人妻无码一区二区三区| 欧美成人精品三级在线观看| 日韩中文字幕亚洲精品| 一区二区不卡99精品日韩| 亚洲午夜香蕉久久精品| 久久丁香五月天综合网| 国产成人精品无码免费看| 免费乱理伦片在线观看| 最新日韩精品中文字幕| 色偷偷一区| 四虎精品视频永久免费| 97se亚洲综合不卡| 亚洲码亚洲码天堂码三区| 脱了老师内裤猛烈进入的软件 | 亚洲欧洲精品成人久久曰| 四虎影视一区二区精品| 国产91丝袜在线播放动漫|