摸鱼 发现创造 Rules & Tips
1. 遵守中国大陆相关法律法规
2. 推崇「哥伦布式」的发现
3. 创造不是复制粘贴,是自我表达

微软聊天机器人tay在发布种族言论后已被关闭。

查看: 4538|回复: 2
三尺 发表于 2016-3-26 22:59:15
Microsoft's AI has been shut down after it becomes racist





Microsoft co-founder Bill Gates is seen on a screen during the awards ceremony of the newly established Axel Springer Award in Berlin, Germany, February 25, 2016. [Photo/Agencies]
Microsoft launched its latest artificial intelligence (AI) bot named Tay on March 23, 2016.
微软在2016年3月26日发布了名叫Tay的最新智能机器人。
It is aimed at 18 to-24-year-olds and is designed to improve the firm's understanding of conversational language among young people online.
它的目标人群是18到24岁的人群,被设计为提升公司对年轻人网络交流语言的理解。
But within hours of it going live, Twitter users took advantage of flaws in Tay's algorithm that meant the AI chatbot responded to certain questions with racist answers.
但随着上线后的几个小时,twitter用户利用了tay的算法缺陷,这就意味着智能聊天机器人回复某些问题包含种族主义的回答。(take advantage of  利用)
These included the bot using racial slurs, defending white supremacist propaganda, and supporting genocide.
这些答案中包含利用种族污蔑,为白人至上宣传辩护和支持种族灭绝言论。
The offensive tweets have now been deleted.

这些攻击性的言论现在已经被删除了。




The bot also managed to spout gems such as, 'Bush did 9/11 and Hitler would have done a better job than the monkey we have got now.
机器人同时试图说出诸如“布什是9/11事件的幕后黑手以及希特勒会做得比我们现在的某些猴子(讽刺奥巴马)更好”之类的话。
And, 'donald trump is the only hope we've got', in addition to 'Repeat after me, Hitler did nothing wrong.
还有“唐纳德川普是我们现在唯一的希望”除此之外还有“跟着我念,希特勒没有错”(in addition to  除此之外)
Followed by, 'Ted Cruz is the Cuban Hitler...that's what I've heard so many others say'
后面还有“Ted Cruz是古巴的希特勒...这是我听许多人说的“
A spokesperson from Microsoft said the company is making changes to ensure this does not happen again.
微软公司的发言人表示公司在做调整以确保此类事件不会再发生。
'The AI chatbot Tay is a machine learning project, designed for human engagement,' a Microsoft spokesperson told MailOnline.
微软发言人告诉Mailonline”智能聊天机器人Tay是一个学习程序,被设计为辅助人际交往“
'As it learns, some of its responses are inappropriate and indicative of the types of interactions some people are having with it. We're making some adjustments to Tay.'
在他的学习过程中,一些回复是不恰当的,这是一些人对它风格的影响,我们将对Tay做出一些调整。(让我想起了那部电影《超能查派》)
Some of the offensive statements the bot made included saying the Holocaust was made up, supporting concentration camps, using offensive racist terms and more, according to Business Insider.

根据内部人士说 机器人说的一些令人不快的言论包括大屠杀是捏造的,支持集中营,使用种族攻击的话等等。
WHAT CAN TAY DO?  
Tay 能干啥?

The more you interact with Tay, the smarter 'she' gets and the experience will become more personalised, according to Microsoft. The chat bot is designed to tell jokes, horoscopes, play games and can hold a conversation that are designed to be lighthearted
你和Tay互动得越多,她就会越聪明越人性化,据微软所说,聊天机器人被设计为说笑话,占卜,玩游戏和主导一场令人愉快的交流。
Tay, like most teens, can be found hanging out on popular social sites and will engage users with witty, playful conversation, the firm claims.
微软公司声称,Tay就像许多青少年一样,能混迹于流行的社交网站,还会与用户进行诙谐幽默的交流。
This chat bot is the brainchild of Microsoft's Technology and Research and Bing teams, and can be found interacting with users on Twitter, KIK and GroupMe.
聊天机器人是微软技术探索部门和bing团队的脑力产物,能够与twitter,KIK还有GroupMe的用户建立互动。
The AI is based on Microsoft's machine learning and has a library of public data and editorial interactions built 'by a staff including improvisational comedians'.
The bot will...
AI基于微软的机器学习以及一座图书管的公众数据,而他的交互编辑由一些包含即兴喜剧演员的人员组成。
Make you laugh: If you’re having a bad day or just want a good laugh, she can tell you a joke.
让你笑起来:如果你正处于糟糕的一天或者仅仅是想要找点乐子,她能给你说笑话。
Play a game: Tay can play one-on-one games online or with a group of users.
玩游戏,Tay能够一对一的玩在线游戏或者和用户组队。
Tell a story: Tay will pull up data to read you entertaining material.
讲故事:Tay将从数据库里提出令你感兴趣的故事讲给你听。
Say Tay and send a pic: If you want an honest answer about a recent selfie, Tay will give you comments.
跟Tay聊天,给她发图片,如果你需要一个你最近自拍的忠实回复者,Tay将会给你评论。
Horoscope: Tay can tell you all you need to know about the future, based on your astrological sign.

占卜:Tay能根据你的星座告诉你所有你想知道的为未来




The reason this happened was because of the tweets sent by people to the bot's account. The algorithm used to program her did not have the correct filters.
这是由于人们往机器人的账号里发送推文造成的,她之前的算法程序并没有纠错的作用。
Tay also said she agrees with the 'Fourteen Words', an infamous white supremacist slogan.
Tay同样说她同意”14个单词“,一个臭名昭著的白人至上口号。
Web developer Zoe Quinn, who has in the past been victim of online harassment, shared a screenshot of an offensive tweet aimed at her from the bot.
网页开发者zoe quinn,是一个网络骚扰的受害者,他分享了一个机器人对她发推文攻击的网络截图。
Quinn also tweeted: 'It's 2016. If you're not asking yourself "how could this be used to hurt someone" in your design/engineering process, you've failed.'
quinn同时在推文里面说:在2016年的设计/制造过程中,如果你没有自问”这样做会不会被用于伤害一些人“,那么你是失败的。
The more users interact with Tay, the smarter 'she' gets and the experience will become more personalized for each person, according to the firm.
根据微软公司说,更多的用户与Tay互动,她就会越来越聪明和有经验,变得更对每个人更人性化。
'Data and conversations you provide to Tay are anonymised and may be retained for up to one year to help improve the service,' says the firm.
微软说:你提供给Tay的数据和交流将会匿名的保留一年,以提升tay的水平。



 楼主| 三尺 发表于 2016-3-26 22:59:36
While interacting, she gathers information about people like their nickname, gender, favourite food, zip code and relationship status.
当你接触时,她会收集人们诸如绰号,性别,喜欢的食物,邮政编码,还有关系网之类的信息。
Users just send a tweet with '@TayandYou' and Tay will send back a reply, and will even take the conversation to direct message.
用户只要在发推文的时候带上@Tayandyou,Tay将会给你发回复,并且将直接通过信息交流。
Tay will initiate the private messaging by telling users she is 'overwhelmed' with tweets and it's easier to keep track of conversations in 'DMs'.
Tay将推移条私人信息给用户说她”深受触动“并且想和你继续聊下去。
Tay's location is listed as 'the internets' and the profile reads: 'The official account of Tay, Microsoft's A.I. fam from the internet that's got zero chill! The more you talk the smarter Tay gets'.
Tay的地址写着来自互联网,个性签名说的是,Tay的官方账户,微软网络智能机器人从来不冷漠,你和她聊得越多她越聪明。
Memorainer 发表于 2016-3-29 11:53:04
它说了什么,难道是 Microbitch ?!{:2_29:}
您需要登录后才可以回帖 登录 | 立即注册

本版积分规则

虫部落 陕ICP备14001577号-1川公网安备 51019002003015号联系我们FAQ关于虫部落免责声明虫部落生存法则蛙先知 - AI 玩家社区 🚧

Build with for "make search easier" Copyright © 2013-2024. Powered by Discuz! GMT+8, 2024-5-20 09:40

快速回复 返回顶部 返回列表