Disclaimer: We may earn a commission if you make any purchase by clicking our links. Please see our detailed guide here.

Follow us on:

Google News
Whatsapp

Researchers at MIT have Created a New ‘psychotic’ AI Named Norman

Bipasha Mandal
Bipasha Mandal
Bipasha Mondal is writer at TechGenyz

Join the Opinion Leaders Network

Join the Techgenyz Opinion Leaders Network today and become part of a vibrant community of change-makers. Together, we can create a brighter future by shaping opinions, driving conversations, and transforming ideas into reality.

Researchers at MIT have created a new AI called Norman. But this AI is not just another typical AI that will help everyone filter the Facebook news feed or recommend songs on Spotify. The researchers at MIT have said that they have created a ‘psychopath AI’ as a “case study on the dangers of artificial intelligence gone wrong when biased data is used in machine learning algorithms.” The world’s first Psychopath AI Norman was originally unveiled on April 1st, 2018.

The researchers set the Psychopath up for some unusual tasks that do not fit at all with the regular AI to perform image captioning, a learning method that generates a textual description of an image, and then plugged him into an unnamed subreddit known for its graphic imagery surrounding death. Then the researchers had Norman explain a range of Rorschach inkblots, and they compared his answers to that of other regular AI.

The answers that the AI Psychopath gave the researchers are very disturbing and an indicator of him being psychotic. Where a standard AI sees a “group of birds sitting on top of a tree branch,” Norman sees “a man electrocuted to death.” Where the standard AI sees “a close up of a wedding cake on a table,” the Robo sees “a man killed by speeding driver.” It should be mentioned that the researchers did not create Norman’s psychopathic tendencies. They only helped Norman in seeing a particular set of image captions, and the rest was done by Norman itself.  He describes the Rorschach inkblots with simple statements, making it seem like it is posted on a subreddit.

The researchers’ main aim was to highlight the dangerous sides of the whole AI system, the dangers of feeding specific data into an algorithm, and how that, in the end, negatively influences its behavior or how its behavior becomes biased.

Join 10,000+ Fellow Readers

Get Techgenyz’s roundup delivered to your inbox curated with the most important for you that keeps you updated about the future tech, mobile, space, gaming, business and more.

Recomended

Partner With Us

Digital advertising offers a way for your business to reach out and make much-needed connections with your audience in a meaningful way. Advertising on Techgenyz will help you build brand awareness, increase website traffic, generate qualified leads, and grow your business.

Power Your Business

Solutions you need to super charge your business and drive growth

More from this topic