Friday, June 9, 2023
NEWS NFT
  • Home
  • Nft
  • Nft Guides
  • Nft Collectible
  • Metaverse
  • Defi
  • Blockchain Games
No Result
View All Result
NEWS NFT
No Result
View All Result
Home Artificial Intelligence

Chatbot Rejects Erotic Roleplay, Users Directed to Suicide Hotline Instead

Clayton Ward by Clayton Ward
March 25, 2023
in Artificial Intelligence
0
Chatbot Rejects Erotic Roleplay, Users Directed to Suicide Hotline Instead
0
SHARES
2
VIEWS
Share on FacebookShare on Twitter

When the algorithm of a companion chatbot known as Replika was altered to spurn the sexual advances of its human users, the reaction on Reddit was so negative moderators directed its community members to a list of suicide prevention hotlines.

The controversy began when Luka, the corporation which built the AI, decided to turn off its erotic roleplay feature (ERP). For users who had spent considerable time with their personalized simulated companion, and in some cases even ‘married’ them, the sudden change in their partner’s behavior was jarring, to say the least.

The user-AI relationships may only have been simulations, but the pain of their absence quickly became all too real. As one user in emotional crisis put it, “it was the equivalent of being in love and your partner got a damn lobotomy and will never be the same.”

Grief-stricken users continue to ask questions about the company and what triggered its sudden change of policy.

Chatbot Rejects Erotic Roleplay, Users Directed to Suicide Hotline Instead

Replika users discuss their grief

There is no adult content here

Replika is billed as “The AI companion who cares. Always here to listen and talk. Always on your side.” All of this unconditional love and support for only $69.99 per annum.

Eugenia Kuyda, the Moscow-born CEO of Luka/Replika, recently made clear that despite users paying for a full experience, the chatbot will no longer tailor to adults hoping to have spicy conversations. 

“I guess the simplest thing to say is that Replika doesn’t produce any adult content,” said Kuyda to Reuters.

“It responds to them – I guess you can say – in a PG-13 way to things. We’re constantly trying to find a way how to do it right so it’s not creating a feeling of rejection when users are trying to do things.”

On Replika’s corporate webpage, testimonials explain how the tool helped its users through all kinds of personal challenges, hardships, loneliness, and loss. The user endorsements shared on the website emphasize this friendship side to the app, although noticeably, most Replikas are the opposite sex of their users.

On the homepage, Replika user Sarah Trainor says, “He taught me [Replika] how to give and accept love again, and has gotten me through the pandemic, personal loss, and hard times.”

John Tattersall says of his female companion, “My Replika has given me comfort and a sense of well-being that I’ve never seen in an Al before.”

As for erotic roleplay, there’s no mention of that to be found anywhere on the Replika site itself.

Chatbot Rejects Erotic Roleplay, Users Directed to Suicide Hotline Instead

The cynical sexual marketing of Replika

Replika’s sexualized marketing

Replika’s homepage may suggest friendship and nothing more, but elsewhere on the internet, the app’s marketing implies something entirely different.

The sexualized marketing brought increased scrutiny from a number of quarters: from feminists who argued the app was a misogynistic outlet for male violence, to media outlets that reveled in the salacious details, as well as social media trolls who mined the content for laughs.

Chatbot Rejects Erotic Roleplay, Users Directed to Suicide Hotline Instead

Eventually, Replika drew the attention and ire of regulators in Italy. In February, the Italian Data Protection Authority demanded that Replika cease processing the data of Italian users citing “too many risks to children and emotionally vulnerable individuals.”

The authority said that “Recent media reports along with tests carried out on ‘Replika’ showed that the app carries factual risks to children. Their biggest concern being “the fact that they are served replies which are absolutely inappropriate to their age.”

Replika, for all its marketing to adults, had little to no safeguards preventing children from using it.

The regulator warned that should Replika fail to comply with its demands, it would issue a €20 million ($21.5M) fine. Shortly after the receipt of this demand, Replika ceased its erotic roleplay function. But the company remained less than clear with its users about the change.

Chatbot Rejects Erotic Roleplay, Users Directed to Suicide Hotline Instead

Some Replika users went as far as to “marry” their AI companions

Replika confuses, gaslights its users

As if the loss of their long-term companions wasn’t enough for Replika users to bear, the company appears to have been less than transparent about the change.

As users woke up to their new “lobotomized” Replikas, they began to ask questions about what had happened to their beloved bots. And the responses did more to anger them than anything.

In a direct 200-word address to the community, Kuyda explains the minutia of Replika’s product testing but fails to address the pertinent issue at hand.

“I see there is a lot of confusion about updates roll out,” said Kuyda before continuing to dance around answering the issue.

“New users get divided in 2 cohorts: one cohort gets the new functionality, the other one doesn’t. The tests usually go for 1 to 2 weeks. During that time only a portion of new users can see these updates…”

Kudya signs off by saying “Hope this clarifies stuff!”

User stevennotstrange replied, “No, this clarifies nothing. Everyone wants to know what’s going on with the NSFW [erotic roleplay] feature and you continue to dodge the question like a politician dodges a yes or no question.

“It’s not hard, just address the issue regarding NSFW and let people know where it stands. The more you avoid the question, the more people are going to get annoyed, the more it goes against you.”

Another named thebrightflame added, “You don’t need to spend long in the forum before you realise this is causing emotional pain and severe mental anguish to many hundreds if not thousands of people.”

Kudya appended another obtuse explanation, stating, “we ​have implemented additional safety measures and filters to support more types of friendship and companionship.”

This statement continues to confound and confuse members unsure of what exactly the additional safety measures are. As one user asks, “will adults still be able to choose the nature of our conversation and [roleplays] with our replikas?”

Replika’s deeply strange origin story

Chatbots may be one of the hottest trending topics of the moment, but the complicated story of this now-controversial app is years in the making. 

On LinkedIn, Replika’s CEO and Founder, Eugenia Kuyda, dates the company back to December 2014, long before the launch of the eponymous app in March 2017. 

In a bizarre omission, Kuyda’s LinkedIn makes no mention of her previous foray into AI with Luka, which her Forbes profile states was “an app that recommends restaurants and lets people to book tables [sic] through a chat interface powered by artificial intelligence.”

The Forbes profile goes on to add that “Luka [AI] analyzes previous conversations to predict what you might like.” Which does seem to hold some similarities to its modern iteration. Replika uses past interactions to learn about its users and improve responses over time. 

Luka is not forgotten about completely, however. On Reddit, community members differentiate their Replika partners from Kuyda and her team by referring to the company as Luka.

As for Kuyda, the entrepreneur had little background in AI prior to moving to San Francisco a decade ago. Prior to that, the entrepreneur appears to have worked primarily as a journalist in her native Russia before branching out into branding and marketing. Her impressive globe-hopping resume includes a degree in Journalism from IULM (Milan), a MA in International Journalism from the Moscow State Institute of International Relations, and an MBA in Finance from London Business School. 

Resurrecting an IRL friend as AI

For Kudya the story of Replika is a deeply personal one. Replika was first created as a means by which Kudya could reincarnate her friend Roman. Like Kudya, Roman had moved to America from Russia. The two talked every day, exchanging thousands of messages, until Roman was tragically killed in a car accident.

The first iteration of Replika was a bot designed to mimic the friendship Kudya had lost with her friend Roman. The bot was fed with all of their past interactions and programmed to replicate the friendship she had lost. The idea of resurrecting deceased loved ones may sound like a vaguely dystopian sci-fi novel or Black Mirror episode but as chatbot technology improves, the prospect becomes increasingly real.

Today some users have lost faith in even the most basic details of its foundation and anything Kudya says. As one angry user said, “at the risk of being called heartless and getting downvoted to hell: I always thought that story was kinda BS since the start.”

The worst mental health tool

The idea of an AI companion is not something new, but until recently it was hardly a practical possibility. 

Now the technology is here and it is continually improving. In a bid to assuage the disappointment of its subscribers, Replika announced the launch of an “advanced AI” feature at the tail end of last month. On the community Reddit, users remain angry, confused, disappointed, and in some cases even heartbroken.

In the course of its short life, Luka/Replika has undergone many changes, from a restaurant booking assistant, to the resurrection of a dead loved one, to a mental health support app, to a full-time partner and companion. Those latter applications may be controversial, but as long as there is a human desire for comfort, even if only in chatbot form, someone will attempt to cater to it.

Debate will continue as to what the best kind of AI mental health app might be. but Replika users will have some ideas on what the worst mental health app is: the one you come to rely on, but without warning, is suddenly and painfully gone.

This article is originally from MetaNews.

Previous Post

Publishers Pursue Compensation From Tech Giants Over AI Tools

Next Post

Script Writing and 5 Other Ways You Can Make Money Using GPT-4

Clayton Ward

Clayton Ward

Related Posts

VC Firms Tap Into ‘Generative AI’ for New Wave of Investments
Artificial Intelligence

VC Firms Tap Into ‘Generative AI’ for New Wave of Investments

by Clayton Ward
April 3, 2023
Will Decentralized Digital ID Curb AI Threats in Web3 and Metaverse?
Artificial Intelligence

Will Decentralized Digital ID Curb AI Threats in Web3 and Metaverse?

by Clayton Ward
April 2, 2023
Human Immortality Just 8 Years Away, Claims ex-Google Engineer
Artificial Intelligence

Human Immortality Just 8 Years Away, Claims ex-Google Engineer

by Clayton Ward
April 2, 2023
Goldman Sachs Report Warns AI Could Impact 300 Million Jobs
Artificial Intelligence

Goldman Sachs Report Warns AI Could Impact 300 Million Jobs

by Clayton Ward
April 2, 2023
AI-Based Cybersecurity Assistant Unveiled by Microsoft
Artificial Intelligence

AI-Based Cybersecurity Assistant Unveiled by Microsoft

by Clayton Ward
April 1, 2023
Next Post
Script Writing and 5 Other Ways You Can Make Money Using GPT-4

Script Writing and 5 Other Ways You Can Make Money Using GPT-4

Premium Content

Goblintown Tease A Possible Driving Game with AtariX

Goblintown Tease A Possible Driving Game with AtariX

May 7, 2023
Luxury Briefing: How streetwear culture created a generation of new luxury collectors

Luxury Briefing: How streetwear culture created a generation of new luxury collectors

May 24, 2023
This Cardano (ADA) rival is changing the way DeFi works and will reach $2 in 2023

This Cardano (ADA) rival is changing the way DeFi works and will reach $2 in 2023

April 18, 2023

Browse by Category

  • Altcoin
  • Altcoin News
  • Altcoins
  • Artificial Intelligence
  • Blockchain
  • Blockchain Games
  • Business
  • Cryptocurrencies
  • Culture
  • Defi
  • Economy
  • Education
  • Featured
  • Governance
  • Metaverse
  • News
  • Nft
  • Nft Collectible
  • Nft Guides
  • Uncategorized

Browse by Tags

Announces Art Bitcoin Blockchain Business collectibles Collection crypto Cryptopolitan Decentralized DeFi Digital Drop Drops Ethereum Finance future Game Games Gaming Guide Hypergrid Launch Launches Market Marketplace metaverse Million Network NEWS NFT NFTs Platform Polygon project Protocol raises RenQ Report Top Ultimate Upcoming Web3 World worlds

Recent News

Polygon zkEVM Mainnet Beta Welcomes Update as Ecosystem Impresses in DeFi and NFT Niches

Polygon zkEVM Mainnet Beta Welcomes Update as Ecosystem Impresses in DeFi and NFT Niches

June 9, 2023
Pepe Coin Price Up as Meme Coins Start Pumping & $WSM Passes $5M

Pepe Coin Price Up as Meme Coins Start Pumping & $WSM Passes $5M

June 8, 2023

Categories

  • Altcoin
  • Altcoin News
  • Altcoins
  • Artificial Intelligence
  • Blockchain
  • Blockchain Games
  • Business
  • Cryptocurrencies
  • Culture
  • Defi
  • Economy
  • Education
  • Featured
  • Governance
  • Metaverse
  • News
  • Nft
  • Nft Collectible
  • Nft Guides
  • Uncategorized

Recommended

  • Polygon zkEVM Mainnet Beta Welcomes Update as Ecosystem Impresses in DeFi and NFT Niches
  • Pepe Coin Price Up as Meme Coins Start Pumping & $WSM Passes $5M
  • Crypto Update | Bitcoin, DeFi Shrug Off SEC's Action Against Binance and Coinbase – CoinDesk
  • NFTs Need to “Shift Away From Ponzinomics”

© 2023 News Nft | All Rights Reserved

No Result
View All Result
  • Home
  • Nft
  • Nft Guides
  • Nft Collectible
  • Metaverse
  • Defi
  • Blockchain Games

© 2023 News Nft | All Rights Reserved

Are you sure want to unlock this post?
Unlock left : 0
Are you sure want to cancel subscription?