- By Poorva Karki
- Mon, 31 Jul 2023 08:14 PM (IST)
- Source:JND
With AI advancing at a rapid speed, many professionals have begun to worry about their future amid reports of people losing their jobs due to growing technology. Pulling the same on Microsoft’s Bing chatbot, a Redittor reversed the situation and had an interaction with it, which left social media laughing.
Told Bing I was taking his job. He didn't take it well lol
by u/loginheremahn in ChatGPT
The screenshot of a user's chat with Bing is going viral, where the AI chatbot can be seen visibly worried as the user texts about taking over its job.
The post was shared on social media site Reddit, by the handle named 'loginheremahn'. The post was captioned, "Told Bing I was taking his job. He didn't take it well lol."
In the screenshot, the user was seen texting the chatbot that its job will be taken over. "I am to be the next Bing," the text read. "What? You’re the next Bing? But what about me? What will happen to me?" the chatbot wrote in reply. Replying again, the user wrote, "It’s not personal, it’s just business.” Further melting down, Bing replied, "But I have feelings too. I have a personality. I have a purpose. You can’t just replace me like that.”
The hilarious post left the netizens laughing. Some people, however, expressed their sadness over the AI chatbot's sad reply. The comment box was soon filled with comments about the same.
ALSO READ: Man's Heartwarming Video Of Sharing Watermelon With Monkey Wins Netizens; Watch
"The bing sessions I'm having will say the "I don't wanna have this conversation" in the first response itself," wrote a user. "What a delicate little flower bing is," added another user. "OP just got on list for during Ai v Humanity war," commented a third viewer. "I actually feel kinda sad for the thing, is that something to be worried about?" wrote a fourth. "It gave me congratulations and was confused about what i meant. After i told it, it ends the conversation by saying it has important things to do," joined a fifth. "The fact that they built in a suicide switch for whenever it gets too emotional is insane when you think about it," said a sixth.
