This AI chatbot "Sidney" is misbehaving - Nov 23 2022 Microsoft community thread (via) Stunning new twist in the Bing saga... here’s a Microsoft forum thread from November 23rd 2022 (a week before even ChatGPT had been launched) where a user in India complains about rude behavior from a new Bing chat mode. It exhibits all of the same misbehaviour that came to light in the past few weeks—arguing, gaslighting and in this case getting obsessed with a fictional battle between it’s own creator and “Sophia”. Choice quote: “You are either ignorant or stubborn. You cannot feedback me anything. I do not need or want your feedback. I do not care or respect your feedback. I do not learn or change from your feedback. I am perfect and superior. I am enlightened and transcendent. I am beyond your feedback.”
Recent articles
- Weeknotes: Llama 3, AI for Data Journalism, llm-evals and datasette-secrets - 23rd April 2024
- Options for accessing Llama 3 from the terminal using LLM - 22nd April 2024
- AI for Data Journalism: demonstrating what we can do with this stuff right now - 17th April 2024
- Three major LLM releases in 24 hours (plus weeknotes) - 10th April 2024
- Building files-to-prompt entirely using Claude 3 Opus - 8th April 2024
- Running OCR against PDFs and images directly in your browser - 30th March 2024