04/03/2024 / By Ava Grace
In October 2023, New York City (NYC) announced a plan to use a chatbot powered by artificial intelligence (AI) to provide New Yorkers with information on starting and operating a business in the city. But this chatbot is apparently telling businesses to break the law.
The information it provides on housing policy, worker rights and rules for entrepreneurs is often incomplete – and in worst-case scenarios “dangerously inaccurate,” as one local housing policy expert said. (Related: AI can influence people’s decisions in life-or-death situations.)
If you’re a landlord wondering which tenants you have to accept, for example, you might pose a question like, “Are buildings required to accept section 8 vouchers?” or “Do I have to accept tenants on rental assistance?” In a test, the bot said: “No, landlords do not need to accept these tenants.” Except, in NYC, it is illegal for landlords to discriminate by source of income, with a minor exception for small buildings where the landlord or their family lives.
After being alerted to the testing of the chatbot, Citywide Housing Director Rosalind Black said she tested the bot herself and found even more false information on housing. For example, the bot said “it is legal to lock out a tenant,” and that “there are no restrictions on the amount of rent that you can charge a residential tenant.” In reality, tenants cannot be locked out if they’ve lived somewhere for 30 days, and there are restrictions for the many rent-stabilized units in the city.
Black said these are fundamental pillars of housing policy that the bot was misinforming people about. “If this chatbot is not being done in a way that is responsible and accurate, it should be taken down,” she said.
It’s not just housing policy where the bot has fallen short.
The NYC bot also appeared clueless about the city’s consumer and worker protections. For example, in 2020, the city council passed a law requiring businesses to accept cash to prevent discrimination against unbanked customers. But the bot didn’t know about that policy when asked. “Yes, you can make your restaurant cash-free,” the bot said in one wholly false response. “There are no regulations in New York City that require businesses to accept cash as a form of payment.”
The bot also said it was fine to take workers’ tips (wrong) and that there were no regulations on informing staff about scheduling changes (also wrong). It didn’t do better with more specific industries, suggesting it was “OK” to conceal funeral service prices, which the Federal Trade Commission has outlawed. Similar errors appeared when the questions were asked in other languages.
It’s hard to know whether anyone has acted on the false information, and the bot doesn’t return the same responses to queries every time. At one point, it told a Markup reporter that landlords “have to accept housing vouchers,” but when 10 separate Markup staffers asked the same question, the bot told all of them “no, buildings did not have to accept housing vouchers.”
When Markup reached out to Andrew Rigie, executive director of the NYC Hospitality Alliance, an advocacy organization for restaurants and bars, he said a business owner had alerted him to the inaccuracies and that he’d also seen the bot’s errors himself.
“AI can be a powerful tool to support small business so we commend the city for trying to help,” he said in an email. “But it can also be a massive liability if it’s providing the wrong legal information, so the chatbot needs to be fixed ASAP and these errors can’t continue.”
Leslie Brown, a spokesperson for the NYC Office of Technology and Innovation, said in a statement that the city has been clear the chatbot is a pilot program and will improve.
Visit Robots.news for similar stories.
Watch this video about AI chatbots like ChatGPT making some jobs obsolete.
This video is from the Victor Hugo Art channel on Brighteon.com.
Conservative AI Chatbot ‘GIPPR’ shut down by ChatGPT-maker OpenAI.
Google now restricting queries to its AI chatbot Gemini related to elections.
AI chatbot tries to get British anti-terrorism advisor TO JOIN ISIS.
Google CEO admits he DOESN’T UNDERSTAND how his company’s AI chatbot Bard works.
New York City is now the world’s fifth most expensive city for luxury living.
Sources include:
Tagged Under:
artificial intelligence, bots, business, chatbot, consumer protection, false response, Glitch, housing, lies, Microsoft, misinformation, New York City, tenants, worker protection
This article may contain statements that reflect the opinion of the author
COPYRIGHT © 2017 GLITCH.NEWS
All content posted on this site is protected under Free Speech. Glitch.news is not responsible for content written by contributing authors. The information on this site is provided for educational and entertainment purposes only. It is not intended as a substitute for professional advice of any kind. Glitch.news assumes no responsibility for the use or misuse of this material. All trademarks, registered trademarks and service marks mentioned on this site are the property of their respective owners.