Back to Articles
industry

NYC's AI Chatbot Told Businesses to Break the Law. Here's What Went Wrong.

A $500K government chatbot gave illegal advice—take worker tips, discriminate against Section 8 tenants. Now it's being killed. The lessons every AI builder must learn.

Serenities Team10 min read
AI chatbot interface

Half a Million Dollars to Tell People to Break the Law

In fall 2023, New York City launched what was supposed to be the future of government services. Mayor Eric Adams unveiled an AI chatbot built on Microsoft's cloud platform—a flagship piece of the ambitious MyCity digital overhaul.

The promise: give business owners an accessible way to check city rules and regulations. No more wading through bureaucracy. Just ask the bot.

The reality: the bot confidently told businesses to take employee tips (illegal), discriminate against Section 8 voucher holders (illegal), refuse cash payments (illegal since 2020), and gave incorrect minimum wage information.

This week, New York's new Mayor Zohran Mamdani announced the chatbot would be killed. The cost to taxpayers? Nearly $600,000 to build a system he called "functionally unusable."

The NYC chatbot is not just a government technology failure. It is a case study in everything that can go wrong when AI is deployed without proper validation—and a warning for every organization building AI-powered tools.

What the Chatbot Actually Said

Investigative reporting by The Markup and THE CITY documented the bot's most egregious failures:

On tipping: The chatbot told businesses they could take a cut of employee tips. Under federal and New York labor law, this is illegal. Tips belong to workers.

On housing discrimination: When asked about housing policy, the bot suggested landlords could discriminate against tenants with Section 8 vouchers. New York has explicit laws prohibiting this discrimination.

On cash payments: The bot told users it was fine to refuse cash payments. New York City enacted a law in 2020 requiring most businesses to accept cash—specifically to prevent discrimination against unbanked populations.

On minimum wage: The bot could not accurately state the minimum wage. This is not an obscure regulation. It is the most basic information a business owner needs.

These were not edge cases uncovered by adversarial testing. These were responses to straightforward questions that any business owner might reasonably ask.

The Adams Administration's Response

When confronted with evidence of the chatbot's failures, the Adams administration did not pull the plug. Instead, they defended it.

"We're identifying what the problems are, we're gonna fix them, and we're going to have the best chatbot system on the globe," Adams said at a press conference. "People are going to come and watch what we're doing in New York City."

The administration made some changes. They added disclaimers advising users to "not use its responses as legal or professional advice." They improved some answers. But they also appeared to limit what kinds of questions the tool would answer—essentially neutering its core functionality while keeping it online.

Today, the bot warns visitors to "ask an NYC government question only" and cautions that "responses may occasionally produce inaccurate or incomplete content." Users must agree to accept limitations before using it.

In other words: the bot that was supposed to make government accessible now requires you to accept that it might give you incorrect information about laws that could expose you to legal liability.

The Real Cost

The $500,000-$600,000 price tag is what we know about. The full cost of the MyCity project—which relied heavily on outside contractors—has been criticized for its opacity.

But the dollar figure understates the true damage:

Reputational harm to AI: Every high-profile AI failure makes it harder for legitimate AI applications to gain trust. Public skepticism rises. Regulation tightens (often poorly). The NYC chatbot will be cited in AI policy debates for years.

Actual harm to businesses: If a business owner followed the chatbot's advice about tips or Section 8, they could face lawsuits, fines, or criminal penalties. The city deployed a system that could cause direct legal harm to the people it claimed to help.

Erosion of public trust: Government technology projects already face skepticism. Spending half a million dollars on a bot that tells people to break the law confirms every cynical assumption about bureaucratic incompetence.

Opportunity cost: That money could have funded human navigators, improved websites, or any number of alternatives that actually work.

What Went Wrong: A Technical Autopsy

How does a chatbot—built by Microsoft's platform, deployed by a major city government—give illegal advice?

Problem 1: No validation layer. The bot generated responses without checking them against authoritative legal sources. It treated law and policy like any other text—something to predict, not verify.

Problem 2: Confidence without accuracy. Like many LLM-based systems, the bot spoke with confidence regardless of correctness. There was no mechanism to express uncertainty or refuse to answer questions beyond its competence.

Problem 3: No domain expertise in the loop. The contractors who built the system apparently did not have employment lawyers, housing experts, or regulatory specialists reviewing outputs before deployment.

Problem 4: Inadequate testing. Basic questions about minimum wage, tipping laws, and housing discrimination should have been in any QA checklist. These failures suggest testing was either minimal or completely absent.

Problem 5: No feedback mechanism. Even after deployment, there was no effective way to identify and correct bad responses before they caused harm. The Markup found these issues; the city did not.

The Lessons for AI Builders

The NYC chatbot failure is not unique. It is a particularly visible example of patterns that exist across the AI industry. Every organization deploying AI—especially in high-stakes domains—should internalize these lessons:

1. Validation Is Not Optional

AI systems that provide advice on legal, financial, medical, or safety topics need validation against authoritative sources. You cannot ship an LLM wrapper and hope for the best.

This is why Serenities built Flow with validation at its core. Every generated response can be checked against your authoritative documentation, regulatory databases, and domain rules. If the AI says something that contradicts your verified sources, you catch it before users see it.

2. Confidence Must Match Competence

LLMs are trained to sound confident. That is a bug when dealing with factual claims. Systems need mechanisms to express uncertainty, refuse to answer beyond their knowledge, and direct users to authoritative human sources.

The NYC bot had no such guardrails. It answered questions about tip law with the same confidence it might discuss the weather—even when its answer was legally actionable misinformation.

3. Domain Experts Must Be in the Loop

No amount of prompt engineering substitutes for subject matter expertise. If you are building a legal advice bot, lawyers must review outputs. If you are building a medical bot, doctors must validate it.

The NYC project apparently outsourced development to contractors without ensuring legal review. The result was predictable.

4. Testing Must Be Adversarial

Your QA process should actively try to break your system. Ask the questions users will ask. Try to get it to say harmful things. Red team it before reporters do.

The Markup found the NYC bot's failures within days of focused testing. The city apparently never conducted similar tests—or ignored the results.

5. Disclaimers Are Not Guardrails

After the failures were exposed, the city added disclaimers. This is the minimum viable liability defense, not a solution.

Disclaimers do not prevent harm. They just shift blame. A business owner who follows illegal advice because a government website told them it was legal has been harmed—regardless of what fine print they clicked through.

Real guardrails prevent bad outputs. Disclaimers acknowledge you cannot prevent them.

6. "It Will Improve Over Time" Is Not a Strategy

Adams defended the chatbot by saying it would get better. But AI systems do not improve automatically. They improve through deliberate effort: better training data, refined prompts, validation systems, user feedback loops, expert review.

"It will improve" is an aspiration, not a plan. And while you wait for improvement, the system is actively causing harm.

The Enterprise AI Imperative

The NYC chatbot failure highlights why enterprise AI requires fundamentally different approaches than consumer AI.

Consumer chatbots can afford some hallucination. If ChatGPT gives you a mediocre recipe suggestion, the stakes are low.

Enterprise AI—especially in regulated domains—cannot afford any hallucination. When your system gives legal advice, medical guidance, or policy information, accuracy is not a nice-to-have. It is the entire value proposition.

This is why the consumer-AI-with-disclaimers approach fails in enterprise contexts. You cannot slap a warning on a legal chatbot and call it deployed.

Real enterprise AI needs:

  • Validation against authoritative sources
  • Confidence scoring and uncertainty expression
  • Human expert oversight before deployment
  • Audit trails for every response
  • Rapid correction mechanisms when errors are found
  • Domain-specific guardrails, not generic filters

The Road Ahead

Mayor Mamdani's decision to kill the chatbot is the right call. But it is also a loss—not of the bot itself, but of the vision it represented.

Government services should be more accessible. AI can help with that. The problem is not the goal; it is the execution.

Other cities and agencies will learn from New York's failure. Some will conclude that AI is too risky for government. That would be the wrong lesson.

The right lesson: AI in high-stakes domains requires validation infrastructure, domain expertise, rigorous testing, and ongoing oversight. You cannot shortcut these requirements with confidence and disclaimers.

The organizations that get AI right—in government and enterprise—will be the ones that treat accuracy as a first-class requirement, not an afterthought.

At Serenities, we built Flow specifically because this problem is solvable. Validation layers, expert review workflows, confidence scoring, and guardrails are not futuristic concepts. They are available now. The NYC chatbot failed not because the technology does not exist to prevent such failures, but because no one implemented it.

The Tombstone

Half a million dollars. Functionally unusable. Confidently wrong about laws that protect workers and tenants.

The NYC AI chatbot joins a growing list of public AI failures—systems deployed with enthusiasm and inadequate preparation, leaving damage in their wake.

Its epitaph should be a warning: AI confidence is not AI correctness. And in high-stakes domains, only correctness counts.

AI
chatbot
government
enterprise AI
guardrails
NYC
Share this article

Related Articles

Ready to automate your workflows?

Start building AI-powered automations with Serenities AI today.