Been testing AI agents for customer support for about a year. Here is the honest breakdown of what actually worked.
So I have been deep in this space for about a year now across our support queue and honestly the conversations I keep seeing online still feel too clean compared to what actually happens in production.
Here is what I have actually learned from running this:
Intercom Fin - strong at deflecting repetitive volume but the setup to get it talking properly about your specific product is more work than they make it sound
Zendesk AI - powerful if you are already deep in the ecosystem, felt clunky to configure outside of it
Ada - serious automation muscle but when it misses it misses confidently which is the worst version of wrong
Chatbase - been on this one the longest, about a year now. The Zendesk integration is what kept us on it. When the agent cannot resolve something the full conversation history transfers with the ticket automatically so agents never pick up cold. 71% resolution rate, CSAT held.
Freshdesk Freddy - fine for getting started, hit its ceiling faster than expected
The thing nobody talks about enough is the maintenance side. Every single one of these tools is only as good as what you feed it and how often you update it. The ones that fell apart on us fell apart because we treated them like infrastructure instead of something that needs a weekly 15 minute review.
The bar has shifted from can it reply to can it actually close the ticket. But I would add a third question now: can it stay accurate six months after you deployed it without someone actively maintaining it. That is where most of them quietly fail.
What are you all running? And genuinely curious if anyone else has had something work great in month one and then slowly fall apart.