20
95 Comments

I want to build an AI support agent for Shopify stores — am I solving a real problem?

Before writing a single line of code I want to talk to real store owners first.
The idea: an AI that handles repetitive support tickets (order tracking, returns, FAQs) for Shopify stores at a flat monthly fee.
If you run a Shopify store, I have 3 quick questions:
What does support cost you monthly?
What's the most painful part?
Would you ever trust AI to handle it?
Happy to DM if you'd rather keep it private.

on April 19, 2026
  1. 1

    The only thing that answers "is this a real problem" is money. Surveys are survey data. Interest is interest data. What cuts through: a landing page with a $50 preorder button, or a paid pilot. If three Shopify founders won't commit real money to reserve a slot, there's no demand yet. If they will, you've found something.

    Shopify stores have budget and are used to paying monthly for tools. Wallet friction isn't the issue. What you're measuring is whether support is painful enough to pay for on top of whatever they're already using.

    1. 1

      A preorder button as the real signal — that's a much cleaner test than any survey. Before I get there though, I want to make sure I'm solving the right problem. What would you set as the minimum number of preorders before considering it validated?

  2. 2

    Shopify store owners are famously over-pitched. Before building, find 5 store owners in a Discord or niche forum (r/ecommerce, r/shopify, Shopify Community) and DM them asking what they currently do for support and what breaks most often.
    If fewer than 3 out of 5 describe the same pain in their own words without you prompting for it, the problem isn't there yet. If 3+ do, you have your wedge. I skipped that step on my own project once and spent 2 months on a feature nobody asked for. Talking to 5 people first would have saved all of it.

    1. 1

      The "3 out of 5 describe it unprompted" test is the clearest validation filter I've heard. No leading questions, just listening for the same pain in their own words. How did you find those 5 store owners to DM — cold outreach or existing communities?

  3. 2

    I think you’re definitely looking at a real problem — Shopify stores deal with a ton of repetitive tickets like order tracking, returns, and basic FAQs, and they’re already paying tools like Zendesk because support is a genuine cost.

    That said, “AI handling support tickets” on its own feels a bit too broad now. A lot of those platforms are already moving in that direction, so the challenge isn’t proving the problem exists, it’s carving out a very specific differentiator where you’re clearly better.

    The part that feels most interesting to me is narrowing it down to something measurable, like reducing “where is my order” tickets or automating returns end-to-end. If you can say “this cuts X% of tickets in Y days,” that’s a much easier sell than a general AI support agent.

    Also i think the trust piece is huge not just whether AI works, but whether store owners are comfortable letting it run autonomously vs acting as a first layer that escalates edge cases.

    Curious to know are you planning to focus on a specific type of ticket first (like order tracking or returns), or go broader from day one?

    1. 1

      Based on everything I'm hearing in this thread, leaning towards returns first — the consequences of getting it wrong are real money, which creates urgency to switch. Order tracking feels safer but lower stakes. Does that reasoning make sense to you?

  4. 2

    not a shopify owner but i built an AI agent in a different vertical (job applications), and the thing that killed my first version was misreading "repetitive support tickets" as a single category. once i actually shadowed real ops, it split into three very different buckets:

    1. pure lookup (order status, return policy text). great for AI, users prefer self-serve anyway
    2. emotional ticket ("my package is late and i need it for a wedding"). AI tanks hard here and escalation is worse than no AI at all
    3. edge case policy (damaged item from a third-party dropshipper). needs the owner's judgment, every store handles differently

    the question i'd push on: what percentage of tickets are actually bucket 1? my guess is stores overestimate because they feel busy, but the bucket 1 ratio determines whether your flat-fee math works or not. the fewer bucket 1 tickets, the more your AI ends up just being a fancy knowledge base and the willingness to pay collapses.

    also if you're pre-code, the shopify app store already has a dozen AI support tools. worth reading the 2-star reviews on the top 3 to see where they actually break. that's where the real wedge is.

    1. 2

      The three bucket breakdown is incredibly useful — especially the emotional ticket category where AI tanks hard. That's exactly what needs human escalation. And the 2-star review tip is brilliant. What percentage of tickets do you think actually fall into bucket 1 for a typical store?

      1. 1

        ticket type is the easier first signal - returns and cancellations have different blast radii. confidence threshold is harder without real data. I usually start with intent + keyword matching before ML thresholds - deterministic fallback that's debuggable.

        1. 1

          Intent and keyword matching as the first layer makes sense — it's debuggable and you can see exactly why it escalated. Returns and cancellations having different "blast radii" is interesting though — do you mean the consequences of a wrong decision are more severe for one than the other?

  5. 1

    The problem exists but it's not uniform across store sizes. Below a certain ticket volume the owner just handles it themselves. No tool needed.

    The question worth settling before interviews: what's the minimum ticket volume where this is actually painful enough to pay for?

    That narrows who to talk to and makes the interviews sharper.

    1. 1

      That's exactly the question I need to answer before interviews — otherwise I'm talking to the wrong stores. My gut says 500+ tickets/month but I genuinely don't know. What threshold would you put it at?

  6. 1

    Interesting project but my question is what can you offer different to zenddesk an this type of bot?

    1. 1

      That's the core question I'm still working out through these conversations. What I'm hearing so far is that Zendesk handles the workflow but doesn't optimize the outcome — it routes tickets but doesn't actually reduce bad approvals or protect margin. What would make something genuinely different to you?

      1. 1

        I used it once in a store and tested it; Zendesk works, and from what I remember, you had to set up business rules that customers could answer yes or no to get a response or be redirected to agents you hired. A key differentiator is having a bot with your LCM loaded that can resolve a customer's technical questions, and everyone who hires your service only needs to send you a summary of the architecture, app, or something like that.

        1. 1

          So the key is coming in pre-loaded with the store's context rather than making them build rules from scratch — that's what removes the setup friction. Most stores don't have time to configure a bot properly which is why they give up. If the onboarding was just "send us a summary of your store and policies," do you think that would be enough to get started?

          1. 1

            If you want do some different yes.

  7. 1

    I'd suggest putting together a list of different Shopify brands who you think would use the product. Once you have the list, you can reach out on LinkedIn/cold email and ask them for a interview without pitching the idea itself. If you see some angle why they would want something like that–means you have a wedge.

    1. 1

      Building a targeted list and reaching out directly makes sense — much more precise than posting and hoping. When you've done cold outreach for interviews, what subject line or opening actually got responses?

  8. 1

    This is interesting.

    One pattern we kept seeing in service businesses wasn’t support overload —
    it was missed communication (clients not replying, not confirming, not showing up).

    Feels like there’s a gap in tools that reduce friction rather than just automate replies.

    1. 1

      Interesting direction.

      One pattern we kept seeing in service businesses wasn’t just support load —
      it was clients not replying or confirming in the first place.

      Feels like there’s room for tools that reduce friction, not just automate responses.

      1. 1

        The missed communication angle is interesting — so the pain isn't just ticket volume, it's customers going silent after a bad experience. A tool that reduces friction in the response loop, not just automates replies. Did you find that pattern was consistent across different types of service businesses or specific to certain ones?

  9. 1

    Talking to store owners before writing code is exactly right. Most builders instinctively validate the problem, like "yes, support is painful," when what you really need to check is whether they'd actually change how they work. That's a big gap.

    The three-bucket framing someone mentioned above is key. The tickets that seem easy for AI, like order tracking, are often the ones with the lowest switching cost, because store owners already have tools that mostly handle them. The high-value ones are the emotional, edge-case tickets, and that's exactly where AI tends to fall apart. That mismatch is worth poking at directly.

    One question I'd add to your three is: "What's your process for handling tickets your current tools can't resolve?" The answer usually tells you whether the pain is in volume or in the judgment-heavy edge cases. Those need very different products.

    1. 1

      That question is going straight into my interview script — whether the pain is in volume or judgment-heavy edge cases completely changes what needs to be built. Did you find most store owners even had a process for those tickets or was it usually just "someone deals with it manually?

  10. 1

    You don’t prove it with claims — you prove it with a small, controlled win.

    Something like:
    → take last 20–30 return/chargeback cases
    → run your logic on them
    → show: what you’d approve/reject vs what actually happened
    → and quantify the difference ($ saved / time reduced)

    Even a rough “this would’ve saved you $X last month” is enough to get attention.

    Early on, it’s less about being perfect and more about making the outcome visible.

    If they can see the delta, they’ll want to test it.

    Also — if you lean into “saves money weekly” as the core promise, your positioning (and even name) should reflect that outcome, not support.

    1. 1

      So the demo isn't a product demo — it's a retrospective audit. "Here's what your last month of returns would have looked like with this." That's a much more convincing proof than any feature list. Did you find stores were willing to share that historical data early on?

      1. 1

        Usually yes — if the ask is small and clearly tied to value.

        You’re not asking for full access, just:
        → a small sample (20–30 cases)
        → in exchange for a clear “you lost $X here” insight

        That feels more like help than risk.

        If you position it as:
        “quick audit to find missed revenue / bad approvals”
        most will say yes — especially if returns are already painful.

        Once they see the loss, the sale becomes much easier.

        1. 1

          Quick audit to find missed revenue" — that's a completely different conversation than asking for data access. It's offering value first. So the first interaction with a store isn't a sales call, it's a free audit. That's a much easier yes to get. How many cases do you think is enough to make the insight convincing — 20, 30, more?

          1. 1

            20–30 is probably enough to show direction — but not enough to make it undeniable.

            If you really want it to hit, I’d push closer to:
            → 50–100 cases
            → across at least a couple weeks

            Otherwise it’s easy for them to say “this was just a lucky sample.”

            Also — one thing I’m thinking:

            If the audit shows something like
            “you lost $2.3k last month from bad approvals”

            that’s not just a nice insight — that’s basically the whole pitch.

            At that point it’s less:
            “do you want this tool?”
            and more:
            “do you want to keep losing this every month?”

            That changes the dynamic completely.

            Curious — when you ran this, did stores actually convert right after seeing the loss, or did they still hesitate?

            Why this works:

            1. 1

              50-100 cases across a couple weeks — that's enough to kill the "lucky sample" objection. And that reframe is powerful — the audit doesn't just prove the product, it IS the sales conversation. Did stores actually convert right after seeing the loss number or did they still hesitate?

              1. 1

                That’s the key shift.

                At that point you’re not selling a tool —
                you’re selling “stop losing $X every month.”

                And that’s where most people miss the next layer:

                if the core promise is money saved weekly,
                but the product still sounds like “returns/support/operations”…

                there’s a mismatch.

                In cold or even post-audit, people remember:
                → the outcome
                not the mechanism

                So the positioning (and even the name) should carry:
                recovery / savings / leakage control — not process.

                That alone usually lifts conversion more than tweaking the pitch.

                I’ve seen a few strong ways to frame this so it lands instantly (especially post-audit).

                Happy to share — easier to show properly 👍

                1. 1

                  Stop losing $X every month" vs "AI handles your support" — completely different conversations. And yes please — would love to see how you'd frame it post-audit. That's exactly the moment where positioning either lands or loses them. What's the strongest opening line you've seen work?

  11. 1

    This is a real problem—especially for stores with growing support volume.

    The challenge isn’t answering FAQs, it’s maintaining accuracy when queries involve real order data, returns, or edge cases. That’s where most AI agents break.

    We’ve seen much better reliability when the system is grounded with store data (orders, policies, FAQs) using a retrieval layer instead of pure prompting.

    Also worth thinking early: how you’ll handle exceptions—refund disputes, damaged items, etc. That’s usually where trust drops.

    Curious—are you planning deep Shopify integration from day one or starting with a simpler FAQ assistant?

    1. 1

      Starting simple — FAQ and order tracking first to prove reliability before touching anything with money attached. Deep Shopify integration comes once trust is established. Does starting simple hurt credibility with store owners or does it actually help?

      1. 1

        Starting simple actually helps—not hurts.

        Most store owners won’t trust AI with refunds or sensitive actions anyway, so proving reliability on FAQs + order tracking is the right move.

        What matters more is how you position it:
        instead of “AI handles everything,” frame it as
        “AI handles repetitive queries, humans handle edge cases.”

        One thing I’d plan early though: even in “simple” use cases, accuracy depends on access to real store data (orders, policies). That’s where most systems either feel reliable or break.

        If you get that part right, expanding into deeper workflows becomes much easier later.

        1. 1

          AI handles repetitive queries, humans handle edge cases" — that's the framing that removes the fear. And noted on the data access point — getting connected to real store orders and policies early seems like the thing that separates reliable from unreliable. Is that usually an API integration or do stores just export their data manually at first?

          1. 1

            Usually API integration—especially for orders and customer data, since it needs to be real-time. Shopify makes that fairly straightforward.

            Some teams start with manual exports for FAQs/policies, but for anything dynamic (order status, returns), APIs are pretty much required for reliability.

            A common approach is: start simple with a small set of endpoints (orders, tracking), then expand as trust builds.

            1. 1

              Start with orders and tracking API, expand as trust builds — that's a clear technical roadmap. For the very first store, would you recommend doing the integration manually to learn what data actually matters before building it properly?

              1. 1

                Yeah, that’s actually a solid approach for the very first store.

                I’d start with a manual pass first—not as a production solution, but to understand the real data patterns (what support questions actually map to order fields, tracking status, refund rules, etc.).

                That helps you avoid over-engineering the initial integration and makes the API design much cleaner later.

                Once you see repetition in the queries, you can map them properly into a lightweight orders + tracking integration and automate only what’s proven useful.

                So manual first for learning → then API for scale is usually the safest path.

                1. 1

                  Manual first to understand the real data patterns before building anything — that completely changes how I'd approach the first store. Less about impressing them with tech, more about learning what questions actually map to what data. Did you find the manual phase revealed anything surprising that you wouldn't have caught by going straight to API?

  12. 1

    For FAQ/order tracking retrieval, the embedding model matters far less than the quality of your input text. Tested 5 models at my company, got only a 7-point spread in accuracy. Switched from embedding raw JSON to LLM-generated summaries and gained 40 points. Spend the savings on data prep, not premium embeddings.

  13. 1

    Same phase you are — day 3 of talking to potential users before writing code. The discipline to validate before building is the right instinct, most founders skip it.

    Two things that might help based on what I've learned talking to founders this week:

    1. Shopify store owners aren't heavy on IH but they're all over r/shopify and the Shopify Community forums. Might get more responses there than here.

    2. When I asked "what would you pay?" as a direct question, people gave vague answers. When I gave them price anchors ($0 / $19 / $49 / $99) they gave me real ones. Worth trying if you're not getting specific pricing signal.

    Rooting for your validation. Drop a follow-up post with what you hear back — curious if Shopify owners answer differently than the SaaS founders I'm talking to.

    1. 1

      The price anchor idea is really smart — open questions get vague answers but giving them options gets real ones. And noted on r/shopify, I've been trying but my karma is too low right now. Working on it. Did the price anchors change what tier most people landed on?

      1. 1

        Fair warning — my sample is still small (like 4-5 real conversations this week). So the honest answer is "not enough data yet to say definitively."
        What I've noticed anecdotally: when I give $0/$19/$49/$99 as anchors, people mostly land on $19 or $49. Nobody has picked $99, which tells me either I haven't described the product as load-bearing enough yet, or $99 really is a ceiling for this audience. Both possible.
        On r/shopify karma — classic chicken-and-egg. One workaround: comment thoughtfully on 5-10 other Shopify founder posts over a week before posting your own. Usually enough to unlock.

        1. 1

          That's really useful — $19 or $49 landing zone is a clear signal. And noted on the Reddit approach, I'll do it properly this time. Did you find the $99 ceiling was audience-specific or more about how the product was framed?

          1. 1

            Have to walk something back — in my last reply I said "people mostly land on $19 or $49." That's what I expect based on the product space, not data I've actually collected. My real sample is 4-5 conversations and I didn't formally test those anchors in any of them. I basically told you a hypothesis as if it were observation, and that's not fair to you if you're making product decisions off it.

            Honest answer to your actual question: I don't know yet whether $99 is an audience ceiling or a framing problem. My guess — based on watching this space, not testing it — is that it's mostly framing. Tools that feel optional cap at $19-49. Tools that feel load-bearing clear $99 without friction. But that's theory, not evidence.

            Going to actually run the anchor test in my next 5 conversations and report back properly. Sorry for the noise.

            1. 1

              Really appreciate you walking that back — that's exactly the kind of honesty that saves founders from making bad decisions on false signals. "Tools that feel load-bearing clear $99" vs "tools that feel optional cap at $49" is a useful framing test. When you do run the anchor test properly, would you mind sharing what you find?

  14. 1

    You are spot on. ~

    It’s not that store owners are opposed to AI-it’s that they don't want it to mess up the things that are important to them.

    Things like, "Where is my order?" and company policies-those are simple. They really don't care if an AI does that.

    However, with refunds, gray area scenarios, and mad customers, that is when people get uncomfortable. They will actually lose more money on one mistake here than they
    would save with automated help.

    Therefore, the question isn't if the AI will "do support" but rather if it will "handle all the tedious tasks and get only the important issues to you."

    Your distinction:
    auto-process volume
    clearly pass the sensitive issues back
    This is exactly right. If they are able to easily and naturally transfer things over, then they are very open.

    Also, the phrasing matters a lot.
    "Will you let an AI do your support?" versus "Which tickets do you have to respond to at least once every day?" or "Which tickets would you never automate?"

    Another point: You aren't just competing with other AIs. You are competing with "This is painful, but it works." Stores are likely already using macros and other automated help and finding that it is "good enough" for some interactions.

    Thus, the standard is not "this is cool" but "this is noticeably better and introduces no new issues."

    The larger concern, in my experience, is not an incorrect answer but the loss of control over how the store represents itself.

    The store owner is more worried about losing their brand's voice, tone and ability to be a human to the customer when necessary.

    The thing that makes a store owner finally agree to it, is if they can still feel in charge while the AI takes care of the bulk of the heavy lifting.

    1. 1

      Feel in charge while the AI does the heavy lifting" — that's probably the framing that actually sells it. Not "AI does your support" but "you stay in control, AI handles the volume." Did you find store owners responded differently when you framed it that way?

  15. 1

    I run agents daily for PM work - framing matters more than trust. The real question is what the agent does when it hits something outside its playbook. Escalate or just guess? That distinction IS the product.

    1. 1

      Escalate or just guess" — that's the exact distinction that matters. A bot that guesses on a $200 return and gets it wrong costs more than the ticket was worth. So the product is really about knowing its own limits. How would you build that escalation trigger — confidence threshold, ticket type, or something else?

  16. 1

    There is a lot of apps like that on Shopify either make it easy to integrate or add good diffentitor

  17. 1

    yes, but it’s already a crowded space—so the opportunity depends entirely on how you differentiate.

  18. 1

    I made something for this already and realised market is flooded. ShopAsk/ PromptMysite, its just the AI can onlu be as good as the content you have

  19. 1

    The instinct to talk to store owners before writing a line of code is the right call. The questions you laid out are good, but I would add one: ask them how they currently handle it. Not "do you want AI support" but "walk me through the last time a customer asked about a delayed order." That answer usually reveals whether the pain is in the volume of tickets or the context-switching cost of managing them, which changes what you actually need to build.

    The trust question you raised is the real one. Most small Shopify operators are the sole face of their brand, so AI handling a response feels different to them than it would to a mid-sized team. Worth probing whether they want it fully autonomous or as a draft-and-review tool first.

    1. 1

      Walk me through the last time a customer asked about a delayed order" — that's a much better question than anything I had planned. It gets to the real workflow instead of a hypothetical. Adding that to my interview script immediately. Did you find people were more honest when you asked it that way?

  20. 1

    Talking to users before building is the right call, but watch out for a trap: store owners will tell you "yes this is painful" and then not pay for it. The real test isn't whether the problem exists — it is whether they're already cobbling together some workaround (spreadsheets, VA, a Zapier nightmare) and paying for that. If they are, that's your signal.

    One angle worth exploring in your interviews: what happens when the AI gets it wrong? Returns and order disputes have real money attached. The fear isn't "will it work 80% of the time" — it's "what is the cost of the 20% when it doesn't." That's the moat question more than the feature question.

    1. 1

      That's the clearest signal filter I've heard — if they're already paying for a VA or Zapier workaround, the pain is real enough. What would you ask in an interview to surface that quickly?

  21. 1

    The "would you trust AI to handle it" question is actually the crux. From what I've seen in adjacent markets: trust isn't given upfront, it's earned through a deliberately narrow starting scope. Start with order tracking only — pure lookup, zero judgment needed. Let the AI deflect the top 30% of tickets there, prove it works, then expand to returns/FAQs.

    The stores that try to automate everything on day one tend to get a bad review from a customer whose sensitive return got mishandled by a bot — and suddenly the whole experiment is over.

    On cost: $8–12k/month for 2 agents + tools is the number I hear for mid-size stores. If you can reliably own even 40% of that ticket volume at $200–400/mo flat, the math works clearly. The challenge is proving the reliability before they'll give you real ticket volume.

    Pre-code customer research is exactly the right first move here. Good luck.

    1. 1

      Starting narrow to earn trust first makes sense — prove it works on low-risk tickets before touching anything with money attached. How would you suggest proving reliability to a store owner before they give you real ticket volume?

  22. 1
    1. $8k–$12k/month for 2 agents + tools.

    2. Repetitive tickets (order tracking, returns, FAQs) eating most of our time and slowing response during peak.

    3. Yes for simple queries, but not for complex or sensitive issues.

    1. 1

      This is exactly the kind of data I was hoping to find. Are you currently using any tool to handle those repetitive tickets or is it still manual?

      1. 1

        Hey — you mentioned spending $8-12K/month on support in my post. I'm doing customer discovery calls this week. Would you be up for 15 minutes? No pitch, just listening. Here's my calendar link: https://calendly.com/mahirafarhan6/30min

  23. 1

    Different path but might be useful: I did the opposite and shipped a $49 product this weekend without talking to a single customer first. 6 weeks from zero coding knowledge, built it in scar tissue from my own agents breaking, launched today.

    Here's what I'd tell you looking back at my own process — talking to users is the smart move, but there's a failure mode where the interviews never end and the product never ships. You get 10 "yes I'd pay" responses and still don't know anything, because people answer hypotheticals differently than they pay.

    One thing that would have helped me: ship the free version first. I have 3 patterns free on GitHub, 9 more behind $49. The free ones are the validation — if nobody stars the repo or files issues, the paid pack doesn't matter. Could you do that with support tickets? Free tier that handles one category (order tracking only), paid for the rest? Validates whether they trust AI with the easy stuff before you build the hard stuff.

    Also — the commenter on Gorgias/Tidio is right. The wedge matters more than the feature set. What's the one thing those two can't do that Shopify owners actually complain about?

    1. 1

      The free tier as validation idea is interesting — let them trust the AI with easy tickets first before asking them to pay for the hard ones. Removes the risk objection completely. How long did it take before you saw real usage signal on your free version?

  24. 1

    the problem is 100% real but the commenter who mentioned Gorgias/Tidio is right — the crowded space means your wedge has to be razor sharp. I'd skip the "we do everything" pitch and pick one painful flow that existing tools handle badly. returns processing maybe? that's where store owners lose the most time and money, and the existing tools basically just route it to a human anyway.

    also +1 on the advice to do 10 stores manually first. you'll discover edge cases no amount of brainstorming would surface — like how different stores have wildly different return policies that the AI needs to handle contextually, not with generic templates.

    1. 1

      The edge cases point is what scares me most about building this — every store has a different return policy and the AI needs to handle that contextually not generically. Did you hit that problem when you were doing stores manually? How different were the policies store to store?

  25. 1

    Good idea, but you would need to add a strong USP. Building an AI assistant bot isn’t a big challenge for a mid-sized business.

    That said, never back down from your idea. I built a Substack Notes scheduler, even though they launched something similar quietly, I’m still adamant about my idea.

    The scheduler is Pubq.io

    1. 1

      You're right — USP is the real challenge. Everyone says "AI support" but that's not a reason to switch. Still figuring out what the one thing is that makes this unmissable. What would a strong USP look like to you in this space?

  26. 1

    Good instinct starting with conversations first.
    The real question isn’t “would you trust AI”, it’s whether it can solve tickets without creating bigger problems.

    1. 1

      That's the real bar isn't it — not "do you trust AI" but "does it actually solve the ticket without making things worse." What do you think is the highest risk failure mode — wrong answer, slow response, or something else?

  27. 1

    My gratitude goes out to the entire team of Revox Credit repair. Someone left information REVOXCREDITREPAIR at GMAIL dotCOM under a comment section on helping to repair credit. I was interested in knowing more and if the hacking team still does this. I had poor credit, an old bankruptcy and problems with getting approved for an apartment due to 2 broken leases from the past which I explained to Revox Credit Repair when I made contact with them. They cleaned my credit records and they boosted my credit score within just few days of contacting them.

  28. 1

    Definitely a real problem — Shopify store owners spend a ton of time on repetitive support queries. The ROI case is easy to make if you can show hours saved per week. What's your target price point — per-ticket or monthly flat?

    1. 1

      Flat monthly fee. The per-ticket model is exactly what makes Gorgias painful for stores doing high volume.

  29. 1

    Interesting project. Shopify store owners are often skeptical about handing over customer conversations to AI. Have anyone thought about how the brand or identity of the agent affects that initial trust? A lot of founders underestimate how much the naming plays into perceived reliability

    1. 1

      The naming point is interesting — I hadn't thought about how much the agent's identity affects trust. Do you think store owners would prefer it to feel like a bot they control, or something that feels more like a human assistant?

      1. 1

        @mahiraaaa both actually. Enterprise clients don't want something that feels like a clunky bot but they also get suspicious if it tries too hard to sound human and then makes mistakes. What they really want is reliability and control. The sweet spot seems to be an agent that feels professional and trustworthy like a competent assistant they can rely on, not a replacement for a human.
        That's why I think the identity layer (naming or branding) matters more than most founders realize in Shopify or any other AI tools.

        1. 1

          Reliability and control" over human-sounding — that's a clear brief for how to position it. So the branding should feel like a competent assistant they manage, not a replacement they hand off to. Does the name matter more than the interface in creating that feeling?

  30. 1

    Makes sense to talk to users first before building.

    Also agree with the point about strong competition comment below. Even if the problem is real, getting people to trust and try something new can still be a challenge, especially with store owners and so many well-known tools already out there.

    1. 1

      That's exactly the plan — talking to store owners before writing a single line of code. Already getting pushed in directions I hadn't considered just from this thread. How did you approach the trust problem when you were early stage

  31. 1

    You really have a great idea 🫡

  32. 1

    The problem is real but the competition is brutal. Gorgias, Tidio, Zendesk all do this already. The question isnt whether store owners need it, it's why theyd pick you over something they already know. Flat monthly fee is smart though, Shopify owners hate perr ticket pricing. If I were you Id find 10 stores manually and do the support yourself before building anything, that's how you learn what the AI actually needs to get right.

    1. 1

      The manual first approach makes a lot of sense — learn the job before automating it. Did you do something similar when validating your own products? And how would you suggest finding those 10 stores willing to let someone unknown handle their support?

  33. 1

    You’re solving a real problem — but it’s also one of the most crowded directions right now.

    The risk isn’t “does this exist” — it’s “why would anyone pick yours over the 20 others doing the same thing?”

    Most tools in this space sound identical:
    AI support, automation, faster replies, etc.

    So the real question early isn’t just validation — it’s:
    can this actually stand out as something memorable, or will it blend in?

    Seen a lot of solid products die here, not because they didn’t work — but because they felt interchangeable.

    If you get a few store owners interested, pay attention to how they describe it back to you — that’s usually where differentiation (and brand) starts.

    1. 1

      That's exactly the kind of honest pushback I needed. You're right — "AI support automation" is everywhere. I'm still in discovery mode so I haven't locked in positioning yet. What do you think makes something in this space actually memorable vs just another tool?

      1. 1

        Usually not features — constraints.

        The ones that stick in this space feel like:
        → “this is ONLY for X”
        → “it ONLY does Y”
        → “and it does it better than anything else”

        Example:
        “handles returns for Shopify stores” is more memorable than “AI support agent”

        Because now it’s clear:
        – who it’s for
        – when to use it
        – and what to expect

        Broad = comparable
        Narrow = ownable

        Most people start broad and try to stand out later — but it’s usually the opposite that works.

        1. 1

          Broad = comparable, narrow = ownable" — that's the clearest way I've heard it put. So instead of AI support agent, something like "handles returns only, better than anyone else" would be more ownable. Does the niche need to be a ticket type like returns, or could it be a store type like fashion or electronics?

          1. 1

            Both can work — but early on, ticket-type usually wins.

            “handles returns for Shopify” is tied to a clear, recurring pain.
            It’s easy to test, easy to measure, and easy for users to say “this solved X for me.”

            Store-type (fashion, electronics) is broader — you still end up needing to define what inside that you’re solving.

            You can layer store-type later once you see patterns.

            Start with:
            → one painful workflow
            → that happens often
            → where speed/accuracy matters

            Own that → then expand.

            Otherwise you risk sounding niche, but still being vague.

            1. 1

              That makes sense — start with one workflow, own it completely, then expand. My gut says order tracking is the most repetitive and high volume ticket for most Shopify stores. Does that feel like a strong starting point or is there a more painful one you'd go after first?

              1. 1

                I wouldn’t start with order tracking at all.

                It’s high volume, but it’s already “good enough” everywhere — so there’s no real reason to switch.

                Early products don’t win on frequency, they win on pain + consequence.

                If nothing breaks when your product is bad, nobody cares enough to change tools.

                I’d go straight to something like returns/refunds/chargebacks — where:
                → money is involved
                → customers get frustrated
                → and mistakes actually cost the business

                That’s where “I need this now” comes from.

                Otherwise you risk building something useful… that nobody feels urgency to adopt.

                1. 1

                  Returns/refunds makes sense — the consequences of getting it wrong are real. Order tracking feels safe but you're right, there's no urgency. So if I focused purely on returns and chargebacks, what would "doing it better than anyone else" actually look like in practice?

                  1. 1

                    “Better” here usually isn’t more automation — it’s removing the parts that cost money or trust.

                    For returns/chargebacks, that tends to look like:

                    → fewer bad approvals (protect margin)
                    → faster resolution (reduce frustration / support load)
                    → and clear, consistent decisions (no back-and-forth)

                    Most tools handle the workflow.

                    Very few actually optimize the outcome:
                    – deciding when to approve vs push back
                    – catching edge cases (policy abuse, repeat claims, etc.)
                    – and making it feel fair to the customer while still protecting the store

                    If you can do that reliably, it stops being “support automation” and starts being:
                    → “this saves me money every week”

                    That’s usually the threshold where people switch.

                    So instead of asking “can this handle returns”
                    I’d frame it as:
                    “does this reduce loss + support overhead in a way I can see quickly?”

                    That’s the version that becomes hard to ignore.

                    1. 1

                      Saves me money every week" vs "does my support" — that's a completely different product conversation. So the pitch isn't features, it's a measurable outcome. How would you suggest I actually prove that outcome to a skeptical store owner before they commit to paying?

Trending on Indie Hackers
I built a tool that shows what a contract could cost you before signing User Avatar 120 comments The coordination tax: six years watching a one-day feature take four months User Avatar 78 comments My users are making my product better without knowing it. Here's how I designed that. User Avatar 65 comments A simple LinkedIn prospecting trick that improved our lead quality User Avatar 58 comments I changed AIagent2 from dashboard-first to chat-first. Does this feel clearer? User Avatar 39 comments Why I built a SaaS for online front-end projects that need more than a playground User Avatar 18 comments