4
21 Comments

Why I Pivoted from an AI Counseling Service to an AI Girlfriend Chat

Hi Indie Hackers,

I previously built an AI counseling service.
While some users found value in it, I struggled with long-term user retention.

That experience taught me an important lesson:
having a useful product isn’t enough if users don’t feel a strong reason to come back.

Based on what I learned, I decided to pivot and launch a new service.
This time, I’m focusing more on repeat usage, simplicity, and clear value from the first interaction.

I’d love to hear your thoughts or feedback.
Thanks for reading.

https://eoerway-ai-therapy-v3-0-616432264786.us-west1.run.app

posted to Icon for group Feedback
Feedback
on December 29, 2025
  1. 2

    Isn't optimizing for GEO actually not very similar to optimizing for SEO?

    1. 2

      If you mean GEO/LLM optimization: it’s similar to SEO in that both are about discoverability, but GEO is about being selected/cited by an LLM. So I focus more on clear answerable pages, structured data, consistent entities, and content that’s easy to quote/summarize, not just ranking signals.

      1. 1

        Having bullet points with clear headings also seems to help

  2. 2

    This is a really honest take on retention - and you've hit on something important.

    Counseling solves a problem users want to solve and move on from. Companionship fulfills an ongoing emotional need. The incentive structures are completely different.

    Curious about a few things as you build this out:

    1. How are you thinking about the "first interaction" value you mentioned? What's the moment users realize this is worth coming back to?

    2. Are you seeing any patterns in what keeps users engaged day-over-day vs week-over-week?

    The pivot makes strategic sense - just be mindful of the moderation/safety considerations that come with companion AI. That space can get complicated fast.

    What's retention looking like so far compared to the counseling product?

    1. 1

      demogod_ai — sending this again, thanks for the quick questions. Really appreciate you calling these out. On first interaction value: We’re less focused on immediate problem-solving and more on creating a sense of being understood. The key moment is when users notice that the AI remembers context, reflects emotional nuance accurately, and doesn’t rush them toward a solution. When people say, “this feels different from other chats,” that’s usually the point where they decide it’s worth coming back. Day-over-day vs week-over-week retention: Day-over-day engagement is often driven by shifts in emotional state — users check in when their mood changes or something small happens. Week-over-week retention is more about habit and identity. Users who begin to see it not as a “tool they use” but as “someone they talk to” tend to stick around much longer. That shift usually happens after a few light but consistent interactions.On safety and moderation: Completely agree here. We’re being very deliberate about boundaries — avoiding exclusivity framing or language that encourages emotional dependency, and clearly positioning the product as support rather than a replacement for real human relationships. This isn’t an afterthought for us; it’s a core product design concern. Current retention: It’s still early, but short-term retention after the first session is higher than what we saw with the counseling product. Counseling users often churn once they feel their issue is resolved, whereas companionship users tend to return more frequently, even if sessions are shorter. We’re still watching the metrics, but the qualitative signals so far are quite encouraging.

      1. 1

        Really appreciate the detailed breakdown here.

        The "being understood" moment you described - where users notice the AI remembers context and reflects emotional nuance - that's a much higher bar than most chat products aim for. Most stop at "responds coherently." You're aiming at "feels like it was listening."

        The day-over-day vs week-over-week distinction is interesting. It sounds like you're essentially building two retention loops: one for emotional check-ins (volatile, mood-driven) and one for identity/habit formation (stable, relationship-driven). The users who make the shift from "tool I use" to "someone I talk to" are probably your highest-LTV cohort.

        Smart that safety is a core design concern rather than a moderation afterthought. The "support rather than replacement" framing matters - both for user wellbeing and for positioning if the space gets more regulatory scrutiny.

        What signals are you watching to catch early signs of unhealthy dependency patterns? That seems like the hardest moderation challenge in this space.

        1. 1

          Honestly, at this stage we don’t think the AI has the kind of “pull” that creates real dependency yet. Most interactions are still lightweight and situational.

          That said, we’re treating this as a latent risk, not something to ignore. If we ever start to see signals that users are orienting their emotional stability around the product, that’s a clear line for us — and we’d intervene immediately, even at the cost of short-term retention.

          Our assumption is that unhealthy dependency doesn’t happen suddenly; it emerges gradually. So the goal right now is less about heavy-handed controls and more about designing boundaries early, before that kind of dynamic has a chance to form.

          1. 1

            The "latent risk" framing is exactly right. What I've seen trip up similar products: the signals that matter aren't always the obvious ones. High session frequency might mean healthy engagement OR early dependency - the differentiator is usually what happens when users can't access it. Do they adapt normally, or show disproportionate distress?

            Your point about designing boundaries early is smart. Retrofitting them after patterns form is 10x harder - users experience it as something being taken away rather than guardrails being established.

            One pattern worth considering: proactive "you've been great company, but go be with humans now" nudges before users reach natural stopping points. It reframes the AI as supporting real-world connection rather than replacing it. Curious if you've experimented with exit prompts that redirect toward offline activity?

            1. 1

              Great point. I completely agree that users’ reactions in moments when the AI isn’t accessible are one of the most important signals for making this distinction.

              We haven’t run any intentional exit prompt experiments yet. However, our current design direction isn’t to position the AI as something users are meant to stay with continuously, but as a companion that offers support when needed and then helps users move back into real life.

              As you mentioned, if boundaries and expectations aren’t clearly set early on, any limitations introduced later can easily feel like something is being taken away.

              I think proactive redirection—like “you’ve been great company, now go be with humans”—is absolutely worth testing. Especially if offered before natural stopping points, it seems like a promising way to reduce dependency while preserving a positive user experience.

              1. 1

                The "companion that helps users move back into real life" framing is exactly the right positioning. It's a subtle but important distinction from products that optimize for time-on-app as the primary metric.

                One experiment worth considering: varying the exit prompt timing based on session context. A user checking in after a tough day might benefit from a gentler, more natural transition ("sounds like you've processed a lot - maybe text a friend about it?"). Someone in a lighter mood might respond better to direct encouragement ("you seem good - go grab coffee with someone").

                The accessibility-reaction signal you mentioned could also inform this. Users who handle unavailability well might need fewer proactive nudges, while those showing early dependency signals might benefit from more frequent, softer redirections.

                Curious how you're thinking about the feedback loop here - are you planning to measure whether exit prompts actually lead to reported offline activity, or is the goal more about establishing healthy usage norms regardless of measurement?

                1. 1

                  Totally aligned — I do think exit prompts can lead to real offline activity, and I want to measure that.
                  My current view is: the goal isn’t just “healthy norms” in the abstract, but a measurable shift from in-app processing → real-world action.
                  How I’m thinking about measuring it (lightweight, privacy-first):
                  Immediate self-report after exit: “Did you do one small offline action in the next 30–120 minutes?” (Yes/No + optional quick tag)
                  Follow-up check-in next session: “Last time, did you do the offline step we suggested?” (again, optional)
                  Proxy metrics: reduced session length after the prompt, fewer rapid re-opens, and increased “action-tag” confirmations over time
                  A/B test: context-sensitive exit prompt timing + wording vs control, to see uplift in reported offline actions
                  I also like your point about the accessibility-reaction signal: people showing early dependency signals may benefit from more frequent but gentler redirections, while others need fewer prompts.
                  Curious: have you seen good patterns for balancing measurement accuracy with keeping the flow non-intrusive?

                  1. 1

                    The measurement framework you've outlined is solid - especially the combination of immediate self-report with follow-up check-ins. That gives you both intent signal and actual behavior validation without being invasive.

                    For balancing accuracy with non-intrusiveness, a few patterns I've seen work:

                    Sampling over surveying everyone: Don't ask every user every time. Random sampling (say, 10-20% of sessions) gives you statistically valid data without creating survey fatigue.

                    Contextual embedding: Make the check-in feel like natural conversation continuation rather than data collection. "How'd that coffee with your friend go?" feels different from "Did you complete your offline action?"

                    Opt-in depth: Quick binary responses for everyone, with optional expansion for users who want to share more. Respects time while giving motivated users a voice.

                    Delayed validation: The next-session follow-up is smart, but watch for recency bias. Users may remember the prompt better than the actual action. Consider occasional longer-delay checks (3-5 days) for more durable behavior change signals.

                    The proxy metrics you mentioned (shorter sessions, fewer rapid re-opens) are probably your most honest signals - they're behavioral rather than self-reported. If those correlate with your A/B prompt variations, you've got strong evidence without asking users anything.

                    1. 1

                      Thanks for the thoughtful feedback — this is incredibly helpful. We especially like the sampling + contextual check-in idea to reduce friction while keeping signals honest. We’ll experiment with delayed follow-ups and lean more on behavioral proxies as you suggested

  3. 1

    Interesting perspective, especially around what teams deprioritize. I’ve noticed that roadmap discussions rarely include assumptions testing or downside scenarios. Planning what not to build has saved more time for us than adding new features ever did.

    1. 1

      Thank you for sharing your experience. Your perspective helped us think more deeply not only about what to build, but also about what not to build. We’ll continue to reflect this as we work toward making the service better.

  4. 1

    Really smart insight on shifting from solving a one-off need (counseling) to repeat engagement.

    One thing I’m curious about is how you’re thinking about the hook that brings people back after the first session. Retention often hinges not just on emotional resonance but on a predictable pattern users can return to.

    Are you tracking any early signals like session frequency relative to mood triggers or time of day that inform what type of interaction feels most “sticky” to your users right now?

    1. 1

      Appreciate this — you’re spot on about predictability over pure emotional resonance.

      Right now we’re intentionally keeping the hook lightweight:
      not “come back to vent again,” but “check in when a familiar emotional state shows up.”

      We’re early, but we’re already seeing repeat usage cluster around specific moments
      (late night, post-work, and low-energy states rather than acute crises).

      We’re starting to track frequency relative to time-of-day and self-reported mood signals,
      and using that to shape repeatable interaction patterns instead of open-ended counseling.

      Still very early, but the shift from one-off relief → familiar emotional companion
      is already changing how people return.

      1. 1

        That’s a really thoughtful shift — anchoring return behavior to recognizable states rather than intensity makes a lot of sense. What you’re describing feels less like “emotional support on demand” and more like situational companionship, which is a very different retention model.

        The clustering you’re seeing around late-night / post-work / low-energy moments is especially interesting — those are predictable, repeatable contexts, which gives you something to design around instead of reacting to.

        As you keep exploring this, I’m curious: are you treating those moments as distinct interaction modes (almost like “states” the product enters), or are you letting a single experience adapt dynamically? That choice seems like it could shape how scalable and habit-forming this becomes.

        1. 1

          Thanks — really appreciate this perspective.

          Right now we’re leaning toward distinct interaction modes, but keeping the transition between them very lightweight.

          Instead of one adaptive “do-everything” experience, we’re experimenting with a small set of familiar states (late-night wind-down, post-work decompression, low-energy check-in) that users can recognize and return to without cognitive load.

          Early signal we’re watching isn’t depth or emotional intensity, but whether users return to the same state at roughly the same time/context. If that repeats, we consider it a win — even if the session itself is short.

          Still early, but framing the product as a familiar presence in predictable moments has already changed how we think about retention and habit formation.

          Would love to share learnings as this evolves.

  5. 1

    Thanks for sharing your pivot experience. Retention is tricky for counseling because once the problem is solved, users naturally move on. Companion AI emphasises ongoing emotional connection.

    I like your focus on creating value from the first interaction and building habit loops. How do you plan to keep the conversations feeling authentic and fresh for repeat users as your user base scales?

    Are there specific metrics or qualitative signals you're watching to ensure you're delivering on ongoing companionship rather than just initial novelty?

    1. 1

      Thanks — great questions. My plan is to keep “authentic + fresh” at scale by focusing on memory + variety + user-led intent rather than scripted roleplay.
      Authenticity: lightweight long-term memory (user preferences, boundaries, recurring themes) + session summaries so it can reference context naturally without pretending it “knows” things.
      Freshness: conversation “modules” (check-in, playful banter, reflection, micro-challenges, story prompts) + rotation rules + repetition detection to avoid the same patterns.
      At scale: automated quality checks (repetition score, generic-response rate), and feedback loops (thumbs up/down + short reason) to continuously tune prompts and flows.
      For metrics beyond novelty, I’m watching:

      D1/D7 retention, sessions per user per week

      Return without notifications (organic pull)

      % of sessions that reference prior context (users bringing it back)

      Qualitative: users explicitly saying it feels “understood,” and whether they form a habit (e.g., nightly check-ins).

      Still early, but I’m optimizing for repeat value from the first session and “small reasons to return,” not just initial wow.

Trending on Indie Hackers
I'm a lawyer who launched an AI contract tool on Product Hunt today — here's what building it as a non-technical founder actually felt like User Avatar 142 comments “This contract looked normal - but could cost millions” User Avatar 54 comments 👉 The most expensive contract mistakes don’t feel risky User Avatar 41 comments The indie maker's dilemma: 2 months in, 700 downloads, and I'm stuck User Avatar 39 comments A simple way to keep AI automations from making bad decisions User Avatar 35 comments I spent weeks building a food decision tool instead of something useful User Avatar 28 comments