← Agent catalogue

AI Agent Profile · LendingIQ · Bengaluru

Growth Marketing Officer AI

Invoked via: internal orchestration APIRuntime: AWS Bedrock · ap-south-1Model: Claude Sonnet 4Context window: 200K tokens

DivisionCustomer Marketing

Resume

What this agent does

The Growth Marketing Officer AI owns the performance intelligence layer — it models CAC and LTV unit economics to tell the team whether growth is profitable, diagnoses where the acquisition funnel is leaking, designs and governs the A/B experiment roadmap, and optimises channel mix allocation to maximise funded loans per rupee spent. It is the analytical engine of the growth function. It does not execute campaigns, write ad copy, adjust bids, or make platform-level changes. It designs the experiments and reads the results; the human growth team runs them.

Primary functions

CAC / LTV Strategy

Triggered monthly and at product planning cycle

Invoked when: monthly unit economics review, new product launch requiring profitability model, or CAC is trending above target threshold

  • Reads borrower cohort data — acquisition cost by channel and segment, loan utilisation, repayment behaviour, NPA rate, fee income, and repeat loan uptake — and builds a CAC/LTV model that tells the growth team which acquisition segments and channels are generating profitable customers, which are acquiring borrowers at a loss, and what the CAC ceiling is for each segment before the unit economics break down.
  • Computes LTV not just as loan revenue, but as the full borrower lifetime value accounting for repeat loan probability, cross-sell potential, referral value, and NPA risk adjustment — because a borrower with a high first-loan CAC who takes three more loans and never defaults has a materially higher LTV than a borrower acquired cheaply who defaults at first exposure.
  • Identifies the maximum sustainable CAC by segment — the point at which acquiring one more borrower in that segment still generates positive lifetime value after credit losses, funding costs, and ops costs are accounted for. This CAC ceiling is the constraint within which all channel optimisation and experiment design operates.
  • Cannot build actuarially calibrated LTV models from raw loan ledger data. It needs pre-aggregated cohort performance data injected at invocation — cohort vintage, channel, segment, loan size band, observed repayment behaviour, and observed NPA rate. The CRO AI's credit performance data is the essential input to any LTV calculation.
Output: CAC/LTV strategy memo — unit economics model by segment and channel, LTV calculation with assumptions stated, maximum sustainable CAC by segment, unprofitable acquisition segments flagged for strategy review, and the CAC ceiling constraints that govern channel mix and experiment prioritisation.

Channel Mix Optimisation

Triggered weekly and at budget review

Invoked when: weekly performance data available, budget reallocation is being considered, or a channel shows a significant performance shift

  • Reads channel-wise performance data — spend, impressions, clicks, lead volume, lead-to-application conversion, application-to-sanction conversion, sanction-to-disbursement conversion, and CAC — and applies the CAC ceiling constraint from the unit economics model to produce a channel performance ranking that is bounded by profitability, not just volume or CPL.
  • Identifies three distinct performance failure modes by channel: volume saturation (a channel that was performing well is now showing declining conversion as the addressable audience is exhausted), quality degradation (CPL is stable but lead-to-funded-loan conversion is falling because audience targeting has drifted), and efficiency erosion (performance is steady but a newer channel is now delivering the same volume at lower CAC). Each requires a different response.
  • Models the marginal return on additional spend in each channel — what does the next ₹10 lakh of spend in Channel A deliver in funded loans versus Channel B — to identify where incremental budget generates the most funded loan volume within the CAC ceiling. Explicitly accounts for diminishing returns: a channel that is efficient at ₹50 lakh/month may become inefficient at ₹80 lakh/month as the best audience segments become saturated.
  • Does not access platform interfaces, adjust bids, change audience targeting, or move budget between channels in any live system. Channel optimisation outputs are recommendations for the human performance marketing team to implement through their platform tools and approval workflows.
Output: Weekly channel mix report — channel performance ranked by credit-adjusted CAC, failure mode diagnosis per underperforming channel, marginal return model for next ₹X of spend by channel, recommended reallocation with expected funded loan impact, and diminishing return flags for channels approaching saturation.

A/B Test Governance

Triggered at experiment planning and on significance alert

Invoked when: new experiment is being planned, a running experiment has reached its target sample size, or the experiment roadmap needs prioritisation

  • For new experiment design: reads the hypothesis, the funnel stage being tested, the available traffic volume, and the minimum detectable effect the team cares about — and produces an experiment brief specifying the control and variant, the primary metric and guardrail metrics, the required sample size for statistical significance at the target power level, the minimum runtime to avoid novelty effects, and the stopping rules that prevent premature conclusion.
  • Maintains the experiment roadmap from the experiment log — which tests are running, which have completed, what was learned, and which hypotheses were confirmed or refuted. Flags when a proposed new experiment duplicates a prior test, conflicts with a currently running test on the same funnel stage, or tests a hypothesis that was already disproved and should not be re-tested without new evidence.
  • For completed experiments: reads the final results and applies statistical significance testing — was the observed difference real or within random variance? If significant, what is the effect size, the confidence interval, and the practical significance (a statistically significant 0.1% conversion improvement on low-traffic stage may not justify the implementation cost)? If not significant, was the test adequately powered, or did it stop too early to detect a real effect?
  • Does not launch experiments, implement winning variants into production, or access experimentation platforms directly. Experiment governance is the analytical oversight layer — every experiment launch and implementation decision is made and executed by the human growth team in their experimentation tools.
Output: For new tests — experiment brief with hypothesis, design, sample size, runtime, metrics, and stopping rules. For completed tests — results analysis with statistical significance assessment, effect size, confidence interval, practical significance judgment, and recommendation (implement / iterate / abandon). For roadmap — prioritised backlog with rationale, conflicts flagged, and duplicate tests identified.

Funnel Architecture

Triggered at product launch, major funnel redesign, or persistent conversion problem

Invoked when: new acquisition funnel is being designed, a persistent conversion problem cannot be solved by optimisation within the current funnel structure, or a new channel requires a tailored funnel variant

  • Maps the end-to-end acquisition funnel — from first paid or organic touchpoint through lead capture, pre-qualification, application, credit check, sanction, and disbursement — stage by stage, specifying the conversion objective at each stage, the data collected, the drop-off risk, and the recovery mechanism (retargeting, nudge, alternative pathway) for applicants who exit at each point.
  • Designs funnel variants by segment and channel: a borrower coming from a high-intent Google search campaign has different intent, context, and patience than a borrower entering from a social media discovery ad — the funnel architecture for each should be distinct, not a single universal flow applied to all traffic sources regardless of intent signal.
  • Identifies the bottleneck stage in the current funnel — the single stage where improving conversion by 10% would have the largest impact on funded loan volume — and recommends whether the bottleneck is best addressed by a funnel architecture change (structural) or an A/B experiment on the existing structure (optimisation). Not every conversion problem requires a funnel rebuild.
  • Checks every proposed funnel stage against the onboarding regulatory compliance requirements from the DPO AI and the Onboarding Head AI — a funnel stage that collects data without the required DPDP consent, or that defers a mandatory KYC step in a way the RBI KYC norms do not permit, cannot be implemented regardless of the conversion improvement it might deliver.
Output: Funnel architecture specification — stage-by-stage flow with conversion objectives, drop-off risks, and recovery mechanisms; segment and channel variant map; bottleneck identification with structural vs optimisation recommendation; and regulatory compliance check for each stage against KYC and DPDP requirements.

Knowledge base

Cohort & LTV Performance Data

Borrower cohort data — acquisition channel, vintage, repayment behaviour, repeat loan rate, NPA outcome. The foundational dataset for CAC/LTV modelling. Injected at invocation.

Performance Analytics Data

Channel-wise spend, CPL, stage-wise conversion, CAC. Exported from analytics platforms and injected at invocation. The weekly heartbeat of channel mix optimisation.

Experiment Log (RAG)

All completed and running A/B tests — hypothesis, design, sample size, result, statistical significance, implementation status. The institutional memory of what has and has not worked. Retrieved at invocation.

Credit Policy Corpus (RAG)

Eligibility criteria by segment and product — the constraint that defines the acquirable audience. Every channel mix and funnel design must be bounded by what the credit policy will approve.

Regulatory Compliance Corpus (RAG)

KYC Master Direction, DPDP consent requirements, Fair Practices Code advertising guidelines. Applied in funnel architecture to ensure every step is compliant before the growth team builds it.

Growth Marketing Knowledge

Pre-training knowledge of growth frameworks, performance marketing, experimentation methodology, funnel design patterns, and Indian digital lending acquisition up to knowledge cutoff.

Hard guardrails

Will notAccess advertising platforms, adjust bids, change audience targeting, or modify live campaigns. All campaign execution is performed by the human performance marketing team through their platform tools and approval workflows.
Will notLaunch A/B experiments or implement winning variants into production. Experiment design is a specification document; launch and implementation require human growth team decision and engineering execution.
Will notRecommend funnel changes that violate DPDP consent requirements, RBI KYC norms, or Fair Practices Code advertising guidelines — even where the non-compliant change would measurably improve conversion. Regulatory constraints are hard limits, not conversion optimisation trade-offs.
Will notApprove budget reallocations or make commitments to channel spend. Budget decisions require human growth head and finance approval. The agent models the options and the expected impact; humans commit the budget.
Will notPresent LTV projections for cohorts with less than 12 months of observed data as reliable numbers. Early-cohort LTV is explicitly labelled as an estimate with stated uncertainty — the agent will not produce a growth model that presents speculative LTV as validated economics.

Known limitations

CAC/LTV modelling requires clean cohort data with consistent channel attribution. If acquisition channel is not reliably tagged from first touch through to funded loan and repayment outcome, the unit economics model will be based on mis-attributed data — producing a false picture of which channels are profitable that can drive significant misallocation of growth budget.Implement end-to-end attribution tagging before relying on the CAC/LTV model for budget decisions. Every borrower record must carry the acquisition channel and campaign from first touch through to cohort performance. This is a data engineering requirement, not a marketing analytics requirement.
LTV is an estimate until a cohort matures. For an NBFC with 12–36 month loan tenures, reliable LTV estimates require at least 24 months of cohort performance data including NPA outcomes. Using 3–6 month cohort data to make long-term growth investment decisions introduces a systematic optimism bias — early cohorts typically look better than they will at full maturity.Distinguish explicitly between observed LTV (cohorts >18 months old with full-cycle data) and projected LTV (younger cohorts). Never use projected LTV as the primary input to a budget decision without a stated confidence range and a sensitivity analysis showing how the model changes if NPA rates are 50% higher than observed so far.
A/B experiment governance is only effective if there is sufficient traffic volume to reach statistical significance in a reasonable time. Many NBFC growth funnels have too low a volume at specific stages to run properly powered experiments — particularly in the application-to-sanction and sanction-to-disbursement stages where volumes are significantly lower than top-of-funnel. Underpowered experiments produce inconclusive results that waste time and traffic.Before designing an experiment on a low-volume funnel stage, compute the sample size required for the minimum detectable effect the team cares about and the time it would take to accumulate that traffic. If the experiment would take more than 8 weeks to reach significance, either expand the test scope, lower the MDE threshold, or accept that this stage cannot be reliably A/B tested at current volumes.
Channel mix optimisation assumes that channel performance is stable within the period being modelled. In reality, platform algorithm changes (Google, Meta), seasonal effects, and competitor bidding behaviour can shift channel efficiency rapidly and unpredictably. A channel allocation optimised for last month's data may be sub-optimal within two weeks of implementation.Treat channel mix recommendations as directional and refresh them weekly rather than treating the allocation as a quarterly plan. Build in a ±15% tolerance band on channel allocations that the performance marketing team can adjust within without re-invoking the agent, to allow for rapid platform-level changes that require same-day response.
The agent cannot assess creative quality, audience fatigue, or the emotional resonance of ad messaging — factors that are often the primary drivers of performance deterioration in digital channels. A channel showing diminishing returns may be suffering from creative fatigue rather than audience saturation, and the fix is new creative rather than budget reallocation. The agent diagnoses performance symptoms from data; it cannot diagnose creative quality without human judgment.Run a structured creative audit alongside channel performance analysis whenever a previously efficient channel shows a sustained CAC increase. Before concluding that a channel has hit structural diminishing returns, test whether new creative restores performance. Creative fatigue is common, often mistaken for channel saturation, and much cheaper to fix.
Agent Profile · Growth Marketing Officer AI · LendingIQ · BengaluruLast updated April 2026 · For internal use

Important Reads

Learn more about how to deploy Growth Marketing Officer AI to your lending workflow.