# What is the “Product” being evaluated?

To understand if users are actually finding value in your product, you must "instrument" your application, setting it up to automatically log specific user actions. The resulting log data allow you to track users as they move through the User Funnel: from their first interaction (Activation), to regular usage (Engagement) to long-term commitment or habit formation (Retention).

In the tech sector, companies might track "clicks" and "purchases" as users move through a website. By analyzing logs, you can then identify which content or products are likely to bring users back to the website over time, or how different web experiences affect browsing time. In the development sector, we need to track actions that signal user intent, and estimate the value returned to users in response. For example:

* For an AI Agronomist, instead of tracking page views you might track whether a farmer uploads a photo of a diseased crop, listens to audio advice to completion, or shares content with another person.
* For an AI Tutor, you might track if a student completes a quiz, how many follow-up questions they ask in a single session, or if they return to the app the night before an exam.

By analyzing these "digital traces," we can identify exactly where users lose interest. Does the farmer drop off because the photo upload takes too long? Does the student quit because the AI’s first response was too complex?

**The good news:** Fortunately, Level 2 evaluation methods are well-scoped; you don't need to reinvent the wheel. The technology sector has spent decades standardizing digital product metrics. Most off-the-shelf analytics tools for web/mobile applications (like Amplitude or Google Analytics) come ready-made to measure the standard metrics you need, such as Daily Active Users (DAU), Time to Activate, and Retention Rates.

The following table defines common Level 2 product metrics for each stage:

<table data-full-width="true"><thead><tr><th width="140.33984375">Stage</th><th>Metric</th><th>Examples</th><th>Notes</th></tr></thead><tbody><tr><td><strong>Acquisition</strong></td><td><strong>New Users (#):</strong><br>Total count of users entering the top of the funnel</td><td># farmers consenting to receive WhatsApp messages; # students downloading an app; # health workers attending a training session.</td><td>There are costs associated with recruitment, so you may wish to target your ideal users effectively and efficiently. Track the source of each new user (e.g., "Referral" vs. "Field Visit") to identify which channels yield the most relevant users, then scale the channel with the highest yield.</td></tr><tr><td><strong>Acquisition</strong></td><td><strong>Cost Per User (CAC)</strong>:<br>Cost of running a recruitment activity, divided by new users acquired. Also called User Recruitment Cost.</td><td>Cost of printing flyers / # of QR code scans.<br>Cost of field agent stipend / # of farmer sign-ups.</td><td>High CAC may be unsustainable for low-margin services (i.e. general advice) but acceptable for high-impact interventions like urgent telemedicine consults.</td></tr><tr><td><strong>Activation</strong></td><td><strong>Activation rate:</strong><br>% of users who complete the "First Value" action after signing up or being recruited.</td><td>% of mothers who complete their health profile<br>% of teachers who create their first lesson plan</td><td>Activation measures if users actually start using the tool or service, not just if they installed it. It is a user's early experience with the product, so important to “get right” to prevent drop-off. Onboarding can also be a critical point for collecting user demographic data used for personalization and subgroup analysis.</td></tr><tr><td><strong>Activation</strong></td><td><strong>Time to activate</strong>:<br>Average time elapsed between sign-up and the first core action.</td><td>Minutes from first WhatsApp message to asking the first medical question<br>Days from initial training to logging the first case data</td><td>New user interest tends to taper off exponentially after sign-up, so it’s important to encourage users to complete onboarding within the first few hours/days.<br>Long activation times usually signal a confusing interface or a lack of trust or value.</td></tr><tr><td><strong>Engagement</strong></td><td><strong>Monthly, weekly, and/or daily active users</strong> (MAU, WAU, DAU):<br>Number of users using the app/feature in a time window.</td><td># Community Health Workers using a reporting tool daily<br># Students using a study bot weekly before exams</td><td>Do not optimize for addiction. In welfare-focused apps, "Time to Success" (i.e., getting to an urgent answer quickly) is better than "Time on Device."</td></tr><tr><td><strong>Engagement</strong></td><td><strong>Interaction depth</strong>:<br>Volume of interaction per session (e.g., turns per conversation).</td><td>Average # of follow-up questions a student asks (signals curiosity) per session<br>Rate at which a user accepts a suggestion (signals trust)</td><td>For a chatbot, the # of conversation turns can be good (deep inquiry) or bad (confusion). Pair engagement metrics with qualitative inquiry (Level 3).<br>Remember that frequent interaction (e.g. page loads, button clicks, form submits, session length) may not translate to meaningful interaction.</td></tr><tr><td><strong>Retention</strong></td><td><strong>Stickiness (DAU/MAU):</strong><br>Ratio of daily users to monthly users.</td><td>A ratio of 0.25 means the average user uses the tool 7-8 days per month.<br>A ratio close to 1 indicates sustained value or habit formation.</td><td>High stickiness indicates the tool is part of a daily workflow. A low ratio suggests the tool is only useful for sporadic problems (e.g., seasonal crop disease) rather than daily habits. Is habit formation critical to your welfare goals?</td></tr></tbody></table>

In addition to these, you may want to specify negative metrics (or safety guardrails) that track problematic user behavior. For example, if a mother uses a health chatbot for an emergency, a long session may indicate confusion and inefficiency. We must aggressively distinguish between intense or extensive engagement, and effective engagement.

***

<details>

<summary>💬 Want to suggest edits or provide feedback?</summary>

{% embed url="<https://tally.so/r/A788l0?originPage=level-2-product-evaluation%2Foverview%2Fwhat-is-the-product-being-evaluated>" %}

</details>
