# Overview

An AI system that produces perfect responses is worthless if no one uses it. Level 2 Evaluation moves beyond technical accuracy to measure the "digital traces" users leave behind. By tracking how users move from their first interaction to long-term habit formation, we can ensure the product actually delivers value in the real world.

***

#### Key Motivation

Technical performance (Level 1) does not guarantee user adoption. Level 2 evaluation is critical because:

* Value Validation: If users stop interacting, they likely see no value, meaning the intervention cannot achieve its intended life outcomes.
* Continuous Improvement: It transforms product development from opinion-driven to data-driven through iterative cycles and A/B testing.
* Safety & Risk Management: Monitoring user signals allows for controlled rollouts of experimental features, preventing negative reactions from reaching your entire user base at once.

<a href="why-is-this-level-of-evaluation-important" class="button primary">Read more -></a>

***

#### Core Concept: The User Funnel

To evaluate the product, we "instrument" the application to track users as they progress through four distinct stages. We prioritize "Time to Success" (solving the user's problem) over "Time on Device" to ensure we are optimizing for welfare rather than just addiction.

| Stage       | Goal                                | Key Metric Example                        |
| ----------- | ----------------------------------- | ----------------------------------------- |
| Acquisition | Bring users into the ecosystem.     | New User Count, Cost Per User (CAC)       |
| Activation  | Ensure users find "First Value."    | Activation Rate, Time to Activate         |
| Engagement  | Measure depth and frequency of use. | Active Users (DAU/WAU), Interaction Depth |
| Retention   | Build long-term habits/commitment.  | Stickiness (DAU/MAU), Retention Rate      |

<a href="what-is-the-product-being-evaluated" class="button primary">Read more -></a>

***

#### How to Evaluate

Level 2 evaluation is performed by integrating 3rd party analytics tools (e.g., Amplitude, Mixpanel) to capture real-time data.

1. Define & Instrument: Map your user journey and identify specific "events" (e.g., "audio advice played") that signal progress.
2. Analyze Trends: Use dashboards to identify friction points where users consistently drop off.
3. Experiment: Run A/B Tests to compare different versions of a feature. By randomly assigning users to "Version A" or "Version B," you can statistically prove which design better supports user goals.
4. Diagnose: If metrics are low, conduct a Process Evaluation (interviews or surveys) to understand the "why" behind the data—such as connectivity constraints or literacy barriers.

<a href="../how-to-evaluate/how-is-level-2-evaluation-performed" class="button primary">Read more -></a>

***

<details>

<summary>💬 Want to suggest edits or provide feedback?</summary>

{% embed url="<https://tally.so/r/A788l0?originPage=level-2-product-evaluation%2Foverview>" %}

</details>


---

# Agent Instructions: Querying This Documentation

If you need additional information that is not directly available in this page, you can query the documentation dynamically by asking a question.

Perform an HTTP GET request on the current page URL with the `ask` query parameter:

```
GET https://eval.playbook.org.ai/product-analytics/level-2-product-evaluation/overview.md?ask=<question>
```

The question should be specific, self-contained, and written in natural language.
The response will contain a direct answer to the question and relevant excerpts and sources from the documentation.

Use this mechanism when the answer is not explicitly present in the current page, you need clarification or additional context, or you want to retrieve related documentation sections.
