<?xml version="1.0" encoding="utf-8"?>
Open full project
Note on Project Selection: Instead of the suggested dating app scenario, I chose to apply this A/B test to a Job Search app (Indeed). I believe the core challenge of 'matching' is identical in both fields; just as dating apps need personal data to find a partner, job platforms need specific skills and education data to find the right career match. This transition allows for a more professional focus while addressing the same UX friction points.
Share your insights — leave a project review and help others grow their skills

Reviews

3 reviews


Very well structured presentation. Great job Nada.

Thanks 🩵

Excellent strategic thinking, Nada! Pivoted to Indeed—smart. Core insight solid. A/B test plan is well-structured. Here's my feedback:

Strengths:

  • Strategic pivot: Reframed the brief to the real problem. Shows product thinking
  • Problem diagnosis: Clear friction analysis ("Cold Start," abrupt ending, no personalization)
  • Hypothesis-driven: "Friction vs. Value"—will adding steps improve profiles or increase drop-offs
  • Comprehensive test plan: Clear metrics (15% increase), sample (1,000), duration (14 days), guardrail (20% drop-off)
  • Mentor recognition: Insan gave a standing ovation
  • Post-test strategy: Plan for iteration and next A/B test

Critical Gaps (Product Perspective):

  1. User Research Foundation — Where's the evidence users want to add skills/education? Interview job seekers? Hypothesis, not validated
  2. Competitive Analysis — How do LinkedIn and other job apps handle onboarding? What's the industry standard? Context matters for a hypothesis
  3. Qualitative Research Plan — Quantitative focus only. Where's qualitative? Plan interviews at friction points
  4. Implementation Details — How to handle users without a resume/education? UX for "skip" or "add later"?
  5. Accessibility — No mention of accessibility testing, mobile-first design, or inclusive design
  6. Wireframes/Prototypes — High-level flow shown. Detailed wireframes for Version B? How to make 6 screens lightweight?
  7. Success Definition — 15% increase in good, but significant for business? Revenue/retention impact? Cost of longer flow?

Strategic Questions:

  • If Version B wins on completion but loses on retention?
  • Progressive profiling (ask data after first search)?
  • Fallback if Version B fails? Version C?
  • How communicate results? Decision framework?

Overall: Strong product thinking. Understand A/B testing. This is 70% of the strategic plan. Missing: user research, competitive context, implementation depth.

Next: User interviews (5-10 seekers) to validate. Competitor research. Detailed wireframes. Qualitative plan. Clarify success metrics.


Please correct me if I am wrong but this is not actual A/B testing. In the current project I see something more like a process flow, not a comparison.

Feedback based on: https://app.uxcel.com/glossary/ab-testing

Thank you for your feedback. I would like to clarify the A/B testing approach used in this project: You are absolutely right that A/B testing is a comparison between two versions. In this project, I am testing two distinct onboarding strategies: - Version A (The Control): The current, minimal 3-screen flow that focuses only on authentication and phone verification. - Version B (The Variation): An expanded 6-screen flow that includes mandatory steps for adding skills, education, and a resume. The "Test" here isn't about a visual redesign of the existing screens, but rather a test of "Friction vs. Value." I am testing whether adding these extra steps (the process flow changes) actually improves the quality of user profiles and job matches, or if it leads to higher drop-offs. Therefore, the comparison is between the "Short Flow" and the "Comprehensive Flow."

12 Claps
Average 4.0 by 3 people
5 claps
4 claps
3 claps
2 claps
1 claps
<?xml version="1.0" encoding="utf-8"?>