How the
test works
The scoring logic, where the Myers-Briggs framework actually comes from, and an honest look at what a personality test can and can't tell you.
The test has 77 questions split across four dimensions — each question presents a real situation and two ways of responding. There's no rating scale, no "somewhat agree." You just pick the one that fits how you'd actually behave.
Each answer maps to one pole of a dimension. The side you chose more often across that dimension's questions determines your letter. Your final result is four letters — one winner from each dimension.
The spectrum bars on your results screen show the percentage split — a 65/35 result means you lean that way but aren't extreme. A 90/10 means that dimension is very consistent for you. Both are useful information.
Most personality tests ask you to rate statements like "I enjoy meeting new people" on a scale of 1 to 5. The problem is that people are genuinely unreliable narrators of their own behaviour. We answer based on who we think we are, who we want to be, and what we think the question is really asking — all at once.
Scenario-based questions sidestep this. When the question is "your colleague asks for honest feedback on a project you think has problems — what do you do?" you're not being asked to self-assess. You're responding to something concrete. The answer still reflects your personality, but it gets there without asking you to introspect accurately, which is a skill most people don't have reliable access to.
MBTI has been studied extensively, and the findings are mixed in ways that are worth being honest about. It's neither the rigorous science its proponents sometimes claim, nor the worthless pseudoscience its critics sometimes dismiss it as.
- The E/I dimension has strong overlap with the well-validated Extraversion dimension in the Big Five
- People generally find their type descriptions meaningful and behaviourally accurate
- Useful as a framework for self-reflection and conversation, even if imprecise as measurement
- Scenario-based formats tend to show better test-retest reliability than self-rating formats
- Test-retest reliability: studies show 30–50% of people get a different type when retested weeks later
- Treating dimensions as binary (J or P) rather than continuous loses real information
- Limited predictive validity for job performance compared to other assessment tools
- The Barnum effect — type descriptions are sometimes broad enough to feel true for almost anyone
The most useful way to approach your result is as a starting point, not a verdict. If the description is accurate in ways that feel specific rather than flattering, it's probably pointing at something real. If it doesn't fit, that's useful information too.
This is a free, independent test built on the publicly documented Myers-Briggs framework. It is not affiliated with, endorsed by, or derived from the official Myers-Briggs Company or CPP Inc. The questions are original, written to test each dimension through behaviour rather than self-assessment. The type and letter descriptions are original writing informed by the broad consensus portrait of each type that has built up across decades of use.
No data is collected. Your answers are scored locally in your browser and never sent anywhere. There is no account, no database, no tracking of results.
the scenarios say?