Prompt IQ Test is a local-first desktop app for assessing how well a person can use LLMs and AI agents.
It is not only about whether someone can write a prompt. The goal is to evaluate whether a user can clearly define goals, structure vague requests, provide enough context and constraints, iterate effectively, and drive an AI system toward a high-quality result.
This repository is currently in the planning stage.
- Product scope is defined
- Core interaction modes are defined
- Public repository README is ready
Prompt IQ Test is designed to answer two practical questions:
- Can a normal user actually use AI well?
- Can an advanced user drive an LLM or Agent at a high level?
The first release focuses on individual users. Hiring, training, and enterprise assessment workflows are future directions, not part of v1.
Fun Quick Test: short, lightweight, lower confidenceFull Assessment: broader coverage, more reliable diagnosisStatic Assessment: user submits a prompt for a given taskDynamic Assessment: user collaborates with an execution model over multiple turnsBuilt-in Question Bank: sample tasks from local question setsAI-Generated Questions: generate tasks from user-provided constraints or random generation
Prompt Only: score the prompt itself without executing itExecute Then Evaluate: run the prompt with an execution model, then score both prompt quality and execution result
- Main chat with an execution AI
- Live side-panel scoring by an evaluator AI
- Trend and current estimated score during the session
- Final report after completion
The first release centers on a fixed core rubric, including:
- clarity of request expression
- ability to define output format
- ability to turn vague needs into structured tasks
- sufficiency of context and constraints
- ability to identify and fill missing conditions
- iteration and acceptance awareness
Prompt IQ Test is designed as a local-first application.
- The program does not upload data to the author's server
- The program is fully free and open source
- Actual cost comes from the user's own AI API usage
- User content sent to configured AI providers may still be visible to those providers
- The author does not assume liability for outcomes caused by use of the software
The current recommended stack for v1 is:
ReactTypeScriptViteTauriSQLite
Target form factor:
- cross-platform desktop
- Windows-first adaptation for the first release
The first release is expected to include:
- Home page
- Assessment setup page
- Static assessment page
- Dynamic assessment page
- Result report page
- History pages
- Settings page
- About page
- Plugin marketplace placeholder page
- Vent-game placeholder page
- Startup privacy and disclaimer dialog
The first release does not aim to include:
- account system
- cloud sync
- real plugin runtime
- real plugin marketplace
- real vent-game chat window
- JSON or image report export
- difficulty system for questions
- enterprise or recruitment scoring mode
- user-editable hidden prompts or scoring rubrics
- License:
Apache-2.0
- Runtime files, local databases, exports, caches, and secrets should not be committed
- Example configuration files can be added later if needed
High-level next steps:
- finish the public repository baseline
- implement the desktop application shell
- implement static and dynamic assessment flows
- implement local persistence and Markdown export
- refine question seeds, prompts, and evaluation contracts