Exam Blueprint & Domain Architecture
This page illustrates the kind of competency architecture an SPCE Associate exam would require — domains, cognitive levels, and standards mappings. This is not a blueprint, a starting point, or a proposal. It is a sample — a demonstration of the structure that a formal Job Task Analysis (JTA) and psychometric validation would produce. The actual domains, weights, and Bloom's distributions would be determined by the JTA study and a qualified psychometrician — not by this document.
To illustrate what a competency framework looks like in practice, here are six sample domains that an Associate-level exam might cover. Each maps to recognized external standards — not SolidProfessor curriculum. Under ISO/IEC 17024, competency definitions must be derived from practitioner-validated job analysis, not vendor content.
Core solid modeling operations: extrude, revolve, sweep, loft. Emphasis on design intent — building models that communicate purpose and survive modification.
Fully-defined sketches, geometric and dimensional constraints, reference planes and axes. The foundation of parametric stability.
Patterns, mirrors, shells, fillets, chamfers — and the judgment calls behind them. When to use a pattern vs. copy. When to add fillets early vs. late. Feature order strategy.
Component insertion, standard mates, advanced mates, assembly design intent. Building assemblies that are stable, maintainable, and communicate relationships between parts.
Engineering drawing creation, dimensioning, tolerancing, section views, detail views. The bridge between 3D model and manufacturing floor — communicating design intent to people who will never open the CAD file.
DFM principles, draft angles, wall thickness, CNC considerations, material selection impact on design. Understanding how what you model affects what gets manufactured.
Bloom's Taxonomy classification based on: Anderson, L. W., & Krathwohl, D. R. (Eds.). (2001). A Taxonomy for Learning, Teaching, and Assessing. Longman.
SPCE domains map to established competency frameworks, not SolidProfessor course structure. This crosswalk demonstrates that the exam blueprint is grounded in industry-recognized standards — a requirement for ISO/IEC 17024 accreditation.
| SPCE Domain | ASME Y14.5 Mapping | NIMS Mapping | Why It Matters |
|---|---|---|---|
| Parametric Modeling | — | — | Core CAD competency; no external standard covers digital modeling directly |
| Sketch Constraint | — | — | Foundational to parametric robustness; assessed through modeling tasks |
| Feature Operations | — | — | Design intent is the bridge between "can model" and "can engineer" |
| Assembly & Mate | — | NIMS: Assembly & Integration | Assembly strategy affects manufacturing sequence and serviceability |
| Drawing & Annotation | Y14.5: Dimensioning & Tolerancing, Y14.100: Engineering Drawing Practices | NIMS: Blueprint Reading | The universal language between design and manufacturing |
| Manufacturing Awareness | Y14.5: GD&T application to manufacturing | NIMS: CNC Programming, Machining, Materials | Validates that the engineer designs for manufacturability, not just geometry |
The exam blueprint isn't just an exam design — it's a data schema. Every domain, every Bloom's level, every score becomes structured intelligence. Here's how that data flows beyond the credential itself.
Verified domain-level skill profiles — not self-reported, not inferred from course completions. Machine-readable competency claims via Open Badges 3.0 that ATS systems can parse. Employers filter by specific capability: "show me engineers who scored 85%+ on Assembly & Mate Strategy." The credential becomes the hiring signal that the market currently lacks.
Domain-level gap analysis transforms certification from a one-time event into a continuous improvement signal. A score of 42% on Drawing & Annotation maps directly to a targeted learning path. Managers see exactly where their team is strong and where the gaps are — not "completed 80% of courses" but "weak in GD&T application, strong in parametric modeling." This is the data engineering managers told us they need.
Aggregate anonymized competency data across thousands of candidates reveals what the industry actually knows — and what it doesn't. Which domains have the widest skill gaps? Where do recent graduates struggle most? What competencies correlate with faster time-to-productivity? This intelligence doesn't exist today because no one is measuring engineering competency at scale with structured, domain-level granularity.
This competency framework is a sample to illustrate the kind of structure a credible certification requires. Before any of it becomes an actual exam blueprint, three things must happen:
See the ISO 17024 Compliance Roadmap for the full timeline and dependencies.
Sources:
Bloom's Taxonomy: Anderson, L. W., & Krathwohl, D. R. (2001). A Taxonomy for Learning, Teaching, and Assessing. Longman. Summary (PDF)
ISO/IEC 17024:2012: iso.org/standard/52993
ASME Y14.5-2018: asme.org
NIMS: nims-skills.org
Open Badges 3.0: 1edtech.org/standards/open-badges
Angoff Method: assess.com | Angoff, W. H. (1971). In Thorndike (Ed.), Educational Measurement, 2nd ed.
Item bank sizing: PSI Exams: Item Bank Blueprint
JTA methodology: Certiverse JTA FAQ
Industry benchmarks (approximate): CompTIA (ISO 17024) · PMI/PMP (ISO 17024) · ISC2/CISSP (ISO 17024)