← Back
TypeScriptNode.jsSQLiteMCPAPI DesignClaude Code

4D AI Fluency Lab

A local Model Context Protocol server that measures, tracks, and develops AI fluency through behavioral assessment of Claude conversations. Classifies prompts against a 4-dimensional rubric, generates HTML reports, trains learners with interactive exercises, and lets coaches run structured programs without leaving Claude Code.

what it does

Classify

Score any Claude conversation against the 4D rubric: every user prompt, all four dimensions, sub-behavior tags.

Report

Self-contained HTML reports with radar charts, score breakdowns, dimension insights, and shareable PNG cards.

Train

Duolingo-style trainer built from the learner's own classified conversations: XP, hearts, streaks.

Coach at Scale

Generate turnkey champion packets a coach can run without education team involvement.

Track Progress

Compare any two profiles, see score deltas per dimension, and visualize trends over time.

Cohort Analytics

Group reporting: tier distribution, weakest shared dimension, outlier detection across a team.

architecture

INPUTPROCESSSTORAGEOUTPUTSrubric.mdsingle source of truthclassify-exportClaude scores each promptSQLite DBprofiles · scores · historyHTML ReportTrainer GameChampion PacketCohort AnalyticsClaude Sessions~/.claude/projects/
rubric.md is the single source of truth, versioned in the database so every profile, artifact, and assessment item is permanently tagged with the rubric version it was scored against. Claude reads the rubric during classify-export and scores each prompt 0–3 per dimension. All results land in SQLite, then fan out to report generation, trainer construction, champion packets, and cohort analytics tools.

the 4d rubric

1.02.03.0DelegationDescriptionDiscernmentDiligenceExpert (avg ≥ 2.0)Foundation (avg < 1.0)

Delegation

What the user delegates vs. keeps, from handing off the full task to setting goals and constraints.

Description

How much the user specifies product, process, and performance constraints.

Discernment

Whether the user builds in evaluation structure: asks for alternatives, flags uncertainty.

Diligence

Responsibility signals: deployment context, privacy, ethics, accountability, transparency.

Based on Anthropic's AI Fluency Index (Feb 2026), measuring 11 observable behaviors across 9,830 real conversations. Rubric version: 1.1.0 (2026-02-26)

scoring & tiers

0.01.02.03.0FoundationPractitionerExpertavg score < 1.0avg score 1.0 – 1.99avg score ≥ 2.0

Foundation

avg < 1.0

Prompt-level thinking. Delegates tasks but provides minimal constraints or evaluation structure.

Practitioner

avg 1.0–1.99

Growing fluency. Starting to specify outputs and occasionally ask for alternatives or flag uncertainty.

Expert

avg ≥ 2.0

Full fluency. Frames work as goals, specifies constraints, builds in accountability and evaluation.

tool suite 20+ MCP tools

Assessment

5
classify_exportstore_profileget_profile_historycompare_profilesget_prompt_scores

Reports & Training

4
generate_reportopen_reportgenerate_trainergenerate_packet

Cohort Analytics

3
store_cohortadd_profile_to_cohortget_cohort_report

Champion Enablement

2
store_champion_feedbackget_content_health

Content Management

2
store_content_artifactlist_stale_artifacts

Session Import

2
list_claude_sessionsimport_claude_session

Rubric

3
check_rubric_currencyget_assessment_itemsupdate_calibration

slash commands

/classify

Browse Claude Code sessions, pick one, score it, view the report

/classify-web

Paste a claude.ai export and classify it against the rubric

/sessions

Browse all Claude Code sessions organized by project

/report

Open a generated HTML report for any stored profile

/progress

Compare two most recent profiles and surface score deltas

/trainer

Generate an interactive Duolingo-style game from a profile

/packet

Generate a turnkey champion success packet for a learner

/feedback

Log champion observations after a coaching session

/health

Content health dashboard: what's fresh vs. stale

/rubric

Check rubric version and surface key changes

generated outputs

Full HTML Report

report-{id}.html

  • -Radar chart across all 4 dimensions
  • -Score bar per dimension (0–3 scale)
  • -Diligence signal breakdown
  • -Insights panel with course links
  • -Monthly prompt volume + score trend chart

Shareable Card

share-{id}.html

  • -Dark theme, 600px wide
  • -Downloadable as PNG via html2canvas
  • -Overall score /3.0
  • -Dimension bars + radar
  • -Monthly trend visualization

Trainer Game

trainer-{id}.html

  • -XP system: Foundation → Practitioner → Expert
  • -3 hearts per session, streak tracking
  • -Mastery: prompts retire after 3 correct
  • -Round types: spot_upgrade, what_changed, tag_dimensions

Champion Packet

packet-{id}.html

  • -Dual-tab: Champion Guide + Learner Deep Dive
  • -Pre-scored conversation examples
  • -Coaching frameworks per dimension
  • -Turnkey, no education team needed

output samples

AI Fluency Assessment report showing dimension scores and overall score

generated HTML report: dimension score bars (Delegation, Description, Discernment, Diligence), overall score out of 3.0, and strongest / focus area cards

Score trend chart and annotated strong prompt examples

monthly score trend line and annotated prompt examples. strong prompts are highlighted by behavior tag showing exactly which signals drove the score

Prompts to strengthen section with before and after rewrites

prompts to strengthen: each weak prompt is annotated with what's missing, paired with a strengthened rewrite that adds delegation structure, output constraints, and diligence signals

claude code · /classify

end-to-end run inside Claude Code: session selection, rubric scoring, and HTML report generation all from a single /classify command

tech stack

LayerTechnologyNote
RuntimeNode.js + TypeScriptcompiled via tsc
MCP@modelcontextprotocol/sdkstdio transport
DatabaseSQLite via better-sqlite3WAL mode, FK constraints on
ReportsSelf-contained HTML + Chart.jsno external dependencies
Share Cardshtml2canvasdownloadable PNG
ValidationZod schemasall tool inputs validated
HTTPlocalhost:3131serves generated HTML reports

database schema

profilesprompt_scoresdimension_distributionsrubric_versionsassessment_itemscohortscohort_memberschampion_feedbackcontent_artifactsuser_calibrations

All multi-table inserts use transactions. WAL mode enabled. Every profile, artifact, and assessment item is permanently tagged with the rubric version it was scored against.