Information Credibility: A Core Skill for AI-Era Performance

Information Credibility is a measurable cognitive construct essential for responsible AI-era performance across organisations and education. It enables structured reasoning, reduces automation risk, and supports defensible decision-making.

Behavioural Indicators

  • Applies structured reasoning under uncertainty
  • Challenges AI outputs before acting
  • Documents rationale for key decisions
  • Identifies bias and flawed assumptions

AI-Era Risk Dimension

Weak information credibility amplifies hallucination risk, automation bias, dashboard misinterpretation, and unverified output dependency. AI systems increase the scale and speed of poor judgement.

Assessment and Measurement

  • Scenario-based reasoning tasks
  • Structured simulations
  • Critical evaluation exercises
  • Rubric-scored written responses

Measurement ensures capability is demonstrated, not assumed.

Bridge Architecture: Corporate and School Pathways

Corporate pathway: This skill underpins AI governance, leadership risk management, and hiring validity.

School pathway: This skill supports AI literacy, exam reasoning, and structured thinking development.

MOSAIC Core Construct Framework

The MOSAIC Core Construct Framework defines twelve psychometrically defensible skill constructs that underpin
AI-era performance across organisations and schools.
Each construct has a corporate expression (RWA-aligned), an education expression (SET-aligned),
and a clear AI-risk dimension.

No. Construct Corporate Expression (RWA) Education Expression (SET) AI Risk Dimension
1 Analytical Reasoning Executive synthesis under complexity. Multi-step reasoning transfer. Over-trusting fluent AI summaries.
2 Inference Evaluation Testing whether conclusions follow from evidence. Distinguishing inference from opinion. Plausible but weak AI conclusions.
3 Assumption Detection Surfacing hidden premises in strategy and vendor claims. Identifying unstated assumptions in arguments. Hidden bias inside AI outputs.
4 AI Output Validation Oversight competence for AI-generated work. Evaluating AI-generated essays and answers. Hallucinations becoming decisions.
5 Ethical Judgement Governance trade-offs in AI use. Responsible AI classroom behaviour. Compliance and reputational exposure.
6 Data Interpretation Accurate dashboard and analytics reading. Graph and table interpretation. Misreading AI analytics outputs.
7 Cognitive Flexibility Adapting judgement under change. Switching reasoning strategies. Automation dependency.
8 Information Credibility Evaluating source reliability. Digital source evaluation. Deepfake and misinformation risk.
9 Structured Decision-Making Explicit criteria and audit trails. Clear argument structuring. Ad hoc AI-driven decision shortcuts.
10 Learning Agility Rapid upskilling and model updating. Feedback-driven improvement. AI workflow stagnation.
11 Attention Control Sustained focus under overload. Exam stamina and focus. Digital distraction bias.
12 Bias Recognition Spotting bias in systems and hiring. Recognising bias in texts and claims. Algorithmic bias blindness.

Rob Williams: 30 Years Designing High-Stakes Assessments

Rob Williams has spent three decades designing, validating, and calibrating:

  • Cognitive ability tests
  • Leadership judgement assessments
  • Situational judgement tests
  • Values and motivational diagnostics
  • High-stakes entrance examinations
  • Executive selection assessments

In each of these, the following AI skills are key:

    • Strategic reasoning
    • Ethical judgement
    • Risk evaluation
    • Applied problem solving
    • Ethical judgement

Hence, these are precisely the skills required to design high-quality psychometric assessments.

Contact

Rob Williams Assessment Ltd

E: rrussellwilliams@hotmail.co.uk

M: 077915 06395