Apply Human Factors to the design of an
AI-based assistant in aviation

    This Path guides in the design of AI-based assistants, with a focus on effective human–AI performance. It is focused on Technology Readiness Levels (TRLs) 1–6, encompassing the conceptualisation, design, development and validation phases. The guidance can be followed linearly, with likely iterations as the AI-based assistant is refined.

How does it work?

1
Identify end-users and stakeholders
  • Who are the key stakeholders and what do they want from the new system?
2
Understand context and identify requirements (Human, AI)
  • How will roles and tasks be split between human and AI?
3
Human–AI teaming CONOPS definition
  • How will humans and AI work together on tasks?
4
Initial Design activity
  • How will human–AI interaction look and feel in practice?
5
Identify Risks and System-Level Issues
  • What could go wrong, and what safeguards need to be put in place?
6
Validate and iterate higher fidelity designs
  • Is the system design working effectively and delivering what we wanted?
7
Deployment & Continuous Improvement
  • What are the steps to take to ensure safe and effective deployment?
EACH PHASE MAY REQUIRE ITERATIONS
Can be mapped onto the “V-MODEL” of Software Development

Choose phases to follow based on the TRL Level of your system

TRLs 1–4

Conceptual Exploration

For projects exploring AI-based system possibilities at a conceptual level, without actually building an AI component.

Recommended for you: First 3-4 phases of the path to flesh out the concept before the code
TRLs 5+

Active Development

For projects actively developing AI-based tools and human-AI interfaces.

Recommended for you: All seven phases should be completed for a full certification evidence trail
HAIKU
PHASE 1

Identify end-users and stakeholders

In this phase, we establish the initial context by identifying all relevant stakeholders and end users. This includes those who will interact directly with the AI assistant, as well as those affected by its outputs or dependent on its performance.

PHASE OUTCOME: Clear understanding of the assistant’s context and purpose.
GATHER CONTEXTUAL KNOWLEDGE
RESOURCE

Documentation review

Review existing research, documentation, and reports to identify stakeholders, users, and operational needs relevant to the AI system.

Outcome
Knowledge Database
IDENTIFY STAKEHOLDERS
METHODOLOGY

Stakeholder mapping

Identify and categorise individuals or groups who affect or are affected by the AI system.

Required HF Expertise
Time required (est.)
Outcome
Visual stakeholder map highlighting key actors, their roles, and relationships.
PHASE 2

Understand context and identify requirements (Human, AI)

This phase establishes a foundational understanding of the work environment, human roles, and key challenges the AI assistant must support. Techniques such as Observation & Walk-throughs, Semi-Structured Interviews, and Focus Groups help ensure that the work as currently done is thoroughly understood, including all the factors, uncertainties, trade-offs and constraints that human end users habitually manage on a day-to-day basis. This in-depth understanding helps to ensure the AI assistant will be able to meet the challenge of assisting the human end user in realistic working conditions. Hierarchical Task Analysis enables a first structured allocation of the split of roles and tasks between human and AI assistant, showing how the future human-AI teaming concept could work in an integrated fashion. The additional consideration of Legal (e.g. EU Act on AI) and Regulatory Requirements (e.g. EASA) can further ensure a good allocation of function between human and AI system elements.

PHASE OUTCOME: Defined AI role and its support to users, with human oversight ensured.
DATA COLLECTION (Initial Exploration)
METHODOLOGY

Observation; Walk/Talk-through; Verbal Protocol

Techniques to elicit expertise to understand what, how and why operators do.

Required HF Expertise
Time required (est.)
Outcome
Understanding of human tasks
DATA COLLECTION Detailed Inquiry
METHODOLOGY

Semi-structured Interviews

Use guided conversations to explore user experiences while allowing flexibility and depth.

Required HF Expertise
Time required (est.)
Outcome
In-depth knowledge of how users experience a certain system/workflow/operation
DATA REPRESENTATION
METHODOLOGY

Hierarchical Task Analysis

Using a diagram to visualise how different tasks interlink to achieve a system goal.

Required HF Expertise
Time required (est.)
Outcome
Clear representation of the human tasks
RUN WORKSHOPS TO ELICIT INITIAL REQUIREMENTS
METHODOLOGY

Focus Groups

Gather collective insights through structured group discussion.

Required HF Expertise
Time required (est.)
Outcome
List of user requirements
ANALYSE THE REGULATORY LANDSCAPE
RESOURCE

Legal documents

Useful external links to regulations, standards, and other background documents to guide your work.

Outcome
List of legal requirements
EXPLORE EASA AI LEVELS
RESOURCE

EASA Guidance

Read about the concept of Human-AI Teaming (HAT) and the main design principles to ensure safe “Human-AI Interaction”

Outcome
List of compliance requirements
PHASE 3

Human–AI teaming CONOPS definition

In this phase it is important to understand how the human and AI will work together on specific tasks in a set of representative scenarios, whether routine, maintenance, emergency or a combination of the three. Scenario-Based Design can identify and explore different scenarios, and Ideation Sessions can then consider how human and AI would respond and interact. This can then be made explicit by mapping the different actors, events and actions in time-based Operations Sequence Diagram. Early application of the HAIQU tool in this Phase. HAIQU (Human-AI QUestionnaire) is a tool designed within the context of HAIKU to be used in a collaborative session to elicit and document teaming-specific requirements. It is envisaged to use HAIQU in multiple phases. In this phase, it is particularly useful to explore the areas of Human Centred Design, Roles and Responsibilities and Teamworking, which will help verify the Conops from a Human Factors Requirements perspective, and avoid problems costly changes in these areas in later Phases.

PHASE OUTCOME: Detailed scenarios outlining human-AI teaming, ready for interface design.
UNDERSTAND THE TASK(S) DISTRIBUTION
METHODOLOGY

Scenario-Based Design

To describe existing or imagine new activities produced by interacting with a new tool.

Required HF Expertise
Time required (est.)
Outcome
Map of task allocation
MAP INTERACTIONS OVER TIME
METHODOLOGY

Operation Sequence Diagram (OSD)

Used for multi-person tasks and maps who does what/when/with which information.

Required HF Expertise
Time required (est.)
Outcome
Map of task allocation
IDEATION
METHODOLOGY

Co-design session(s)

Collaboratively shape initial Human-AI teaming concepts with users and stakeholders.

Required HF Expertise
Time required (est.)
Outcome
Conceptualisation
VERIFY CONOPS FROM A HF PERSPECTIVE focus on Human Centred Design, Roles and Responsibilities and Teamworking
TOOL

Human-AI Teaming Questionnaire (HAIQU)

A web app for collaborative expert analysis of AI systems against Human Factors requirements.

Required HF Expertise
Time required (est.)
Outcome
Refinement of concept with Human-AI Teaming requirements
PHASE 4

Initial design Activity

In this phase a prototype, whether static (e.g. successive screenshots) or dynamic (an interactive interface) is developed and tested with end users using Low Fidelity Prototyping. To represent the AI part of the interaction, either a ‘Wizard of Oz’ approach is used in which a human pretends to be the AI (e.g. responding using text messages), or else a programme gives scripted answers that the ‘real’ AI would generate in the real situation. In some cases, an early prototype of the AI assistant itself may be ready, in which case it can be used (this can also help the ‘training’ of the AI). A key ingredient in human-AI teaming is sense-making, ensuring that the human and AI are ‘on the same page’. If prototyping suggests that the human may need to better understand what the AI is doing and why, this is where Explainability Generation should be applied. The HAIQU tool should be reapplied in this phase, focusing on Sense-Making (which includes displays, interactions and explainability) and Communications (particularly if speech interfaces are to be used). This phase is likely to go through several iterations and should enlist end-user feedback at each iteration.

PHASE OUTCOME: Preliminary design of the human-AI interaction interface.
DEVELOP LOW-FIDELITY INTERACTIVE PROTOTYPES
METHODOLOGY

Low-Fidelity Prototyping

A cost-effective and efficient approach for assessing different design solutions and making informed decisions, at multiple levels of fidelity.

Required HF Expertise
Time required (est.)
Outcome
Mock-up
EXPLAINABILITY GENERATION
METHODOLOGY

Construal Level Theory (CLT) for XAI Generation

A psychological framework applied to design layered, context-specific explanations for AI systems.

Required HF Expertise
Time required (est.)
Outcome
Levels of explainability
SIMULATE AND REFINE BEHAVIOUR
METHODOLOGY

Wizard of OZ

A method for testing complex systems by simulating functionality with a human behind the scenes, avoiding costly development while gathering realistic feedback.

Required HF Expertise
Time required (est.)
Outcome
List of requirements to improve mock-up
VERIFY CONOPS FROM A HF PERSPECTIVE focus on Sensemaking and communications
TOOL

Human-AI Teaming Questionnaire (HAIQU)

2nd application

A web app for collaborative expert analysis of AI systems against Human Factors requirements.

Required HF Expertise
Time required (est.)
Outcome
Refinement of concept with Human-AI Teaming requirements
PHASE 5

Identify Risks and System-Level Issues

In this phase risks are identified related to the use of the AI-based assistant. Techniques such as Human HAZOP, SHELL and Expert Walk-through can be applied to low and higher-TRL projects, whereas techniques such as STPA are for later TRL projects (TRL 6+). These techniques all aim to identify and mitigate vulnerabilities in the human-AI teaming operation, and as such, normally only one technique needs to be applied, though the application of more than one may flag differently nuanced risks. Legal Case Methodology, usually applied to later TRL projects, considers legal aspects. The HAIQU tool can be applied again with a focus on Errors and Resilience, and the project progressing towards TRLs 5 and beyond may wish to revisit the EASA Regulations, specifically those relating to human-AI teaming, explainability and ethics.

PHASE OUTCOME:Identified hazards, risks, and proposed mitigations.
IDENTIFY HAZARDS
METHODOLOGY

Human Hazard and Operability Study (HAZOP)

Structured workshop-based technique to anticipate deviations in human-AI interaction through systematic application of guidewords across operational sequences.

Required HF Expertise
Time required (est.)
Outcome
List of hazards from human and AI perspective
IDENTIFY GENERAL RISKS AND FURTHER REQUIREMENTS
METHODOLOGY

SHELL Model

Shows key human factors components of sociol-technical systems. Allows problem analysis.

Required HF Expertise
Time required (est.)
Outcome
List of hazards from human and AI perspective
EVALUATE CONCEPT ROBUSTNESS
METHODOLOGY

Expert Walkthroughs

Assess concept robustness by gathering expert feedback on systemlevel functionality

Required HF Expertise
Time required (est.)
Outcome
Identification of operational weak spots
EVALUATE ACCIDENTS FROM A SYSTEM LEVEL POINT OF VIEW
METHODOLOGY

System-Theoretic Accident Model & Processes (STAMP)

A systemic accident analysis method focusing on control processes and dysfunctional interactions within socio-technical systems.

Required HF Expertise
Time required (est.)
Outcome
List of compliance requirements
IDENTIFY LEGAL RISKS
METHODOLOGY

Legal Case Methodology

Identifying, analysing, and mitigating legal risks, particularly liability issues, arising from AI and advanced automation in safety-critical domains.

Required HF Expertise
Time required (est.)
Outcome
Document outlining all legal risks
MONITOR SYSTEM EVOLUTION Focus on Errors and Resilience
TOOL

Human-AI Teaming Questionnaire (HAIQU)

3rd application

A web app for collaborative expert analysis of AI systems against Human Factors requirements.

Required HF Expertise
Time required (est.)
Outcome
Refinement of concept with Human-AI Teaming requirements
REVISIT EASA
RESOURCE

EASA Guidance

2nd application

Read about the concept of Human-AI Teaming (HAT) and the main design principles to ensure safe “Human-AI Interaction”

Outcome
Updated list of compliance requirements
PHASE 6

Validate and iterate higher fidelity designs

In this phase a dynamically interactive system is available and tested with licensed end users in a high-fidelity simulation across one or more scenarios. Performance is measured, including overall system performance as well as the performance of the constituent components (human and AI). Since the AI assistant is there to support the human end user, measures such as workload and situation awareness may be measured, as well as canvassing simulation participants’ views on the degree of support afforded by the AI assistant, via Qualitative Debriefings and post-simulation questionnaires such as the System Usability Scale. In some cases, more advanced ‘Neuro-ID’ psycho-physiological measures (e.g. heart rate, galvanic skin response, EEG, etc.) may be used to infer impacts on the human user. Eye Tracking may also be used to determine effects on pilot or air traffic controller visual patterns and sense-making of the scenario, or to better track the dynamic interaction between human and AI. System Logs can often help understand the detail of such interactions. Following such simulations (often there are more than one, to allow at least one design iteration), the HAIQU tool can be run again for the previous six areas, updating earlier responses where new insights or information have arisen. As mentioned under Phase 4, towards the end of Phase 5 hazard analyses should be repeated to see if mitigations identified in Phase 4 worked, and if any new hazards have been discovered.

PHASE OUTCOME: Most HF requirements addressed; initial user approval achieved.
PRODUCE ROBUST PROTOTYPES

Build robust prototype

Translate your validated design concepts into working code to create a functional and testable prototype.

Outcome
Working prototype for evaluation
CONDUCT SIMULATIONS
METHODOLOGY

Real-time simulations

To test the future product and simulate the environment under assessment in real time.

Required HF Expertise
Time required (est.)
Outcome
Prototype integrated with simulator
COLLECT VISUAL DATA During the experiment
METHODOLOGY

Eye tracking analysis technique

To test the future product and simulate the environment under assessment in real time.

Required HF Expertise
Time required (est.)
Outcome
Data of usage during the experiment
COLLECT COGNITIVE DATA During the experiment
METHODOLOGY

Neuro-ID

To test the future product and simulate the environment under assessment in real time.

Required HF Expertise
Time required (est.)
Outcome
Data of usage during the experiment
COLLECT DATA During the experiment, qualitative data
METHODOLOGY

Observation; Walk/Talk-through; Verbal Protocol

2nd application

Techniques to elicit expertise to understand what, how and why operators do.

Required HF Expertise
Time required (est.)
Outcome
Qualitative data to explore the UX
COLLECT DATA During/post experiment, quantitative data
METHODOLOGY

System Log Analysis

Analyse recorded system data to understand user actions, AI responses, and interaction patterns over time.

Required HF Expertise
Time required (est.)
Outcome
Quantitative data of usage
COLLECT DATA Post-experiment, qualitative data
METHODOLOGY

Qualitative Debriefings

Gather in-depth feedback after users interact with a prototype or system.

Required HF Expertise
Time required (est.)
Outcome
Qualitative data to explore the UX
COLLECT DATA VIA QUESTIONNAIRES Post-experiment, mixed
METHODOLOGY

Post-exercise questionnaire

Gather structured user feedback after system use to assess usability, workload, trust, and more.

Required HF Expertise
Time required (est.)
Outcome
Actionable data to use for future iterations
MONITOR SYSTEM EVOLUTION Full evaluation
TOOL

Human-AI Teaming Questionnaire (HAIQU)

4th application

A web app for collaborative expert analysis of AI systems against Human Factors requirements.

Required HF Expertise
Time required (est.)
Outcome
Refinement of concept with Human-AI Teaming requirements
LOOPING BACK

HAZARD Analysis

Repeat steps from previous phases to see if the potential hazards persists

Outcome
Improved concept
PHASE 7

Deployment and Continuous Improvement

This phase consists of preparation for transition to deployment into the intended operational, organisational and social environment for which the AI-based assistant will be used. The HAIQU tool contains two areas relevant to Phase 7, namely Competencies and Training, and Organisational Readiness. Additionally, two questionnaires on Societal Acceptance and Safety Culture help determine the readiness of the user population to accept the new technology, and any concerns over impacts on individual or organisational safety culture. When the tool is first being deployed and people are being trained to use it and working with the tool, it can be useful to apply the User Journey Map technique on a representative sample of end users. This tool picks up annoyances (called ‘pain points’), whether related to the tool itself, the way it is being released and deployed into the system, or lack of smooth integration into legacy systems. Such problems can detract from the tool’s effective usage, such that its full benefits are never realised, no matter how well it was designed. Lastly, Error Reporting on the use of the tool is critical in the early deployment phase in case of errors (human or AI) or misunderstandings or other problems. If such problems are not detected quickly and corrected, the AI assistant will rapidly fall into disuse. This Phase does not end until decommissioning, and so is a continuous learning and adaptation phase, and hopefully one in which the AI assistant becomes a valued part of the aviation system in which it serves. This phase implies successive AI Maintenance activities (such as continuous monitoring and benchmarking, and model retraining whenever necessary).

PHASE OUTCOME: AI assistant deployed, monitored, and continuously improved for effective long-term use.
FOCUS ON COMPETENCIES AND ORGANISATIONAL READINESS
TOOL

Human-AI Teaming Questionnaire (HAIQU)

5th application

A web app for collaborative expert analysis of AI systems against Human Factors requirements.

Required HF Expertise
Time required (est.)
Outcome
Refinement of concept with new competencies/organisational requirements
ASSESS BROADER IMPACT
METHODOLOGY

Societal Acceptance Questionnaire

A questionnaire to measure perceptions and attitudes toward use of an AI-based system

Required HF Expertise
Time required (est.)
Outcome
Concrete numbers related to the Societal Acceptane for your IA
ASSESS SAFETY CULTURE
METHODOLOGY

Safety Culture Debrief

A short questionnaire for aviation workers (pilots and ATCOs) to elicit perceptions and judgements about an IAs potential impact on Safety Culture.

Required HF Expertise
Time required (est.)
Outcome
Deep understanding of how safety culture might be concretely impacted by the IA
IDENTIFY PAIN POINTS FROM ACTUAL USAGE
METHODOLOGY

User Journey Map

Map real user experiences to uncover gaps, pain points, and improvement opportunities.

Required HF Expertise
Time required (est.)
Outcome
Clear overview on the UX of your system
MAINTAIN THE AI SYSTEM

AI Maintenance activities

  • Monitor and benchmark
  • Model retraining
Outcome
Updated and optimised AI model
REPEAT STEPS

Design Iteration

Redesign mockup and review concept based on first simulations. Go back to previous steps in the path if necessary

Outcome
Improved concept