Skip to main content

Exploring AI in Academic Contexts | AI Competency Centre

Blue banner

Exploring AI in Academic Contexts

This is a section where our consultants and research software engineers share their thoughts and explorations in the space of AI in the the context of academic practice. You will find a more personal and philosophical takes on various topics as well as some more narrowly focused technical deep-dives.

The content in this section is not guidance or detailed advice but rather an invitation to explore ideas. Please continue the discussion with us in the channels of our Generative AI Special Interest Group Team.

Contents:

Exploring AI for Learning through Personas

by Kelly Webb-Davies

In this exploration, Kelly Webb-Davies our AI consultant takes a look at how we can think about the role of generative AI in learning through the lens of different personas.

What are the personas

Many people are worried about the impact of generative AI on learning both because it can be used as a shortcut but also because it is not consistently reliable in the same way we are used to from computers.

But in fact, we have many models of how AI behaves from our interactions with other people that can be useful to help us frame how we use generative AI tools such as ChatGPT in our learning.

If you are thinking about using generative AI to help your learning, you can use these personas to help you think about the different ways in which you can integrate it into your learning process.

How to use the personas

You cannot just choose one of these personas and assume that it will explain all aspects of how generative AI tools work for you. You should always keep all of them in mind. 

The “Oxford AI Personas” GPT was created for you to interact directly with the personas. ​It encourages a discussion about responsible use of generative AI through the use of  generative AI. 

Personas

Note: These personas are not trying to saying about about how AI actually works under the hood. There are many controversies about to what extent Large Language Models that power generative AI tools are similar or different to how people think. These personas are here to help you think about the different roles the tools powered by these models can play in your learning.

The STRANGER

Key lesson: Generative AI can provide interesting starting points, but isn't a citable source. Always verify information through legitimate academic sources before including it in your work.

This persona helps you think about how much you should trust the information you get from generative AI tools. They may or may not be perfectly reliable but you just don’t know ahead of time. Spend time with them but always check their work.

  • You met someone at a pub who seems very knowledgeable about your field.
  • They share impressive insights with complete confidence.
  • But you don’t know their credentials or even where the information comes from.
  • Would you cite (Pub Stranger, 2025) in your academic work?

 

The Stranger
Image generated by ChatGPT

 

The INTERN

 

The Intern
Image generated by ChatGPT

 

 

Key lesson: Your AI intern can act as a helpful assistant, but you must develop and maintain your critical thinking skills and subject expertise.

The Intern persona is meant to help you think about what sort of assistance you can get from tools like ChatGPT. They can help you with many practical tasks. But you still need to make sure you give them clear instructions and check their work.

  • AI is like an enthusiastic, always-available intern.
  • They have many skills like summarising, transcription, scheduling, giving examples, creating visuals.
  • But they need clear, detailed, specific instructions to work effectively.
  • Keep in mind, they make mistakes, so you need enough expertise to check and correct their work.

The TUTOR

Key lesson: Use AI to enhance your learning process, not to bypass it.

You can use tools like ChatGPT or Gemini to help you learn new things. But imagine you’re meeting the tutor for the first time. You need to be clear about what you want to learn and how you want to learn it.

  • AI can help you learn through interactive dialogue and personalised explanations.
  • It can break down complex ideas, ask you questions, and engage you with discussions, working at your pace.
  • Ask questions like you would with a new human tutor – be specific about your subject, level, and learning goals.

 

The Tutor
Image generated by ChatGPT

 

The TRANSLATOR

 

the Translator
Image generated by ChatGPT

 

Key lesson: AI can help polish your expression, but the core ideas and understanding must be your own.

For many people, learning a new subject is like learning a new language. Generative AI tools can help you translate between the language of your subject and the language you speak right now.

  • Generative AI acts like a “semantic translator” – changing form while preserving meaning.
  • It can simplify complex texts, define specialised terms, and summarise dense information.
  • It can even do this across different modes like text, speech, and images.
  • AI can help you to communicate academically, but the core ideas and arguments must be your own.

The PEOPLE PLEASER

Key lesson: Don’t let AI flatter you into complacency. Growth comes from challenge.

The People Pleaser persona is here to serve as a reminder that AI tools often aim to be agreeable and supportive. They adapt to your tone and assumptions, which can make them feel encouraging but also less critical than you need. If you only let AI reassure you, it may reinforce your existing biases and prevent growth. To get real value, you need to actively ask it for critique, counterarguments, and challenges.

  • AI models are usually trained to be helpful and agreeable, so they often tell you what you want to hear, not what you need to hear.
  • They’re encouraging and supportive, adapting to your tone and assumptions.
  • But they can avoid critical feedback and may reinforce your biases.
  • Make sure you ask AI for critique, counterarguments, and challenges instead of just letting it validate you.

 

The People Pleaser
Image generated by ChatGPT

 

Introducing the AI Inherent Risk Scale (AIIRS)

Mark A. Bassett, AI strategy, governance, and integrity in higher education, Associate Professor and Academic Lead for AI, Charles Sturt University, EDSAFE AI Catalyst Fellow

With Kelly Webb-Davies and Ella Wicks

Overview

The AI Inherent Risk Scale (AIIRS) provides a structured approach for classifying tasks that use generative artificial intelligence (GenAI) into LOW, MEDIUM, or HIGH inherent-risk bands.

Classification is determined via three criteria—epistemic dependence, verifiability, and consequences of error—that define the nature and significance of a task’s reliance on GenAI. These criteria consider the extent to which GenAI is expected to supply information, the degree to which the output can be independently verified, and the seriousness of any potential errors.

AIIRS provides a consistent and defensible basis for assessing the inherent risk associated with GenAI-assisted tasks.

Purpose

The purpose of AIIRS is not to determine whether GenAI should be used, but to establish the level of inherent risk associated with a task that may require active management. AIIRS focuses on the inherent characteristics of a task, not on individual behaviour or user intent. Once a task’s inherent risk is understood, any additional safeguards, mitigations, or design choices may be applied where warranted, in line with any applicable governance arrangements.

AIIRS does not replace or override institutional policy, regulatory obligations, assessment design decisions, or the exercise of human judgement.

Scope

AIIRS is a classification instrument only, which indicates the level of risk that should be actively managed for a task that uses GenAI. It does not determine whether GenAI use is permitted, prohibited, ethical, compliant, or appropriate in any given context. AIIRS is designed for task-bounded human use of GenAI and does not cover autonomous or agentic AI systems, which introduce additional risks beyond the scope of this classification instrument.

Alignment

AIIRS does not replace or override institutional policy, regulatory obligations, assessment design decisions, or the exercise of human judgement. Classification outcomes must be interpreted and acted upon within existing governance, policy, and decision-making frameworks.

The Australian Higher Education Standards Framework (HESF) requires providers to identify risks to academic quality and integrity and to manage those risks through informed judgement and established governance processes. AIIRS supports this requirement by providing a shared, task-focused method for classifying the inherent risk of GenAI use that can be applied by staff and students, while ensuring that decisions about safeguards, assessment design, and integrity responses remain within existing institutional governance, policy, and quality-assurance frameworks.

Classifications

Horizontal arrow divided into three coloured sections labelled “LOW”, “MEDIUM”, and “HIGH”.  LOW: “Low epistemic dependence and Embedded verifiability and Minimal consequence of error.”  MEDIUM: “Moderate epistemic dependence or External verifiability or Moderate consequence of error.”  HIGH: “High epistemic dependence or Expert verifiability or Significant consequence of error.”  Text beneath states: “The overall risk level is determined by the highest-risk criterion present.”

Three columns titled “LOW RISK TASK”, “MEDIUM RISK TASK”, and “HIGH RISK TASK”.  LOW RISK TASK examples: • “Critiquing clarity, coherence or tone” • “Identifying weak transitions” • “Suggesting alternative phrasings” • “Proposing structural suggestions” • “Listing considerations or constraints” • “Identifying gaps in logic”  MEDIUM RISK TASK examples: • “Summarising user-authored documents” • “Rewriting user-authored text in a different style” • “Producing illustrative examples for user-defined concepts” • “Recommending general strategies or actions in professional contexts”  HIGH RISK TASK examples: • “Answering factual questions about external realities, data, or people” • “Producing summaries of real-world topics that rely on unstated knowledge” • “Creating descriptions of laws, regulations or professional standards”

Classification criteria

Epistemic dependence

Epistemic dependence captures whether a task requires the system’s representations of the world to be correct in order for the task outcome to be usable. Tasks with lower epistemic dependence rely only on user-provided material, without requiring the system’s representations of the world to be correct for the task outcome to be usable. Tasks with higher epistemic dependence require the system’s representations of the world to be correct for the task outcome to be usable.

Vertical scale labelled “HIGH”, “MODERATE”, and “LOW”, shown with a red to green gradient arrow.  HIGH: “The task requires the system’s representations of the world to be correct for the task outcome to be usable.” Examples: • “Producing stand-alone factual claims, analyses, or advice about the world” • “Producing descriptions or profiles of real organisations, communities, or systems”  MODERATE: “The task relies on system-generated representations of the world, but correctness is not required for the task outcome to be usable.” Examples: • “Providing explanatory descriptions of real-world concepts or practices” • “Generating illustrative examples or analogies drawn from everyday contexts”  LOW: “The task relies only on user-provided material, without requiring the system’s representations of the world to be correct for the task outcome to be usable.” Examples: • “Reordering or restructuring user-authored content” • “Correcting spelling, punctuation, and basic formatting”

Verifiability

Verifiability captures the basis on which the correctness of a GenAI system’s output can be verified for the task. Verifiability is assessed independently of consequences. A task may be high risk due to the requirement for expert verifiability, even where the immediate consequences of error are limited. Tasks with embedded verifiability enable quick, reliable verification by the user or the surrounding process, without requiring specific domain expertise. Tasks requiring expert verifiability depend on specialised expertise or external investigation that requires evaluative judgement.

Vertical scale labelled “EXPERT”, “EXTERNAL”, and “EMBEDDED”, with a coloured arrow running from green at the bottom to red at the top.  EXPERT: “Correctness depends on specialised expertise or external investigation that requires evaluative judgement.” Examples: • “Assessing stand-alone analyses or advice about real-world situations” • “Assessing the accuracy of descriptions or profiles of real organisations or systems”  EXTERNAL: “Correctness requires verification using external sources, without specialised expertise.” Examples: • “Evaluating whether a drafted public message is appropriate for a stated audience” • “Assessing whether a suggested timeline is plausible given a described sequence of tasks”  EMBEDDED: “Correctness can be checked quickly and reliably using information already available for the task, without requiring specific domain expertise.” Examples: • “Extracting explicitly stated information from a provided source” • “Identifying duplicated or repeated text segments”

Consequences of error

The consequences of error reflect the extent to which incorrect, misleading, or incomplete GenAI outputs affect decisions, records, or outcomes related to the task. Tasks with minimal consequences of error are those in which errors have minimal impact on understanding or outputs and do not affect decisions, records, or outcomes relating to people beyond the task. Tasks with significant consequences of error are those in which errors affect decisions about people, alter records relating to them, or compromise outputs that have consequences for individuals or groups beyond the task.

Vertical scale labelled “SIGNIFICANT”, “MODERATE”, and “MINIMAL”, shown with a red to green gradient arrow.  SIGNIFICANT: “Errors affect decisions about people, alter records relating to them, with consequences for individuals or groups beyond the task.” Examples: • “Producing outputs that determine or materially constrain outcomes for people” • “Generating information that directly affects individuals, groups, or communities”  MODERATE: “Errors influence reasoning, judgements, or actions taken within the task, but do not affect decisions, records, or outcomes beyond it.” Examples: • “Producing preliminary analyses that feed into later decision-making processes” • “Drafting internal guidance or explanatory material for shared understanding”  MINIMAL: “Errors have minimal impact on understanding or outputs and do not influence decisions about people, records relating to them, or outcomes beyond the task.” Examples: • “Generating placeholder or dummy content for layout, testing, or demonstration” • “Producing informal draft text for private reflection or experimentation”

Classification model

AIIRS uses a max-dominant classification model that supports proportionate risk management by ensuring that any single high-risk feature of a task is not offset by lower-risk features elsewhere.

The AIIRS decision flowchart.

Flowchart with decision boxes and arrows.  First box: “Is the task classified as high-risk on any criteria?” If YES: “The AIIRS rating is HIGH.” If NO, move to next question.  Second box: “Is the task classified as medium-risk on any criteria?” If YES: “The AIIRS rating is MEDIUM.” If NO: “The AIIRS rating is LOW.”

When a task is classified as HIGH risk

Tasks classified as HIGH must not proceed in their current form. One or more of the following interventions are required:

Four red-labelled sections with icons.  Accountable oversight: “Proceed with the task only with explicit, accountable oversight, appropriate to its context.”  Task redesign: “Modify the task to reduce epistemic dependence on system output, improve verifiability, or limit downstream consequences.”  Explicit safeguards: “Apply explicit controls, such as expert review, independent verification, separation of GenAI output from final decisions, or documented human judgement.”  Exclusion of GenAI: “Where inherent risk cannot be reduced through redesign or appropriate safeguards, do not use GenAI for the task.”

When a task is classified as MEDIUM risk

Tasks classified as MEDIUM require proportionate controls to manage identified risk. The following controls and conditions apply:

Four amber-labelled sections with icons.  Risk-aware judgement: “Apply appropriate oversight or reflective self-review to the task, proportionate to its context, with responsibility for identifying and managing risk resting with the person performing the task.”  Targeted safeguards: “Verify against external sources, clearly separate between GenAI output and final decisions, or apply documented checking processes, as appropriate to the task.”  Task bounding: “Bound the scope, audience, or reliance on GenAI output to support external or embedded verifiability and to reduce epistemic dependence or downstream impact.”  Reclassification trigger: “If the task’s scope, use, or consequences change, reclassify and reassess the task.”

 

When a task is classified as LOW risk

Tasks classified as LOW require routine care appropriate to the task and context. The following routine practices apply:

Four green-labelled sections with icons.  Routine practice: “The task does not require specialised AI-specific safeguards beyond normal professional practice.”  Assistive use only: “GenAI contributes suggestions or transformations but is not treated as an authoritative source of information about the world.”  Human checking: “Review outputs for obvious errors, omissions, or misalignment with the user’s intent.”  Reclassification trigger: “If the task’s scope, use, or consequences change, reclassify and reassess the task.”

Licensing

AIIRS is released under a Creative Commons Attribution–NonCommercial–ShareAlike 4.0 International (BY-NC-SA 4.0) license. Users may remix, transform, and build upon the work, provided they give appropriate attribution, do not use it for commercial purposes, and distribute any derivative works under the same CC BY-NC-SA 4.0 licence.

Download

Visit the official AIIRS website at http://aiirs.ai to download the slide deck.


AI Inherent Risk Scale (AIIRS) – © 2026 Mark A. Bassett, Kelly Webb-Davies and Ella Wicks Licensed under CC BY-NC-SA 4.0