Skip to main content
Navigate back to homepage
Open search bar.
Open main navigation menu

Main navigation

  • Study
    UCL Portico statue
    Study at UCL

    Being a student at UCL is about so much more than just acquiring knowledge. Studying here gives you the opportunity to realise your potential as an individual, and the skills and tools to thrive.

    • Undergraduate courses
    • Graduate courses
    • Short courses
    • Study abroad
    • Centre for Languages & International Education
  • Research
    Tree-of-Life-MehmetDavrandi-UCL-EastmanDentalInstitute-042_2017-18-800x500-withborder (1)
    Research at UCL

    Find out more about what makes UCL research world-leading, how to access UCL expertise, and teams in the Office of the Vice-Provost (Research, Innovation and Global Engagement).

    • Engage with us
    • Explore our Research
    • Initiatives and networks
    • Research news
  • Engage
    UCL Print room
    Engage with UCL

    Discover the many ways you can connect with UCL, and how we work with industry, government and not-for-profit organisations to tackle tough challenges.

    • Alumni
    • Business partnerships and collaboration
    • Global engagement
    • News and Media relations
    • Public Policy
    • Schools and priority groups
    • Visit us
  • About
    UCL welcome quad
    About UCL

    Founded in 1826 in the heart of London, UCL is London's leading multidisciplinary university, with more than 16,000 staff and 50,000 students from 150 different countries.

    • Who we are
    • Faculties
    • Governance
    • President and Provost
    • Strategy
  • Active parent page: UCL Engineering
    • Study
    • Research
    • Collaborate
    • Departments
    • Active parent page: News and Events
    • People
    • About

How do we teach AI to be moral? UCL research offers a new approach

Researchers from UCL Computer Science have developed a novel method for aligning artificial intelligence (AI) agents with explicit human moral values.

Breadcrumb trail

  • UCL Engineering
  • News and Events

Faculty menu

  • Events
  • Current page: Faculty news
  • UCL News: Engineering
Thirteenth International Conference on Learning representations with Elizaveta Tennant

Large language models (LLMs) are increasingly deployed as agents capable of making decisions and taking actions independently. These agents are used in tools like AI assistants, chatbots, decision-making systems, and negotiation tools-where they may need to act on behalf of a user, reason strategically, or interact with others. Ensuring that their behaviour aligns with human values is therefore a critical challenge.

Traditionally, LLM alignment has relied on preference data from human raters, but this method can be opaque, inconsistent, and expensive.

Instead, the UCL-led team explored a method that directly encodes moral principles into an AI agent’s reward function. Drawing on established philosophical frameworks -Deontological ethics and Utilitarianism - the team trained a language model to make decisions in a series of moral dilemma games. Using a classic environment known as the Iterated Prisoner’s Dilemma, they demonstrated that agents could learn ethical behaviours such as maximising collective good, reciprocating cooperation or avoiding betrayal.

They also showed that:

  • Agents trained to behave selfishly could later be “untrained” and learn to act more morally;
  • These moral behaviours could generalise to new environments, suggesting broad applicability of the approach.
Elizaveta Tennant pointing at a research poster

“This work shows that it is possible to specify and teach values to AI agents in a transparent manner, rather than relying on indirect or potentially biased human feedback,” said lead author Elizaveta Tennant, a PhD student at UCL Computer Science funded by the Leverhulme Doctoral Training Programme for the Ecological Study of the Brain.

The research has potential implications for any context where LLM agents act on behalf of users - from virtual customer service agents and productivity tools to multi-agent simulations and collaborative AI systems.  It also contributes to UCL’s broader work on Responsible AI and the safe deployment of frontier AI technologies. 

The study, led by Elizaveta Tennant, a final-year PhD student and supervised by Professor Stephen Hailes and Professor Mirco Musolesi, was presented at the International Conference on Learning Representations (ICLR 2025), held in Singapore from 24–28 April 2025. 

Conference:
International Conference on Learning Representations (ICLR 2025), 24–28 April, Singapore

Publication:
Moral Alignment for LLM Agents (ICLR 2025)

More from UCL Engineering...

Engineering Foundation Year
UCL East Marshgate building at dusk

Programme Spotlight

Engineering Foundation Year

We'll help you to gain new knowledge, learn academic and study skills, and develop your confidence levels so you'll have what it takes to transform your life.

Inaugural Lectures
Farhaneen Mazlan delivering a talk at UCL

Event series

Inaugural Lectures

An opportunity to explore ground-breaking research that is shaping the future and transforming the world.

Disruptive Thinkers Video Series
Dr Claire Walsh looking at a human organ in an imaging facility

Watch Now

Disruptive Thinkers Video Series

From making cities more inclusive to using fibre optics in innovative medical procedures, explore the disruptive thinking taking place across UCL Engineering.

UCL footer

Visit

  • Bloomsbury Theatre and Studio
  • Library, Museums and Collections
  • UCL Maps
  • UCL Shop
  • Contact UCL

Students

  • Accommodation
  • Current Students
  • Moodle
  • Students' Union

Staff

  • Inside UCL
  • Staff Intranet
  • Work at UCL
  • Human Resources

UCL social media menu

  • Link to Soundcloud
  • Link to Flickr
  • Link to TikTok
  • Link to Youtube
  • Link to Instagram
  • Link to Facebook
  • Link to Twitter

University College London, Gower Street, London, WC1E 6BT

Tel: +44 (0) 20 7679 2000

© 2025 UCL

Essential

  • Disclaimer
  • Freedom of Information
  • Accessibility
  • Cookies
  • Privacy
  • Slavery statement
  • Log in