Alexander F. Spies

London, UK | [email protected] | linkedin.com/in/afspies | afspies.com | +44 7854 494 600

Summary

I work on making language models safer and more reliable. During my PhD I published four papers on causal world models & interpretability and received multiple AI-safety grants. At Epic Games I design pipelines that fine-tune and evalute frontier LLMs on low-resource data. I'm now looking to bring that blend of research rigour and production know-how to teams shipping reliable, scalable AI.

Education

Imperial College London

Oct 2020 - Present
PhD in Computer Science, AI London, UK

Thesis: Interpretable Representations in Artificial Neural Networks

  • Research on object-centric representations and interpretability of LLMs
  • Advisors – Prof. Alessandra Russo & Prof. Michael Shanahan

Imperial College London

Sep 2019 - Sep 2020
MSc in Computing (AI & ML) London, UK
  • Thesis: Learning World Models in the Animal-AI Environment
  • Independent project: Neurosymbolic Learning & Neurally Weighted dSILP

University of California, Berkeley

Aug 2017 - May 2018
Study Abroad Year, Major: Physics Berkeley, CA, USA
  • Completed graduate-level courses as an undergrad, alongside research

University of Manchester

Sep 2015 - Jun 2019
MPhys in Theoretical Physics Manchester, UK
  • Thesis: AI for the Automated Diagnosis of Atrial Fibrillation

Professional Experience

Epic Games

Jan 2025 - Present
Research Engineer Intern London, UK

Research‑engineer intern focused on large‑scale fine‑tuning of LLMs on low-resource languages.

  • Implemented finetuning pipeline for local as well as cloud-based training (UnSloth, SageMaker, etc.).
  • Deployed evaluation suite with W&B sweeps, vLLM serving and multiple LLM APIs.

UnSearch (AI Safety Camp)

Mar 2023 - Oct 2024
Research Team Lead Remote

Led independent research groups on language model behaviour and interpretability.

  • Developed research agenda on mechanistic interpretability for understanding maze-solving LLMs.
  • Trained transformers models and Sparse Autoencoders and developed interpretability pipelines.
  • Managed 9 researchers across 2 projects resulting in 2 workshop papers and a best-poster award.

National Institute of Informatics

Aug 2023 - Jun 2024
JSPS Doctoral Fellow Tokyo, Japan

Mechanistic analysis of Transformers trained on maze‑solving tasks.

Lawrence Berkeley National Laboratory

Feb 2018 - Jul 2018
Undergraduate Researcher Berkeley, CA, USA

Investigated non‑local thresholds in pixel detectors; co‑authored JINST publication.

German Electron Synchrotron (DESY)

Jul 2018 - Sep 2018
Research Intern Hamburg, Germany

Exclusion analysis of Higgs decay channels in MSSM.

Publications

Detailed List

(*indicates equal contributions)

Selected Publications

[1]

Transformers Use Causal World Models in Maze‑Solving Tasks

A.F. Spies, W. Edwards, M.I. Ivanitskiy, et al.

World Models Workshop (ICLR 2025), Oct 2024

[2]

Structured World Representations in Maze‑Solving Transformers

M.I. Ivanitskiy*, A.F. Spies*, T. Räuker*, et al.

Unifying Representations in Neural Models Workshop (NeurIPS 2023), Dec 2023

[3]

Sparse Relational Reasoning with Object‑Centric Representations

A.F. Spies, A. Russo, M. Shanahan

Dynamic Neural Networks Workshop (ICML 2022) — spotlight, Jul 2022

Skills

Frameworks & MLOps

Hugging Face Transformers, PyTorch, Jax, Weights & Biases, Pandas

Programming

Python, C++, Java, Git, HTML, CSS, JavaScript

Technical

Language Model Training, Deep Learning, Interpretability, Representation Learning, AI Safety

Languages

English (native), German (native), Japanese (beginner)

Awards & Grants

Long‑Term Future Fund Grant — Safe AI Research

Jul 2024

FAR Labs Residency

Jun 2024

Best Poster — Technical AI Safety Conference

Apr 2024

JSPS Postdoctoral Fellowship

May 2023

Google Cloud Research Grant

Aug 2022

1st Place — AIHack 2022

Mar 2022

Full PhD Scholarship (UKRI)

Sep 2020

Leadership & Service

Pivotal Fellowship

Jan 2025 - Apr 2025
Technical Research Advisor London, UK
  • Provided technical guidance on AI Safety Research to 8+ Research Fellows

Imperial College London

Jan 2021 - Dec 2024
Co‑founder — ICARL Seminar Series London, UK

Reviewer

Jan 2022 - Present
Journals & Top ML conferences
  • NeurIPS, ICLR, ICML, AAAI, UAI, Artificial Intelligence Journal

Imperial College London & Manchester

Sep 2021 - Feb 2025
Teaching Assistant
  • Led technical coursework for Deep Learning, ML Math, Data Structures & Algorithms, and Python
  • Engineered GPU-backed autograding pipeline for 120+ students using Otter Grader and Paperspace

ML Alignment & Theory Scholars

Jun 2024 - Jul 2024
Research Proposal Reviewer
  • Evaluated research proposals for alignment‑focused projects