Alexander F. Spies
London, UK | [email protected] | linkedin.com/in/afspies | afspies.com | +44 7854 494 600
Summary
I work on making language models safer and more reliable. During my PhD I published four papers on causal world models & interpretability and received multiple AI-safety grants. At Epic Games I design pipelines that fine-tune and evalute frontier LLMs on low-resource data. I'm now looking to bring that blend of research rigour and production know-how to teams shipping reliable, scalable AI.
Education
Imperial College London
Oct 2020 - PresentThesis: Interpretable Representations in Artificial Neural Networks
- Research on object-centric representations and interpretability of LLMs
- Advisors – Prof. Alessandra Russo & Prof. Michael Shanahan
Imperial College London
Sep 2019 - Sep 2020- Thesis: Learning World Models in the Animal-AI Environment
- Independent project: Neurosymbolic Learning & Neurally Weighted dSILP
University of California, Berkeley
Aug 2017 - May 2018- Completed graduate-level courses as an undergrad, alongside research
University of Manchester
Sep 2015 - Jun 2019- Thesis: AI for the Automated Diagnosis of Atrial Fibrillation
Professional Experience
Epic Games
Jan 2025 - PresentResearch‑engineer intern focused on large‑scale fine‑tuning of LLMs on low-resource languages.
- Implemented finetuning pipeline for local as well as cloud-based training (UnSloth, SageMaker, etc.).
- Deployed evaluation suite with W&B sweeps, vLLM serving and multiple LLM APIs.
UnSearch (AI Safety Camp)
Mar 2023 - Oct 2024Led independent research groups on language model behaviour and interpretability.
- Developed research agenda on mechanistic interpretability for understanding maze-solving LLMs.
- Trained transformers models and Sparse Autoencoders and developed interpretability pipelines.
- Managed 9 researchers across 2 projects resulting in 2 workshop papers and a best-poster award.
National Institute of Informatics
Aug 2023 - Jun 2024Mechanistic analysis of Transformers trained on maze‑solving tasks.
Lawrence Berkeley National Laboratory
Feb 2018 - Jul 2018Investigated non‑local thresholds in pixel detectors; co‑authored JINST publication.
German Electron Synchrotron (DESY)
Jul 2018 - Sep 2018Exclusion analysis of Higgs decay channels in MSSM.
Publications
Detailed List →(*indicates equal contributions)
Selected Publications
[1]
Transformers Use Causal World Models in Maze‑Solving Tasks
World Models Workshop (ICLR 2025), Oct 2024
[2]
Structured World Representations in Maze‑Solving Transformers
Unifying Representations in Neural Models Workshop (NeurIPS 2023), Dec 2023
[3]
Sparse Relational Reasoning with Object‑Centric Representations
Dynamic Neural Networks Workshop (ICML 2022) — spotlight, Jul 2022
Skills
Frameworks & MLOps
Programming
Technical
Languages
Awards & Grants
Long‑Term Future Fund Grant — Safe AI Research
Jul 2024FAR Labs Residency
Jun 2024Best Poster — Technical AI Safety Conference
Apr 2024JSPS Postdoctoral Fellowship
May 2023Google Cloud Research Grant
Aug 20221st Place — AIHack 2022
Mar 2022Full PhD Scholarship (UKRI)
Sep 2020Leadership & Service
Pivotal Fellowship
Jan 2025 - Apr 2025- Provided technical guidance on AI Safety Research to 8+ Research Fellows
Imperial College London
Jan 2021 - Dec 2024Reviewer
Jan 2022 - Present- NeurIPS, ICLR, ICML, AAAI, UAI, Artificial Intelligence Journal
Imperial College London & Manchester
Sep 2021 - Feb 2025- Led technical coursework for Deep Learning, ML Math, Data Structures & Algorithms, and Python
- Engineered GPU-backed autograding pipeline for 120+ students using Otter Grader and Paperspace
ML Alignment & Theory Scholars
Jun 2024 - Jul 2024- Evaluated research proposals for alignment‑focused projects