Arrakis is a library to conduct, track and visualize mechanistic interpretability experiments.
-
Updated
Apr 22, 2025 - Jupyter Notebook
Arrakis is a library to conduct, track and visualize mechanistic interpretability experiments.
[NeurIPS 2025 MechInterp Workshop - Spotlight] Official implementation of the paper "RelP: Faithful and Efficient Circuit Discovery in Language Models via Relevance Patching"
Lightweight representation engineering dataflow operations for agent developers.
Implementation and analysis of Sparse Autoencoders for neural network interpretability research. Features interactive visualization dashboard and W&B integration.
Investigating whether language models encode anticipated social consequences in their activations. Uses a 2x2 factorial design crossing truth Ă— social valence to show that models are more sensitive to expected approval/disapproval than to truth itself.
Training and exploration of linear probes into Othello-GPT by Li et al. (2022)
Testing role-based pathways on small LLMs
A Flax-based library for examining transformers, based on TransformerLens.
Reverse engineering the circuit responsible for the "greater than" capability in a language model
(a1)Mechanistic Interpretability using Transformer Lens (a2) PEFT
Add a description, image, and links to the transformerlens topic page so that developers can more easily learn about it.
To associate your repository with the transformerlens topic, visit your repo's landing page and select "manage topics."