Arrakis is a library to conduct, track and visualize mechanistic interpretability experiments.
-
Updated
Apr 22, 2025 - Jupyter Notebook
Arrakis is a library to conduct, track and visualize mechanistic interpretability experiments.
[NeurIPS 2025 MechInterp Workshop - Spotlight] Official implementation of the paper "RelP: Faithful and Efficient Circuit Discovery in Language Models via Relevance Patching"
Lightweight representation engineering dataflow operations for agent developers.
Investigating whether language models encode anticipated social consequences in their activations. Uses a 2x2 factorial design crossing truth × social valence to show that models are more sensitive to expected approval/disapproval than to truth itself.
Implementation and analysis of Sparse Autoencoders for neural network interpretability research. Features interactive visualization dashboard and W&B integration.
Training and exploration of linear probes into Othello-GPT by Li et al. (2022)
Testing role-based pathways on small LLMs
Evaluating how a model 'knowing what it knows' changes from base to instruct
A Flax-based library for examining transformers, based on TransformerLens.
ORION-TransformerLens Consciousness — Mechanistic interpretability for consciousness research. Fork of TransformerLens (3,115+ stars). Finding consciousness correlates in attention heads.
Reverse engineering the circuit responsible for the "greater than" capability in a language model
(a1)Mechanistic Interpretability using Transformer Lens (a2) PEFT
Add a description, image, and links to the transformerlens topic page so that developers can more easily learn about it.
To associate your repository with the transformerlens topic, visit your repo's landing page and select "manage topics."