alan-turing-institute / robots-in-disguiseLinks
Information and materials for the Turing's "robots-in-disguise" reading group on fundamental AI research.
☆33Updated 5 months ago
Alternatives and similar repositories for robots-in-disguise
Users that are interested in robots-in-disguise are comparing it to the libraries listed below
Sorting:
- Zennit is a high-level framework in Python using PyTorch for explaining/exploring neural networks using attribution methods like LRP.☆230Updated last month
- LENS Project☆49Updated last year
- PyTorch code corresponding to my blog series on adversarial examples and (confidence-calibrated) adversarial training.☆67Updated 2 years ago
- MetaQuantus is an XAI performance tool to identify reliable evaluation metrics☆39Updated last year
- An eXplainable AI toolkit with Concept Relevance Propagation and Relevance Maximization☆130Updated last year
- ☆74Updated 2 years ago
- 🧠 Starter templates for doing interpretability research☆73Updated 2 years ago
- Starting kit for the NeurIPS 2023 unlearning challenge☆376Updated 2 years ago
- OpenXAI : Towards a Transparent Evaluation of Model Explanations☆247Updated last year
- ☆81Updated last year
- Layer-wise Relevance Propagation for Large Language Models and Vision Transformers [ICML 2024]☆184Updated last month
- ☆81Updated 6 months ago
- ViT Prisma is a mechanistic interpretability library for Vision and Video Transformers (ViTs).☆301Updated last month
- Notebooks accompanying Anthropic's "Toy Models of Superposition" paper☆129Updated 2 years ago
- we got you bro☆36Updated last year
- ☆68Updated 2 years ago
- A toolkit for quantitative evaluation of data attribution methods.☆53Updated last month
- 👋 Overcomplete is a Vision-based SAE Toolbox☆79Updated last month
- LLM finetuning in resource-constrained environments.☆51Updated last year
- Benchmarks for the Evaluation of LLM Supervision☆32Updated last month
- PyTorch-centric library for evaluating and enhancing the robustness of AI technologies☆57Updated last year
- The M2L school 2022 tutorials☆36Updated 2 years ago
- Efficient LLM inference on Slurm clusters using vLLM.☆77Updated this week
- Discount jupyter.☆51Updated 5 months ago
- Tools for studying developmental interpretability in neural networks.☆101Updated 2 months ago
- Dataset and code for the CLEVR-XAI dataset.☆31Updated last year
- Lecture on Automated Machine Learning☆77Updated 2 years ago
- Sparse Autoencoder for Mechanistic Interpretability☆260Updated last year
- Quantus is an eXplainable AI toolkit for responsible evaluation of neural network explanations☆621Updated last month
- NeuroSurgeon is a package that enables researchers to uncover and manipulate subnetworks within models in Huggingface Transformers☆41Updated 6 months ago