James Campbell

I am an incoming CS PhD student at CMU.

Currently, my interests include: LLM agents, evaluation, and alignment.

I am very eager to meet other ambitious people! If you'd like to chat, please don't hesitate to get in touch.

Headshot of James Campbell

About

In Fall 2024, I will be starting a CS PhD at Carnegie Mellon University.

As an undergrad at Cornell, I worked a lot on LLM interpretability and honesty, and was a primary contributor to the papers Representation Engineering and Localizing Lying in Llama. I have also done work on semantic representations in the brain and LLM robustness.

Papers

Image for paper 1

Localizing Lying in Llama: Understanding Instructed Dishonesty on True-False Questions Through Prompting, Probing, and Patching

James Campbell*, Phillip Guo*, Richard Ren*

Accepted at NeurIPS 2023 SoLaR Workshop

Summary: We prompt Llama-2-70B-chat to lie and localize mechanisms involved using activation patching and linear probing.

ArXiv | PDF | Code | Thread
Image for paper 2

Representation Engineering: A Top-Down Approach to AI Transparency

Andy Zou, Long Phan*, Sarah Chen*, James Campbell*, Phillip Guo*, Richard Ren*, Alexander Pan, Xuwang Yin, Mantas Mazeika, Ann-Kathrin Dombrowski, Shashwat Goel, Nathaniel Li, Michael J. Byun, Zifan Wang, Alex Mallen, Steven Basart, Sanmi Koyejo, Dawn Song, Matt Fredrikson, J. Zico Kolter, Dan Hendrycks

Submitted to ICLR 2024

Summary: We introduce the field of Representation Engineering which seeks to understand and control LLM's using a top-down approach.

ArXiv | Website | Code | Coverage
Image for previous paper

Considerations of Biological Plausibility in Deep Learning

James Campbell

Published front page of the Cornell Undergraduate Research Journal (CURJ)
Winner of the $300 James E. Rice Award

A literature review of learning algorithms and their biological plausibility

Paper

Projects

Past Research

Image

Theoretical Bound on the Weights of a Neural Network under Nesterov's Accelerated Gradient Flow

I solved an "open question" posed in a previous paper by deriving this bound on the weights of a neural network. This was done in summer of 2021 during an REU at Johns Hopkins under Rene Vidal.

PDF
Image

Neural Network Linearization for Out-of-Distribution Robustness

In summer of 2022, I worked with Yaodong Yu on improving out-of-distribution robustness using the empirical neural tangent kernel.

Poster | Slides | Code
Image

Data-driven clustering of neural responses to a large set of natural images

I investigated the clusterability of visual representations in the brain. This work was done in the Computational Connectomics Lab at Cornell and was presented as a poster at OHBM.

Abtract | Poster
Image

Exploring Additive Compositionality in the Brain

I conducted experiments attempting to understand the extent to which semantic representations in the brain exhibit additive compositionality, i.e. does representation(A) + representation(B) = representation(A+B).

Slides | Code
Image

Cornell Machine Learning Kaggle Competition

I came in first out of 155 participants in a Kaggle competition hosted by Cornell's big machine learning class.

Kaggle | Code
Image

CereBERTo: Improving Distributional Robustness with Brain-Like Language Representations

Improving the out-of-distribution robustness of BERT and GPT-2 by pretraining them to predict brain fMRI data.

Code

Blog

Coming Soon