Skip to content
@ippolito-cmu

ippolito-cmu

👋 Welcome to ippolito-cmu!

Research projects from Daphne Ippolito's lab at Carnegie Mellon University's Language Technologies Institute.

Last updated: July 2025

🔬 Projects

"Forcing Diffuse Distributions out of Language Models"
Yiming Zhang, Avi Schwarzschild, Nicholas Carlini, Zico Kolter, Daphne Ippolito
Paper

Addresses the problem that instruct LLMs exhibit strong biases when asked to generate random outputs, proposing a fine-tuing method to make models produce more uniform distributions for synthetic dataset generation.

"Chasing Random: Instruction Selection Strategies Fail to Generalize"
Harshita Diddee, Daphne Ippolito
Paper

Demonstrates that popular instruction selection strategies for LLMs generalize poorly across different datasets and evaluation setups, often failing to consistently outperform random baselines while incurring higher computational costs than training on full datasets.

"Human-Aligned Chess With a Bit of Search"
Yiming Zhang, Athul Paul Jacob, Vivian Lai, Daniel Fried, Daphne Ippolito
Paper

Introduces Allie, a chess AI that learns to ponder lke humans with a time-adaptive MCTS algorithm. Achieves state-of-the-art prediction of human behavior and strong skill adaptation against online players from 1000-2600 Elo.

"Persistent Pre-Training Poisoning of LLMs"
Yiming Zhang*, Javier Rando*, Ivan Evtimov, Jianfeng Chi, Eric Michael Smith, Nicholas Carlini, Florian Tramèr, Daphne Ippolito Paper

Shows that 0.1% poisoning of LLM's pre-training data persists after alignment (SFT and DPO), and an attacker can achieve various goals such as denial-of-service, context extraction and belief manipulation with access to pre-training data alone.

"NoveltyBench: Evaluating Creativity and Diversity in Language Models"
Yiming Zhang, Harshita Diddee, Susan Holm, Hanchen Liu, Xinyue Liu, Vinay Samuel, Barry Wang, Daphne Ippolito
Paper Website

Benchmark for measuring how well language models generate multiple diverse, high-quality responses to queries involving subjectivity, randomness, and creativity. Evaluates 20+ frontier models including GPT-4o, Claude 3.5 Sonnet, and Gemini 2.0 Pro.

"CIE: Controlling Language Model Text Generations Using Continuous Signals"
Vinay Samuel, Harshita Diddee, Yiming Zhang, Daphne Ippolito
Paper

Continuous control signal approach for controlling attributes in LM outputs. Experiments ran for response word count attribute but designed to be extended to other attributes as well. We show stronger control of desired word count compared to previous approaches including prompting approaches.

"Command-V: Pasting LLM Behaviors via Activation Profiles"
Barry Wang, Avi Schwarzschild, Alexander Robey, Ali Payani, Charles Fleming, Mingjie Sun, Daphne Ippolito
arXiv

"Finetune once, use on many LLMs" - the first method to demonstrate cross-architecture transfer of finetuned adapters without backpropagation. Works between Llama, Qwen, and other model families.

Popular repositories Loading

  1. allie allie Public

    Human-Aligned Chess With a Bit of Search

    Python 17 3

  2. Command-V Command-V Public

    ⌘V: Approximate another LLM’s activations on the fly, and transplant its finetuning to apply.

    Python 3

  3. ChasingRandom ChasingRandom Public

    Official Repository for the Paper: Chasing Random: Instruction Selection Strategies Fail to Generalize

    Python 2

  4. CIE CIE Public

    [EMNLP 2025] Dataset and Code of "CIE: Controlling Language Model Text Generations Using Continuous Signals"

    Python 2

  5. diffuse-distributions diffuse-distributions Public

    Forked from y0mingzhang/diffuse-distributions

    Forcing Diffuse Distributions out of Language Models

    Python

  6. .github .github Public

    ippolito-cmu landing page customisations

Repositories

Showing 6 of 6 repositories

People

This organization has no public members. You must be a member to see who’s a part of this organization.

Top languages

Loading…

Most used topics

Loading…