Diffusion-based Code Language Model

Context

While standard code models generate text one token at a time (autoregressive), Diffusion Language Models (DLMs) generate and refine the entire block of code simultaneously. This allows the model to look ahead and fix structural errors in a non-linear fashion, where inference becomes an online optimization problem.

Goal

The student will fine-tune a Discrete Diffusion Model like LLaDA-Instruct specifically for reasoning about code execution on a custom tracing dataset for python code.

Requirements

Pointers

Contact