Machine Learning

  •  Agent-Based Code Repair |  Current Topics

    Context

    Large language models (LLMs) have became popular over the last few years, one of the reason being the quality of the outputs these models generate. LLM Agents are a step further, they allow LLMs to use memory, tools and sequential thinking.

  •  Exploration of Self-Reflective LLMs for Code |  Current Topics

    Context

    Large language models (LLMs) have became popular over the last few years, one of the reason being the quality of the outputs these models generate. Recent advancements try to make models think more, by either utilizing simple prompts or by training them using self-reflection via reinforcement learning.

  •  RL-based Training for Code in LLMs |  Current Topics

    Context

    Large Language Models (LLMs) have shown strong performance in code generation, completion, and repair tasks. However, supervised pretraining on massive code corpora is limited by data quality, lack of explicit feedback, and the inability to capture correctness beyond next-token prediction. Recent research has explored Reinforcement Learning (RL) based training approaches to refine LLMs for code. By leveraging feedback signals—such as compilation success, test case execution, or static analysis warnings—models can be trained to better align with correctness and developer intent.