| --- |
| license: apple-amlr |
| base_model: |
| - Qwen/Qwen2.5-Coder-7B |
| tags: |
| - code |
| - text-diffusion-model |
| - diffusion large language model |
| --- |
| |
| ### DiffuCoder-7B-Base |
|
|
| The DiffuCoder-7B-Base model is our foundational masked diffusion LLM for code generation. |
|
|
| - Training recipe: Using [DiffuLLaMA](https://github.com/HKUNLP/DiffuLLaMA)'s adaptation approach, trained on a large corpus of code: with Stage 1 65B tokens and Stage 2 65B tokens. |
|
|
| - Benchmarks: Strong baseline performance on HumanEval, MBPP and BigCodeBench. |
|
|
| #### More details and usage examples: |
|
|
| - Paper: [DiffuCoder: Understanding and Improving Masked Diffusion Models for Code Generation](https://arxiv.org/abs/2506.20639) |
|
|
| - GitHub: https://github.com/apple/ml-diffucoder |
|
|
| ``` |
| import torch |
| from transformers import AutoModel, AutoTokenizer |
| |
| model_path = "apple/DiffuCoder-7B-Base" |
| model = AutoModel.from_pretrained(model_path, torch_dtype=torch.bfloat16, trust_remote_code=True) |
| tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True) |
| model = model.to("cuda").eval() |
| |
| prompt = """ |
| from typing import List |
| |
| def has_close_elements(numbers: List[float], threshold: float) -> bool: |
| \"\"\" |
| Check if in given list of numbers, are any two numbers closer to each other than given threshold. |
| >>> has_close_elements([1.0, 2.0, 3.0], 0.5) |
| False |
| >>> has_close_elements([1.0, 2.8, 3.0, 4.0, 5.0, 2.0], 0.3) |
| True |
| \"\"\" |
| """ |
| |
| TOKEN_PER_STEP = 1 # diffusion timesteps * TOKEN_PER_STEP = total new tokens |
| |
| inputs = tokenizer(prompt, return_tensors="pt") |
| input_ids = inputs.input_ids.to(device="cuda") |
| attention_mask = inputs.attention_mask.to(device="cuda") |
| |
| output = model.diffusion_generate( |
| input_ids, |
| attention_mask=attention_mask, |
| max_new_tokens=256, |
| output_history=True, |
| return_dict_in_generate=True, |
| steps=256//TOKEN_PER_STEP, |
| temperature=0.2, |
| top_p=0.95, |
| alg="entropy", |
| alg_temp=0., |
| ) |
| generations = [ |
| tokenizer.decode(g[len(p) :].tolist()) |
| for p, g in zip(input_ids, output.sequences) |
| ] |
| |
| print(generations[0].split(tokenizer.eos_token)[0]) |
| ``` |
|
|
| #### Acknowledgement |
| To power this HuggingFace model release, we reuse [Dream](https://huggingface.co/Dream-org/Dream-v0-Base-7B)'s modeling architecture and generation utils. |