Skip to content

mdlm.predictor_mdlm

MDLMPredictor

Bases: Module, Predictor[MDLMBatch, MDLMPredictionDict]

Base predictor for MLM. Stochastically selects positions to unmask based on max_steps and max_new_tokens.

__init__(max_steps, max_new_tokens=None, tokenizer=None, model=None, noise_schedule=None, top_k=None, top_p=None)

Initialize MDLM Predictor.

Parameters:

Name Type Description Default
max_steps int

Maximum number of prediction steps.

required
tokenizer Optional[Tokenizer]

Tokenizer for encoding/decoding.

None
noise_schedule Optional[NoiseSchedule]

Noise schedule for the diffusion process.

None
top_k Optional[int]

Top-k sampling parameter.

None
top_p Optional[float]

Top-p sampling parameter.

None
model Optional[MDLMModel]

The MDLM model to use for predictions.

None

decode(results)

Parameters:

Name Type Description Default
results MDLMStepResults

x: Integer[TT, " batch seq_len"] Current predicted sequence.

required

Returns: out: List[str] Decoded sequence with special tokens. x: Integer[TT, " batch seq_len"] Current predicted sequence.