splade-code-06B / modeling_splade.py
Tom Aarsen
Attempt to integrate with Sentence Transformers; simplify implementation
d71d855
raw
history blame
326 Bytes
"""
This file exists solely to allow loading the Qwen3ForCausalLM via the AutoModelForMaskedLM class.
Compared to standard Qwen3, we're using bidirectional attention and not causal attention, but it's specified
with `is_causal=False` in the config.
"""
from transformers import Qwen3ForCausalLM
__all__ = ["Qwen3ForCausalLM"]