LFM2.5-VL-450M-MLX-6bit

MLX export of LFM2.5-VL-450M for Apple Silicon inference.

LFM2.5-VL-450M is a vision-language model built on the LFM2.5-350M backbone with a SigLIP2 NaFlex vision encoder (86M). It supports OCR, document comprehension, multilingual vision understanding, bounding box prediction, and function calling.

Model Details

Property Value
Parameters 450M
Precision 6-bit
Group Size 64
Size 0.43 GB
Context Length 32K
Vision Encoder SigLIP2 NaFlex (86M)
Native Resolution up to 512x512

Quickstart

uv pip install 'mlx-vlm==0.3.9'
from mlx_vlm import load, generate
from mlx_vlm.utils import load_image

model, processor = load("LiquidAI/LFM2.5-VL-450M-MLX-6bit")

image = load_image("photo.jpg")

# Apply chat template (required for LFM2.5-VL)
messages = [{"role": "user", "content": [
    {"type": "image"},
    {"type": "text", "text": "What do you see in this image?"},
]}]
prompt = processor.apply_chat_template(messages, add_generation_prompt=True)

result = generate(
    model,
    processor,
    prompt,
    [image],
    temp=0.1,
    min_p=0.15,
    repetition_penalty=1.05,
    verbose=True,
)
print(result.text)

Recommended Sampling Parameters

Parameter Value
temperature 0.1
min_p 0.15
repetition_penalty 1.05

License

This model is released under the LFM 1.0 License.

Downloads last month
-
Safetensors
Model size
0.2B params
Tensor type
BF16
·
U32
·
MLX
Hardware compatibility
Log In to add your hardware

6-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for LiquidAI/LFM2.5-VL-450M-MLX-6bit

Quantized
(6)
this model

Collection including LiquidAI/LFM2.5-VL-450M-MLX-6bit