Skip to content

Pull requests: Blaizzy/mlx-vlm

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Assigned to nobody Loading
Sort

Pull requests list

Improve text-only decode performance
#1105 opened May 4, 2026 by lucasnewman Collaborator Loading…
Fix Gemma4 chat templating for Python generate()
#1099 opened May 1, 2026 by Girish011 Loading…
minicpmo / fastvlm: fix pixel cast on quantized language models
#1098 opened May 1, 2026 by contrapuntal Loading…
3 tasks done
fix: Batch generation breaks top-p sampling
#1094 opened Apr 30, 2026 by spicyneuron Contributor Loading…
fix: Prevent batched cache metadata lazy graph buildup
#1093 opened Apr 30, 2026 by spicyneuron Contributor Loading…
Add Sapiens2 + RTMDet (top-down pose pipeline)
#1081 opened Apr 26, 2026 by Blaizzy Owner Loading…
8 tasks done
fix: remap per-layer quantization keys for mixed-bit models
#1078 opened Apr 26, 2026 by ivanfioravanti Contributor Loading…
Fix Unicode byte-fallback decoding in server streaming responses
#1076 opened Apr 26, 2026 by spicyneuron Contributor Loading…
Fix stale test_utils.py regressions + extract get_class_predicate
#1071 opened Apr 25, 2026 by mdstaff Contributor Loading…
2 tasks done
Handle OpenAI tool_choice requests
#1070 opened Apr 25, 2026 by eloe Draft
Support OpenAI stop sequences in server
#1069 opened Apr 25, 2026 by eloe Loading…
Add MiniCPM-V 4.6 support
#1058 opened Apr 24, 2026 by pzc163 Loading…
Add vision feature caching to all models
#1028 opened Apr 16, 2026 by Blaizzy Owner Loading…
6 tasks done
fix: propagate the verbose to the Prefill tqdm
#1015 opened Apr 12, 2026 by PeterStaar-IBM Loading…
server: indicate finish reason properly when model made a tool call.
#1014 opened Apr 12, 2026 by viktike Contributor Loading…
ProTip! What’s not been updated in a month: updated:<2026-04-04.