-
Notifications
You must be signed in to change notification settings - Fork 417
Pull requests: ml-explore/mlx-lm
Author
Label
Projects
Milestones
Reviews
Assignee
Sort
Pull requests list
Fix dynamic_quant for MoE and VL models
#870
opened Feb 10, 2026 by
Taderich73
Loading…
3 tasks done
server: add usage.prompt_tokens_details.cached_tokens to json response
#849
opened Feb 6, 2026 by
percontation
Loading…
refactor: use time.perf_counter() for duration measurements
#848
opened Feb 6, 2026 by
m92y
Loading…
feat: enhance chat CLI with readline history, line editing, and distributed support
#841
opened Feb 3, 2026 by
Vlor999
Loading…
moved more activation functions to activations module
#774
opened Jan 19, 2026 by
Goekdeniz-Guelmez
Loading…
feat: allow callers to cancel stream generation via callback check, and ensure prompt cache consistency
#710
opened Dec 30, 2025 by
zhutao100
Loading…
Add tokenizer_config support to LoRA YAML configuration
#688
opened Dec 17, 2025 by
breitburg
Loading…
Add server logging and stop generation on client disconnect
#677
opened Dec 15, 2025 by
otarkhan
Loading…
Added optional accuracy reporting while training and evaluating models
#595
opened Nov 7, 2025 by
jyork03
Loading…
Fix : mlx-server for chunked request (to support one-api, curl)
#589
opened Nov 3, 2025 by
yiakwy-xpu-ml-framework-team
Loading…
Normalize scheduler/warmup step-like arguments by grad_accumulation_steps; warmup-only configuration, sane argument defaults and tests
#582
opened Nov 1, 2025 by
jyork03
Loading…
Previous Next
ProTip!
Mix and match filters to narrow down what you’re looking for.