- 
                Notifications
    You must be signed in to change notification settings 
- Fork 836
Open
Description
This document includes the features in LMFlow's roadmap. We welcome any discuss or contribute to the specific features at related Issues/PRs. 🤗
Main Features
- Data
- DPO dataset format [Feature] reward model inferencer and dpov2 aligner #867
- Conversation template in DPO [Feature] Iterative DPO #883
- jinja template [usability] support qwen2.5 and deepseek #931
- Tools in conversation dataset function-call-finetune #884 Modify the formatter of function and observation #892 [usability] support qwen2.5 and deepseek #931
- Packing with block diagonal attention
- Add a tokenize-only script, allowing tokenization separately without gpu environment. (For those who has large datasets but limited gpu hrs)
 
- Model
- Backend
- 🏗️ Accelerate support [usability] Accelerate Support #936
 
- Tokenization
- Tokenization update, using hf method [usability] support qwen2.5 and deepseek #931
 
 
- Backend
- Pipeline
- Train/Finetune/Align
- DPO (multi-gpu) [Feature] reward model inferencer and dpov2 aligner #867
- Iterative DPO [Feature] reward model inferencer and dpov2 aligner #867 [Feature] Iterative DPO #883
- PPO
- LISA (multi-gpu, qwen2, chatglm)
- Batch size and learning rate recommendation (arxiv)
- No trainer version pipelines, allowing users to customize/modify based on their needs
- Sparse training for moe models [New Feature] Is Mixtral supported? #879
 
- Inference
- vllm inference [Feature] vllm inferencer and memory safe vllm inferencer #860 [Feature] Add vllm inference example #863
- Reward model scoring [Feature] reward model inferencer and dpov2 aligner #867
- Multiple instances inference (vllm, rm, others) [Feature] Iterative DPO #883
- Inference checkpointing and resume from checkpoints
- Inference accelerate EAGLE
-  Inferencer for chat/instruction models, and chatbot.pyupgrade Conversation_template #917
 
 
- Train/Finetune/Align
Usability
- Make some packages/functions (gradio, vllm, ray, etc.) optional, add conditional import. [usability] deps streamlining #905
-  Inference method auto-downgrading (vllm>ds, etc.), and make vllmpackage optional. [usability] deps streamlining #905
-  Merging similar model methods into hf_model_mixin
-  Set torch_dtype='bfloat16'whenbf16is specified, etc. (bf16is inFinetunerArgumentsbuttorch_dtypeis inModelArguments, thus cannot handle in__post_init__(). )
Bug fixes
-   model.generate()with dsz3 [BUG] The text cannot be generated successfully during the Raft step #861
-  merge_loralora with abs path merging
-  load_datasetlong data fix [Bug Fix] update load_dataset to support long data #878
-  src/lmflow/utils/common.py create_copied_dataclasscompatibility when python version >= 3.10 (kw_onlyissue) [BUG]TypeError: Field.__init__() missing 1 required positional argument: 'kw_only' #903 [usability] deps streamlining #905
Issues left over from history
-  use_accelerator->use_acceleratetypo fix (with Accelerate support PR) [usability] Accelerate Support #936
-   model_args.use_loraleads to truncation of the sequence, mentioned in [Feature] reward model inferencer and dpov2 aligner #867
- Make ports, addresses, and all other settings in distributed training tidy and clear (with Accelerate support PR)
Documentation
- Approx GPU memory requirement w.r.t model size & pipeline
- Dev handbook, indicating styles, test list, etc.
research4pan and shizhediao
Metadata
Metadata
Assignees
Labels
No labels