Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up
Appearance settings

0.5.dev pr #84

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
chencyudel merged 15 commits into main from 0.5.dev
Nov 13, 2024
Merged
Show file tree
Hide file tree
Changes from 1 commit
Commits
Show all changes
15 commits
Select commit Hold shift + click to select a range
e827775
0.5.dev, first refactor, add offline_tokneizion for pretraining, add ...
chencyudel Oct 28, 2024
7898fa6
remove default writer
chencyudel Oct 28, 2024
a68a716
remove default writer
chencyudel Oct 28, 2024
08ab5ee
update coba loss
cocoocoder Oct 29, 2024
65f4511
update
cocoocoder Oct 29, 2024
a4cd0b9
Merge pull request #81 from codefuse-ai/support_coba_loss
chencyudel Oct 29, 2024
912e792
update the tutorial of CoBa
cocoocoder Oct 29, 2024
1a10e49
project readme
chencyudel Oct 30, 2024
1dc28f5
Merge pull request #82 from codefuse-ai/support_coba_loss
chencyudel Oct 30, 2024
96aaa1e
update tutorial of CoBa arguments
cocoocoder Oct 30, 2024
b658546
Merge pull request #83 from codefuse-ai/support_coba_loss
GoneZ5 Oct 30, 2024
46daf5d
mannual readme
chencyudel Oct 30, 2024
6de58d3
readme
chencyudel Oct 30, 2024
2cd7712
readme
chencyudel Oct 30, 2024
ae61b4d
readme
chencyudel Oct 31, 2024
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Prev Previous commit
Next Next commit
project readme
  • Loading branch information
chencyudel committed Oct 30, 2024
commit 1a10e4902c33e5da3d16c5af55fbfcb96126ef6c
4 changes: 3 additions & 1 deletion README.md
View file Open in desktop
Original file line number Diff line number Diff line change
Expand Up @@ -46,7 +46,9 @@


## News
🔥🔥🔥 [2024年10月29日] Our paper [CoBa: Convergence Balancer for Multitask Finetuning of Large Language Models](https://arxiv.org/abs/2410.06741) has been accepted by EMNLP-2024, which achieves balanced convergence across various tasks.
🔥🔥🔥 [2024年11月01日] We released **MFTCoder v0.5** mainly for MFTCoder-accelerate, which is now supporting preference alignment methods like **DPO/RPO/ORPO** in the new **xxpo** module, adding full-parameter continue-training in the additional **mpt** module along with its **offline_tokenization** module, updating selfpaced method to new convergence balance(CoBa) method for MFT in the original **pefts** module.

🔥🔥🔥 [2024年11月01日] Our paper [CoBa: Convergence Balancer for Multitask Finetuning of Large Language Models](https://arxiv.org/abs/2410.06741) has been accepted by EMNLP-2024, which achieves balanced convergence across various tasks.

🔥🔥🔥 [2024年05月20日] We released **MFTCoder v0.4**, mainly for MFTCoder-accelerate. It supports **QLoRA + DeepSpeed Zero3** and **QLoRA + FSDP** as options allowing you training very large models. It now supports new models like Qwen2, Qwen2-MoE, Starcoder2, Gemma, etc.

Expand Down
4 changes: 3 additions & 1 deletion README_cn.md
View file Open in desktop
Original file line number Diff line number Diff line change
Expand Up @@ -45,7 +45,9 @@


## 新闻
🔥🔥🔥 [2024年10月29日] 我们的论文 [CoBa: Convergence Balancer for Multitask Finetuning of Large Language Models](https://arxiv.org/abs/2410.06741) 已被 EMNLP 2024 接收,可以实现多任务收敛均衡。
🔥🔥🔥 [2024年11月01日] **MFTCoder-v0.5**发布,新增**xxpo**模块支持偏好对齐DPO/RPO/ORPO;新增**mpt**和**offline_tokenization**模块支持全量参数的加训;在原本的**pefts**模块(MFT)更新selfpaced收敛均衡技术并更名CoBa。

🔥🔥🔥 [2024年11月01日] 我们的论文 [CoBa: Convergence Balancer for Multitask Finetuning of Large Language Models](https://arxiv.org/abs/2410.06741) 已被 EMNLP 2024 接收,可以实现多任务收敛均衡。

🔥🔥🔥 [2024年05月20日] **MFTCoder-v0.4**发布。新增支持**QLoRA+ DeepSpeed Zero3**, **QLoRA + FSDP**训练模式,可以更好的支持微调更大的模型,比如Qwen1.5-70B等。新增对Qwen2, Qwen2-MoE, Starcoder2, Gemma等模型的支持。

Expand Down
4 changes: 2 additions & 2 deletions requirements.txt
View file Open in desktop
Original file line number Diff line number Diff line change
Expand Up @@ -3,8 +3,8 @@ pandas==2.2.1
torch==2.1.0
tensorboard==2.11.0
deepspeed==0.14.0
transformers==4.40.2
accelerate==0.28.0
transformers==4.44.2
accelerate==0.31.0
peft==0.10.0
BitsAndBytes==0.43.0
xformers==0.0.22.post7
Expand Down

AltStyle によって変換されたページ (->オリジナル) /