Paddlenlp

Latest version: v2.8.1

Safety actively analyzes 688600 Python packages for vulnerabilities to keep your Python projects secure.

Scan your dependencies

Page 1 of 9

3.0

* finetune support continue_training by tianhaodongbd in https://github.com/PaddlePaddle/PaddleNLP/pull/8615
* [PaddleNLP 3.0] Refactor/3 part1- remove fast tokenizer. by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/8613
* Repo adjustment by wtmlon in https://github.com/PaddlePaddle/PaddleNLP/pull/8605
* [PaddleNLP 3.0] Refactor, merge examples/language_model model_zoo to legacy/model_zoo by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/8614
* [PaddleNLP 3.0] Refactor RLHF by gongel in https://github.com/PaddlePaddle/PaddleNLP/pull/8617
* Remove delay_scale_loss and release_grads for llama-2 13B's benchmark. by Xreki in https://github.com/PaddlePaddle/PaddleNLP/pull/8623
* [PaddleNLP 3.0] Fix dead link by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/8626
* Update PaddleNLP to fix PPO by sneaxiy in https://github.com/PaddlePaddle/PaddleNLP/pull/8618
* [LLM] support sparse attention for LLAMA by GuoxiaWang in https://github.com/PaddlePaddle/PaddleNLP/pull/8592
* remove fast generation by wtmlon in https://github.com/PaddlePaddle/PaddleNLP/pull/8625
* fix npu llama by zhink in https://github.com/PaddlePaddle/PaddleNLP/pull/8628
* [PaddleNLP 3.0] Refactor/3 part3, move pipelines. by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/8619
* [PaddleNLP 3.0] update dataset preprocess by DrownFish19 in https://github.com/PaddlePaddle/PaddleNLP/pull/8629
* [LLM] Support prefix tuning and lora for qwen2 by DrownFish19 in https://github.com/PaddlePaddle/PaddleNLP/pull/8601
* modify path of model_zoo in ci_case_auto.sh and ci_case_dy.sh by jeff41404 in https://github.com/PaddlePaddle/PaddleNLP/pull/8633
* 【benchmark】 fix model_zoo path by mmglove in https://github.com/PaddlePaddle/PaddleNLP/pull/8643
* [PaddleNLP 3.0] [LLM] change llm content by lugimzzz in https://github.com/PaddlePaddle/PaddleNLP/pull/8627
* [LLM] Add sequence_parallel support for qwen by Difers in https://github.com/PaddlePaddle/PaddleNLP/pull/8558
* [NPU][LLM] add README & reformat llama scripts by SylarTiaNII in https://github.com/PaddlePaddle/PaddleNLP/pull/8642
* align llama auto_parallel dataloader with manual_parallel by zhiqiu in https://github.com/PaddlePaddle/PaddleNLP/pull/8639
* fix fast_ln compile error by deepllz in https://github.com/PaddlePaddle/PaddleNLP/pull/8650
* Apache License by DrownFish19 in https://github.com/PaddlePaddle/PaddleNLP/pull/8658
* Fix different length for numpy>=1.24.x by DrownFish19 in https://github.com/PaddlePaddle/PaddleNLP/pull/8655
* [LLM][NPU] fix on readme by SylarTiaNII in https://github.com/PaddlePaddle/PaddleNLP/pull/8659
* [DOC] Fix dead link by DrownFish19 in https://github.com/PaddlePaddle/PaddleNLP/pull/8662
* fix benchmark dir because of PR8627 by fightfat in https://github.com/PaddlePaddle/PaddleNLP/pull/8649
* fix llama alibi pretrain by lugimzzz in https://github.com/PaddlePaddle/PaddleNLP/pull/8668
* inference support llama3(wint8|4/a8w8) by yuanlehome in https://github.com/PaddlePaddle/PaddleNLP/pull/8630
* 【benchmark】 fix benchmark script by mmglove in https://github.com/PaddlePaddle/PaddleNLP/pull/8648
* [cpu]llama avx model inference supports by bukejiyu in https://github.com/PaddlePaddle/PaddleNLP/pull/8634
* 【AutoParallel】Change benchmark config for llama2-7b by heavyrain-lzy in https://github.com/PaddlePaddle/PaddleNLP/pull/8667
* support flashmask by lugimzzz in https://github.com/PaddlePaddle/PaddleNLP/pull/8670
* [PaddleNLP 3.0] Update README.md by DrownFish19 in https://github.com/PaddlePaddle/PaddleNLP/pull/8666
* adjust llm readme by lugimzzz in https://github.com/PaddlePaddle/PaddleNLP/pull/8672
* Update export model by DesmonDay in https://github.com/PaddlePaddle/PaddleNLP/pull/8671
* Update version by gongel in https://github.com/PaddlePaddle/PaddleNLP/pull/8675
* Sft flash mask by wtmlon in https://github.com/PaddlePaddle/PaddleNLP/pull/8664
* Update version by gongel in https://github.com/PaddlePaddle/PaddleNLP/pull/8676

New Contributors
* Southpika made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/8082
* cxa-unique made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/8331
* dynamicheart made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/8282
* EnflameGCU made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/8445
* cqulilujia made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/8459
* yinfan98 made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/8440
* zhangyuqin1998 made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/8396
* ming1753 made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/8456
* asr-sheep1 made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/8472
* NeroLoh made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/8515
* bukejiyu made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/8634

**Full Changelog**: https://github.com/PaddlePaddle/PaddleNLP/compare/v2.8.1...v3.0.0-beta0

3.0.0beta2

本次更新强化了PaddleNLP的基础设施,新增了Qwen2.5、Mixtral 8*22B模型并升级了Tokenizer功能,同时重命名了数据索引工具。

此外,还修复了MoE模型参数保存与加载等问题,提升了文本处理准确性,并更新了文档与测试用例。在推理性能、硬件支持及自动并行方面也进行了优化,包括支持更多模型与参数配置、多GPU推理、国产硬件支持增强以及分布式训练流程优化等。


核心变更与增强功能

1. **基础设施强化**:
- 新增Qwen2.5模型(9157 ),Mixtral 8*22B。进一步丰富模型库。
- Tokenizer功能升级,现支持加载额外解码标记added_tokens_decoder(8997 ),提升灵活性。
- 数据索引工具`tool_helpers`重命名为`fast_dataindex`(9134 ),以更直观反映其功能特性。
- 实现训练过程中数据间隔跳过的功能(8989 ),优化数据处理效率。
- **Unified Checkpoint优化**:
- 更新优化器异步保存信号(8975 ),保证保存稳定。
- 修复统一检查点中的多项问题(9082 ),确保功能正确性。

3. **问题修复**:
- 解决了MoE模型参数保存与加载的问题(9045 )。
- 修正Tokenizer中空格与特殊符号处理的不足(9010 , 9144 ),提升文本处理准确性。

4. **文档与测试更新**:
- 更新多个文档,涵盖LLM模型文档(如8990 , 8999 )及量化文档(9057 )等,确保信息的时效性与准确性。
- 新增测试用例,如针对PIR模式序列并行的测试(9015 ),强化测试覆盖度。
- 修复文档中的链接错误(如9127 ),提升用户体验。

5. **其他关键变更**:
- **推理性能优化**:
- LLM推理代码得到优化,支持更多模型与参数配置(如8986 , 8995 ),拓宽应用场景。
- 实现Qwen2_Moe多GPU推理(9121 )及wint4量化(9129 ),提升推理效率。
- 加强LLM推理对FP8与INT8的支持(如9032 , 9151 ),满足多样化精度需求。
- **硬件支持拓展**:
- 增强对DCU、XPU、MLU等国产硬件的支持(如8983 , 8504 , 9075 ),促进国产化替代。
- 优化上述硬件上的模型训练与推理性能,提升整体运算效率。
- **自动并行优化**:
- 修复训练过程中数据重复跳过的问题(8980 ),确保数据处理的正确性。
- 更新自动并行配置与检查点转换器(如8847 , 9136 ),提升并行训练的灵活性与稳定性。
- 新增损失NaN/Inf检查器(8943 ),及时发现并处理潜在数值问题。
- 优化分布式训练中的数据加载与梯度合并流程(如9120 , 9179 ),提升训练速度与稳定性。


What's Changed
* [Unified checkpoint] update optimizer async save signal by DesmonDay in https://github.com/PaddlePaddle/PaddleNLP/pull/8975
* 更正run_dpo.py文件路径 by Mangodadada in https://github.com/PaddlePaddle/PaddleNLP/pull/8952
* fix the loss base in llama_align_dygraph_dy2st_auto_bs2_bf16_DP2-MP1-… by winter-wang in https://github.com/PaddlePaddle/PaddleNLP/pull/8986
* [Bug fix] fix skip consumed_samples twice bug by zhangyuqin1998 in https://github.com/PaddlePaddle/PaddleNLP/pull/8980
* fix pip error in legacy benchmarks by fightfat in https://github.com/PaddlePaddle/PaddleNLP/pull/8978
* 【auto_parallel】Add checkpoint convertor by xingmingyyj in https://github.com/PaddlePaddle/PaddleNLP/pull/8847
* [llm]update finetune.md by lugimzzz in https://github.com/PaddlePaddle/PaddleNLP/pull/8990
* tool_helpers升级后可以支持32766个数据集. by JunnYu in https://github.com/PaddlePaddle/PaddleNLP/pull/8994
* add DCU inference docs by YanhuiDua in https://github.com/PaddlePaddle/PaddleNLP/pull/8983
* [Distributed]Add loss nan/inf checker by ForFishes in https://github.com/PaddlePaddle/PaddleNLP/pull/8943
* 【llm】update docs by lugimzzz in https://github.com/PaddlePaddle/PaddleNLP/pull/8999
* [Feature] Fused Mixtral support by penPenf28 in https://github.com/PaddlePaddle/PaddleNLP/pull/8901
* [XPU] Add README.md for llama2-7b by xiguapipi in https://github.com/PaddlePaddle/PaddleNLP/pull/8979
* Add gcu llama readme by EnflameGCU in https://github.com/PaddlePaddle/PaddleNLP/pull/8950
* fix qwen model use_casual_mask by deepllz in https://github.com/PaddlePaddle/PaddleNLP/pull/9009
* [ZeroPadding] revert zero_padding 8973 by DrownFish19 in https://github.com/PaddlePaddle/PaddleNLP/pull/9003
* [LLM Inference] Fix step.cu bug by yuanlehome in https://github.com/PaddlePaddle/PaddleNLP/pull/8995
* Refine checkpoint converter by zhangbo9674 in https://github.com/PaddlePaddle/PaddleNLP/pull/9001
* [Feature] fused mixtral wint4 by penPenf28 in https://github.com/PaddlePaddle/PaddleNLP/pull/9013
* llm inference docs by Sunny-bot1 in https://github.com/PaddlePaddle/PaddleNLP/pull/8976
* [LLM Inference] Support Qwen2_Moe Inference Model by CJ77Qi in https://github.com/PaddlePaddle/PaddleNLP/pull/8892
* fix llama3 static run by yuanlehome in https://github.com/PaddlePaddle/PaddleNLP/pull/8849
* [paddle inference cpu]update cpu inference by bukejiyu in https://github.com/PaddlePaddle/PaddleNLP/pull/8984
* fix the tipc ce case by wawltor in https://github.com/PaddlePaddle/PaddleNLP/pull/8748
* [Cherry-pick] Add is_distributed field in sharding reshard param_meta by sneaxiy in https://github.com/PaddlePaddle/PaddleNLP/pull/9028
* [Tokenizer] Support for loading added_tokens_decoder by DrownFish19 in https://github.com/PaddlePaddle/PaddleNLP/pull/8997
* [Inference] Add a8w8(fp8) a8w8c8(int8) quant_type support by lixcli in https://github.com/PaddlePaddle/PaddleNLP/pull/9032
* Fix checker of nan/inf by ForFishes in https://github.com/PaddlePaddle/PaddleNLP/pull/9029
* [Cherry-pick] add comm buffer size (8963) by ForFishes in https://github.com/PaddlePaddle/PaddleNLP/pull/9031
* [Unified Checkpoint] Update async save info by DesmonDay in https://github.com/PaddlePaddle/PaddleNLP/pull/8982
* [llm]support pad to max_length & fix sp bug by lugimzzz in https://github.com/PaddlePaddle/PaddleNLP/pull/9040
* [Bugfix] fix bias optional by penPenf28 in https://github.com/PaddlePaddle/PaddleNLP/pull/9037
* fix setup.py for llm inference by yuanlehome in https://github.com/PaddlePaddle/PaddleNLP/pull/9041
* [Inference] Add cutlass gemm dequant op by gzy19990617 in https://github.com/PaddlePaddle/PaddleNLP/pull/8909
* [Inference] update fakequant support by lixcli in https://github.com/PaddlePaddle/PaddleNLP/pull/9047
* add test for pir sequence parallel on llama model by liym27 in https://github.com/PaddlePaddle/PaddleNLP/pull/9015
* Fix moe save load by Meiyim in https://github.com/PaddlePaddle/PaddleNLP/pull/9045
* Update quantization.md by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/9057
* 【Fix】Initialize dp degree in single GPU by greycooker in https://github.com/PaddlePaddle/PaddleNLP/pull/9056
* fix bos download by westfish in https://github.com/PaddlePaddle/PaddleNLP/pull/9023
* [Inference] Update fakequant script by lixcli in https://github.com/PaddlePaddle/PaddleNLP/pull/9054
* [AutoParallel][PIR] Fit pir grad merge by AndSonder in https://github.com/PaddlePaddle/PaddleNLP/pull/8985
* [MLU] Support rms_norm_mlu by PeiyuLau in https://github.com/PaddlePaddle/PaddleNLP/pull/8504
* [Inference] support llama3 a8w8c8_fp8 inference and cutlass_fp8_gemm by ckl117 in https://github.com/PaddlePaddle/PaddleNLP/pull/8953
* [Inference] Qwen2 support fp8 inference by ckl117 in https://github.com/PaddlePaddle/PaddleNLP/pull/8954
* [Version] update version info by DrownFish19 in https://github.com/PaddlePaddle/PaddleNLP/pull/9060
* [NPU] Fix baichuan2-13b-chat infer by ronny1996 in https://github.com/PaddlePaddle/PaddleNLP/pull/9070
* [MLU] Fix Llama attrntion_mask in npu and mlu by DrownFish19 in https://github.com/PaddlePaddle/PaddleNLP/pull/9075
* Fix the memory overflow bug of the tune_cublaslt_gemm operator by Hanyonggong in https://github.com/PaddlePaddle/PaddleNLP/pull/9076
* [Inference] Fix weight_only_int4 bug by lixcli in https://github.com/PaddlePaddle/PaddleNLP/pull/9073
* [Auto Parallel] fix data stream bug of dist.to_static by zhangyuqin1998 in https://github.com/PaddlePaddle/PaddleNLP/pull/9077
* fix hang when Flag_dataloader_use_file_descriptor=True by deepllz in https://github.com/PaddlePaddle/PaddleNLP/pull/9080
* fix llm predict install error by fightfat in https://github.com/PaddlePaddle/PaddleNLP/pull/9088
* [PIR] add pir grad merge test by AndSonder in https://github.com/PaddlePaddle/PaddleNLP/pull/9074
* Update readme by EnflameGCU in https://github.com/PaddlePaddle/PaddleNLP/pull/9046
* [LLM] Add tensor parallel for chatglmv2 by SevenSamon in https://github.com/PaddlePaddle/PaddleNLP/pull/9014
* [data] update tool_helpers version and add unittest by JunnYu in https://github.com/PaddlePaddle/PaddleNLP/pull/9093
* fix baseline because of PR8769 by fightfat in https://github.com/PaddlePaddle/PaddleNLP/pull/9092
* fix use paddle.incubate.jit.inference(model) errors by chang-wenbin in https://github.com/PaddlePaddle/PaddleNLP/pull/9016
* [CI] Fix paddlepaddle install by DesmonDay in https://github.com/PaddlePaddle/PaddleNLP/pull/9102
* [LLM] fix train on npu by SylarTiaNII in https://github.com/PaddlePaddle/PaddleNLP/pull/9101
* Disable ut by zhangbo9674 in https://github.com/PaddlePaddle/PaddleNLP/pull/9108
* [AutoParallel] Enable CI for gradclip by JZ-LIANG in https://github.com/PaddlePaddle/PaddleNLP/pull/9059
* [Inference] Remove ceval from run_finetune by lixcli in https://github.com/PaddlePaddle/PaddleNLP/pull/9100
* [Bugfix] fix multi-gpu infer by penPenf28 in https://github.com/PaddlePaddle/PaddleNLP/pull/9107
* 【Inference】fix step kernel by gzy19990617 in https://github.com/PaddlePaddle/PaddleNLP/pull/9122
* [DCU] fix DCU w8a8c8 GEMM shape by YanhuiDua in https://github.com/PaddlePaddle/PaddleNLP/pull/9115
* [Inference] FP8 gemm auto-tune by ckl117 in https://github.com/PaddlePaddle/PaddleNLP/pull/9094
* Open ut llama_align_dygraph_dy2st_pir_auto_grad_merge_bs2_fp32_DP1-MP1-PP1 by zhangbo9674 in https://github.com/PaddlePaddle/PaddleNLP/pull/9120
* [LLM Inference] Support Qwen2_Moe Inference with MultiGPU by CJ77Qi in https://github.com/PaddlePaddle/PaddleNLP/pull/9121
* [Unified Checkpoint] Fix uc lora config, fix release_grads by DesmonDay in https://github.com/PaddlePaddle/PaddleNLP/pull/9082
* [Inference]qwen2-a8w8c8 support use_fake_parameter by ckl117 in https://github.com/PaddlePaddle/PaddleNLP/pull/9109
* Add fast_ln spmd rules by From00 in https://github.com/PaddlePaddle/PaddleNLP/pull/9125
* fix pir dtype by wanghuancoder in https://github.com/PaddlePaddle/PaddleNLP/pull/9130
* Remove ring_flash_attention warning by DrownFish19 in https://github.com/PaddlePaddle/PaddleNLP/pull/9119
* [DOC] Fix LLM page 404 Not Found by DrRyanHuang in https://github.com/PaddlePaddle/PaddleNLP/pull/9127
* Add hardware flops for pretraining by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/9069
* [Benchmark] Fix amp level bug in some gpt tests by zhangbo9674 in https://github.com/PaddlePaddle/PaddleNLP/pull/9116
* [Auto Parallel] Fix ckpt_converter for auto_parallel by zhangyuqin1998 in https://github.com/PaddlePaddle/PaddleNLP/pull/9136
* [Inference] Update fakequant by lixcli in https://github.com/PaddlePaddle/PaddleNLP/pull/9140
* [DOC] Update docs by DrownFish19 in https://github.com/PaddlePaddle/PaddleNLP/pull/9141
* [LLM Inference] Qwen2_Moe Support wint4 by CJ77Qi in https://github.com/PaddlePaddle/PaddleNLP/pull/9129
* add multy devices supported models by a31413510 in https://github.com/PaddlePaddle/PaddleNLP/pull/9079
* [fix] freeze 参数冗余存储 兼容shard-reshard (9067) by bo-ke in https://github.com/PaddlePaddle/PaddleNLP/pull/9148
* [Docs] Update LLM docs by DrownFish19 in https://github.com/PaddlePaddle/PaddleNLP/pull/9143
* fix llm ce predict run error by fightfat in https://github.com/PaddlePaddle/PaddleNLP/pull/9149
* [Tokenizer] Add replace_additional_special_tokens parameter to add_special_tokens by lvdongyi in https://github.com/PaddlePaddle/PaddleNLP/pull/9144
* [Tokenizer] Fix decode output with space in decode_token by DrownFish19 in https://github.com/PaddlePaddle/PaddleNLP/pull/9010
* 【Inference】Optimize top_p kernel performance by gzy19990617 in https://github.com/PaddlePaddle/PaddleNLP/pull/9132
* [Models] Add Qwen2.5 by DrownFish19 in https://github.com/PaddlePaddle/PaddleNLP/pull/9157
* Update README.md by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/9160
* [Inference] FP8 dual gemm auto-tune and support compile parallelization by ckl117 in https://github.com/PaddlePaddle/PaddleNLP/pull/9151
* [AutoParallel] enable ci for dp amp clip by JZ-LIANG in https://github.com/PaddlePaddle/PaddleNLP/pull/9062
* [llm]support dpo pp by lugimzzz in https://github.com/PaddlePaddle/PaddleNLP/pull/9039
* [Tools] Rename tool_helpers to fast_dataindex. by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/9134
* [Trainer] Support skip data intervals by greycooker in https://github.com/PaddlePaddle/PaddleNLP/pull/8989
* remove run_pretrain_auto_static.py CI when open PIR by fightfat in https://github.com/PaddlePaddle/PaddleNLP/pull/9177
* [Tokenizer] Enable padding_side as call time kwargs by lvdongyi in https://github.com/PaddlePaddle/PaddleNLP/pull/9161
* Revert "[Tokenizer] Enable padding_side as call time kwargs" by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/9192
* [XPU] add xpu support for llama sft by tizhou86 in https://github.com/PaddlePaddle/PaddleNLP/pull/9152
* [AutoParallel] Add FLAGS_enable_fused_ffn_qkv_pass for llama by zhangbo9674 in https://github.com/PaddlePaddle/PaddleNLP/pull/9182
* [AutoParallel] Fix ckpt convert bug for sharding v2 by zhangbo9674 in https://github.com/PaddlePaddle/PaddleNLP/pull/9179
* [Test] Disable dynamic to static test case for paddle PIR by DrownFish19 in https://github.com/PaddlePaddle/PaddleNLP/pull/9196
* Fix ppt eval hang by gongel in https://github.com/PaddlePaddle/PaddleNLP/pull/9218
* Update branch version to 3.0.0b2 by gongel in https://github.com/PaddlePaddle/PaddleNLP/pull/9220
* Update branch version to 3.0.0b2 by gongel in https://github.com/PaddlePaddle/PaddleNLP/pull/9221
* Revert "Fix ppt eval hang" by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/9229

New Contributors
* Mangodadada made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/8952
* xingmingyyj made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/8847
* penPenf28 made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/8901
* xiguapipi made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/8979
* Sunny-bot1 made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/8976
* CJ77Qi made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/8892
* lixcli made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/9032
* gzy19990617 made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/8909
* SevenSamon made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/9014
* chang-wenbin made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/9016
* DrRyanHuang made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/9127
* a31413510 made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/9079
* lvdongyi made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/9144
* tizhou86 made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/9152

**Full Changelog**: https://github.com/PaddlePaddle/PaddleNLP/compare/v3.0.0-beta1...v3.0.0-beta2

3.0.0beta1

PaddleNLP从v3.0.0-beta0升级至v3.0.0-beta1版本,带来了多项重要更新与增强。新引入了Yuan、mamba和jamba模型,并优化了LLM推理代码,提升了兼容性和效率。

基础性能优化方面,添加了快速分词器,实现了MoE优化器参数广播,加速了层归一化。同时,修复了多个bug,包括safetensors shape切片问题和Windows下mmap问题,提升了系统稳定性和兼容性。

文档与测试方面,进行了全面更新和优化,确保了文档的准确性和代码的可读性。此外,还增强了国产硬件支持,包括DCU和XPU的优化,以及PIR模式和自动并行的配置更新。


主要变更与新增功能

1. 新模型与特性引入
- **新模型**:在8654 中引入了Yuan模型;在8513 和8517 中分别添加了mamba和jamba新模型,并在后续Pull Request中修复了相关bug,确保了模型的稳定运行。
- **LLM推理优化**:通过多个Pull Request,我们优化了LLM推理代码,并新增了对新模型和参数的支持,进一步提升了推理效率和兼容性。

2. 基础性能优化
- **快速分词器**:在8832 中,我们添加了基于`tokenizers`库的快速分词器,显著提升了分词速度和性能。
- **MoE优化**:在8810 中,我们实现了MoE(Mixture of Experts)优化器参数的广播,有效增强了模型训练的效率。
- **层归一化加速**:通过多个Pull Request,我们添加了fast_rmsnorm,启用了use_fast_layer_norm,并更新了基准测试配置,进一步加速了模型训练过程。特别是在8717 中,我们支持了在微调过程中使用use_fast_layer_norm,为用户提供了更多灵活性。
- **训练性能优化**:在8803 中,我们添加了`enable_sp_async_reduce_scatter`选项,有效优化了训练性能。
- **字典参数支持**:在8446 中,我们为trainer的argparser添加了支持字典参数的新特性,增强了参数传递的灵活性。同时,在8904 中,我们更新了tensorboard的要求,确保了与最新版本的兼容性。

3. Bug修复
- **safetensors修复**:在8702 中,我们修复了safetensors的形状问题。
- **Windows系统mmap修复**:在8734 中修复了mmap问题,提升了windows的兼容性。
- **其他Bug修复**:包括8687 、8730 等多个Pull Request中的bug修复。

4. 文档与测试更新
- **文档优化**:在多个Pull Request中,我们进行了文档更新、代码风格清理和版本信息更新,确保了文档的准确性和可读性。
- **README修复与增强**:在8741 中,我们修复了README中的断链问题;同时,多个贡献者更新了README文档,添加了新的测试用例,确保了文档与代码的同步更新。

5. 其他重要变更

国产硬件支持增强
- **DCU支持**:在8580 中,我们实现了针对DCU的高性能LLM训练和推理,拓展了PaddleNLP的硬件支持范围。
- **XPU优化**:在8527 中,我们为XPU添加了LoRA优化;在8697 和8710 中,我们分别实现了XPU的allgather功能和修复了统一检查点的gather问题,进一步提升了XPU上的模型训练效率。

PIR模式支持
- **导出与加载优化**:在8689 中,我们修改了PIR模式下llama模型的导出方式;在8712 和8766 中,我们支持了以三种模式(旧IR、PIR模型文件、PIR JSON文件)加载或保存Llama2-7b模型,为用户提供了更多灵活性和兼容性。

自动并行优化
- **配置更新**:在8679 中,我们更改了Llama2-7b配置中的`max_steps`以适应自动并行;在8767 和8828 中,我们优化了自动训练器的保存和加载功能;在8750 中,我们更新了全局剪切的损失函数,进一步提升了自动并行的效率和准确性。


What's Changed
* [DCU] high performance LLM train and inference for DCU by yuguo-Jack in https://github.com/PaddlePaddle/PaddleNLP/pull/8580
* fix benchmark dir and add CUDA_DEVICE_MAX_CONNECTIONS to qwen by fightfat in https://github.com/PaddlePaddle/PaddleNLP/pull/8678
* bug fix by wtmlon in https://github.com/PaddlePaddle/PaddleNLP/pull/8687
* [XPU] add lora optimization by dynamicheart in https://github.com/PaddlePaddle/PaddleNLP/pull/8527
* [pir save] Modiy export llama model file in pir mode by xiaoguoguo626807 in https://github.com/PaddlePaddle/PaddleNLP/pull/8689
* [AutoParallel]Change `max_steps` in Llama2-7b config for auto-parallel. by heavyrain-lzy in https://github.com/PaddlePaddle/PaddleNLP/pull/8679
* [benchmark] Change the mirror source for pip by mmglove in https://github.com/PaddlePaddle/PaddleNLP/pull/8699
* update loss base of auto-parallel tests by zhiqiu in https://github.com/PaddlePaddle/PaddleNLP/pull/8701
* Add new mistral by wtmlon in https://github.com/PaddlePaddle/PaddleNLP/pull/7425
* [Safetensors] Fix safetensors shape by DesmonDay in https://github.com/PaddlePaddle/PaddleNLP/pull/8702
* [BUG] num_samples 向下去整, 防止prefrech预取时候超过数据集最大长度... by JunnYu in https://github.com/PaddlePaddle/PaddleNLP/pull/8690
* xpu use allgather by FeixLiu in https://github.com/PaddlePaddle/PaddleNLP/pull/8697
* add fast_rmsnorm by deepllz in https://github.com/PaddlePaddle/PaddleNLP/pull/8680
* enable use_fast_layer_norm for llama2 benchmark by deepllz in https://github.com/PaddlePaddle/PaddleNLP/pull/8714
* fix xpu gather for unified ckpt by FeixLiu in https://github.com/PaddlePaddle/PaddleNLP/pull/8710
* [inference] support load or save Llama2-7b in three patterns by lizexu123 in https://github.com/PaddlePaddle/PaddleNLP/pull/8712
* fix fast_ln backward by deepllz in https://github.com/PaddlePaddle/PaddleNLP/pull/8719
* finetune support use_fast_layer_norm by tianhaodongbd in https://github.com/PaddlePaddle/PaddleNLP/pull/8717
* bug fix by FeixLiu in https://github.com/PaddlePaddle/PaddleNLP/pull/8730
* disable lora by lugimzzz in https://github.com/PaddlePaddle/PaddleNLP/pull/8674
* [Safetensors] Fix mmap for Windows system by DrownFish19 in https://github.com/PaddlePaddle/PaddleNLP/pull/8734
* correct broken links in readme by jzhang533 in https://github.com/PaddlePaddle/PaddleNLP/pull/8741
* revert benchmark fix by ronny1996 in https://github.com/PaddlePaddle/PaddleNLP/pull/8747
* [LLM] Add Yuan model by zhaogf01 in https://github.com/PaddlePaddle/PaddleNLP/pull/8654
* fix nlp dir and auto_parallel_ci exit -6 by fightfat in https://github.com/PaddlePaddle/PaddleNLP/pull/8744
* [LLM] Update sequence parallel linear import by DrownFish19 in https://github.com/PaddlePaddle/PaddleNLP/pull/8706
* [Bug fixes] Fix ring attention by zhangyuqin1998 in https://github.com/PaddlePaddle/PaddleNLP/pull/8740
* update a100 loss by zhiqiu in https://github.com/PaddlePaddle/PaddleNLP/pull/8708
* [PaddleNLP 3.0] Update README by DrownFish19 in https://github.com/PaddlePaddle/PaddleNLP/pull/8681
* [AutoParallel] update loss for global clip by JZ-LIANG in https://github.com/PaddlePaddle/PaddleNLP/pull/8750
* [NPU] Fix sequence parallel lib import by DrownFish19 in https://github.com/PaddlePaddle/PaddleNLP/pull/8760
* [DEV] Update develop version show by DrownFish19 in https://github.com/PaddlePaddle/PaddleNLP/pull/8754
* [inference] support load or save Llama2-7b in three patterns by lizexu123 in https://github.com/PaddlePaddle/PaddleNLP/pull/8766
* add benchmark baichuan2 scripts by fightfat in https://github.com/PaddlePaddle/PaddleNLP/pull/8683
* Add the missing truncation=True in llm/predictor.py by lszxb in https://github.com/PaddlePaddle/PaddleNLP/pull/8768
* fix the ce for the unittest by wawltor in https://github.com/PaddlePaddle/PaddleNLP/pull/8772
* Enable parallel_config to use commas as delimiters. by Difers in https://github.com/PaddlePaddle/PaddleNLP/pull/8677
* fix incorrect token counting in `llm/predictor.py` by lszxb in https://github.com/PaddlePaddle/PaddleNLP/pull/8769
* Refine savable by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/8758
* [CodeStyle] remove markdownlint-cli by DrownFish19 in https://github.com/PaddlePaddle/PaddleNLP/pull/8779
* [XPU] use allgather and fp32 multinomial for XPU by houj04 in https://github.com/PaddlePaddle/PaddleNLP/pull/8787
* fix version show by DrownFish19 in https://github.com/PaddlePaddle/PaddleNLP/pull/8791
* [BUG] Add 20 redundant data in post pretrain by JunnYu in https://github.com/PaddlePaddle/PaddleNLP/pull/8789
* vera-pissa method added by TranscenderNing in https://github.com/PaddlePaddle/PaddleNLP/pull/8722
* update version by DrownFish19 in https://github.com/PaddlePaddle/PaddleNLP/pull/8792
* [Inference LLM] refine some code in llama wint8/4 by yuanlehome in https://github.com/PaddlePaddle/PaddleNLP/pull/8796
* [DCU] Llama a8w8 inference performance optimization by Deleter-D in https://github.com/PaddlePaddle/PaddleNLP/pull/8800
* [Prediction] Update LLM prediction. by DesmonDay in https://github.com/PaddlePaddle/PaddleNLP/pull/8778
* [Trainer] Add enable_sp_async_reduce_scatter by DesmonDay in https://github.com/PaddlePaddle/PaddleNLP/pull/8803
* [AutoParallel] Refine auto_trainer save load by zhangbo9674 in https://github.com/PaddlePaddle/PaddleNLP/pull/8767
* [MoE] Optimizer parameter broadcast by DesmonDay in https://github.com/PaddlePaddle/PaddleNLP/pull/8810
* [Doc] Update README by DrownFish19 in https://github.com/PaddlePaddle/PaddleNLP/pull/8817
* support Llama3.1 8B 128K generation on single GPU 80GB by GuoxiaWang in https://github.com/PaddlePaddle/PaddleNLP/pull/8811
* add paddle nv-embed-v1 by Li-Z-Q in https://github.com/PaddlePaddle/PaddleNLP/pull/8785
* fix pad_token_id bug by yuanlehome in https://github.com/PaddlePaddle/PaddleNLP/pull/8814
* [DCU] fix llama inference bug on DCU by Deleter-D in https://github.com/PaddlePaddle/PaddleNLP/pull/8815
* [Doc] Add LLaMA3.1 by DrownFish19 in https://github.com/PaddlePaddle/PaddleNLP/pull/8824
* [BUG] Fix build train valid test datasets by JunnYu in https://github.com/PaddlePaddle/PaddleNLP/pull/8826
* Add tune_cublaslt_gemm operator by cublaslt gemm algorithm and generate algo cache file by Hanyonggong in https://github.com/PaddlePaddle/PaddleNLP/pull/8799
* fix tune_cublaslt_gemm compile bug by yuanlehome in https://github.com/PaddlePaddle/PaddleNLP/pull/8844
* [AutoParallel] Refine save and load ckpt for auto_trainer by zhangbo9674 in https://github.com/PaddlePaddle/PaddleNLP/pull/8828
* [Unified Checkpoint] update merge tensor parallel by DesmonDay in https://github.com/PaddlePaddle/PaddleNLP/pull/8856
* [Trainer] update clear_grad by DesmonDay in https://github.com/PaddlePaddle/PaddleNLP/pull/8829
* [Unified Checkpoint] Fix tie_word_embeddings by DesmonDay in https://github.com/PaddlePaddle/PaddleNLP/pull/8795
* [Inference LLM] support static c8 by yuanlehome in https://github.com/PaddlePaddle/PaddleNLP/pull/8833
* support sft mapdataset by greycooker in https://github.com/PaddlePaddle/PaddleNLP/pull/8840
* Cherry pick some changes from incubate branch by sneaxiy in https://github.com/PaddlePaddle/PaddleNLP/pull/8862
* support nested list of dict inputs by deepllz in https://github.com/PaddlePaddle/PaddleNLP/pull/8876
* Fix the bug with issues code 8641. by smallbenxiong in https://github.com/PaddlePaddle/PaddleNLP/pull/8880
* Fix the issue of P-tuning official sample error by guangyunms in https://github.com/PaddlePaddle/PaddleNLP/pull/8884
* modify Paddlemix qwen dytostatic by xiaoguoguo626807 in https://github.com/PaddlePaddle/PaddleNLP/pull/8869
* [llm]fix zeropadding by lugimzzz in https://github.com/PaddlePaddle/PaddleNLP/pull/8895
* 修复fast_ln算子动半开启后报错 by Wennie396 in https://github.com/PaddlePaddle/PaddleNLP/pull/8891
* enable_sp_async_reduce_scatter for qwen_72b && llama2_70b by deepllz in https://github.com/PaddlePaddle/PaddleNLP/pull/8897
* Update run_pretrain.py by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/8902
* [doc] Update readme by DrownFish19 in https://github.com/PaddlePaddle/PaddleNLP/pull/8905
* [AutoParallel] Bugfix auto parallel FA by JZ-LIANG in https://github.com/PaddlePaddle/PaddleNLP/pull/8903
* [Readme] Update README.md by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/8908
* [cherry-pick] Optimize async save by ForFishes in https://github.com/PaddlePaddle/PaddleNLP/pull/8878
* [LLM Inference] Refactor BlockInferencePredictor by yuanlehome in https://github.com/PaddlePaddle/PaddleNLP/pull/8879
* [Fix] modify tensorboard requirements by greycooker in https://github.com/PaddlePaddle/PaddleNLP/pull/8904
* [LLM Inference] Support qwen2 by yuanlehome in https://github.com/PaddlePaddle/PaddleNLP/pull/8893
* modify dict include none to aviod pir dytostatic bug in while op by xiaoguoguo626807 in https://github.com/PaddlePaddle/PaddleNLP/pull/8898
* [LLM]Update yuan model by zhaogf01 in https://github.com/PaddlePaddle/PaddleNLP/pull/8786
* update qwen && baichuan benchmark config by deepllz in https://github.com/PaddlePaddle/PaddleNLP/pull/8920
* [doc] Update README by DrownFish19 in https://github.com/PaddlePaddle/PaddleNLP/pull/8922
* [ New features]Trainer support dict parameter by greycooker in https://github.com/PaddlePaddle/PaddleNLP/pull/8446
* set logging_step to 5 with baichuan && qwen benchmark by deepllz in https://github.com/PaddlePaddle/PaddleNLP/pull/8928
* [Cherry-pick]fix pipeline eval by gongel in https://github.com/PaddlePaddle/PaddleNLP/pull/8924
* fix test_wint8 ut by yuanlehome in https://github.com/PaddlePaddle/PaddleNLP/pull/8930
* [LLM Inference] support llama3.1 by yuanlehome in https://github.com/PaddlePaddle/PaddleNLP/pull/8929
* Fix tokens count for benchmark by DrownFish19 in https://github.com/PaddlePaddle/PaddleNLP/pull/8938
* [bug fix] fix create_optimizer_and_scheduler for auto_parallel by zhangyuqin1998 in https://github.com/PaddlePaddle/PaddleNLP/pull/8937
* [LLM Inference] fix _get_tensor_parallel_mappings in llama by yuanlehome in https://github.com/PaddlePaddle/PaddleNLP/pull/8939
* [Unified Checkpoint] Fix load best checkpoint by DesmonDay in https://github.com/PaddlePaddle/PaddleNLP/pull/8935
* fix bug by yuanlehome in https://github.com/PaddlePaddle/PaddleNLP/pull/8947
* [LLM Inference] move llm.utils.utils.py to paddlenlp.utils.llm_utils.py by yuanlehome in https://github.com/PaddlePaddle/PaddleNLP/pull/8946
* support amp in pir dy2st mode. by winter-wang in https://github.com/PaddlePaddle/PaddleNLP/pull/8485
* [Trainer] Fix distributed dataloader by DesmonDay in https://github.com/PaddlePaddle/PaddleNLP/pull/8932
* [Tokenizer] Add Fast Tokenizer by DrownFish19 in https://github.com/PaddlePaddle/PaddleNLP/pull/8832
* [ZeroPadding] add greedy_zero_padding by DesmonDay in https://github.com/PaddlePaddle/PaddleNLP/pull/8933
* [NEW Model] Add mamba by JunnYu in https://github.com/PaddlePaddle/PaddleNLP/pull/8513
* [BUG] fix mamba tokenizer by JunnYu in https://github.com/PaddlePaddle/PaddleNLP/pull/8958
* [NEW Model] add jamba by JunnYu in https://github.com/PaddlePaddle/PaddleNLP/pull/8517
* [LLM Inference] add --use_fake_parameter option for ptq fake scales and fix compute error of total_max_length by yuanlehome in https://github.com/PaddlePaddle/PaddleNLP/pull/8955
* [LLM Inference] support qwen2 a8w8c8 inference by ckl117 in https://github.com/PaddlePaddle/PaddleNLP/pull/8925
* fix JambaModelIntegrationTest by JunnYu in https://github.com/PaddlePaddle/PaddleNLP/pull/8965
* [Fix] Enable tensor parallel tests. by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/8757
* [CI] Fix by DrownFish19 in https://github.com/PaddlePaddle/PaddleNLP/pull/8793
* [Unified Checkpoint] update async save by DesmonDay in https://github.com/PaddlePaddle/PaddleNLP/pull/8801
* [AutoParallel] Support save model for auto trainer by zhangbo9674 in https://github.com/PaddlePaddle/PaddleNLP/pull/8927
* fix qwen benchmark by deepllz in https://github.com/PaddlePaddle/PaddleNLP/pull/8969
* [ZeroPadding] padding to max_length for sequence parallel by DrownFish19 in https://github.com/PaddlePaddle/PaddleNLP/pull/8973
* add amp unit test case for auto_parallel ci. by winter-wang in https://github.com/PaddlePaddle/PaddleNLP/pull/8966
* [New Version] Upgrade to 3.0 b1 by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/8977

New Contributors
* yuguo-Jack made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/8580
* ruisunyc made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/8698
* xiaoguoguo626807 made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/8689
* lizexu123 made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/8712
* jzhang533 made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/8741
* zhaogf01 made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/8654
* lszxb made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/8768
* TranscenderNing made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/8722
* Deleter-D made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/8800
* Li-Z-Q made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/8785
* Hanyonggong made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/8799
* smallbenxiong made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/8880
* guangyunms made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/8884
* winter-wang made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/8485
* ckl117 made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/8925

**Full Changelog**: https://github.com/PaddlePaddle/PaddleNLP/compare/v3.0.0-beta0...v3.0.0-beta1

3.0.0beta0

很高兴地通知大家,飞桨大模型套件发布v3.0.0beat版本:拥抱大模型,体验全升级。具体工作如下:
* 统一大模型工具链,实现国产计算芯片全流程接入;
* 全面支持飞桨4D并行配置、高效精调策略、高效对齐算法、高性能推理等大模型产业级应用流程;
* 自研极致收敛的RsLoRA+算法、自动扩缩容存储机制Unified Checkpoint和通用化支持FastFFN、FusedQKV助力大模型训推;
* 主流模型持续支持更新,提供高效解决方案。

大模型精调对齐训推优化

* PEFT:
* 新增scaling策略,支持rslora, pissa算法 in https://github.com/PaddlePaddle/PaddleNLP/pull/8256
* 适配FusedQKV和FastFFN参数 in https://github.com/PaddlePaddle/PaddleNLP/pull/8372 https://github.com/PaddlePaddle/PaddleNLP/pull/8526
* DPO:
* 支持DPO(llama,qwen)in https://github.com/PaddlePaddle/PaddleNLP/pull/8474
* 支持序列并行 in https://github.com/PaddlePaddle/PaddleNLP/pull/7953
* 国产芯片支持:
* 适配NPU in https://github.com/PaddlePaddle/PaddleNLP/pull/8303 https://github.com/PaddlePaddle/PaddleNLP/pull/8342 https://github.com/PaddlePaddle/PaddleNLP/pull/8359 https://github.com/PaddlePaddle/PaddleNLP/pull/8399 https://github.com/PaddlePaddle/PaddleNLP/pull/8409 https://github.com/PaddlePaddle/PaddleNLP/pull/8401 https://github.com/PaddlePaddle/PaddleNLP/pull/8431 https://github.com/PaddlePaddle/PaddleNLP/pull/8439 https://github.com/PaddlePaddle/PaddleNLP/pull/8438 https://github.com/PaddlePaddle/PaddleNLP/pull/8442 https://github.com/PaddlePaddle/PaddleNLP/pull/8528 https://github.com/PaddlePaddle/PaddleNLP/pull/8642
* 适配XPU in https://github.com/PaddlePaddle/PaddleNLP/pull/8282 https://github.com/PaddlePaddle/PaddleNLP/pull/8505 https://github.com/PaddlePaddle/PaddleNLP/pull/8515 https://github.com/PaddlePaddle/PaddleNLP/pull/8588 https://github.com/PaddlePaddle/PaddleNLP/pull/8595 https://github.com/PaddlePaddle/PaddleNLP/pull/8598
* 适配GCU in https://github.com/PaddlePaddle/PaddleNLP/pull/8445 https://github.com/PaddlePaddle/PaddleNLP/pull/8470

* 性能优化:
* 优化Unified Checkpoint机制 in https://github.com/PaddlePaddle/PaddleNLP/pull/8204 https://github.com/PaddlePaddle/PaddleNLP/pull/8409 https://github.com/PaddlePaddle/PaddleNLP/pull/8422 https://github.com/PaddlePaddle/PaddleNLP/pull/8512
* 模型并行优化 in https://github.com/PaddlePaddle/PaddleNLP/pull/8370
* 序列并行优化 in https://github.com/PaddlePaddle/PaddleNLP/pull/8551
* 支持llama3 (wint8|4/a8w8) in https://github.com/PaddlePaddle/PaddleNLP/pull/8630

* 其他
* 新增模型内存监控 in https://github.com/PaddlePaddle/PaddleNLP/pull/8269

模型新增

* 新增Gemma模型 in https://github.com/PaddlePaddle/PaddleNLP/pull/8082
* google/gemma-7b
* google/gemma-7b-it
* google/gemma-2b
* google/gemma-2b-it

* 新增llama3模型 in https://github.com/PaddlePaddle/PaddleNLP/pull/8307 https://github.com/PaddlePaddle/PaddleNLP/pull/8371
* meta-llama/Meta-Llama-3-8B
* meta-llama/Meta-Llama-3-8B-Instruct
* meta-llama/Meta-Llama-3-70B
* meta-llama/Meta-Llama-3-70B-Instruct

* 新增Qwen2模型 in https://github.com/PaddlePaddle/PaddleNLP/pull/8338 https://github.com/PaddlePaddle/PaddleNLP/pull/8584 https://github.com/PaddlePaddle/PaddleNLP/pull/8601
* Qwen/Qwen1.5-0.5B
* Qwen/Qwen1.5-0.5B-Chat
* Qwen/Qwen1.5-1.8B
* Qwen/Qwen1.5-1.8B-Chat
* Qwen/Qwen1.5-4B
* Qwen/Qwen1.5-4B-Chat
* Qwen/Qwen1.5-7B
* Qwen/Qwen1.5-7B-Chat
* Qwen/Qwen1.5-14B
* Qwen/Qwen1.5-14B-Chat
* Qwen/Qwen1.5-32B
* Qwen/Qwen1.5-32B-Chat
* Qwen/Qwen1.5-72B
* Qwen/Qwen1.5-72B-Chat
* Qwen/Qwen1.5-110B
* Qwen/Qwen1.5-110B-Chat
* Qwen/Qwen1.5-MoE-A2.7B
* Qwen/Qwen1.5-MoE-A2.7B-Chat
* Qwen/Qwen2-0.5B
* Qwen/Qwen2-0.5B-Instruct
* Qwen/Qwen2-1.5B
* Qwen/Qwen2-1.5B-Instruct
* Qwen/Qwen2-7B
* Qwen/Qwen2-7B-Instruct
* Qwen/Qwen2-72B
* Qwen/Qwen2-72B-Instruct
* Qwen/Qwen2-57B-A14B
* Qwen/Qwen2-57B-A14B-Instruct

基础框架升级

* 功能优化:
* 支持FusedQKV和FastFFN权重自动融合分割 in https://github.com/PaddlePaddle/PaddleNLP/pull/8202 https://github.com/PaddlePaddle/PaddleNLP/pull/8378 https://github.com/PaddlePaddle/PaddleNLP/pull/8432
* 支持模型并行参数同步设置 in https://github.com/PaddlePaddle/PaddleNLP/pull/8311
* 支持RoPE算子设定theta in https://github.com/PaddlePaddle/PaddleNLP/pull/8440
* 通信overlap优化 in https://github.com/PaddlePaddle/PaddleNLP/pull/8276 https://github.com/PaddlePaddle/PaddleNLP/pull/8473 https://github.com/PaddlePaddle/PaddleNLP/pull/8499 https://github.com/PaddlePaddle/PaddleNLP/pull/8594

* AutoParallel优化
* llama支持recompute机制 in https://github.com/PaddlePaddle/PaddleNLP/pull/8265
* 适配llama3 in https://github.com/PaddlePaddle/PaddleNLP/pull/8395
* position_ids优化 in https://github.com/PaddlePaddle/PaddleNLP/pull/8363
* 支持流水线并行split_backward in https://github.com/PaddlePaddle/PaddleNLP/pull/8479
* 适配qwen in https://github.com/PaddlePaddle/PaddleNLP/pull/8312


* 分布式能力优化:
* 修复流水线并行中enable_sharding_comm_overlap中参数错误问题 in https://github.com/PaddlePaddle/PaddleNLP/pull/8333
* MoE并行支持 in https://github.com/PaddlePaddle/PaddleNLP/pull/8498 https://github.com/PaddlePaddle/PaddleNLP/pull/8522

* chat能力优化:
* 增加Chat template in https://github.com/PaddlePaddle/PaddleNLP/pull/8226

* 其他
* 文档 in https://github.com/PaddlePaddle/PaddleNLP/pull/8336 https://github.com/PaddlePaddle/PaddleNLP/pull/8393
* 更新nested操作 in https://github.com/PaddlePaddle/PaddleNLP/pull/8380
* 随机性更新 in https://github.com/PaddlePaddle/PaddleNLP/pull/8450 https://github.com/PaddlePaddle/PaddleNLP/pull/8396
* 算子更新 in https://github.com/PaddlePaddle/PaddleNLP/pull/8472
* example更新 in https://github.com/PaddlePaddle/PaddleNLP/pull/8538

问题修复

* 修复sharding数量小于100的bug in https://github.com/PaddlePaddle/PaddleNLP/pull/8146
* 修复TP/PP参数合并问题 in https://github.com/PaddlePaddle/PaddleNLP/pull/8239
* 修复tensor.shape与paddle.shape(tensor)不一致问题 in https://github.com/PaddlePaddle/PaddleNLP/pull/8260
* 修复fp16+delay_scale_loss_scale+sharding_stage1_overlap的bug in https://github.com/PaddlePaddle/PaddleNLP/pull/8314
* 增加pipelines运行文档及提示 in https://github.com/PaddlePaddle/PaddleNLP/pull/8292 https://github.com/PaddlePaddle/PaddleNLP/pull/8308 https://github.com/PaddlePaddle/PaddleNLP/pull/8202 https://github.com/PaddlePaddle/PaddleNLP/pull/8353
* 修复text feature extraction任务中tokenizer输入 in https://github.com/PaddlePaddle/PaddleNLP/pull/8331
* 修复import error in https://github.com/PaddlePaddle/PaddleNLP/pull/8332 https://github.com/PaddlePaddle/PaddleNLP/pull/8367

结构调整

PaddleNLP文件结构调整 in https://github.com/PaddlePaddle/PaddleNLP/pull/8609 https://github.com/PaddlePaddle/PaddleNLP/pull/8613 https://github.com/PaddlePaddle/PaddleNLP/pull/8605 https://github.com/PaddlePaddle/PaddleNLP/pull/8614 https://github.com/PaddlePaddle/PaddleNLP/pull/8617 https://github.com/PaddlePaddle/PaddleNLP/pull/8626 https://github.com/PaddlePaddle/PaddleNLP/pull/8618 https://github.com/PaddlePaddle/PaddleNLP/pull/8625 https://github.com/PaddlePaddle/PaddleNLP/pull/8619 https://github.com/PaddlePaddle/PaddleNLP/pull/8629 https://github.com/PaddlePaddle/PaddleNLP/pull/8601 https://github.com/PaddlePaddle/PaddleNLP/pull/8627 https://github.com/PaddlePaddle/PaddleNLP/pull/8666

What's Changed
* [dist]pip requirements-dev.txt by Liujie0926 in https://github.com/PaddlePaddle/PaddleNLP/pull/8258
* add scaling by lugimzzz in https://github.com/PaddlePaddle/PaddleNLP/pull/8256
* [LLM]Support Gemma model by Southpika in https://github.com/PaddlePaddle/PaddleNLP/pull/8082
* [BugFix] Try except sequence parallel utils by DesmonDay in https://github.com/PaddlePaddle/PaddleNLP/pull/8189
* Update CodeCov GitHub Action by sijunhe in https://github.com/PaddlePaddle/PaddleNLP/pull/8268
* [AutoParallel] Open recompute strategy for llama model by zhangbo9674 in https://github.com/PaddlePaddle/PaddleNLP/pull/8265
* Fix sharding < 100 limitation bug by sneaxiy in https://github.com/PaddlePaddle/PaddleNLP/pull/8146
* use tensor.shape bug not paddle.shape(tensor) by wanghuancoder in https://github.com/PaddlePaddle/PaddleNLP/pull/8260
* [dist CI]update paddlenlp install for CI by Liujie0926 in https://github.com/PaddlePaddle/PaddleNLP/pull/8267
* [Bug Fix]Fix merge parameters in pp by Southpika in https://github.com/PaddlePaddle/PaddleNLP/pull/8239
* [LLM] add memory stats to logger of trainer by SylarTiaNII in https://github.com/PaddlePaddle/PaddleNLP/pull/8269
* Add p2p_comm_overlap for Llama-2-70b benchmark. by Xreki in https://github.com/PaddlePaddle/PaddleNLP/pull/8276
* add a100 test ground truth by zhiqiu in https://github.com/PaddlePaddle/PaddleNLP/pull/8249
* [paddle-pipelines] faq semantic search question answering reamde by w5688414 in https://github.com/PaddlePaddle/PaddleNLP/pull/8292
* [paddle-pipelines] Add pipelines documentation by w5688414 in https://github.com/PaddlePaddle/PaddleNLP/pull/8308
* Support llama-3 by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/8307
* [Distributed] [CustomDevices] Adapt SP on lora && polish MC2 APIs by SylarTiaNII in https://github.com/PaddlePaddle/PaddleNLP/pull/8303
* fix bug for fp16 + delay_scale_loss_scale + sharding_stage1_overlap by FeixLiu in https://github.com/PaddlePaddle/PaddleNLP/pull/8314
* [paddle-pipelines] Update mkdocs by w5688414 in https://github.com/PaddlePaddle/PaddleNLP/pull/8310
* [benchmark]update llama2_ips by Liujie0926 in https://github.com/PaddlePaddle/PaddleNLP/pull/8322
* [dist CI]fix before_hook by Liujie0926 in https://github.com/PaddlePaddle/PaddleNLP/pull/8283
* benchmark llama worker=1 by wanghuancoder in https://github.com/PaddlePaddle/PaddleNLP/pull/8305
* 【AutoParallel】Add llama2 UT for auto-parallel by heavyrain-lzy in https://github.com/PaddlePaddle/PaddleNLP/pull/8300
* Add system env log for llama test by zhangbo9674 in https://github.com/PaddlePaddle/PaddleNLP/pull/8321
* [LLM] Support fuse attention q, k, v weights by DrownFish19 in https://github.com/PaddlePaddle/PaddleNLP/pull/8202
* [Distributed] fix lora by SylarTiaNII in https://github.com/PaddlePaddle/PaddleNLP/pull/8325
* fix try import by w5688414 in https://github.com/PaddlePaddle/PaddleNLP/pull/8332
* [DEV] Support sync params in tensor parallel config by From00 in https://github.com/PaddlePaddle/PaddleNLP/pull/8311
* cherry pick paddlenlp 2.8 by w5688414 in https://github.com/PaddlePaddle/PaddleNLP/pull/8323
* textfeature_queryinput by cxa-unique in https://github.com/PaddlePaddle/PaddleNLP/pull/8331
* [BugFix] Fix gpu ci by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/8337
* [Trainer] Fix sharding overlap bug by DesmonDay in https://github.com/PaddlePaddle/PaddleNLP/pull/8333
* [Tokenizer]Add Chat template by Southpika in https://github.com/PaddlePaddle/PaddleNLP/pull/8226
* [AutoParallel]Refine lr warm_up configuration strategy for llama by zhangbo9674 in https://github.com/PaddlePaddle/PaddleNLP/pull/8329
* Add num_hidden_layer config for llama run_pretrain by zhangbo9674 in https://github.com/PaddlePaddle/PaddleNLP/pull/8288
* [XPU] llama add xpu support by dynamicheart in https://github.com/PaddlePaddle/PaddleNLP/pull/8282
* add eliminate_transpose arg by zhiqiu in https://github.com/PaddlePaddle/PaddleNLP/pull/8339
* change llama/modeling.py to opt npu performence by Galaxy1458 in https://github.com/PaddlePaddle/PaddleNLP/pull/8342
* Update llm docs requirements by w5688414 in https://github.com/PaddlePaddle/PaddleNLP/pull/8336
* Disable eval and predict for llama-2 benchmark. by Xreki in https://github.com/PaddlePaddle/PaddleNLP/pull/8366
* update by Galaxy1458 in https://github.com/PaddlePaddle/PaddleNLP/pull/8359
* [LLM] fix lora target modules on llama by SylarTiaNII in https://github.com/PaddlePaddle/PaddleNLP/pull/8372
* [paddle-pipelines] Update offline ann by w5688414 in https://github.com/PaddlePaddle/PaddleNLP/pull/8353
* refine benchmard bert ips stat by wanghuancoder in https://github.com/PaddlePaddle/PaddleNLP/pull/8361
* [BugFix] Update truncate in distributed training by KB-Ding in https://github.com/PaddlePaddle/PaddleNLP/pull/8362
* [dist benchmark]Fix llama2 benchmark by Liujie0926 in https://github.com/PaddlePaddle/PaddleNLP/pull/8376
* Revert "update" by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/8389
* Fix test init by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/8377
* [Performance] Optimize unified checkpoint save/load speed. by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/8204
* [npu model bug]fix_global_bug by Galaxy1458 in https://github.com/PaddlePaddle/PaddleNLP/pull/8399
* [Bugfix] Fix fast tokenizer import error by w5688414 in https://github.com/PaddlePaddle/PaddleNLP/pull/8367
* [bugfix] fix uie by w5688414 in https://github.com/PaddlePaddle/PaddleNLP/pull/8379
* fit for llama3 for auto_parallel by zhiqiu in https://github.com/PaddlePaddle/PaddleNLP/pull/8395
* [DistDataloader] Update implementation, add nested.py by DesmonDay in https://github.com/PaddlePaddle/PaddleNLP/pull/8380
* [LLM] Fix fuse or split with same key by DrownFish19 in https://github.com/PaddlePaddle/PaddleNLP/pull/8378
* [UC] Fix compatible with npu by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/8409
* pre copy pinned data to gpu by wanghuancoder in https://github.com/PaddlePaddle/PaddleNLP/pull/8386
* Refine position_ids for auto parallel training of llama by zhangbo9674 in https://github.com/PaddlePaddle/PaddleNLP/pull/8363
* [Distributed] enable tensor_parallel_output for finetuning by SylarTiaNII in https://github.com/PaddlePaddle/PaddleNLP/pull/8370
* fix type promotion problem. by zxcd in https://github.com/PaddlePaddle/PaddleNLP/pull/8414
* Fix ckpt done by gongel in https://github.com/PaddlePaddle/PaddleNLP/pull/8402
* [LLM] rename logits_tensor_parallel_output to avoid conflict by SylarTiaNII in https://github.com/PaddlePaddle/PaddleNLP/pull/8419
* [Trainer] fix distdataloader by DesmonDay in https://github.com/PaddlePaddle/PaddleNLP/pull/8420
* fix safe open. by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/8422
* adapter new type promotion rule for Paddle 2.6 by zxcd in https://github.com/PaddlePaddle/PaddleNLP/pull/8421
* [BugFix] Fix llama3 `eot_id` by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/8371
* add npu-llama-opt0-script by Galaxy1458 in https://github.com/PaddlePaddle/PaddleNLP/pull/8401
* [LLM] add assertion for enable_stage1_overlap in lora mode by SylarTiaNII in https://github.com/PaddlePaddle/PaddleNLP/pull/8425
* [NPU]Custom fusion operator unification by Galaxy1458 in https://github.com/PaddlePaddle/PaddleNLP/pull/8431
* delete csrc/generation/reset_need_stop_value.cc by yuanlehome in https://github.com/PaddlePaddle/PaddleNLP/pull/8413
* Update llama_npu_opt_lora.sh by Galaxy1458 in https://github.com/PaddlePaddle/PaddleNLP/pull/8439
* [CI]add scripts for unittest by Liujie0926 in https://github.com/PaddlePaddle/PaddleNLP/pull/8433
* fix npu sft ckpt load bug and no FA bug by NINGBENZHE in https://github.com/PaddlePaddle/PaddleNLP/pull/8438
* Fix CI bugs by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/8430
* Fix/test gpu by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/8452
* Support fused_attention_qkv for auto_parallel llama by zhangbo9674 in https://github.com/PaddlePaddle/PaddleNLP/pull/8432
* [BugFix] Fix load rng compatibility. by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/8450
* update by Galaxy1458 in https://github.com/PaddlePaddle/PaddleNLP/pull/8448
* [GCU] Support llama for GCU by EnflameGCU in https://github.com/PaddlePaddle/PaddleNLP/pull/8445
* [bugfix] fix erniedoc by w5688414 in https://github.com/PaddlePaddle/PaddleNLP/pull/8393
* [benchmark]Add llama2 auto by Liujie0926 in https://github.com/PaddlePaddle/PaddleNLP/pull/8424
* Add llama2-70b for test_tipc by zhangbo9674 in https://github.com/PaddlePaddle/PaddleNLP/pull/8455
* Fix ci tests. by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/8471
* [NPU] support npu llama2-13B export & inference by ronny1996 in https://github.com/PaddlePaddle/PaddleNLP/pull/8442
* [LLM] fix bug when loss is None in llama modeling.py by cqulilujia in https://github.com/PaddlePaddle/PaddleNLP/pull/8459
* fix rotary_emb for llama by EnflameGCU in https://github.com/PaddlePaddle/PaddleNLP/pull/8470
* [Ops] RoPE kernel support theta input by yinfan98 in https://github.com/PaddlePaddle/PaddleNLP/pull/8440
* Support Sharding Overlap by iosmers in https://github.com/PaddlePaddle/PaddleNLP/pull/8473
* Revert "Support Sharding Overlap (8473)" by SylarTiaNII in https://github.com/PaddlePaddle/PaddleNLP/pull/8491
* fix run_benchmark for llama2_70b in auto_parallel by fightfat in https://github.com/PaddlePaddle/PaddleNLP/pull/8484
* 【AutoParallel】Add split_backward for vpp by heavyrain-lzy in https://github.com/PaddlePaddle/PaddleNLP/pull/8479
* Quick fix from_pretrained. by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/8486
* Fix rng_state in llm models by zhangyuqin1998 in https://github.com/PaddlePaddle/PaddleNLP/pull/8396
* [AutoParallel] Support qwen for auto_parallel by GhostScreaming in https://github.com/PaddlePaddle/PaddleNLP/pull/8312
* modify block_multihead_attention api by ming1753 in https://github.com/PaddlePaddle/PaddleNLP/pull/8456
* [LLM] disable part of MC2 in lora by SylarTiaNII in https://github.com/PaddlePaddle/PaddleNLP/pull/8505
* Update model_utils.py by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/8509
* Update merge_lora_params.py by Galaxy1458 in https://github.com/PaddlePaddle/PaddleNLP/pull/8514
* [fea] moe support by bo-ke in https://github.com/PaddlePaddle/PaddleNLP/pull/8498
* Add Sharding V1 broadcast and V2 allgather overlap optimize by iosmers in https://github.com/PaddlePaddle/PaddleNLP/pull/8499
* [fix] Broadcast optimizer state using broadcast_dp without shard-resh… by bo-ke in https://github.com/PaddlePaddle/PaddleNLP/pull/8522
* Update README.md by wawltor in https://github.com/PaddlePaddle/PaddleNLP/pull/8524
* [Safetensors] Fix fast safe open slice. by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/8512
* Update Benchmark scripts by iosmers in https://github.com/PaddlePaddle/PaddleNLP/pull/8519
* fix eval. by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/8529
* [BugFix][NPU] fix llama attn_mask astype error by tianhaodongbd in https://github.com/PaddlePaddle/PaddleNLP/pull/8528
* fused_ln:Added implementation for the HIP platform by asr-sheep1 in https://github.com/PaddlePaddle/PaddleNLP/pull/8472
* [CI] Update pip source. by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/8540
* [PIP] Update run_ci.sh by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/8552
* add mteb evaluation by cxa-unique in https://github.com/PaddlePaddle/PaddleNLP/pull/8538
* [Cherry-pick] Add release grad & sharding format & decorate_exclude_layers by ForFishes in https://github.com/PaddlePaddle/PaddleNLP/pull/8545
* Add RingFlashAttention for context parallel by zhangyuqin1998 in https://github.com/PaddlePaddle/PaddleNLP/pull/8383
* fix codecov conflicts by greycooker in https://github.com/PaddlePaddle/PaddleNLP/pull/8555
* support fused weights for export_model by ronny1996 in https://github.com/PaddlePaddle/PaddleNLP/pull/8554
* 【benchmark】 add llama-7b_auto_dp2mp2pp2 benchmark script for cinn by mmglove in https://github.com/PaddlePaddle/PaddleNLP/pull/8423
* Fix memory leak bug by sneaxiy in https://github.com/PaddlePaddle/PaddleNLP/pull/8546
* Update sequence_parallel for predict by DesmonDay in https://github.com/PaddlePaddle/PaddleNLP/pull/8551
* [GPT][CE] Update modeling.py by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/8548
* add fuse_attention_ffn support for qwen by deepllz in https://github.com/PaddlePaddle/PaddleNLP/pull/8526
* Update generation_utils.py by carryyu in https://github.com/PaddlePaddle/PaddleNLP/pull/8502
* fix llama export by ronny1996 in https://github.com/PaddlePaddle/PaddleNLP/pull/8561
* Update llama_npu_opt_lora.sh by Galaxy1458 in https://github.com/PaddlePaddle/PaddleNLP/pull/8562
* [FIX DDP] fix ddp by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/8549
* [AutoParallel] Add benchmark for llama-7b-dy2st. by GhostScreaming in https://github.com/PaddlePaddle/PaddleNLP/pull/8559
* [Cherry pick] Sharding reshard function enhancement by sneaxiy in https://github.com/PaddlePaddle/PaddleNLP/pull/8544
* [BugFix] Fix test_long_sequence_strategies by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/8568
* Fix/ci pip by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/8541
* Add async save for optimizer by ForFishes in https://github.com/PaddlePaddle/PaddleNLP/pull/8557
* add llama & qwen dpo by lugimzzz in https://github.com/PaddlePaddle/PaddleNLP/pull/8474
* [LLM] support Qwen2 by DrownFish19 in https://github.com/PaddlePaddle/PaddleNLP/pull/8338
* [LLM] Fix Qwen2 by DrownFish19 in https://github.com/PaddlePaddle/PaddleNLP/pull/8584
* fix autotunner benchmark error and fix llama2 dy2st benchmark by fightfat in https://github.com/PaddlePaddle/PaddleNLP/pull/8587
* fix autoruner resume case by Difers in https://github.com/PaddlePaddle/PaddleNLP/pull/8259
* Enable test with re-try. by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/8590
* [xpu] add xpu custom ops support for llama2-7b by NeroLoh in https://github.com/PaddlePaddle/PaddleNLP/pull/8515
* xpu devices support llama-7b basic mode inference (turn on BlockAtten… by zhink in https://github.com/PaddlePaddle/PaddleNLP/pull/8588
* Add Pipeline Parallel for PPO training and support generation with InferenceModel by guoshengCS in https://github.com/PaddlePaddle/PaddleNLP/pull/7953
* [xpu] change xpu setup.py to paddlenlp_ops by NeroLoh in https://github.com/PaddlePaddle/PaddleNLP/pull/8595
* Clean RLHF main script by guoshengCS in https://github.com/PaddlePaddle/PaddleNLP/pull/8596
* Fix dataset with empty char. by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/8469
* XPU open ir pass by zhink in https://github.com/PaddlePaddle/PaddleNLP/pull/8598
* [bug fix] fix sharding stage1 allgather overlap bug, which needs to forbiden pin memory by iosmers in https://github.com/PaddlePaddle/PaddleNLP/pull/8594
* Add main process print function by ForFishes in https://github.com/PaddlePaddle/PaddleNLP/pull/8604
* [Feature] Optimize config saving. by ZHUI in https://github.com/PaddlePaddle/PaddleNLP/pull/8490
* to_json_string兼容性升级 by sneaxiy in https://github.com/PaddlePaddle/PaddleNLP/pull/8608

2.52.6

New Contributors
* zws-2019 made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/5167
* qiuwenbogdut made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/5098
* kuizhiqing made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/5347
* 46319943 made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/5419
* jiaohuix made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/5465
* kangguangli made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/5438
* vivienfanghuagood made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/5563
* zhiboniu made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/5470
* cyber-pioneer made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/5598
* invokerbyxv made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/5622
* megemini made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/5658
* zhenyun-li made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/5683
* solrex made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/5736
* nemonameless made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/5487
* Yulv-git made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/5709
* wangxinxin08 made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/5773
* AlphaHinex made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/5815
* houj04 made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/5820
* Joker1718 made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/5816
* pkuzyc made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/5538
* jadepeng made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/5841
* KB-Ding made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/5886
* parap1uie-s made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/5775
* zirui made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/5866
* GOH-Gu made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/5951
* yangjianfengo1 made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/6069
* zhangting2020 made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/5922
* rogerserper made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/6192
* wtmlon made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/6258
* qingzhong1 made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/6251
* BeingGod made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/6307
* zhiqiu made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/6347
* DesmonDay made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/6435
* cyk1337 made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/6447
* lxp521125 made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/6491
* littsk made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/6425
* RachelXu7 made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/6572
* wanghuancoder made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/6539
* DrownFish19 made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/6570
* GhostScreaming made their first contribution in https://github.com/PaddlePaddle/PaddleNLP/pull/6673

**Full Changelog**: https://github.com/PaddlePaddle/PaddleNLP/compare/v2.5.2...v2.6.0

2.8.1

**Full Changelog**: https://github.com/PaddlePaddle/PaddleNLP/compare/v2.8.0...v2.8.1

Page 1 of 9

© 2024 Safety CLI Cybersecurity Inc. All Rights Reserved.