Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

【MetaX】Merge Metax's modifications to mxmaca/2.6 branch #68534

Merged

Conversation

idontkonwher
Copy link
Contributor

PR Category

Environment Adaptation

PR Types

Others

Description

Open source the modifications made by Metax to the Paddle repository when adapting to the MXMACA software stack.

risemeup1 and others added 30 commits December 26, 2023 17:15
* fix windows bug

* fix windows bug

* fix windows bug

* fix windows bug

* fix windows bug

* fix windows bug

* Update inference_lib.cmake
…#60324)

Co-authored-by: gouzil <66515297+gouzil@users.noreply.github.com>
…addle#60184)

* fix weight-only quant kernel error for n div 64 !=0

* code style fix
…0208) (PaddlePaddle#60495)

* fix chunk allocator posix_memalign return value check;test=develop

* fix chunk allocator posix_memalign return value check;test=develop

* fix chunk allocator posix_memalign return value check;test=develop
…e#60620)

* fix fleetutil get_online_pass_interval bug3; test=develop

* fix fleetutil get_online_pass_interval bug3; test=develop

* fix fleetutil get_online_pass_interval bug3; test=develop
* update 2023 security advisory, test=document_fix

* update pdsa-2023-019, test=document_fix
…EADME (PaddlePaddle#60786)

* [Dy2St][2.6] Disable `test_transformer` on release/2.6 and update README

* [Docs] Update latest release version in README (PaddlePaddle#60691)

* restore order
* Fix set value grad (PaddlePaddle#59034)

* first fix the UT

* fix set value grad

* polish code

* add static mode backward test

* always has input valuetensor

* add dygraph test

* Fix shape error in combined-indexing setitem (PaddlePaddle#60447)

* add ut

* fix shape error in combine-indexing

* fix ut

* Set value with scalar (PaddlePaddle#60452)

* set_value with scalar

* fix ut

* remove test_pir

* remove one test since 2.6 not support uint8-add
This uses shlex for safe command parsing to fix arbitrary code injection

Co-authored-by: ndren <andreien@proton.me>
…ePaddle#61382)

* OS Command Injection prune_by_memory_estimation fix

* Fix StyleCode
…addlePaddle#60774) (PaddlePaddle#61045)

Co-authored-by: Tian <121000916+SylarTiaNII@users.noreply.github.com>
* fix issue 60092

* update

* update

* update
* fix unique kernel, row to num_out
* remove _wget

* remove _wget

* remove wget test
Zhao Wu and others added 25 commits August 28, 2024 12:27
(cherry picked from commit fe4655e86b92f5053fa886af49bf199307960a05)
Change-Id: I35003420292359f8a41b19b7ca2cbaae17dc5b45
Signed-off-by: m00891 <Zequn.Yang@metax-tech.com>
…move ldg up.

(cherry picked from commit a7cb0ed275a3488f79445ef31456ab6560e9de43)
Change-Id: Ia89df4e5a26de64baae4152837d2ce3076c56df1
Signed-off-by: m00891 <Zequn.Yang@metax-tech.com>
…stDivMod;3.move ldg up.

(cherry picked from commit 4fb857655d09f55783d9445b91a2d953ed14d0b8)
Change-Id: I7df7f3af7b4615e5e96d33b439e5276be6ddb732
Signed-off-by: m00891 <Zequn.Yang@metax-tech.com>
(cherry picked from commit 333cba7aca1edf7a0e87623a0e55e230cd1e9451)
Change-Id: Ic808d42003677ed543621eb22a797f0ab7751baa
Signed-off-by: m00891 <Zequn.Yang@metax-tech.com>
…onzero and masked_select (forward only) OP.

(cherry picked from commit c907b40eb3f9ded6ee751e522c2a97a353ac93bd)
Change-Id: I7f4845405e64e7599134a8c497f464ac04dead88
Signed-off-by: m00891 <Zequn.Yang@metax-tech.com>
1. 256 Blocksize launch for small shape inputgrad;
2. FastDivMod in inputgrad and filtergrad;
3. shared memory to put output_grad_data in small shape.

(cherry picked from commit f9f29bf7b8d929fb95eb1153a79d8a6b96d5b6d2)
Change-Id: I1a3818201784031dbedc320286ea5f4802dbb6b1
Signed-off-by: m00891 <Zequn.Yang@metax-tech.com>
…iple tensors.

(cherry picked from commit 3bd200f262271a333b3947326442b86af7fb6da1)
Change-Id: I57c94cc5e709be8926e1b21da14b653cb18eabc3
Signed-off-by: m00891 <Zequn.Yang@metax-tech.com>
…nto multiple tensors."

This reverts commit 3bd200f262271a333b3947326442b86af7fb6da1.

(cherry picked from commit 86ed8adaa8c20d3c824eecb0ee1e10d365bcea37)
Change-Id: I5b8b7819fdf99255c65fe832d5d77f8e439bdecb
Signed-off-by: m00891 <Zequn.Yang@metax-tech.com>
(cherry picked from commit cddb01a83411c45f68363248291c0c4685e60b24)
Change-Id: Ie106ff8d65c21a8545c40636f021b73f3ad84587
Signed-off-by: m00891 <Zequn.Yang@metax-tech.com>
(cherry picked from commit 07ea3acf347fda434959c8c9cc3533c0686d1836)
Change-Id: Id7a727fd18fac4a662f8af1bf6c6b5ebc6233c9f
Signed-off-by: m00891 <Zequn.Yang@metax-tech.com>
Use tmp to store ldg data in the loop so calculate and ldg time
can fold each other.

(cherry picked from commit 7ddab49d868cdb6deb7c3e17c5ef9bbdbab86c3e)
Change-Id: I46399594d1d7f76b78b9860e483716fdae8fc7d6
Signed-off-by: m00891 <Zequn.Yang@metax-tech.com>
…ed memory and making single thread do more tasks.

(cherry picked from commit 631ffdda2847cda9562e591dc87b3f529a51a978)
Change-Id: Ie9ffdd872ab06ff34d4daf3134d6744f5221e41e
Signed-off-by: m00891 <Zequn.Yang@metax-tech.com>
1.LayerNormBackward: remove if statement, now will always loop VPT
times for ldg128 in compiler, bool flag to control if write action
will be taken or not;
2.ContiguousCaseOneFunc: tmp saving division result for less division

(cherry picked from commit 422d676507308d26f6107bed924424166aa350d3)
Change-Id: I37aab7e2f97ae6b61c0f50ae4134f5eb1743d429
Signed-off-by: m00891 <Zequn.Yang@metax-tech.com>
Set BlockDim.z to make blockSize always be 512, each block can
handle several batches.
Then all threads will loop 4 times for better performance.

(cherry picked from commit 7550c90ca29758952fde13eeea74857ece41908b)
Change-Id: If24de87a0af19ee07e29ac2e7e237800f0181148
Signed-off-by: m00891 <Zequn.Yang@metax-tech.com>
…ange it to 64 warp reduce.

(cherry picked from commit a346af182b139dfc7737e5f6473dc394b21635d7)
Change-Id: I6c8d8105fd77947c662e6d22a0d15d7bad076bde
Signed-off-by: m00891 <Zequn.Yang@metax-tech.com>
Might have lossdiff with old optimization without atomicAdd.

(cherry picked from commit 80b0bcaa9a307c94dbeda658236fd75e104ccccc)
Change-Id: I4a7c4ec2a0e885c2d581dcebc74464830dae7637
Signed-off-by: m00891 <Zequn.Yang@metax-tech.com>
(cherry picked from commit cc421d7861c359740de0d2870abcfde4354d8c71)
Change-Id: I55c049e951f93782af1c374331f44b521ed75dfe
Signed-off-by: m00891 <Zequn.Yang@metax-tech.com>
…oat16>.

Change-Id: I5788c73a9c45f65e60ed5a88d16a473bbb888927
Change-Id: I8b34f02958ddccb3467f639daaac8044022f3d34
Change-Id: I77730da567903f43ef7a9992925b90ed4ba179c7
Change-Id: I1b7eb58e7959daff8660ce7889ba390cdfae0c1a
Change-Id: I94d422c969bdb83ad74262e03efe38ca85ffa673
Change-Id: I8ece364d926596a40f42d973190525d9b8224d99
Copy link

paddle-bot bot commented Sep 30, 2024

你的PR提交成功,感谢你对开源项目的贡献!
请关注后续CI自动化测试结果,详情请参考Paddle-CI手册
Your PR has been submitted. Thanks for your contribution!
Please wait for the result of CI firstly. See Paddle CI Manual for details.

@CLAassistant
Copy link

CLA assistant check
Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.


zequn yang seems not to be a GitHub user. You need a GitHub account to be able to sign the CLA. If you have already a GitHub account, please add the email address used for this commit to your account.
You have signed the CLA already but the status is still pending? Let us recheck it.

@idontkonwher idontkonwher changed the title 【Metax】Merge Metax's modifications to mxmaca/2.6 branch 【MetaX】Merge Metax's modifications to mxmaca/2.6 branch Sep 30, 2024
@xiaoguoguo626807 xiaoguoguo626807 merged commit b102bc4 into PaddlePaddle:release-mxmaca/2.6 Sep 30, 2024
0 of 27 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.