Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[shardformer] add Dropout layer support different dropout pattern #3856

Merged
merged 6 commits into from
Jun 1, 2023

Conversation

FoolPlayer
Copy link
Contributor

📌 Checklist before creating the PR

  • I have created an issue for this PR for traceability
  • The title follows the standard format: [doc/gemini/tensor/...]: A concise description
  • I have added relevant tags if possible for us to better distinguish different PRs

🚨 Issue number

Link this PR to your issue with words like fixed to automatically close the linked issue upon merge

e.g. fixed #1234, closed #1234, resolved #1234

📝 What does this PR do?

  1. add dropout layer to support different drop pattern when training, and just modify the local random state when dropout
  2. add dropout test in test.py

💥 Checklist before requesting a review

  • I have linked my PR to an issue (instruction)
  • My issue clearly describes the problem/feature/proposal, with diagrams/charts/table/code if possible
  • I have performed a self-review of my code
  • I have added thorough tests.
  • I have added docstrings for all the functions/methods I implemented

⭐️ Do you enjoy contributing to Colossal-AI?

  • 🌝 Yes, I do.
  • 🌚 No, I don't.

Tell us more if you don't enjoy contributing to Colossal-AI.

@github-actions
Copy link
Contributor

The code coverage for the changed files is 3%.

Click me to view the complete report
Name                                            Stmts   Miss  Cover
-------------------------------------------------------------------
colossalai/nn/layer/parallel_1d/_operation.py      53     26    51%
colossalai/shardformer/__init__.py                  0      0   100%
colossalai/shardformer/layer/__init__.py            0      0   100%
colossalai/shardformer/layer/_operation.py         54     54     0%
colossalai/shardformer/layer/dropout.py            32     32     0%
colossalai/shardformer/layer/layers.py            464    464     0%
colossalai/shardformer/model/__init__.py            0      0   100%
colossalai/shardformer/model/modeling_bert.py      20     20     0%
colossalai/shardformer/policies/__init__.py         0      0   100%
colossalai/shardformer/policies/autopolicy.py      16     16     0%
colossalai/shardformer/policies/basepolicy.py      49     49     0%
colossalai/shardformer/policies/bert.py            41     41     0%
colossalai/shardformer/shard/__init__.py            4      4     0%
colossalai/shardformer/shard/shard_config.py       14     14     0%
colossalai/shardformer/shard/sharder.py           109    109     0%
colossalai/shardformer/shard/slicer.py             35     35     0%
colossalai/shardformer/utils/__init__.py            0      0   100%
colossalai/shardformer/utils/utils.py              28     28     0%
-------------------------------------------------------------------
TOTAL                                             919    892     3%

@github-actions
Copy link
Contributor

The code coverage for the changed files is 2%.

Click me to view the complete report
Name                                                Stmts   Miss  Cover
-----------------------------------------------------------------------
colossalai/nn/layer/parallel_1d/_operation.py          53     35    34%
colossalai/shardformer/__init__.py                      0      0   100%
colossalai/shardformer/layer/__init__.py                0      0   100%
colossalai/shardformer/layer/_operation.py             54     54     0%
colossalai/shardformer/layer/dist_crossentropy.py      49     49     0%
colossalai/shardformer/layer/dropout.py                32     32     0%
colossalai/shardformer/layer/layers.py                464    464     0%
colossalai/shardformer/model/__init__.py                0      0   100%
colossalai/shardformer/model/modeling_bert.py          20     20     0%
colossalai/shardformer/policies/__init__.py             0      0   100%
colossalai/shardformer/policies/autopolicy.py          16     16     0%
colossalai/shardformer/policies/basepolicy.py          49     49     0%
colossalai/shardformer/policies/bert.py                41     41     0%
colossalai/shardformer/shard/__init__.py                4      4     0%
colossalai/shardformer/shard/shard_config.py           14     14     0%
colossalai/shardformer/shard/sharder.py               109    109     0%
colossalai/shardformer/shard/slicer.py                 35     35     0%
colossalai/shardformer/utils/__init__.py                0      0   100%
colossalai/shardformer/utils/utils.py                  28     28     0%
-----------------------------------------------------------------------
TOTAL                                                 968    950     2%

@github-actions
Copy link
Contributor

The code coverage for the changed files is 3%.

Click me to view the complete report
Name                                                Stmts   Miss  Cover
-----------------------------------------------------------------------
colossalai/nn/layer/parallel_1d/_operation.py          53     26    51%
colossalai/shardformer/__init__.py                      0      0   100%
colossalai/shardformer/layer/__init__.py                0      0   100%
colossalai/shardformer/layer/_operation.py             54     54     0%
colossalai/shardformer/layer/dist_crossentropy.py      49     49     0%
colossalai/shardformer/layer/dropout.py                32     32     0%
colossalai/shardformer/layer/layers.py                464    464     0%
colossalai/shardformer/model/__init__.py                0      0   100%
colossalai/shardformer/model/modeling_bert.py          20     20     0%
colossalai/shardformer/policies/__init__.py             0      0   100%
colossalai/shardformer/policies/autopolicy.py          16     16     0%
colossalai/shardformer/policies/basepolicy.py          49     49     0%
colossalai/shardformer/policies/bert.py                41     41     0%
colossalai/shardformer/shard/__init__.py                4      4     0%
colossalai/shardformer/shard/shard_config.py           14     14     0%
colossalai/shardformer/shard/sharder.py               109    109     0%
colossalai/shardformer/shard/slicer.py                 35     35     0%
colossalai/shardformer/utils/__init__.py                0      0   100%
colossalai/shardformer/utils/utils.py                  28     28     0%
-----------------------------------------------------------------------
TOTAL                                                 968    941     3%

@FrankLeeeee FrankLeeeee merged commit 2cdb688 into hpcaitech:feature/shardformer Jun 1, 2023
@FoolPlayer FoolPlayer deleted the dropout branch June 1, 2023 09:12
FrankLeeeee pushed a commit to FrankLeeeee/ColossalAI that referenced this pull request Jun 8, 2023
…caitech#3856)

* add dropout layer, add dropout test

* modify seed manager as context manager

* add a copy of col_nn.layer

* add dist_crossentropy loss; separate module test

* polish the code

* fix dist crossentropy loss
FrankLeeeee pushed a commit to FrankLeeeee/ColossalAI that referenced this pull request Jun 8, 2023
…caitech#3856)

* add dropout layer, add dropout test

* modify seed manager as context manager

* add a copy of col_nn.layer

* add dist_crossentropy loss; separate module test

* polish the code

* fix dist crossentropy loss
FrankLeeeee pushed a commit to FrankLeeeee/ColossalAI that referenced this pull request Jun 8, 2023
…caitech#3856)

* add dropout layer, add dropout test

* modify seed manager as context manager

* add a copy of col_nn.layer

* add dist_crossentropy loss; separate module test

* polish the code

* fix dist crossentropy loss
FrankLeeeee pushed a commit that referenced this pull request Jun 26, 2023
)

* add dropout layer, add dropout test

* modify seed manager as context manager

* add a copy of col_nn.layer

* add dist_crossentropy loss; separate module test

* polish the code

* fix dist crossentropy loss
flybird11111 pushed a commit to flybird11111/ColossalAI that referenced this pull request Jul 3, 2023
…caitech#3856)

* add dropout layer, add dropout test

* modify seed manager as context manager

* add a copy of col_nn.layer

* add dist_crossentropy loss; separate module test

* polish the code

* fix dist crossentropy loss
FrankLeeeee pushed a commit that referenced this pull request Jul 4, 2023
)

* add dropout layer, add dropout test

* modify seed manager as context manager

* add a copy of col_nn.layer

* add dist_crossentropy loss; separate module test

* polish the code

* fix dist crossentropy loss
ver217 pushed a commit to ver217/ColossalAI that referenced this pull request Jul 13, 2023
…caitech#3856)

* add dropout layer, add dropout test

* modify seed manager as context manager

* add a copy of col_nn.layer

* add dist_crossentropy loss; separate module test

* polish the code

* fix dist crossentropy loss
ver217 pushed a commit to ver217/ColossalAI that referenced this pull request Jul 13, 2023
…caitech#3856)

* add dropout layer, add dropout test

* modify seed manager as context manager

* add a copy of col_nn.layer

* add dist_crossentropy loss; separate module test

* polish the code

* fix dist crossentropy loss
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants