-
Notifications
You must be signed in to change notification settings - Fork 1.1k
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
2872 implementation of mixup, cutmix and cutout (#7198)
Fixes #2872 ### Description Implementation of mixup, cutmix and cutout as described in the original papers. Current implementation support both, the dictionary-based batches and tuples of tensors. ### Types of changes <!--- Put an `x` in all the boxes that apply, and remove the not applicable items --> - [x] Non-breaking change (fix or new feature that would not break existing functionality). - [ ] Breaking change (fix or new feature that would cause existing functionality to change). - [x] New tests added to cover the changes. - [ ] Integration tests passed locally by running `./runtests.sh -f -u --net --coverage`. - [x] Quick tests passed locally by running `./runtests.sh --quick --unittests --disttests`. - [x] In-line docstrings updated. - [x] Documentation updated, tested `make html` command in the `docs/` folder. --------- Signed-off-by: Juan Pablo de la Cruz Gutiérrez <juampatronics@gmail.com> Signed-off-by: monai-bot <monai.miccai2019@gmail.com> Signed-off-by: elitap <elias.tappeiner@gmx.at> Signed-off-by: Felix Schnabel <f.schnabel@tum.de> Signed-off-by: YanxuanLiu <yanxuanl@nvidia.com> Signed-off-by: ytl0623 <david89062388@gmail.com> Signed-off-by: Dženan Zukić <dzenan.zukic@kitware.com> Signed-off-by: KumoLiu <yunl@nvidia.com> Signed-off-by: YunLiu <55491388+KumoLiu@users.noreply.github.com> Signed-off-by: Ishan Dutta <ishandutta0098@gmail.com> Signed-off-by: dependabot[bot] <support@github.com> Signed-off-by: kaibo <ktang@unc.edu> Signed-off-by: heyufan1995 <heyufan1995@gmail.com> Signed-off-by: binliu <binliu@nvidia.com> Signed-off-by: axel.vlaminck <axel.vlaminck@gmail.com> Signed-off-by: Ibrahim Hadzic <ibrahimhadzic45@gmail.com> Signed-off-by: Behrooz <3968947+drbeh@users.noreply.github.com> Signed-off-by: Timothy Baker <bakertim@umich.edu> Signed-off-by: Mathijs de Boer <m.deboer-41@umcutrecht.nl> Signed-off-by: Fabian Klopfer <fabian.klopfer@ieee.org> Signed-off-by: Lucas Robinet <robinet.lucas@iuct-oncopole.fr> Signed-off-by: Lucas Robinet <67736918+Lucas-rbnt@users.noreply.github.com> Signed-off-by: chaoliu <chaoliu@nvidia.com> Signed-off-by: cxlcl <chaoliucxl@gmail.com> Signed-off-by: chaoliu <chaoliucxl@gmail.com> Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com> Co-authored-by: monai-bot <64792179+monai-bot@users.noreply.github.com> Co-authored-by: elitap <elitap@users.noreply.github.com> Co-authored-by: Felix Schnabel <f.schnabel@tum.de> Co-authored-by: YanxuanLiu <104543031+YanxuanLiu@users.noreply.github.com> Co-authored-by: ytl0623 <david89062388@gmail.com> Co-authored-by: Dženan Zukić <dzenan.zukic@kitware.com> Co-authored-by: Eric Kerfoot <17726042+ericspod@users.noreply.github.com> Co-authored-by: YunLiu <55491388+KumoLiu@users.noreply.github.com> Co-authored-by: Ishan Dutta <ishandutta0098@gmail.com> Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> Co-authored-by: Kaibo Tang <ktang@unc.edu> Co-authored-by: Yufan He <59374597+heyufan1995@users.noreply.github.com> Co-authored-by: binliunls <107988372+binliunls@users.noreply.github.com> Co-authored-by: Ben Murray <ben.murray@gmail.com> Co-authored-by: axel.vlaminck <axel.vlaminck@gmail.com> Co-authored-by: Mingxin Zheng <18563433+mingxin-zheng@users.noreply.github.com> Co-authored-by: Ibrahim Hadzic <ibrahimhadzic45@gmail.com> Co-authored-by: Dr. Behrooz Hashemian <3968947+drbeh@users.noreply.github.com> Co-authored-by: Timothy J. Baker <62781117+tim-the-baker@users.noreply.github.com> Co-authored-by: Mathijs de Boer <8137653+MathijsdeBoer@users.noreply.github.com> Co-authored-by: Mathijs de Boer <m.deboer-41@umcutrecht.nl> Co-authored-by: Fabian Klopfer <fabian.klopfer@ieee.org> Co-authored-by: Yiheng Wang <68361391+yiheng-wang-nv@users.noreply.github.com> Co-authored-by: Lucas Robinet <67736918+Lucas-rbnt@users.noreply.github.com> Co-authored-by: Lucas Robinet <robinet.lucas@iuct-oncopole.fr> Co-authored-by: cxlcl <chaoliucxl@gmail.com>
- Loading branch information
1 parent
97678fa
commit e5bebfc
Showing
7 changed files
with
434 additions
and
0 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,10 @@ | ||
# Copyright (c) MONAI Consortium | ||
# Licensed under the Apache License, Version 2.0 (the "License"); | ||
# you may not use this file except in compliance with the License. | ||
# You may obtain a copy of the License at | ||
# http://www.apache.org/licenses/LICENSE-2.0 | ||
# Unless required by applicable law or agreed to in writing, software | ||
# distributed under the License is distributed on an "AS IS" BASIS, | ||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
# See the License for the specific language governing permissions and | ||
# limitations under the License. |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,173 @@ | ||
# Copyright (c) MONAI Consortium | ||
# Licensed under the Apache License, Version 2.0 (the "License"); | ||
# you may not use this file except in compliance with the License. | ||
# You may obtain a copy of the License at | ||
# http://www.apache.org/licenses/LICENSE-2.0 | ||
# Unless required by applicable law or agreed to in writing, software | ||
# distributed under the License is distributed on an "AS IS" BASIS, | ||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
# See the License for the specific language governing permissions and | ||
# limitations under the License. | ||
|
||
from __future__ import annotations | ||
|
||
from abc import abstractmethod | ||
from math import ceil, sqrt | ||
|
||
import torch | ||
|
||
from ..transform import RandomizableTransform | ||
|
||
__all__ = ["MixUp", "CutMix", "CutOut", "Mixer"] | ||
|
||
|
||
class Mixer(RandomizableTransform): | ||
def __init__(self, batch_size: int, alpha: float = 1.0) -> None: | ||
""" | ||
Mixer is a base class providing the basic logic for the mixup-class of | ||
augmentations. In all cases, we need to sample the mixing weights for each | ||
sample (lambda in the notation used in the papers). Also, pairs of samples | ||
being mixed are picked by randomly shuffling the batch samples. | ||
Args: | ||
batch_size (int): number of samples per batch. That is, samples are expected tp | ||
be of size batchsize x channels [x depth] x height x width. | ||
alpha (float, optional): mixing weights are sampled from the Beta(alpha, alpha) | ||
distribution. Defaults to 1.0, the uniform distribution. | ||
""" | ||
super().__init__() | ||
if alpha <= 0: | ||
raise ValueError(f"Expected positive number, but got {alpha = }") | ||
self.alpha = alpha | ||
self.batch_size = batch_size | ||
|
||
@abstractmethod | ||
def apply(self, data: torch.Tensor): | ||
raise NotImplementedError() | ||
|
||
def randomize(self, data=None) -> None: | ||
""" | ||
Sometimes you need may to apply the same transform to different tensors. | ||
The idea is to get a sample and then apply it with apply() as often | ||
as needed. You need to call this method everytime you apply the transform to a new | ||
batch. | ||
""" | ||
self._params = ( | ||
torch.from_numpy(self.R.beta(self.alpha, self.alpha, self.batch_size)).type(torch.float32), | ||
self.R.permutation(self.batch_size), | ||
) | ||
|
||
|
||
class MixUp(Mixer): | ||
"""MixUp as described in: | ||
Hongyi Zhang, Moustapha Cisse, Yann N. Dauphin, David Lopez-Paz. | ||
mixup: Beyond Empirical Risk Minimization, ICLR 2018 | ||
Class derived from :py:class:`monai.transforms.Mixer`. See corresponding | ||
documentation for details on the constructor parameters. | ||
""" | ||
|
||
def apply(self, data: torch.Tensor): | ||
weight, perm = self._params | ||
nsamples, *dims = data.shape | ||
if len(weight) != nsamples: | ||
raise ValueError(f"Expected batch of size: {len(weight)}, but got {nsamples}") | ||
|
||
if len(dims) not in [3, 4]: | ||
raise ValueError("Unexpected number of dimensions") | ||
|
||
mixweight = weight[(Ellipsis,) + (None,) * len(dims)] | ||
return mixweight * data + (1 - mixweight) * data[perm, ...] | ||
|
||
def __call__(self, data: torch.Tensor, labels: torch.Tensor | None = None): | ||
self.randomize() | ||
if labels is None: | ||
return self.apply(data) | ||
return self.apply(data), self.apply(labels) | ||
|
||
|
||
class CutMix(Mixer): | ||
"""CutMix augmentation as described in: | ||
Sangdoo Yun, Dongyoon Han, Seong Joon Oh, Sanghyuk Chun, Junsuk Choe, Youngjoon Yoo. | ||
CutMix: Regularization Strategy to Train Strong Classifiers with Localizable Features, | ||
ICCV 2019 | ||
Class derived from :py:class:`monai.transforms.Mixer`. See corresponding | ||
documentation for details on the constructor parameters. Here, alpha not only determines | ||
the mixing weight but also the size of the random rectangles used during for mixing. | ||
Please refer to the paper for details. | ||
The most common use case is something close to: | ||
.. code-block:: python | ||
cm = CutMix(batch_size=8, alpha=0.5) | ||
for batch in loader: | ||
images, labels = batch | ||
augimg, auglabels = cm(images, labels) | ||
output = model(augimg) | ||
loss = loss_function(output, auglabels) | ||
... | ||
""" | ||
|
||
def apply(self, data: torch.Tensor): | ||
weights, perm = self._params | ||
nsamples, _, *dims = data.shape | ||
if len(weights) != nsamples: | ||
raise ValueError(f"Expected batch of size: {len(weights)}, but got {nsamples}") | ||
|
||
mask = torch.ones_like(data) | ||
for s, weight in enumerate(weights): | ||
coords = [torch.randint(0, d, size=(1,)) for d in dims] | ||
lengths = [d * sqrt(1 - weight) for d in dims] | ||
idx = [slice(None)] + [slice(c, min(ceil(c + ln), d)) for c, ln, d in zip(coords, lengths, dims)] | ||
mask[s][idx] = 0 | ||
|
||
return mask * data + (1 - mask) * data[perm, ...] | ||
|
||
def apply_on_labels(self, labels: torch.Tensor): | ||
weights, perm = self._params | ||
nsamples, *dims = labels.shape | ||
if len(weights) != nsamples: | ||
raise ValueError(f"Expected batch of size: {len(weights)}, but got {nsamples}") | ||
|
||
mixweight = weights[(Ellipsis,) + (None,) * len(dims)] | ||
return mixweight * labels + (1 - mixweight) * labels[perm, ...] | ||
|
||
def __call__(self, data: torch.Tensor, labels: torch.Tensor | None = None): | ||
self.randomize() | ||
augmented = self.apply(data) | ||
return (augmented, self.apply_on_labels(labels)) if labels is not None else augmented | ||
|
||
|
||
class CutOut(Mixer): | ||
"""Cutout as described in the paper: | ||
Terrance DeVries, Graham W. Taylor. | ||
Improved Regularization of Convolutional Neural Networks with Cutout, | ||
arXiv:1708.04552 | ||
Class derived from :py:class:`monai.transforms.Mixer`. See corresponding | ||
documentation for details on the constructor parameters. Here, alpha not only determines | ||
the mixing weight but also the size of the random rectangles being cut put. | ||
Please refer to the paper for details. | ||
""" | ||
|
||
def apply(self, data: torch.Tensor): | ||
weights, _ = self._params | ||
nsamples, _, *dims = data.shape | ||
if len(weights) != nsamples: | ||
raise ValueError(f"Expected batch of size: {len(weights)}, but got {nsamples}") | ||
|
||
mask = torch.ones_like(data) | ||
for s, weight in enumerate(weights): | ||
coords = [torch.randint(0, d, size=(1,)) for d in dims] | ||
lengths = [d * sqrt(1 - weight) for d in dims] | ||
idx = [slice(None)] + [slice(c, min(ceil(c + ln), d)) for c, ln, d in zip(coords, lengths, dims)] | ||
mask[s][idx] = 0 | ||
|
||
return mask * data | ||
|
||
def __call__(self, data: torch.Tensor): | ||
self.randomize() | ||
return self.apply(data) |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,97 @@ | ||
# Copyright (c) MONAI Consortium | ||
# Licensed under the Apache License, Version 2.0 (the "License"); | ||
# you may not use this file except in compliance with the License. | ||
# You may obtain a copy of the License at | ||
# http://www.apache.org/licenses/LICENSE-2.0 | ||
# Unless required by applicable law or agreed to in writing, software | ||
# distributed under the License is distributed on an "AS IS" BASIS, | ||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
# See the License for the specific language governing permissions and | ||
# limitations under the License. | ||
|
||
from __future__ import annotations | ||
|
||
from monai.config import KeysCollection | ||
from monai.utils.misc import ensure_tuple | ||
|
||
from ..transform import MapTransform | ||
from .array import CutMix, CutOut, MixUp | ||
|
||
__all__ = ["MixUpd", "MixUpD", "MixUpDict", "CutMixd", "CutMixD", "CutMixDict", "CutOutd", "CutOutD", "CutOutDict"] | ||
|
||
|
||
class MixUpd(MapTransform): | ||
""" | ||
Dictionary-based version :py:class:`monai.transforms.MixUp`. | ||
Notice that the mixup transformation will be the same for all entries | ||
for consistency, i.e. images and labels must be applied the same augmenation. | ||
""" | ||
|
||
def __init__( | ||
self, keys: KeysCollection, batch_size: int, alpha: float = 1.0, allow_missing_keys: bool = False | ||
) -> None: | ||
super().__init__(keys, allow_missing_keys) | ||
self.mixup = MixUp(batch_size, alpha) | ||
|
||
def __call__(self, data): | ||
self.mixup.randomize() | ||
result = dict(data) | ||
for k in self.keys: | ||
result[k] = self.mixup.apply(data[k]) | ||
return result | ||
|
||
|
||
class CutMixd(MapTransform): | ||
""" | ||
Dictionary-based version :py:class:`monai.transforms.CutMix`. | ||
Notice that the mixture weights will be the same for all entries | ||
for consistency, i.e. images and labels must be aggregated with the same weights, | ||
but the random crops are not. | ||
""" | ||
|
||
def __init__( | ||
self, | ||
keys: KeysCollection, | ||
batch_size: int, | ||
label_keys: KeysCollection | None = None, | ||
alpha: float = 1.0, | ||
allow_missing_keys: bool = False, | ||
) -> None: | ||
super().__init__(keys, allow_missing_keys) | ||
self.mixer = CutMix(batch_size, alpha) | ||
self.label_keys = ensure_tuple(label_keys) if label_keys is not None else [] | ||
|
||
def __call__(self, data): | ||
self.mixer.randomize() | ||
result = dict(data) | ||
for k in self.keys: | ||
result[k] = self.mixer.apply(data[k]) | ||
for k in self.label_keys: | ||
result[k] = self.mixer.apply_on_labels(data[k]) | ||
return result | ||
|
||
|
||
class CutOutd(MapTransform): | ||
""" | ||
Dictionary-based version :py:class:`monai.transforms.CutOut`. | ||
Notice that the cutout is different for every entry in the dictionary. | ||
""" | ||
|
||
def __init__(self, keys: KeysCollection, batch_size: int, allow_missing_keys: bool = False) -> None: | ||
super().__init__(keys, allow_missing_keys) | ||
self.cutout = CutOut(batch_size) | ||
|
||
def __call__(self, data): | ||
result = dict(data) | ||
self.cutout.randomize() | ||
for k in self.keys: | ||
result[k] = self.cutout(data[k]) | ||
return result | ||
|
||
|
||
MixUpD = MixUpDict = MixUpd | ||
CutMixD = CutMixDict = CutMixd | ||
CutOutD = CutOutDict = CutOutd |
Oops, something went wrong.