diff --git a/CHANGELOG.md b/CHANGELOG.md
index 06e25cc914b..e8cf6dd90de 100644
--- a/CHANGELOG.md
+++ b/CHANGELOG.md
@@ -9,7 +9,7 @@
* Build FFmpeg with zimg to enable zscale filter ([pull #1481](https://github.com/bytedeco/javacpp-presets/pull/1481))
* Enable PulseAudio support for FFmpeg on Linux ([pull #1472](https://github.com/bytedeco/javacpp-presets/pull/1472))
* Virtualize `btCollisionWorld`, `btOverlapFilterCallback`, `btOverlapCallback` from Bullet Physics SDK ([pull #1475](https://github.com/bytedeco/javacpp-presets/pull/1475))
- * Upgrade presets for OpenCV 4.10.0, FFmpeg 7.1, Spinnaker 4.0.0.116 ([pull #1524](https://github.com/bytedeco/javacpp-presets/pull/1524)), DNNL 3.5.3, OpenBLAS 0.3.28, CMINPACK 1.3.9, GSL 2.8, CPython 3.13.0, NumPy 2.1.2, SciPy 1.14.1, LLVM 19.1.2, LibRaw 0.21.2 ([pull #1520](https://github.com/bytedeco/javacpp-presets/pull/1520)), Tesseract 5.4.1, libffi 3.4.6, CUDA 12.6.0, cuDNN 9.3.0, NCCL 2.22.3, nvCOMP 4.0.0, OpenCL 3.0.16, NVIDIA Video Codec SDK 12.2.72, PyTorch 2.4.0 ([pull #1466](https://github.com/bytedeco/javacpp-presets/pull/1466)), SentencePiece 0.2.0, TensorFlow Lite 2.17.0, TensorRT 10.3.0.26, Triton Inference Server 2.48.0, ONNX 1.17.0, ONNX Runtime 1.19.2, TVM 0.17.0, and their dependencies
+ * Upgrade presets for OpenCV 4.10.0, FFmpeg 7.1, Spinnaker 4.0.0.116 ([pull #1524](https://github.com/bytedeco/javacpp-presets/pull/1524)), DNNL 3.5.3, OpenBLAS 0.3.28, CMINPACK 1.3.9, GSL 2.8, CPython 3.13.0, NumPy 2.1.2, SciPy 1.14.1, LLVM 19.1.2, LibRaw 0.21.2 ([pull #1520](https://github.com/bytedeco/javacpp-presets/pull/1520)), Tesseract 5.4.1, libffi 3.4.6, CUDA 12.6.0, cuDNN 9.3.0, NCCL 2.22.3, nvCOMP 4.0.0, OpenCL 3.0.16, NVIDIA Video Codec SDK 12.2.72, PyTorch 2.5.0 ([pull #1466](https://github.com/bytedeco/javacpp-presets/pull/1466)), SentencePiece 0.2.0, TensorFlow Lite 2.17.0, TensorRT 10.3.0.26, Triton Inference Server 2.48.0, ONNX 1.17.0, ONNX Runtime 1.19.2, TVM 0.17.0, and their dependencies
### January 29, 2024 version 1.5.10
* Introduce `macosx-arm64` builds for PyTorch ([pull #1463](https://github.com/bytedeco/javacpp-presets/pull/1463))
diff --git a/README.md b/README.md
index 6d2c7defd3c..faef55e0e89 100644
--- a/README.md
+++ b/README.md
@@ -223,7 +223,7 @@ Each child module in turn relies by default on the included [`cppbuild.sh` scrip
* NVIDIA Video Codec SDK 12.2.x https://developer.nvidia.com/nvidia-video-codec-sdk
* OpenCL 3.0.x https://github.com/KhronosGroup/OpenCL-ICD-Loader
* MXNet 1.9.x https://github.com/apache/incubator-mxnet
- * PyTorch 2.4.x https://github.com/pytorch/pytorch
+ * PyTorch 2.5.x https://github.com/pytorch/pytorch
* SentencePiece 0.2.0 https://github.com/google/sentencepiece
* TensorFlow 1.15.x https://github.com/tensorflow/tensorflow
* TensorFlow Lite 2.17.x https://github.com/tensorflow/tensorflow
diff --git a/platform/pom.xml b/platform/pom.xml
index db4f43e431b..937907d9633 100644
--- a/platform/pom.xml
+++ b/platform/pom.xml
@@ -292,7 +292,7 @@
org.bytedeco
pytorch-platform
- 2.4.0-${project.version}
+ 2.5.0-${project.version}
org.bytedeco
diff --git a/pytorch/README.md b/pytorch/README.md
index e5cccb5525b..80a65ffa627 100644
--- a/pytorch/README.md
+++ b/pytorch/README.md
@@ -9,7 +9,7 @@ Introduction
------------
This directory contains the JavaCPP Presets module for:
- * PyTorch 2.4.0 https://pytorch.org/
+ * PyTorch 2.5.0 https://pytorch.org/
Please refer to the parent README.md file for more detailed information about the JavaCPP Presets.
@@ -48,14 +48,14 @@ We can use [Maven 3](http://maven.apache.org/) to download and install automatic
org.bytedeco
pytorch-platform
- 2.4.0-1.5.11-SNAPSHOT
+ 2.5.0-1.5.11-SNAPSHOT
org.bytedeco
pytorch-platform-gpu
- 2.4.0-1.5.11-SNAPSHOT
+ 2.5.0-1.5.11-SNAPSHOT
diff --git a/pytorch/cppbuild.sh b/pytorch/cppbuild.sh
index 3512f84a4de..7434bfa9eeb 100755
--- a/pytorch/cppbuild.sh
+++ b/pytorch/cppbuild.sh
@@ -38,7 +38,7 @@ if [[ $PLATFORM == windows* ]]; then
export PYTHON_BIN_PATH=$(which python.exe)
fi
-PYTORCH_VERSION=2.4.1
+PYTORCH_VERSION=2.5.0
export PYTORCH_BUILD_VERSION="$PYTORCH_VERSION"
export PYTORCH_BUILD_NUMBER=1
@@ -129,7 +129,7 @@ mkdir -p "$PYTHON_INSTALL_PATH"
export CFLAGS="-I$CPYTHON_PATH/include/ -I$PYTHON_LIB_PATH/include/python/ -L$CPYTHON_PATH/lib/ -L$CPYTHON_PATH/libs/"
export PYTHONNOUSERSITE=1
-$PYTHON_BIN_PATH -m pip install --target=$PYTHON_LIB_PATH setuptools==67.6.1 pyyaml==6.0.1 typing_extensions==4.8.0
+$PYTHON_BIN_PATH -m pip install --target=$PYTHON_LIB_PATH setuptools==67.6.1 pyyaml==6.0.2 typing_extensions==4.8.0
case $PLATFORM in
linux-x86)
@@ -184,6 +184,7 @@ sedinplace 's/ build_deps()/ build_deps(); sys.exit()/g' setup.py
sedinplace 's/AND NOT DEFINED ENV{CUDAHOSTCXX}//g' cmake/public/cuda.cmake
sedinplace 's/CMAKE_CUDA_FLAGS "/CMAKE_CUDA_FLAGS " --use-local-env /g' CMakeLists.txt
+sedinplace '/pycore_opcode.h/d' torch/csrc/dynamo/cpython_defs.c functorch/csrc/dim/dim*
sedinplace 's/using ExpandingArrayDouble/public: using ExpandingArrayDouble/g' ./torch/csrc/api/include/torch/nn/options/pooling.h
# allow setting the build directory and passing CUDA options
@@ -192,6 +193,7 @@ sedinplace 's/var.startswith(("BUILD_", "USE_", "CMAKE_"))/var.startswith(("BUIL
# allow resizing std::vector
sedinplace 's/TensorIndex(c10::nullopt_t)/TensorIndex(c10::nullopt_t none = None)/g' aten/src/ATen/TensorIndexing.h
+sedinplace 's/TensorIndex(std::nullopt_t)/TensorIndex(std::nullopt_t none = None)/g' aten/src/ATen/TensorIndexing.h
# add missing declarations
sedinplace '/using ExampleType = ExampleType_;/a\
diff --git a/pytorch/platform/gpu/pom.xml b/pytorch/platform/gpu/pom.xml
index 387956fdb79..0d8bd29ae81 100644
--- a/pytorch/platform/gpu/pom.xml
+++ b/pytorch/platform/gpu/pom.xml
@@ -12,7 +12,7 @@
org.bytedeco
pytorch-platform-gpu
- 2.4.0-${project.parent.version}
+ 2.5.0-${project.parent.version}
JavaCPP Presets Platform GPU for PyTorch
diff --git a/pytorch/platform/pom.xml b/pytorch/platform/pom.xml
index a3ab5725b2d..4ff46efa3fd 100644
--- a/pytorch/platform/pom.xml
+++ b/pytorch/platform/pom.xml
@@ -12,7 +12,7 @@
org.bytedeco
pytorch-platform
- 2.4.0-${project.parent.version}
+ 2.5.0-${project.parent.version}
JavaCPP Presets Platform for PyTorch
diff --git a/pytorch/pom.xml b/pytorch/pom.xml
index 3b1a9856d95..5f697e12b53 100644
--- a/pytorch/pom.xml
+++ b/pytorch/pom.xml
@@ -11,7 +11,7 @@
org.bytedeco
pytorch
- 2.4.0-${project.parent.version}
+ 2.5.0-${project.parent.version}
JavaCPP Presets for PyTorch
diff --git a/pytorch/samples/pom.xml b/pytorch/samples/pom.xml
index ef136d7088d..1add9a6b25c 100644
--- a/pytorch/samples/pom.xml
+++ b/pytorch/samples/pom.xml
@@ -12,14 +12,14 @@
org.bytedeco
pytorch-platform
- 2.4.0-1.5.11-SNAPSHOT
+ 2.5.0-1.5.11-SNAPSHOT
org.bytedeco
pytorch-platform-gpu
- 2.4.0-1.5.11-SNAPSHOT
+ 2.5.0-1.5.11-SNAPSHOT
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/AcceleratorHooksInterface.java b/pytorch/src/gen/java/org/bytedeco/pytorch/AcceleratorHooksInterface.java
index 9fb904a24ff..7699b290118 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/AcceleratorHooksInterface.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/AcceleratorHooksInterface.java
@@ -45,4 +45,8 @@ public class AcceleratorHooksInterface extends Pointer {
public native @Cast("c10::DeviceIndex") byte exchangeDevice(@Cast("c10::DeviceIndex") byte device);
public native @Cast("c10::DeviceIndex") byte maybeExchangeDevice(@Cast("c10::DeviceIndex") byte device);
+
+ public native @Cast("bool") boolean isPinnedPtr(@Const Pointer data);
+
+ public native Allocator getPinnedMemoryAllocator();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/AdaptiveAvgPool1dImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/AdaptiveAvgPool1dImplCloneable.java
index 6cd0d41375f..08eb0049aea 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/AdaptiveAvgPool1dImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/AdaptiveAvgPool1dImplCloneable.java
@@ -33,6 +33,6 @@ public class AdaptiveAvgPool1dImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/AdaptiveAvgPool2dImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/AdaptiveAvgPool2dImplCloneable.java
index 014f7184f1b..39c0ce609aa 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/AdaptiveAvgPool2dImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/AdaptiveAvgPool2dImplCloneable.java
@@ -33,6 +33,6 @@ public class AdaptiveAvgPool2dImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/AdaptiveAvgPool3dImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/AdaptiveAvgPool3dImplCloneable.java
index 680a8c29567..d35872413fe 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/AdaptiveAvgPool3dImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/AdaptiveAvgPool3dImplCloneable.java
@@ -33,6 +33,6 @@ public class AdaptiveAvgPool3dImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/AdaptiveLogSoftmaxWithLossImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/AdaptiveLogSoftmaxWithLossImplCloneable.java
index a1dfcbd3a2d..8dc7c4a65a1 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/AdaptiveLogSoftmaxWithLossImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/AdaptiveLogSoftmaxWithLossImplCloneable.java
@@ -33,6 +33,6 @@ public class AdaptiveLogSoftmaxWithLossImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/AdaptiveMaxPool1dImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/AdaptiveMaxPool1dImplCloneable.java
index 7182a820e06..96c3dea0bb2 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/AdaptiveMaxPool1dImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/AdaptiveMaxPool1dImplCloneable.java
@@ -33,6 +33,6 @@ public class AdaptiveMaxPool1dImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/AdaptiveMaxPool2dImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/AdaptiveMaxPool2dImplCloneable.java
index 1145fd54eb9..f35a31b6e95 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/AdaptiveMaxPool2dImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/AdaptiveMaxPool2dImplCloneable.java
@@ -33,6 +33,6 @@ public class AdaptiveMaxPool2dImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/AdaptiveMaxPool3dImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/AdaptiveMaxPool3dImplCloneable.java
index c73ce111401..f42c7694a04 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/AdaptiveMaxPool3dImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/AdaptiveMaxPool3dImplCloneable.java
@@ -33,6 +33,6 @@ public class AdaptiveMaxPool3dImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/AlphaDropoutImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/AlphaDropoutImplCloneable.java
index db38d7884c4..3d96932be20 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/AlphaDropoutImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/AlphaDropoutImplCloneable.java
@@ -33,6 +33,6 @@ public class AlphaDropoutImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/AnyModule.java b/pytorch/src/gen/java/org/bytedeco/pytorch/AnyModule.java
index 3991d3564c1..89761c5e532 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/AnyModule.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/AnyModule.java
@@ -391,7 +391,7 @@ public class AnyModule extends Pointer {
/** Creates a deep copy of an {@code AnyModule} if it contains a module, else an
* empty {@code AnyModule} if it is empty. */
- public native @ByVal AnyModule clone(@ByVal(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ public native @ByVal AnyModule clone(@ByVal(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @ByVal AnyModule clone();
/** Assigns a module to the {@code AnyModule} (to circumvent the explicit
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/Argument.java b/pytorch/src/gen/java/org/bytedeco/pytorch/Argument.java
index 9a807c22489..141bc6808ab 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/Argument.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/Argument.java
@@ -37,50 +37,50 @@ public class Argument extends Pointer {
public Argument(
@StdString BytePointer name/*=""*/,
@Const @ByRef(nullValue = "c10::TypePtr(nullptr)") Type.TypePtr type,
- @ByVal(nullValue = "std::optional(c10::nullopt)") IntOptional N,
- @ByVal(nullValue = "std::optional(c10::nullopt)") IValueOptional default_value,
+ @ByVal(nullValue = "std::optional(std::nullopt)") IntOptional N,
+ @ByVal(nullValue = "std::optional(std::nullopt)") IValueOptional default_value,
@Cast("bool") boolean kwarg_only/*=false*/,
- @ByVal(nullValue = "std::optional(c10::nullopt)") AliasInfoOptional alias_info) { super((Pointer)null); allocate(name, type, N, default_value, kwarg_only, alias_info); }
+ @ByVal(nullValue = "std::optional(std::nullopt)") AliasInfoOptional alias_info) { super((Pointer)null); allocate(name, type, N, default_value, kwarg_only, alias_info); }
private native void allocate(
@StdString BytePointer name/*=""*/,
@Const @ByRef(nullValue = "c10::TypePtr(nullptr)") Type.TypePtr type,
- @ByVal(nullValue = "std::optional(c10::nullopt)") IntOptional N,
- @ByVal(nullValue = "std::optional(c10::nullopt)") IValueOptional default_value,
+ @ByVal(nullValue = "std::optional(std::nullopt)") IntOptional N,
+ @ByVal(nullValue = "std::optional(std::nullopt)") IValueOptional default_value,
@Cast("bool") boolean kwarg_only/*=false*/,
- @ByVal(nullValue = "std::optional(c10::nullopt)") AliasInfoOptional alias_info);
+ @ByVal(nullValue = "std::optional(std::nullopt)") AliasInfoOptional alias_info);
public Argument() { super((Pointer)null); allocate(); }
private native void allocate();
public Argument(
@StdString String name/*=""*/,
@Const @ByRef(nullValue = "c10::TypePtr(nullptr)") Type.TypePtr type,
- @ByVal(nullValue = "std::optional(c10::nullopt)") IntOptional N,
- @ByVal(nullValue = "std::optional(c10::nullopt)") IValueOptional default_value,
+ @ByVal(nullValue = "std::optional(std::nullopt)") IntOptional N,
+ @ByVal(nullValue = "std::optional(std::nullopt)") IValueOptional default_value,
@Cast("bool") boolean kwarg_only/*=false*/,
- @ByVal(nullValue = "std::optional(c10::nullopt)") AliasInfoOptional alias_info) { super((Pointer)null); allocate(name, type, N, default_value, kwarg_only, alias_info); }
+ @ByVal(nullValue = "std::optional(std::nullopt)") AliasInfoOptional alias_info) { super((Pointer)null); allocate(name, type, N, default_value, kwarg_only, alias_info); }
private native void allocate(
@StdString String name/*=""*/,
@Const @ByRef(nullValue = "c10::TypePtr(nullptr)") Type.TypePtr type,
- @ByVal(nullValue = "std::optional(c10::nullopt)") IntOptional N,
- @ByVal(nullValue = "std::optional(c10::nullopt)") IValueOptional default_value,
+ @ByVal(nullValue = "std::optional(std::nullopt)") IntOptional N,
+ @ByVal(nullValue = "std::optional(std::nullopt)") IValueOptional default_value,
@Cast("bool") boolean kwarg_only/*=false*/,
- @ByVal(nullValue = "std::optional(c10::nullopt)") AliasInfoOptional alias_info);
+ @ByVal(nullValue = "std::optional(std::nullopt)") AliasInfoOptional alias_info);
public Argument(
@StdString BytePointer name,
@ByVal Type.TypePtr fake_type,
@ByVal Type.TypePtr real_type,
- @ByVal(nullValue = "std::optional(c10::nullopt)") IntOptional N,
- @ByVal(nullValue = "std::optional(c10::nullopt)") IValueOptional default_value,
+ @ByVal(nullValue = "std::optional(std::nullopt)") IntOptional N,
+ @ByVal(nullValue = "std::optional(std::nullopt)") IValueOptional default_value,
@Cast("bool") boolean kwarg_only/*=false*/,
- @ByVal(nullValue = "std::optional(c10::nullopt)") AliasInfoOptional alias_info) { super((Pointer)null); allocate(name, fake_type, real_type, N, default_value, kwarg_only, alias_info); }
+ @ByVal(nullValue = "std::optional(std::nullopt)") AliasInfoOptional alias_info) { super((Pointer)null); allocate(name, fake_type, real_type, N, default_value, kwarg_only, alias_info); }
private native void allocate(
@StdString BytePointer name,
@ByVal Type.TypePtr fake_type,
@ByVal Type.TypePtr real_type,
- @ByVal(nullValue = "std::optional(c10::nullopt)") IntOptional N,
- @ByVal(nullValue = "std::optional(c10::nullopt)") IValueOptional default_value,
+ @ByVal(nullValue = "std::optional(std::nullopt)") IntOptional N,
+ @ByVal(nullValue = "std::optional(std::nullopt)") IValueOptional default_value,
@Cast("bool") boolean kwarg_only/*=false*/,
- @ByVal(nullValue = "std::optional(c10::nullopt)") AliasInfoOptional alias_info);
+ @ByVal(nullValue = "std::optional(std::nullopt)") AliasInfoOptional alias_info);
public Argument(
@StdString BytePointer name,
@ByVal Type.TypePtr fake_type,
@@ -93,18 +93,18 @@ public Argument(
@StdString String name,
@ByVal Type.TypePtr fake_type,
@ByVal Type.TypePtr real_type,
- @ByVal(nullValue = "std::optional(c10::nullopt)") IntOptional N,
- @ByVal(nullValue = "std::optional(c10::nullopt)") IValueOptional default_value,
+ @ByVal(nullValue = "std::optional(std::nullopt)") IntOptional N,
+ @ByVal(nullValue = "std::optional(std::nullopt)") IValueOptional default_value,
@Cast("bool") boolean kwarg_only/*=false*/,
- @ByVal(nullValue = "std::optional(c10::nullopt)") AliasInfoOptional alias_info) { super((Pointer)null); allocate(name, fake_type, real_type, N, default_value, kwarg_only, alias_info); }
+ @ByVal(nullValue = "std::optional(std::nullopt)") AliasInfoOptional alias_info) { super((Pointer)null); allocate(name, fake_type, real_type, N, default_value, kwarg_only, alias_info); }
private native void allocate(
@StdString String name,
@ByVal Type.TypePtr fake_type,
@ByVal Type.TypePtr real_type,
- @ByVal(nullValue = "std::optional(c10::nullopt)") IntOptional N,
- @ByVal(nullValue = "std::optional(c10::nullopt)") IValueOptional default_value,
+ @ByVal(nullValue = "std::optional(std::nullopt)") IntOptional N,
+ @ByVal(nullValue = "std::optional(std::nullopt)") IValueOptional default_value,
@Cast("bool") boolean kwarg_only/*=false*/,
- @ByVal(nullValue = "std::optional(c10::nullopt)") AliasInfoOptional alias_info);
+ @ByVal(nullValue = "std::optional(std::nullopt)") AliasInfoOptional alias_info);
public Argument(
@StdString String name,
@ByVal Type.TypePtr fake_type,
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/AutogradCompilerCall.java b/pytorch/src/gen/java/org/bytedeco/pytorch/AutogradCompilerCall.java
index 6a5e475e9f5..4bbcad83cba 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/AutogradCompilerCall.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/AutogradCompilerCall.java
@@ -43,6 +43,7 @@ public class AutogradCompilerCall extends Pointer {
public native @ByRef @NoOffset TensorArgs tensor_args(); public native AutogradCompilerCall tensor_args(TensorArgs setter);
public native @StdVector @NoOffset SizeInput all_size_inputs(); public native AutogradCompilerCall all_size_inputs(SizeInput setter);
+ public native @ByRef LiftedIValueArgs lifted_ivalue_args(); public native AutogradCompilerCall lifted_ivalue_args(LiftedIValueArgs setter);
public native @ByRef @Cast("std::vector*") @NoOffset LongVector dyn_size_inputs(); public native AutogradCompilerCall dyn_size_inputs(LongVector setter);
public native @ByRef @NoOffset NodeCalls node_calls(); public native AutogradCompilerCall node_calls(NodeCalls setter);
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/AutogradContext.java b/pytorch/src/gen/java/org/bytedeco/pytorch/AutogradContext.java
index 18f6e7a5b8d..67ef2a58781 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/AutogradContext.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/AutogradContext.java
@@ -43,7 +43,6 @@ public class AutogradContext extends Pointer {
/** Can be used to save non-variable data for {@code backward}. */
- // NOLINTNEXTLINE(cppcoreguidelines-non-private-member-variables-in-classes)
/** Saves the list of variables for a future call to {@code backward}. This
* should be called at most once from inside of {@code forward}. */
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/AvgPool1dImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/AvgPool1dImplCloneable.java
index 7dae207aa06..43036e23092 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/AvgPool1dImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/AvgPool1dImplCloneable.java
@@ -33,6 +33,6 @@ public class AvgPool1dImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/AvgPool2dImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/AvgPool2dImplCloneable.java
index 9426de9839a..b1d2c5868f9 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/AvgPool2dImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/AvgPool2dImplCloneable.java
@@ -33,6 +33,6 @@ public class AvgPool2dImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/AvgPool3dImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/AvgPool3dImplCloneable.java
index 49b36b2077f..ddcc3b31ae6 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/AvgPool3dImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/AvgPool3dImplCloneable.java
@@ -33,6 +33,6 @@ public class AvgPool3dImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/BCELossImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/BCELossImplCloneable.java
index 6c4b5a0eb53..d5659cbb7ed 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/BCELossImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/BCELossImplCloneable.java
@@ -33,6 +33,6 @@ public class BCELossImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/BCEWithLogitsLossImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/BCEWithLogitsLossImplCloneable.java
index b1b8dcbec0d..66ed07f193a 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/BCEWithLogitsLossImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/BCEWithLogitsLossImplCloneable.java
@@ -33,6 +33,6 @@ public class BCEWithLogitsLossImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/BFloat16.java b/pytorch/src/gen/java/org/bytedeco/pytorch/BFloat16.java
index 4a93c34caac..d2bce48148c 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/BFloat16.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/BFloat16.java
@@ -44,12 +44,14 @@ public class BFloat16 extends Pointer {
public BFloat16(@Cast("unsigned short") short bits, @ByVal from_bits_t arg1) { super((Pointer)null); allocate(bits, arg1); }
private native void allocate(@Cast("unsigned short") short bits, @ByVal from_bits_t arg1);
- public BFloat16(float value) { super((Pointer)null); allocate(value); }
- private native void allocate(float value);
+ /* implicit */ public BFloat16(float value) { super((Pointer)null); allocate(value); }
+private native void allocate(float value);
public native @Name("operator float") float asFloat();
// #if defined(__CUDACC__) && !defined(USE_ROCM)
// #endif
+// #if defined(__HIPCC__) && defined(USE_ROCM)
+// #endif
// #if defined(SYCL_EXT_ONEAPI_BFLOAT16_MATH_FUNCTIONS)
// #endif
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/BatchNorm1dImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/BatchNorm1dImplCloneable.java
index a487c9c2454..b71e14f0cd7 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/BatchNorm1dImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/BatchNorm1dImplCloneable.java
@@ -33,6 +33,6 @@ public class BatchNorm1dImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/BatchNorm2dImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/BatchNorm2dImplCloneable.java
index e07f2e7eb04..d0918ec9f81 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/BatchNorm2dImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/BatchNorm2dImplCloneable.java
@@ -33,6 +33,6 @@ public class BatchNorm2dImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/BatchNorm3dImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/BatchNorm3dImplCloneable.java
index 0e0c66b1df1..8ab453e2a3c 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/BatchNorm3dImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/BatchNorm3dImplCloneable.java
@@ -33,6 +33,6 @@ public class BatchNorm3dImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/BilinearImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/BilinearImplCloneable.java
index 7bda99227c6..36424e488c1 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/BilinearImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/BilinearImplCloneable.java
@@ -33,6 +33,6 @@ public class BilinearImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/BuiltinModule.java b/pytorch/src/gen/java/org/bytedeco/pytorch/BuiltinModule.java
index bb6afe3ba45..bf470adbd4d 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/BuiltinModule.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/BuiltinModule.java
@@ -25,12 +25,12 @@ public class BuiltinModule extends SugaredValue {
/** Pointer cast constructor. Invokes {@link Pointer#Pointer(Pointer)}. */
public BuiltinModule(Pointer p) { super(p); }
- public BuiltinModule(@StdString BytePointer name, @ByVal(nullValue = "std::optional(at::nullopt)") LongOptional version) { super((Pointer)null); allocate(name, version); }
- private native void allocate(@StdString BytePointer name, @ByVal(nullValue = "std::optional(at::nullopt)") LongOptional version);
+ public BuiltinModule(@StdString BytePointer name, @ByVal(nullValue = "std::optional(std::nullopt)") LongOptional version) { super((Pointer)null); allocate(name, version); }
+ private native void allocate(@StdString BytePointer name, @ByVal(nullValue = "std::optional(std::nullopt)") LongOptional version);
public BuiltinModule(@StdString BytePointer name) { super((Pointer)null); allocate(name); }
private native void allocate(@StdString BytePointer name);
- public BuiltinModule(@StdString String name, @ByVal(nullValue = "std::optional(at::nullopt)") LongOptional version) { super((Pointer)null); allocate(name, version); }
- private native void allocate(@StdString String name, @ByVal(nullValue = "std::optional(at::nullopt)") LongOptional version);
+ public BuiltinModule(@StdString String name, @ByVal(nullValue = "std::optional(std::nullopt)") LongOptional version) { super((Pointer)null); allocate(name, version); }
+ private native void allocate(@StdString String name, @ByVal(nullValue = "std::optional(std::nullopt)") LongOptional version);
public BuiltinModule(@StdString String name) { super((Pointer)null); allocate(name); }
private native void allocate(@StdString String name);
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/CELUImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/CELUImplCloneable.java
index c5755d11b03..d6b0878acc0 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/CELUImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/CELUImplCloneable.java
@@ -33,6 +33,6 @@ public class CELUImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/CTCLossImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/CTCLossImplCloneable.java
index a10426829c4..e9bbaa3e6d4 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/CTCLossImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/CTCLossImplCloneable.java
@@ -33,6 +33,6 @@ public class CTCLossImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/CUDAHooksInterface.java b/pytorch/src/gen/java/org/bytedeco/pytorch/CUDAHooksInterface.java
index 2ce28032170..61b0355be09 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/CUDAHooksInterface.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/CUDAHooksInterface.java
@@ -65,7 +65,7 @@ public class CUDAHooksInterface extends AcceleratorHooksInterface {
public native @ByVal Device getDeviceFromPtr(Pointer arg0);
- public native @Cast("bool") boolean isPinnedPtr(@Const Pointer arg0);
+ public native @Cast("bool") boolean isPinnedPtr(@Const Pointer data);
public native @Cast("bool") boolean hasCUDA();
@@ -119,5 +119,8 @@ public class CUDAHooksInterface extends AcceleratorHooksInterface {
public native int getNumGPUs();
+// #ifdef USE_ROCM
+// #endif
+
public native void deviceSynchronize(@Cast("c10::DeviceIndex") byte arg0);
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/ChunkBatchDataset.java b/pytorch/src/gen/java/org/bytedeco/pytorch/ChunkBatchDataset.java
index 3756cba2e7b..16262b3cd70 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/ChunkBatchDataset.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/ChunkBatchDataset.java
@@ -30,7 +30,8 @@ public class ChunkBatchDataset extends Pointer {
/** Returns a batch of data given an index. */
public native @ByVal ExampleVectorOptional get_batch(@Cast("size_t") long request);
- /** Returns the size of the dataset, or an empty optional if it is unsized. */
+ /** Returns the size of the dataset, or an empty std::optional if it is
+ * unsized. */
public native @ByVal SizeTOptional size();
/** Creates a {@code MapDataset} that applies the given {@code transform} to this dataset. */
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/ChunkBatchSharedBatchDataset.java b/pytorch/src/gen/java/org/bytedeco/pytorch/ChunkBatchSharedBatchDataset.java
index 225630df0ab..14e40403032 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/ChunkBatchSharedBatchDataset.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/ChunkBatchSharedBatchDataset.java
@@ -30,7 +30,8 @@ public class ChunkBatchSharedBatchDataset extends Pointer {
/** Returns a batch of data given an index. */
public native @ByVal ExampleVectorOptional get_batch(@Cast("size_t") long request);
- /** Returns the size of the dataset, or an empty optional if it is unsized. */
+ /** Returns the size of the dataset, or an empty std::optional if it is
+ * unsized. */
public native @ByVal SizeTOptional size();
/** Creates a {@code MapDataset} that applies the given {@code transform} to this dataset. */
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/ChunkBatchSharedTensorBatchDataset.java b/pytorch/src/gen/java/org/bytedeco/pytorch/ChunkBatchSharedTensorBatchDataset.java
index 25f7ef669d2..b1949c1ee90 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/ChunkBatchSharedTensorBatchDataset.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/ChunkBatchSharedTensorBatchDataset.java
@@ -30,7 +30,8 @@ public class ChunkBatchSharedTensorBatchDataset extends Pointer {
/** Returns a batch of data given an index. */
public native @ByVal TensorExampleVectorOptional get_batch(@Cast("size_t") long request);
- /** Returns the size of the dataset, or an empty optional if it is unsized. */
+ /** Returns the size of the dataset, or an empty std::optional if it is
+ * unsized. */
public native @ByVal SizeTOptional size();
/** Creates a {@code MapDataset} that applies the given {@code transform} to this dataset. */
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/ChunkMapBatchDataset.java b/pytorch/src/gen/java/org/bytedeco/pytorch/ChunkMapBatchDataset.java
index 3a8b2770d2f..d6a4b57f3c9 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/ChunkMapBatchDataset.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/ChunkMapBatchDataset.java
@@ -31,7 +31,8 @@ public class ChunkMapBatchDataset extends Pointer {
public native @ByVal ExampleVector get_batch(@ByVal SizeTArrayRef request);
public native @ByVal ExampleVector get_batch(@ByVal @Cast({"size_t*", "c10::ArrayRef", "std::vector&"}) @StdVector("size_t") long... request);
- /** Returns the size of the dataset, or an empty optional if it is unsized. */
+ /** Returns the size of the dataset, or an empty std::optional if it is
+ * unsized. */
public native @ByVal SizeTOptional size();
/** Creates a {@code MapDataset} that applies the given {@code transform} to this dataset. */
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/ChunkMapTensorBatchDataset.java b/pytorch/src/gen/java/org/bytedeco/pytorch/ChunkMapTensorBatchDataset.java
index b0e5062563b..2daf765c87e 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/ChunkMapTensorBatchDataset.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/ChunkMapTensorBatchDataset.java
@@ -31,7 +31,8 @@ public class ChunkMapTensorBatchDataset extends Pointer {
public native @ByVal TensorExampleVector get_batch(@ByVal SizeTArrayRef request);
public native @ByVal TensorExampleVector get_batch(@ByVal @Cast({"size_t*", "c10::ArrayRef", "std::vector&"}) @StdVector("size_t") long... request);
- /** Returns the size of the dataset, or an empty optional if it is unsized. */
+ /** Returns the size of the dataset, or an empty std::optional if it is
+ * unsized. */
public native @ByVal SizeTOptional size();
/** Creates a {@code MapDataset} that applies the given {@code transform} to this dataset. */
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/ChunkTensorBatchDataset.java b/pytorch/src/gen/java/org/bytedeco/pytorch/ChunkTensorBatchDataset.java
index 81fdaf27545..97882153239 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/ChunkTensorBatchDataset.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/ChunkTensorBatchDataset.java
@@ -30,7 +30,8 @@ public class ChunkTensorBatchDataset extends Pointer {
/** Returns a batch of data given an index. */
public native @ByVal TensorExampleVectorOptional get_batch(@Cast("size_t") long request);
- /** Returns the size of the dataset, or an empty optional if it is unsized. */
+ /** Returns the size of the dataset, or an empty std::optional if it is
+ * unsized. */
public native @ByVal SizeTOptional size();
/** Creates a {@code MapDataset} that applies the given {@code transform} to this dataset. */
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/ClassType.java b/pytorch/src/gen/java/org/bytedeco/pytorch/ClassType.java
index 4118b5e9e37..e3108caf0e5 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/ClassType.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/ClassType.java
@@ -227,10 +227,10 @@ public static class Property extends Pointer {
public native @Const @ByRef FunctionVector getForwardPreHooks();
public native void checkForwardPreHookSchema(
- int pre_hook_idx,
+ @Cast("size_t") long pre_hook_idx,
@Const @ByRef FunctionSchema pre_hook_schema);
public native void checkForwardHookSchema(
- int hook_idx,
+ @Cast("size_t") long hook_idx,
@Const @ByRef FunctionSchema hook_schema);
public native void addMethod(Function method);
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/CompilationUnit.java b/pytorch/src/gen/java/org/bytedeco/pytorch/CompilationUnit.java
index 954bb5cc311..09fe8c60f68 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/CompilationUnit.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/CompilationUnit.java
@@ -72,7 +72,7 @@ public enum FunctionType { Method(0), Hook(1), PreHook(2);
@Const @ByRef ResolverVector defResolvers,
@Const Self self,
@Cast("bool") boolean shouldMangle/*=false*/,
- @ByVal(nullValue = "std::optional(c10::nullopt)") SizeTOptional operator_set_version);
+ @ByVal(nullValue = "std::optional(std::nullopt)") SizeTOptional operator_set_version);
public native @ByVal FunctionVector define(
@Const @ByRef QualifiedNameOptional prefix,
@Const @ByRef PropertyVector properties,
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/CompiledNodeArgs.java b/pytorch/src/gen/java/org/bytedeco/pytorch/CompiledNodeArgs.java
index 1e6ede7a4dd..a8fb9170e18 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/CompiledNodeArgs.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/CompiledNodeArgs.java
@@ -29,6 +29,8 @@ public class CompiledNodeArgs extends Pointer {
public native void collect(@Const @ByRef Tensor t);
public native void collect(@Const @ByRef SymInt t);
+ public native void collect(@Cast("const c10::ArrayRef*") @ByRef Pointer t, @Cast("bool") boolean is_output);
+ public native void collect(@Const @ByRef IValue iv, @Cast("bool") boolean nested/*=false*/);
public native void collect(@Const @ByRef IValue iv);
public native void collect(@Const @ByRef Scalar t);
public native void collect(@Const @ByRef TensorOptions t);
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/ConstantPad1dImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/ConstantPad1dImplCloneable.java
index 6879bed85d7..1b0e82b95b7 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/ConstantPad1dImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/ConstantPad1dImplCloneable.java
@@ -33,6 +33,6 @@ public class ConstantPad1dImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/ConstantPad2dImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/ConstantPad2dImplCloneable.java
index 34d9fde521d..7d715830bf8 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/ConstantPad2dImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/ConstantPad2dImplCloneable.java
@@ -33,6 +33,6 @@ public class ConstantPad2dImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/ConstantPad3dImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/ConstantPad3dImplCloneable.java
index 35b21d86868..c8b5791727f 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/ConstantPad3dImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/ConstantPad3dImplCloneable.java
@@ -33,6 +33,6 @@ public class ConstantPad3dImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/Context.java b/pytorch/src/gen/java/org/bytedeco/pytorch/Context.java
index 148e6fd68be..0eb75a84e05 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/Context.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/Context.java
@@ -39,11 +39,18 @@ public class Context extends Pointer {
public native @Const @ByRef Generator defaultGenerator(@ByVal Device device);
public native @Const @ByRef AcceleratorHooksInterface getAcceleratorHooksInterface(
- @ByVal(nullValue = "std::optional(c10::nullopt)") DeviceTypeOptional opt_device_type);
+ @ByVal(nullValue = "std::optional(std::nullopt)") DeviceTypeOptional opt_device_type);
public native @Const @ByRef AcceleratorHooksInterface getAcceleratorHooksInterface();
public native @ByVal Device getDeviceFromPtr(Pointer data, DeviceType device_type);
public native @ByVal Device getDeviceFromPtr(Pointer data, @Cast("c10::DeviceType") byte device_type);
- public static native @Cast("bool") boolean isPinnedPtr(@Const Pointer data);
+ public native @Cast("bool") boolean isPinnedPtr(
+ @Const Pointer data,
+ @ByVal(nullValue = "std::optional(std::nullopt)") DeviceTypeOptional device_type);
+ public native @Cast("bool") boolean isPinnedPtr(
+ @Const Pointer data);
+ public native Allocator getPinnedMemoryAllocator(
+ @ByVal(nullValue = "std::optional(std::nullopt)") DeviceTypeOptional device_type);
+ public native Allocator getPinnedMemoryAllocator();
public static native @Cast("bool") boolean hasOpenMP();
public static native @Cast("bool") boolean hasMKL();
public static native @Cast("bool") boolean hasLAPACK();
@@ -89,6 +96,8 @@ public class Context extends Pointer {
public native void setBenchmarkLimitCuDNN(int arg0);
public native @Cast("bool") boolean deterministicCuDNN();
public native void setDeterministicCuDNN(@Cast("bool") boolean arg0);
+ public native @Cast("bool") boolean deterministicMkldnn();
+ public native void setDeterministicMkldnn(@Cast("bool") boolean arg0);
public native @Cast("bool") boolean userEnabledNNPACK();
public native void setUserEnabledNNPACK(@Cast("bool") boolean e);
@@ -115,6 +124,12 @@ public class Context extends Pointer {
public native void setSDPUseCuDNN(@Cast("bool") boolean arg0);
public native @Cast("bool") boolean userEnabledCuDNNSDP();
+ public native void setAllowFP16BF16ReductionMathSDP(@Cast("bool") boolean arg0);
+ public native @Cast("bool") boolean allowFP16BF16ReductionMathSDP();
+
+ public native void setSDPUseOverrideable(@Cast("bool") boolean arg0);
+ public native @Cast("bool") boolean userEnabledOverrideableSDP();
+
public native LinalgBackend linalgPreferredBackend();
public native void setLinalgPreferredBackend(LinalgBackend arg0);
public native void setLinalgPreferredBackend(@Cast("at::LinalgBackend") byte arg0);
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/Conv1dImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/Conv1dImplCloneable.java
index c074380c569..887123cd398 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/Conv1dImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/Conv1dImplCloneable.java
@@ -33,6 +33,6 @@ public class Conv1dImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/Conv2dImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/Conv2dImplCloneable.java
index fddb42dce8a..17f5d769968 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/Conv2dImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/Conv2dImplCloneable.java
@@ -33,6 +33,6 @@ public class Conv2dImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/Conv3dImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/Conv3dImplCloneable.java
index 81aa2a34c37..ab62d30f438 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/Conv3dImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/Conv3dImplCloneable.java
@@ -33,6 +33,6 @@ public class Conv3dImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/ConvTranspose1dImpl.java b/pytorch/src/gen/java/org/bytedeco/pytorch/ConvTranspose1dImpl.java
index 0d180745418..da8b16ca85a 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/ConvTranspose1dImpl.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/ConvTranspose1dImpl.java
@@ -52,10 +52,10 @@ public ConvTranspose1dImpl(
@SharedPtr @Name("std::make_shared") private native void allocate(@ByVal ConvTranspose1dOptions options_);
public native @ByVal Tensor forward(
@Const @ByRef Tensor input,
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") LongArrayRefOptional output_size);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") LongArrayRefOptional output_size);
public native @ByVal Tensor forward(
@Const @ByRef Tensor input);
public native @ByVal Tensor forward(
@Const @ByRef Tensor input,
- @ByRef(nullValue = "std::optional(c10::nullopt)") @Cast({"int64_t*", "c10::ArrayRef", "std::vector&"}) @StdVector long... output_size);
+ @ByRef(nullValue = "std::optional(std::nullopt)") @Cast({"int64_t*", "c10::ArrayRef", "std::vector&"}) @StdVector long... output_size);
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/ConvTranspose1dImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/ConvTranspose1dImplCloneable.java
index fab0daed646..8cf46437e02 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/ConvTranspose1dImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/ConvTranspose1dImplCloneable.java
@@ -33,6 +33,6 @@ public class ConvTranspose1dImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/ConvTranspose2dImpl.java b/pytorch/src/gen/java/org/bytedeco/pytorch/ConvTranspose2dImpl.java
index e00009621c0..65f9b0eee95 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/ConvTranspose2dImpl.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/ConvTranspose2dImpl.java
@@ -52,10 +52,10 @@ public ConvTranspose2dImpl(
@SharedPtr @Name("std::make_shared") private native void allocate(@ByVal ConvTranspose2dOptions options_);
public native @ByVal Tensor forward(
@Const @ByRef Tensor input,
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") LongArrayRefOptional output_size);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") LongArrayRefOptional output_size);
public native @ByVal Tensor forward(
@Const @ByRef Tensor input);
public native @ByVal Tensor forward(
@Const @ByRef Tensor input,
- @ByRef(nullValue = "std::optional(c10::nullopt)") @Cast({"int64_t*", "c10::ArrayRef", "std::vector&"}) @StdVector long... output_size);
+ @ByRef(nullValue = "std::optional(std::nullopt)") @Cast({"int64_t*", "c10::ArrayRef", "std::vector&"}) @StdVector long... output_size);
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/ConvTranspose2dImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/ConvTranspose2dImplCloneable.java
index 47ec0dc5f2a..814903c8188 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/ConvTranspose2dImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/ConvTranspose2dImplCloneable.java
@@ -33,6 +33,6 @@ public class ConvTranspose2dImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/ConvTranspose3dImpl.java b/pytorch/src/gen/java/org/bytedeco/pytorch/ConvTranspose3dImpl.java
index 6eb214538c3..9b4e1ca1437 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/ConvTranspose3dImpl.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/ConvTranspose3dImpl.java
@@ -52,10 +52,10 @@ public ConvTranspose3dImpl(
@SharedPtr @Name("std::make_shared") private native void allocate(@ByVal ConvTranspose3dOptions options_);
public native @ByVal Tensor forward(
@Const @ByRef Tensor input,
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") LongArrayRefOptional output_size);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") LongArrayRefOptional output_size);
public native @ByVal Tensor forward(
@Const @ByRef Tensor input);
public native @ByVal Tensor forward(
@Const @ByRef Tensor input,
- @ByRef(nullValue = "std::optional(c10::nullopt)") @Cast({"int64_t*", "c10::ArrayRef", "std::vector&"}) @StdVector long... output_size);
+ @ByRef(nullValue = "std::optional(std::nullopt)") @Cast({"int64_t*", "c10::ArrayRef", "std::vector&"}) @StdVector long... output_size);
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/ConvTranspose3dImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/ConvTranspose3dImplCloneable.java
index 15f840cae2e..cd1e70c1989 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/ConvTranspose3dImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/ConvTranspose3dImplCloneable.java
@@ -33,6 +33,6 @@ public class ConvTranspose3dImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/CosineEmbeddingLossImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/CosineEmbeddingLossImplCloneable.java
index 51c7b9676c1..69d461b1b55 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/CosineEmbeddingLossImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/CosineEmbeddingLossImplCloneable.java
@@ -33,6 +33,6 @@ public class CosineEmbeddingLossImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/CosineSimilarityImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/CosineSimilarityImplCloneable.java
index a4ca54aa408..a84940a5c31 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/CosineSimilarityImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/CosineSimilarityImplCloneable.java
@@ -33,6 +33,6 @@ public class CosineSimilarityImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/CrossEntropyLossImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/CrossEntropyLossImplCloneable.java
index 5681332dba9..61709c87a6e 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/CrossEntropyLossImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/CrossEntropyLossImplCloneable.java
@@ -33,6 +33,6 @@ public class CrossEntropyLossImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/CrossMapLRN2dImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/CrossMapLRN2dImplCloneable.java
index 6142160122d..c8690fb4c1e 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/CrossMapLRN2dImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/CrossMapLRN2dImplCloneable.java
@@ -33,6 +33,6 @@ public class CrossMapLRN2dImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/DistributedBackend.java b/pytorch/src/gen/java/org/bytedeco/pytorch/DistributedBackend.java
index 2087980d594..edeef500c1d 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/DistributedBackend.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/DistributedBackend.java
@@ -259,10 +259,10 @@ public native void registerOnCompletionHook(
public native @Cast("bool") boolean hasHooks();
// Do not call this directly, use ProcessGroup::setGroupName instead.
- public native void setGroupName(@StdString BytePointer name);
- public native void setGroupName(@StdString String name);
+ public native void setGroupUid(@StdString BytePointer pg_uid);
+ public native void setGroupUid(@StdString String pg_uid);
- public native @StdString BytePointer getGroupName();
+ public native @StdString BytePointer getGroupUid();
public native void setGroupDesc(@StdString BytePointer desc);
public native void setGroupDesc(@StdString String desc);
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/DistributedRandomSampler.java b/pytorch/src/gen/java/org/bytedeco/pytorch/DistributedRandomSampler.java
index d20c7d8db7a..7d1f33ff5fc 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/DistributedRandomSampler.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/DistributedRandomSampler.java
@@ -43,7 +43,7 @@ private native void allocate(
@Cast("size_t") long size);
/** Resets the {@code DistributedRandomSampler} to a new set of indices. */
- public native void reset(@ByVal(nullValue = "std::optional(c10::nullopt)") SizeTOptional new_size);
+ public native void reset(@ByVal(nullValue = "std::optional(std::nullopt)") SizeTOptional new_size);
public native void reset();
/** Returns the next batch of indices. */
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/DistributedSequentialSampler.java b/pytorch/src/gen/java/org/bytedeco/pytorch/DistributedSequentialSampler.java
index 0c175158825..4ec8f3ce8ef 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/DistributedSequentialSampler.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/DistributedSequentialSampler.java
@@ -42,7 +42,7 @@ private native void allocate(
@Cast("size_t") long size);
/** Resets the {@code DistributedSequentialSampler} to a new set of indices. */
- public native void reset(@ByVal(nullValue = "std::optional(c10::nullopt)") SizeTOptional new_size);
+ public native void reset(@ByVal(nullValue = "std::optional(std::nullopt)") SizeTOptional new_size);
public native void reset();
/** Returns the next batch of indices. */
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/Dropout2dImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/Dropout2dImplCloneable.java
index 6069088dde1..ebdea8ac907 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/Dropout2dImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/Dropout2dImplCloneable.java
@@ -33,6 +33,6 @@ public class Dropout2dImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/Dropout3dImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/Dropout3dImplCloneable.java
index e45b7160813..ca8ac329608 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/Dropout3dImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/Dropout3dImplCloneable.java
@@ -33,6 +33,6 @@ public class Dropout3dImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/DropoutImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/DropoutImplCloneable.java
index 4a0a0ebfea8..9ac64515924 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/DropoutImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/DropoutImplCloneable.java
@@ -33,6 +33,6 @@ public class DropoutImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/ELUImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/ELUImplCloneable.java
index 4bf29a55373..e0ee274e3f6 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/ELUImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/ELUImplCloneable.java
@@ -33,6 +33,6 @@ public class ELUImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/EmbeddingBagImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/EmbeddingBagImplCloneable.java
index c94ed7aaf50..29128c11376 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/EmbeddingBagImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/EmbeddingBagImplCloneable.java
@@ -33,6 +33,6 @@ public class EmbeddingBagImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/EmbeddingImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/EmbeddingImplCloneable.java
index 9d051b7ccef..72eb6424c5a 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/EmbeddingImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/EmbeddingImplCloneable.java
@@ -33,6 +33,6 @@ public class EmbeddingImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/FeatureAlphaDropoutImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/FeatureAlphaDropoutImplCloneable.java
index fdc119d70d6..54063936189 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/FeatureAlphaDropoutImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/FeatureAlphaDropoutImplCloneable.java
@@ -33,6 +33,6 @@ public class FeatureAlphaDropoutImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/FlattenImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/FlattenImplCloneable.java
index ba5fecfc3fc..5eac88e01ab 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/FlattenImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/FlattenImplCloneable.java
@@ -33,6 +33,6 @@ public class FlattenImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/FoldImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/FoldImplCloneable.java
index 6f9386cd223..3130a900b55 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/FoldImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/FoldImplCloneable.java
@@ -33,6 +33,6 @@ public class FoldImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/FractionalMaxPool2dImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/FractionalMaxPool2dImplCloneable.java
index bac58df5b66..8f80bd2860e 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/FractionalMaxPool2dImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/FractionalMaxPool2dImplCloneable.java
@@ -33,6 +33,6 @@ public class FractionalMaxPool2dImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/FractionalMaxPool3dImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/FractionalMaxPool3dImplCloneable.java
index 024f8a373f8..d09ad17105f 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/FractionalMaxPool3dImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/FractionalMaxPool3dImplCloneable.java
@@ -33,6 +33,6 @@ public class FractionalMaxPool3dImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/FunctionCrossMapLRN2d.java b/pytorch/src/gen/java/org/bytedeco/pytorch/FunctionCrossMapLRN2d.java
index a7d27c6917c..5c3e413ba04 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/FunctionCrossMapLRN2d.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/FunctionCrossMapLRN2d.java
@@ -56,7 +56,7 @@
* static variable_list forward(AutogradContext *ctx, int n, Variable var) {
* // Save data for backward in context
* ctx->saved_data["n"] = n;
- * var.mul_(2);
+ * var.mul_(n);
* // Mark var as modified by inplace operation
* ctx->mark_dirty({var});
* return {var};
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/FunctionSchema.java b/pytorch/src/gen/java/org/bytedeco/pytorch/FunctionSchema.java
index cd220694cc9..05bb778a862 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/FunctionSchema.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/FunctionSchema.java
@@ -230,16 +230,16 @@ public FunctionSchema(
public native @StdString BytePointer formatTypeMismatchMsg(
@Const @ByRef Argument expected,
@StdString BytePointer actual_type,
- @ByVal(nullValue = "std::optional(c10::nullopt)") SizeTOptional _position,
- @ByVal(nullValue = "std::optional(c10::nullopt)") StringOptional value);
+ @ByVal(nullValue = "std::optional(std::nullopt)") SizeTOptional _position,
+ @ByVal(nullValue = "std::optional(std::nullopt)") StringOptional value);
public native @StdString BytePointer formatTypeMismatchMsg(
@Const @ByRef Argument expected,
@StdString BytePointer actual_type);
public native @StdString String formatTypeMismatchMsg(
@Const @ByRef Argument expected,
@StdString String actual_type,
- @ByVal(nullValue = "std::optional(c10::nullopt)") SizeTOptional _position,
- @ByVal(nullValue = "std::optional(c10::nullopt)") StringOptional value);
+ @ByVal(nullValue = "std::optional(std::nullopt)") SizeTOptional _position,
+ @ByVal(nullValue = "std::optional(std::nullopt)") StringOptional value);
public native @StdString String formatTypeMismatchMsg(
@Const @ByRef Argument expected,
@StdString String actual_type);
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/Future.java b/pytorch/src/gen/java/org/bytedeco/pytorch/Future.java
index 6ab36a9fcc0..b3634b6a1b5 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/Future.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/Future.java
@@ -54,7 +54,7 @@ public class Future extends Pointer {
*/
public native void markCompleted(
@ByVal IValue value,
- @ByVal(nullValue = "std::optional > >(c10::nullopt)") WeakStorageVectorOptional storages);
+ @ByVal(nullValue = "std::optional > >(std::nullopt)") WeakStorageVectorOptional storages);
public native void markCompleted(
@ByVal IValue value);
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/GELUImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/GELUImplCloneable.java
index 205c9c28a55..6d9fb4d436d 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/GELUImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/GELUImplCloneable.java
@@ -33,6 +33,6 @@ public class GELUImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/GLUImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/GLUImplCloneable.java
index d15a9829227..1c1af5bb3dd 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/GLUImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/GLUImplCloneable.java
@@ -33,6 +33,6 @@ public class GLUImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/GRUCellImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/GRUCellImplCloneable.java
index 1900809c054..bb4008fef44 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/GRUCellImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/GRUCellImplCloneable.java
@@ -33,6 +33,6 @@ public class GRUCellImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/GRUImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/GRUImplCloneable.java
index 0961973ff5d..b83b30a5e92 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/GRUImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/GRUImplCloneable.java
@@ -33,6 +33,6 @@ public class GRUImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/GenericDict.java b/pytorch/src/gen/java/org/bytedeco/pytorch/GenericDict.java
index 620a92bde70..c4cd32f5415 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/GenericDict.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/GenericDict.java
@@ -160,7 +160,7 @@ public class GenericDict extends Pointer {
public native @Cast("bool") boolean is(@Const @ByRef GenericDict rhs);
// private API for now because the return type will change to TypePtr
- // instead of optional once types are mandatory.
+ // instead of std::optional once types are mandatory.
public native @ByVal Type.TypePtr keyType();
public native @ByVal Type.TypePtr valueType();
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/GenericList.java b/pytorch/src/gen/java/org/bytedeco/pytorch/GenericList.java
index 28f562288d7..92a671ea699 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/GenericList.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/GenericList.java
@@ -228,6 +228,4 @@ public class GenericList extends Pointer {
// See [unsafe set type] for why this exists.
public native void unsafeSetElementType(@ByVal Type.TypePtr t);
- private static native @Namespace @Const @Name("c10::impl::ptr_to_first_element") IValue ptr_to_first_element(@Const @ByRef GenericList list);
- public IValue ptr_to_first_element() { return ptr_to_first_element(this); }
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/Graph.java b/pytorch/src/gen/java/org/bytedeco/pytorch/Graph.java
index 88709059cef..6ca933d1b8a 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/Graph.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/Graph.java
@@ -174,8 +174,8 @@ public native JitNode createClone(
// Insert constant IValue into the graph.
public native Value insertConstant(
@Const @ByRef IValue val,
- @ByVal(nullValue = "std::optional(c10::nullopt)") SourceRangeOptional loc,
- @ByVal(nullValue = "std::optional(c10::nullopt)") @Cast("std::optional*") ScopeOptional scope);
+ @ByVal(nullValue = "std::optional(std::nullopt)") SourceRangeOptional loc,
+ @ByVal(nullValue = "std::optional(std::nullopt)") @Cast("std::optional*") ScopeOptional scope);
public native Value insertConstant(
@Const @ByRef IValue val);
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/GraphExecutor.java b/pytorch/src/gen/java/org/bytedeco/pytorch/GraphExecutor.java
index bc03282ad5f..d6294719eee 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/GraphExecutor.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/GraphExecutor.java
@@ -77,7 +77,7 @@ private native void allocate(
// current global fusion strategy settings.
public native @Const @ByRef ExecutionPlan getPlanFor(
@ByRef IValueVector inputs,
- @ByVal(nullValue = "std::optional(c10::nullopt)") SizeTOptional remaining_bailout_depth);
+ @ByVal(nullValue = "std::optional(std::nullopt)") SizeTOptional remaining_bailout_depth);
public native @Const @ByRef ExecutionPlan getPlanFor(
@ByRef IValueVector inputs);
public native @ByVal GraphExecutorState getDebugState();
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/GraphFunction.java b/pytorch/src/gen/java/org/bytedeco/pytorch/GraphFunction.java
index dcfd1622b2f..5501ee26907 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/GraphFunction.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/GraphFunction.java
@@ -25,17 +25,16 @@ public class GraphFunction extends Function {
/** Pointer cast constructor. Invokes {@link Pointer#Pointer(Pointer)}. */
public GraphFunction(Pointer p) { super(p); }
- // NOLINTNEXTLINE(cppcoreguidelines-pro-type-member-init)
public GraphFunction(
@ByVal QualifiedName name,
@SharedPtr("torch::jit::Graph") @ByVal Graph graph,
@ByVal GraphFunctionCreator function_creator,
- @ByVal(nullValue = "std::optional(c10::nullopt)") ExecutorExecutionModeOptional executor_execution_mode) { super((Pointer)null); allocate(name, graph, function_creator, executor_execution_mode); }
+ @ByVal(nullValue = "std::optional(std::nullopt)") ExecutorExecutionModeOptional executor_execution_mode) { super((Pointer)null); allocate(name, graph, function_creator, executor_execution_mode); }
private native void allocate(
@ByVal QualifiedName name,
@SharedPtr("torch::jit::Graph") @ByVal Graph graph,
@ByVal GraphFunctionCreator function_creator,
- @ByVal(nullValue = "std::optional(c10::nullopt)") ExecutorExecutionModeOptional executor_execution_mode);
+ @ByVal(nullValue = "std::optional(std::nullopt)") ExecutorExecutionModeOptional executor_execution_mode);
public GraphFunction(
@ByVal QualifiedName name,
@SharedPtr("torch::jit::Graph") @ByVal Graph graph,
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/GroupNormImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/GroupNormImplCloneable.java
index 78c3d08255c..eb5d6ea8078 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/GroupNormImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/GroupNormImplCloneable.java
@@ -33,6 +33,6 @@ public class GroupNormImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/HIPHooksInterface.java b/pytorch/src/gen/java/org/bytedeco/pytorch/HIPHooksInterface.java
index 8da8d17f9c6..2b33aa12303 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/HIPHooksInterface.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/HIPHooksInterface.java
@@ -24,7 +24,7 @@
// dispatched, to allow for separate compilation of HIP code). See
// CUDAHooksInterface for more detailed motivation.
@Namespace("at") @Properties(inherit = org.bytedeco.pytorch.presets.torch.class)
-public class HIPHooksInterface extends Pointer {
+public class HIPHooksInterface extends AcceleratorHooksInterface {
static { Loader.load(); }
/** Default native constructor. */
public HIPHooksInterface() { super((Pointer)null); allocate(); }
@@ -53,9 +53,13 @@ public class HIPHooksInterface extends Pointer {
public native @Cast("c10::DeviceIndex") byte current_device();
+ public native @Cast("bool") boolean isPinnedPtr(@Const Pointer data);
+
public native Allocator getPinnedMemoryAllocator();
public native void registerHIPTypes(Context arg0);
public native int getNumGPUs();
+
+ public native @Cast("bool") boolean hasPrimaryContext(@Cast("c10::DeviceIndex") byte device_index);
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/Half.java b/pytorch/src/gen/java/org/bytedeco/pytorch/Half.java
index 42213b00ffe..5689b1f562f 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/Half.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/Half.java
@@ -44,7 +44,7 @@ public class Half extends Pointer {
public Half(@Cast("unsigned short") short bits, @ByVal from_bits_t arg1) { super((Pointer)null); allocate(bits, arg1); }
private native void allocate(@Cast("unsigned short") short bits, @ByVal from_bits_t arg1);
-// #if defined(__aarch64__) && !defined(C10_MOBILE) && !defined(__CUDACC__)
+// #if defined(__aarch64__) && !defined(__CUDACC__)
// #else
public Half(float value) { super((Pointer)null); allocate(value); }
private native void allocate(float value);
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/HardshrinkImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/HardshrinkImplCloneable.java
index 426499c46c4..a5bcfee780a 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/HardshrinkImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/HardshrinkImplCloneable.java
@@ -33,6 +33,6 @@ public class HardshrinkImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/HardtanhImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/HardtanhImplCloneable.java
index 4989af3c526..ccacda645f0 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/HardtanhImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/HardtanhImplCloneable.java
@@ -33,6 +33,6 @@ public class HardtanhImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/HingeEmbeddingLossImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/HingeEmbeddingLossImplCloneable.java
index 38ac8b2ff47..ec34852cd66 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/HingeEmbeddingLossImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/HingeEmbeddingLossImplCloneable.java
@@ -33,6 +33,6 @@ public class HingeEmbeddingLossImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/HuberLossImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/HuberLossImplCloneable.java
index 076cd6e982d..792ebde777f 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/HuberLossImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/HuberLossImplCloneable.java
@@ -33,6 +33,6 @@ public class HuberLossImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/IPUHooksInterface.java b/pytorch/src/gen/java/org/bytedeco/pytorch/IPUHooksInterface.java
index 83cbe8f2edf..b0ff5e80d34 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/IPUHooksInterface.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/IPUHooksInterface.java
@@ -39,9 +39,9 @@ public class IPUHooksInterface extends Pointer {
public native @Const @ByRef Generator getDefaultIPUGenerator(
- @Cast("c10::DeviceIndex") byte device_index/*=-1*/);
+ @Cast("c10::DeviceIndex") byte device_index/*[[maybe_unused]] = -1*/);
public native @Const @ByRef Generator getDefaultIPUGenerator();
- public native @ByVal Generator newIPUGenerator(@Cast("c10::DeviceIndex") byte device_index/*=-1*/);
+ public native @ByVal Generator newIPUGenerator(@Cast("c10::DeviceIndex") byte device_index/*[[maybe_unused]] = -1*/);
public native @ByVal Generator newIPUGenerator();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/IValue.java b/pytorch/src/gen/java/org/bytedeco/pytorch/IValue.java
index 3e8da1cd5ca..9d88e6b219c 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/IValue.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/IValue.java
@@ -549,11 +549,11 @@ public static class CompIdentityIValues extends Pointer {
// TODO: There are several places that recurse over IValue. This is fragile.
// This visitor should be used to recurse over ivalues.
- public native @ByVal IValue deepcopy(@ByVal(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ public native @ByVal IValue deepcopy(@ByVal(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @ByVal IValue deepcopy();
public native @ByVal IValue deepcopy(
@ByRef HashIdentityIValueMap memo,
- @ByVal(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @ByVal(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @ByVal IValue deepcopy(
@ByRef HashIdentityIValueMap memo);
// Don't edit this just to add results for new tags; edit
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/IdentityImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/IdentityImplCloneable.java
index 6dfe9b10f23..023eb5322f7 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/IdentityImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/IdentityImplCloneable.java
@@ -33,6 +33,6 @@ public class IdentityImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/InputArchive.java b/pytorch/src/gen/java/org/bytedeco/pytorch/InputArchive.java
index 13d0b08664b..8b649b67251 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/InputArchive.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/InputArchive.java
@@ -93,12 +93,12 @@ public class InputArchive extends Pointer {
* is not specified, the module is loaded to the original device. */
public native void load_from(
@StdString BytePointer filename,
- @ByVal(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @ByVal(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native void load_from(
@StdString BytePointer filename);
public native void load_from(
@StdString String filename,
- @ByVal(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @ByVal(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native void load_from(
@StdString String filename);
@@ -107,7 +107,7 @@ public native void load_from(
* is not specified, the module is loaded to the original device. */
public native void load_from(
@Cast("std::istream*") @ByRef Pointer stream,
- @ByVal(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @ByVal(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native void load_from(
@Cast("std::istream*") @ByRef Pointer stream);
@@ -115,14 +115,14 @@ public native void load_from(
public native void load_from(
@Cast("const char*") BytePointer data,
@Cast("size_t") long size,
- @ByVal(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @ByVal(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native void load_from(
@Cast("const char*") BytePointer data,
@Cast("size_t") long size);
public native void load_from(
String data,
@Cast("size_t") long size,
- @ByVal(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @ByVal(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native void load_from(
String data,
@Cast("size_t") long size);
@@ -131,7 +131,7 @@ public native void load_from(
public native void load_from(
@Const @ByRef Reader read_func,
@Const @ByRef SizeTSupplier size_func,
- @ByVal(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @ByVal(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native void load_from(
@Const @ByRef Reader read_func,
@Const @ByRef SizeTSupplier size_func);
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/InstanceNorm1dImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/InstanceNorm1dImplCloneable.java
index a32a79ac70c..b2fb2df1d12 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/InstanceNorm1dImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/InstanceNorm1dImplCloneable.java
@@ -33,6 +33,6 @@ public class InstanceNorm1dImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional(c10::nullopt)") DeviceOptional device);
+ @Const @ByRef(nullValue = "std::optional(std::nullopt)") DeviceOptional device);
public native @SharedPtr("torch::nn::Module") @ByVal Module clone();
}
diff --git a/pytorch/src/gen/java/org/bytedeco/pytorch/InstanceNorm2dImplCloneable.java b/pytorch/src/gen/java/org/bytedeco/pytorch/InstanceNorm2dImplCloneable.java
index 6d1e5157004..5eafaa7f59a 100644
--- a/pytorch/src/gen/java/org/bytedeco/pytorch/InstanceNorm2dImplCloneable.java
+++ b/pytorch/src/gen/java/org/bytedeco/pytorch/InstanceNorm2dImplCloneable.java
@@ -33,6 +33,6 @@ public class InstanceNorm2dImplCloneable extends Module {
* and submodules in the cloned module are different from those in the
* original module. */
public native @SharedPtr("torch::nn::Module") @ByVal Module clone(
- @Const @ByRef(nullValue = "std::optional