Skip to content

Commit

Permalink
* Upgrade presets for CUDA 11.4.2 + NvToolsExt (NVTX), cuDNN 8.2.4, …
Browse files Browse the repository at this point in the history
…NCCL 2.11.4
  • Loading branch information
saudet committed Sep 15, 2021
1 parent e803686 commit 02c27e4
Show file tree
Hide file tree
Showing 361 changed files with 14,589 additions and 9,543 deletions.
12 changes: 6 additions & 6 deletions .github/actions/deploy-centos/action.yml
Original file line number Diff line number Diff line change
Expand Up @@ -62,16 +62,16 @@ runs:
if [[ "$CI_DEPLOY_PLATFORM" == "linux-x86_64" ]] && [[ -n ${CI_DEPLOY_NEED_CUDA:-} ]]; then
echo Installing CUDA, cuDNN, etc
curl -LO https://developer.download.nvidia.com/compute/cuda/11.4.0/local_installers/cuda-repo-rhel7-11-4-local-11.4.0_470.42.01-1.x86_64.rpm
curl -LO https://developer.download.nvidia.com/compute/redist/cudnn/v8.2.2/cudnn-11.4-linux-x64-v8.2.2.26.tgz
curl -LO https://developer.download.nvidia.com/compute/redist/nccl/v2.10/nccl_2.10.3-1+cuda11.4_x86_64.txz
curl -LO https://developer.download.nvidia.com/compute/cuda/11.4.2/local_installers/cuda-repo-rhel7-11-4-local-11.4.2_470.57.02-1.x86_64.rpm
curl -LO https://developer.download.nvidia.com/compute/redist/cudnn/v8.2.4/cudnn-11.4-linux-x64-v8.2.4.15.tgz
curl -LO https://developer.download.nvidia.com/compute/redist/nccl/v2.11/nccl_2.11.4-1+cuda11.4_x86_64.txz
rpm -i --force --ignorearch --nodeps cuda-repo-rhel7-11-4-local-11.4.0_470.42.01-1.x86_64.rpm
rpm -i --force --ignorearch --nodeps cuda-repo-rhel7-11-4-local-11.4.2_470.57.02-1.x86_64.rpm
pushd /var/cuda-repo-rhel7-11-4-local/; rpm -i --force --ignorearch --nodeps cuda*.rpm libc*.rpm libn*.rpm; rm *.rpm; popd
ln -sf /usr/local/cuda/lib64/stubs/libcuda.so /usr/local/cuda/lib64/libcuda.so
ln -sf /usr/local/cuda/lib64/stubs/libnvidia-ml.so /usr/local/cuda/lib64/libnvidia-ml.so
tar -hxvf cudnn-11.4-linux-x64-v8.2.2.26.tgz -C /usr/local/
tar -hxvf nccl_2.10.3-1+cuda11.4_x86_64.txz --strip-components=1 -C /usr/local/cuda/
tar -hxvf cudnn-11.4-linux-x64-v8.2.4.15.tgz -C /usr/local/
tar -hxvf nccl_2.11.4-1+cuda11.4_x86_64.txz --strip-components=1 -C /usr/local/cuda/
mv /usr/local/cuda/lib/* /usr/local/cuda/lib64/
# Work around issues with CUDA 10.2/11.x
Expand Down
24 changes: 12 additions & 12 deletions .github/actions/deploy-ubuntu/action.yml
Original file line number Diff line number Diff line change
Expand Up @@ -13,21 +13,21 @@ runs:
elif [[ "$CI_DEPLOY_PLATFORM" == "linux-arm64" ]]; then
export ARCH=arm64
export PREFIX=aarch64-linux-gnu
export CUDA=cuda-repo-rhel8-11-4-local-11.4.0_470.42.01-1.aarch64.rpm
export CUDNN=cudnn-11.4-linux-aarch64sbsa-v8.2.2.26.tgz
export NCCL=nccl_2.10.3-1+cuda11.4_aarch64.txz
export CUDA=cuda-repo-rhel8-11-4-local-11.4.2_470.57.02-1.aarch64.rpm
export CUDNN=cudnn-11.4-linux-aarch64sbsa-v8.2.4.15.tgz
export NCCL=nccl_2.11.4-1+cuda11.4_aarch64.txz
elif [[ "$CI_DEPLOY_PLATFORM" == "linux-ppc64le" ]]; then
export ARCH=ppc64el
export PREFIX=powerpc64le-linux-gnu
export CUDA=cuda-repo-rhel8-11-4-local-11.4.0_470.42.01-1.ppc64le.rpm
export CUDNN=cudnn-11.4-linux-ppc64le-v8.2.2.26.tgz
export NCCL=nccl_2.10.3-1+cuda11.4_ppc64le.txz
export CUDA=cuda-repo-rhel8-11-4-local-11.4.2_470.57.02-1.ppc64le.rpm
export CUDNN=cudnn-11.4-linux-ppc64le-v8.2.4.15.tgz
export NCCL=nccl_2.11.4-1+cuda11.4_ppc64le.txz
elif [[ "$CI_DEPLOY_PLATFORM" == "linux-x86_64" ]]; then
export ARCH=amd64
export PREFIX=x86_64-linux-gnu
export CUDA=cuda-repo-rhel7-11-4-local-11.4.0_470.42.01-1.x86_64.rpm
export CUDNN=cudnn-11.4-linux-x64-v8.2.2.26.tgz
export NCCL=nccl_2.10.3-1+cuda11.4_x86_64.txz
export CUDA=cuda-repo-rhel7-11-4-local-11.4.2_470.57.02-1.x86_64.rpm
export CUDNN=cudnn-11.4-linux-x64-v8.2.4.15.tgz
export NCCL=nccl_2.11.4-1+cuda11.4_x86_64.txz
fi
echo "ARCH=$ARCH" >> $GITHUB_ENV
echo "PREFIX=$PREFIX" >> $GITHUB_ENV
Expand Down Expand Up @@ -107,9 +107,9 @@ runs:
if [[ "$CI_DEPLOY_PLATFORM" == "linux-arm64" ]] || [[ "$CI_DEPLOY_PLATFORM" == "linux-ppc64le" ]] && [[ -n ${CI_DEPLOY_NEED_CUDA:-} ]]; then
echo Installing CUDA, cuDNN, etc
curl -LO https://developer.download.nvidia.com/compute/cuda/11.4.0/local_installers/$CUDA
curl -LO https://developer.download.nvidia.com/compute/redist/cudnn/v8.2.2/$CUDNN
curl -LO https://developer.download.nvidia.com/compute/redist/nccl/v2.10/$NCCL
curl -LO https://developer.download.nvidia.com/compute/cuda/11.4.2/local_installers/$CUDA
curl -LO https://developer.download.nvidia.com/compute/redist/cudnn/v8.2.4/$CUDNN
curl -LO https://developer.download.nvidia.com/compute/redist/nccl/v2.11/$NCCL
rpm -i --force --ignorearch --nodeps $CUDA
pushd /var/cuda-repo-rhel8-11-4-local/; rpm -i --force --ignorearch --nodeps cuda*.rpm libc*.rpm libn*.rpm; rm *.rpm; popd
Expand Down
8 changes: 4 additions & 4 deletions .github/actions/deploy-windows/action.yml
Original file line number Diff line number Diff line change
Expand Up @@ -68,10 +68,10 @@ runs:
if "%CI_DEPLOY_PLATFORM%"=="windows-x86_64" if not "%CI_DEPLOY_NEED_CUDA%"=="" (
echo Installing CUDA, cuDNN, etc
curl -LO https://developer.download.nvidia.com/compute/cuda/11.4.0/local_installers/cuda_11.4.0_471.11_win10.exe
curl -LO https://developer.download.nvidia.com/compute/redist/cudnn/v8.2.2/cudnn-11.4-windows-x64-v8.2.2.26.zip
cuda_11.4.0_471.11_win10.exe -s
unzip cudnn-11.4-windows-x64-v8.2.2.26.zip
curl -LO https://developer.download.nvidia.com/compute/cuda/11.4.2/local_installers/cuda_11.4.2_471.41_win10.exe
curl -LO https://developer.download.nvidia.com/compute/redist/cudnn/v8.2.4/cudnn-11.4-windows-x64-v8.2.4.15.zip
cuda_11.4.2_471.41_win10.exe -s
unzip cudnn-11.4-windows-x64-v8.2.4.15.zip
move cuda\bin\*.dll "%ProgramFiles%\NVIDIA GPU Computing Toolkit\CUDA\v11.4\bin"
move cuda\include\*.h "%ProgramFiles%\NVIDIA GPU Computing Toolkit\CUDA\v11.4\include"
move cuda\lib\x64\*.lib "%ProgramFiles%\NVIDIA GPU Computing Toolkit\CUDA\v11.4\lib\x64"
Expand Down
3 changes: 2 additions & 1 deletion CHANGELOG.md
Original file line number Diff line number Diff line change
@@ -1,9 +1,10 @@

* Add presets for the NvToolsExt (NVTX) module of CUDA ([issue #1068](https://github.com/bytedeco/javacpp-presets/issues/1068))
* Increase the amount of function pointers available for callbacks in presets for Qt ([pull #1080](https://github.com/bytedeco/javacpp-presets/pull/1080))
* Map C++ JIT classes and functions of TorchScript in presets for PyTorch ([issue #1068](https://github.com/bytedeco/javacpp-presets/issues/1068))
* Synchronize `cachePackage()` and prevent repeated package caching in all presets ([pull #1071](https://github.com/bytedeco/javacpp-presets/pull/1071))
* Build FFmpeg with VA-API enabled and bundle its libraries to avoid loading issues ([issue bytedeco/javacv#1188](https://github.com/bytedeco/javacv/issues/1188))
* Upgrade presets for Arrow 5.0.0, DNNL 2.3.2, NumPy 1.21.2, SciPy 1.7.1, Gym 0.19.0, TensorFlow Lite 2.6.0, DepthAI 2.10.0, ONNX 1.10.1, ONNX Runtime 1.8.2, and their dependencies
* Upgrade presets for Arrow 5.0.0, DNNL 2.3.2, NumPy 1.21.2, SciPy 1.7.1, Gym 0.19.0, CUDA 11.4.2, cuDNN 8.2.4, NCCL 2.11.4, TensorFlow Lite 2.6.0, DepthAI 2.10.0, ONNX 1.10.1, ONNX Runtime 1.8.2, and their dependencies

### August 2, 2021 version 1.5.6
* Change `opencv_core.Mat` constructors to create column vectors out of arrays for consistency ([issue #1064](https://github.com/bytedeco/javacpp-presets/issues/1064))
Expand Down
2 changes: 1 addition & 1 deletion README.md
Original file line number Diff line number Diff line change
Expand Up @@ -206,7 +206,7 @@ Each child module in turn relies by default on the included [`cppbuild.sh` scrip
* OpenPose 1.7.0 https://github.com/CMU-Perceptual-Computing-Lab/openpose
* CUDA 11.4.x https://developer.nvidia.com/cuda-downloads
* cuDNN 8.2.x https://developer.nvidia.com/cudnn
* NCCL 2.10.x https://developer.nvidia.com/nccl
* NCCL 2.11.x https://developer.nvidia.com/nccl
* NVIDIA Video Codec SDK 11.1.x https://developer.nvidia.com/nvidia-video-codec-sdk
* OpenCL 3.0 https://github.com/KhronosGroup/OpenCL-ICD-Loader
* MXNet 1.8.0 https://github.com/apache/incubator-mxnet
Expand Down
6 changes: 3 additions & 3 deletions cuda/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -23,9 +23,9 @@ Introduction
------------
This directory contains the JavaCPP Presets module for:

* CUDA 11.4.0 https://developer.nvidia.com/cuda-zone
* cuDNN 8.2.2 https://developer.nvidia.com/cudnn
* NCCL 2.10.3 https://developer.nvidia.com/nccl
* CUDA 11.4.2 https://developer.nvidia.com/cuda-zone
* cuDNN 8.2.4 https://developer.nvidia.com/cudnn
* NCCL 2.11.4 https://developer.nvidia.com/nccl

Please refer to the parent README.md file for more detailed information about the JavaCPP Presets.

Expand Down
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
// Targeted by JavaCPP version 1.5.6: DO NOT EDIT THIS FILE
// Targeted by JavaCPP version 1.5.7-SNAPSHOT: DO NOT EDIT THIS FILE

package org.bytedeco.cuda.cublas;

Expand Down
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
// Targeted by JavaCPP version 1.5.6: DO NOT EDIT THIS FILE
// Targeted by JavaCPP version 1.5.7-SNAPSHOT: DO NOT EDIT THIS FILE

package org.bytedeco.cuda.cublas;

Expand Down
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
// Targeted by JavaCPP version 1.5.6: DO NOT EDIT THIS FILE
// Targeted by JavaCPP version 1.5.7-SNAPSHOT: DO NOT EDIT THIS FILE

package org.bytedeco.cuda.cublas;

Expand Down
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
// Targeted by JavaCPP version 1.5.6: DO NOT EDIT THIS FILE
// Targeted by JavaCPP version 1.5.7-SNAPSHOT: DO NOT EDIT THIS FILE

package org.bytedeco.cuda.cublas;

Expand All @@ -22,9 +22,5 @@ public class cublasLtLoggerCallback_t extends FunctionPointer {
public cublasLtLoggerCallback_t(Pointer p) { super(p); }
protected cublasLtLoggerCallback_t() { allocate(); }
private native void allocate();
public native void call(
int logLevel,
@Cast("const char*") BytePointer functionName,
@Cast("const char*") BytePointer message
);
public native void call(int logLevel, @Cast("const char*") BytePointer functionName, @Cast("const char*") BytePointer message);
}
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
// Targeted by JavaCPP version 1.5.6: DO NOT EDIT THIS FILE
// Targeted by JavaCPP version 1.5.7-SNAPSHOT: DO NOT EDIT THIS FILE

package org.bytedeco.cuda.cublas;

Expand Down Expand Up @@ -36,6 +36,6 @@ public class cublasLtMatmulAlgo_t extends Pointer {
return new cublasLtMatmulAlgo_t((Pointer)this).offsetAddress(i);
}

public native @Cast("uint64_t") long data(int i); public native cublasLtMatmulAlgo_t data(int i, long setter);
@MemberGetter public native @Cast("uint64_t*") LongPointer data();
public native @Cast("uint64_t") long data(int i); public native cublasLtMatmulAlgo_t data(int i, long setter);
@MemberGetter public native @Cast("uint64_t*") LongPointer data();
}
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
// Targeted by JavaCPP version 1.5.6: DO NOT EDIT THIS FILE
// Targeted by JavaCPP version 1.5.7-SNAPSHOT: DO NOT EDIT THIS FILE

package org.bytedeco.cuda.cublas;

Expand Down Expand Up @@ -33,6 +33,6 @@ public class cublasLtMatmulDescOpaque_t extends Pointer {
return new cublasLtMatmulDescOpaque_t((Pointer)this).offsetAddress(i);
}

public native @Cast("uint64_t") long data(int i); public native cublasLtMatmulDescOpaque_t data(int i, long setter);
@MemberGetter public native @Cast("uint64_t*") LongPointer data();
public native @Cast("uint64_t") long data(int i); public native cublasLtMatmulDescOpaque_t data(int i, long setter);
@MemberGetter public native @Cast("uint64_t*") LongPointer data();
}
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
// Targeted by JavaCPP version 1.5.6: DO NOT EDIT THIS FILE
// Targeted by JavaCPP version 1.5.7-SNAPSHOT: DO NOT EDIT THIS FILE

package org.bytedeco.cuda.cublas;

Expand Down Expand Up @@ -35,27 +35,28 @@ public class cublasLtMatmulHeuristicResult_t extends Pointer {
return new cublasLtMatmulHeuristicResult_t((Pointer)this).offsetAddress(i);
}

/** Matmul algorithm descriptor.
*
* Must be initialized with cublasLtMatmulAlgoInit() if preferences' CUBLASLT_MATMUL_PERF_SEARCH_MODE is set to
* CUBLASLT_SEARCH_LIMITED_BY_ALGO_ID
*/
public native @ByRef cublasLtMatmulAlgo_t algo(); public native cublasLtMatmulHeuristicResult_t algo(cublasLtMatmulAlgo_t setter);

/** Actual size of workspace memory required.
*/
public native @Cast("size_t") long workspaceSize(); public native cublasLtMatmulHeuristicResult_t workspaceSize(long setter);

/** Result status, other fields are only valid if after call to cublasLtMatmulAlgoGetHeuristic() this member is set to CUBLAS_STATUS_SUCCESS.
*/
public native @Cast("cublasStatus_t") int state(); public native cublasLtMatmulHeuristicResult_t state(int setter);

/** Waves count - a device utilization metric.
*
* wavesCount value of 1.0f suggests that when kernel is launched it will fully occupy the GPU.
*/
public native float wavesCount(); public native cublasLtMatmulHeuristicResult_t wavesCount(float setter);

public native int reserved(int i); public native cublasLtMatmulHeuristicResult_t reserved(int i, int setter);
@MemberGetter public native IntPointer reserved();
/** Matmul algorithm descriptor.
*
* Must be initialized with cublasLtMatmulAlgoInit() if preferences' CUBLASLT_MATMUL_PERF_SEARCH_MODE is set to
* CUBLASLT_SEARCH_LIMITED_BY_ALGO_ID
*/
public native @ByRef cublasLtMatmulAlgo_t algo(); public native cublasLtMatmulHeuristicResult_t algo(cublasLtMatmulAlgo_t setter);

/** Actual size of workspace memory required.
*/
public native @Cast("size_t") long workspaceSize(); public native cublasLtMatmulHeuristicResult_t workspaceSize(long setter);

/** Result status, other fields are only valid if after call to cublasLtMatmulAlgoGetHeuristic() this member is set to
* CUBLAS_STATUS_SUCCESS.
*/
public native @Cast("cublasStatus_t") int state(); public native cublasLtMatmulHeuristicResult_t state(int setter);

/** Waves count - a device utilization metric.
*
* wavesCount value of 1.0f suggests that when kernel is launched it will fully occupy the GPU.
*/
public native float wavesCount(); public native cublasLtMatmulHeuristicResult_t wavesCount(float setter);

public native int reserved(int i); public native cublasLtMatmulHeuristicResult_t reserved(int i, int setter);
@MemberGetter public native IntPointer reserved();
}
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
// Targeted by JavaCPP version 1.5.6: DO NOT EDIT THIS FILE
// Targeted by JavaCPP version 1.5.7-SNAPSHOT: DO NOT EDIT THIS FILE

package org.bytedeco.cuda.cublas;

Expand Down Expand Up @@ -33,6 +33,6 @@ public class cublasLtMatmulPreferenceOpaque_t extends Pointer {
return new cublasLtMatmulPreferenceOpaque_t((Pointer)this).offsetAddress(i);
}

public native @Cast("uint64_t") long data(int i); public native cublasLtMatmulPreferenceOpaque_t data(int i, long setter);
@MemberGetter public native @Cast("uint64_t*") LongPointer data();
public native @Cast("uint64_t") long data(int i); public native cublasLtMatmulPreferenceOpaque_t data(int i, long setter);
@MemberGetter public native @Cast("uint64_t*") LongPointer data();
}
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
// Targeted by JavaCPP version 1.5.6: DO NOT EDIT THIS FILE
// Targeted by JavaCPP version 1.5.7-SNAPSHOT: DO NOT EDIT THIS FILE

package org.bytedeco.cuda.cublas;

Expand Down Expand Up @@ -33,6 +33,6 @@ public class cublasLtMatrixLayoutOpaque_t extends Pointer {
return new cublasLtMatrixLayoutOpaque_t((Pointer)this).offsetAddress(i);
}

public native @Cast("uint64_t") long data(int i); public native cublasLtMatrixLayoutOpaque_t data(int i, long setter);
@MemberGetter public native @Cast("uint64_t*") LongPointer data();
public native @Cast("uint64_t") long data(int i); public native cublasLtMatrixLayoutOpaque_t data(int i, long setter);
@MemberGetter public native @Cast("uint64_t*") LongPointer data();
}
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
// Targeted by JavaCPP version 1.5.6: DO NOT EDIT THIS FILE
// Targeted by JavaCPP version 1.5.7-SNAPSHOT: DO NOT EDIT THIS FILE

package org.bytedeco.cuda.cublas;

Expand Down Expand Up @@ -33,6 +33,6 @@ public class cublasLtMatrixTransformDescOpaque_t extends Pointer {
return new cublasLtMatrixTransformDescOpaque_t((Pointer)this).offsetAddress(i);
}

public native @Cast("uint64_t") long data(int i); public native cublasLtMatrixTransformDescOpaque_t data(int i, long setter);
@MemberGetter public native @Cast("uint64_t*") LongPointer data();
public native @Cast("uint64_t") long data(int i); public native cublasLtMatrixTransformDescOpaque_t data(int i, long setter);
@MemberGetter public native @Cast("uint64_t*") LongPointer data();
}
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
// Targeted by JavaCPP version 1.5.6: DO NOT EDIT THIS FILE
// Targeted by JavaCPP version 1.5.7-SNAPSHOT: DO NOT EDIT THIS FILE

package org.bytedeco.cuda.cublas;

Expand Down
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
// Targeted by JavaCPP version 1.5.6: DO NOT EDIT THIS FILE
// Targeted by JavaCPP version 1.5.7-SNAPSHOT: DO NOT EDIT THIS FILE

package org.bytedeco.cuda.cudart;

Expand Down
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
// Targeted by JavaCPP version 1.5.6: DO NOT EDIT THIS FILE
// Targeted by JavaCPP version 1.5.7-SNAPSHOT: DO NOT EDIT THIS FILE

package org.bytedeco.cuda.cudart;

Expand Down
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
// Targeted by JavaCPP version 1.5.6: DO NOT EDIT THIS FILE
// Targeted by JavaCPP version 1.5.7-SNAPSHOT: DO NOT EDIT THIS FILE

package org.bytedeco.cuda.cudart;

Expand Down
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
// Targeted by JavaCPP version 1.5.6: DO NOT EDIT THIS FILE
// Targeted by JavaCPP version 1.5.7-SNAPSHOT: DO NOT EDIT THIS FILE

package org.bytedeco.cuda.cudart;

Expand Down
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
// Targeted by JavaCPP version 1.5.6: DO NOT EDIT THIS FILE
// Targeted by JavaCPP version 1.5.7-SNAPSHOT: DO NOT EDIT THIS FILE

package org.bytedeco.cuda.cudart;

Expand Down
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
// Targeted by JavaCPP version 1.5.6: DO NOT EDIT THIS FILE
// Targeted by JavaCPP version 1.5.7-SNAPSHOT: DO NOT EDIT THIS FILE

package org.bytedeco.cuda.cudart;

Expand Down
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
// Targeted by JavaCPP version 1.5.6: DO NOT EDIT THIS FILE
// Targeted by JavaCPP version 1.5.7-SNAPSHOT: DO NOT EDIT THIS FILE

package org.bytedeco.cuda.cudart;

Expand Down
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
// Targeted by JavaCPP version 1.5.6: DO NOT EDIT THIS FILE
// Targeted by JavaCPP version 1.5.7-SNAPSHOT: DO NOT EDIT THIS FILE

package org.bytedeco.cuda.cudart;

Expand Down
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
// Targeted by JavaCPP version 1.5.6: DO NOT EDIT THIS FILE
// Targeted by JavaCPP version 1.5.7-SNAPSHOT: DO NOT EDIT THIS FILE

package org.bytedeco.cuda.cudart;

Expand Down
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
// Targeted by JavaCPP version 1.5.6: DO NOT EDIT THIS FILE
// Targeted by JavaCPP version 1.5.7-SNAPSHOT: DO NOT EDIT THIS FILE

package org.bytedeco.cuda.cudart;

Expand Down
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
// Targeted by JavaCPP version 1.5.6: DO NOT EDIT THIS FILE
// Targeted by JavaCPP version 1.5.7-SNAPSHOT: DO NOT EDIT THIS FILE

package org.bytedeco.cuda.cudart;

Expand Down
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
// Targeted by JavaCPP version 1.5.6: DO NOT EDIT THIS FILE
// Targeted by JavaCPP version 1.5.7-SNAPSHOT: DO NOT EDIT THIS FILE

package org.bytedeco.cuda.cudart;

Expand Down
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
// Targeted by JavaCPP version 1.5.6: DO NOT EDIT THIS FILE
// Targeted by JavaCPP version 1.5.7-SNAPSHOT: DO NOT EDIT THIS FILE

package org.bytedeco.cuda.cudart;

Expand Down
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
// Targeted by JavaCPP version 1.5.6: DO NOT EDIT THIS FILE
// Targeted by JavaCPP version 1.5.7-SNAPSHOT: DO NOT EDIT THIS FILE

package org.bytedeco.cuda.cudart;

Expand Down
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
// Targeted by JavaCPP version 1.5.6: DO NOT EDIT THIS FILE
// Targeted by JavaCPP version 1.5.7-SNAPSHOT: DO NOT EDIT THIS FILE

package org.bytedeco.cuda.cudart;

Expand Down
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
// Targeted by JavaCPP version 1.5.6: DO NOT EDIT THIS FILE
// Targeted by JavaCPP version 1.5.7-SNAPSHOT: DO NOT EDIT THIS FILE

package org.bytedeco.cuda.cudart;

Expand Down
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
// Targeted by JavaCPP version 1.5.6: DO NOT EDIT THIS FILE
// Targeted by JavaCPP version 1.5.7-SNAPSHOT: DO NOT EDIT THIS FILE

package org.bytedeco.cuda.cudart;

Expand Down
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
// Targeted by JavaCPP version 1.5.6: DO NOT EDIT THIS FILE
// Targeted by JavaCPP version 1.5.7-SNAPSHOT: DO NOT EDIT THIS FILE

package org.bytedeco.cuda.cudart;

Expand Down
Loading

0 comments on commit 02c27e4

Please sign in to comment.