Skip to content

Commit 187ff96

Browse files
authored
Merge pull request #768 from JuliaGPU/tb/cuda
Update to CUDA 11.2 Update 2.
2 parents 5e3d31b + 02199ff commit 187ff96

File tree

12 files changed

+73
-27
lines changed

12 files changed

+73
-27
lines changed

Artifacts.toml

Lines changed: 12 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -151,35 +151,35 @@ lazy = true
151151

152152
[[CUDA112]]
153153
arch = "powerpc64le"
154-
git-tree-sha1 = "ef3928da3f9b68a5213a93f91da0d27e32c01e50"
154+
git-tree-sha1 = "a0b84558b5a0c78f4d489f7b421ec4722899d6c0"
155155
libc = "glibc"
156-
os = "linux"
157156
lazy = true
157+
os = "linux"
158158

159159
[[CUDA112.download]]
160-
sha256 = "770235b69868b88e6db4efc30a8659e9708f3b432028e2032ba589cf2c3efaf8"
161-
url = "https://github.com/JuliaBinaryWrappers/CUDA_jll.jl/releases/download/CUDA-v11.2.1+1/CUDA.v11.2.1.powerpc64le-linux-gnu.tar.gz"
160+
sha256 = "23e3e4b0d0b325e223edd55e5b7dcc8d2a11722d0169b1dcd1a4a548e3f4d9c6"
161+
url = "https://github.com/JuliaBinaryWrappers/CUDA_jll.jl/releases/download/CUDA-v11.2.2+0/CUDA.v11.2.2.powerpc64le-linux-gnu.tar.gz"
162162

163163
[[CUDA112]]
164164
arch = "x86_64"
165-
git-tree-sha1 = "18f4e83091aec02d8229c2b009a45a5c22b47664"
165+
git-tree-sha1 = "53345f55f74b563606017c90cbafd05b61cd1768"
166166
libc = "glibc"
167-
os = "linux"
168167
lazy = true
168+
os = "linux"
169169

170170
[[CUDA112.download]]
171-
sha256 = "6da495c82fae19e0aae8691addc72829376547543324358f39e16835cb208e6e"
172-
url = "https://github.com/JuliaBinaryWrappers/CUDA_jll.jl/releases/download/CUDA-v11.2.1+1/CUDA.v11.2.1.x86_64-linux-gnu.tar.gz"
171+
sha256 = "a644bda2c4d7936d220ddbbe5719a2db5047659f23cc35b53bc4c9f7f6211711"
172+
url = "https://github.com/JuliaBinaryWrappers/CUDA_jll.jl/releases/download/CUDA-v11.2.2+0/CUDA.v11.2.2.x86_64-linux-gnu.tar.gz"
173173

174174
[[CUDA112]]
175175
arch = "x86_64"
176-
git-tree-sha1 = "4765905e93e1e93ca8d2eb52a1e8cec5de4627b1"
177-
os = "windows"
176+
git-tree-sha1 = "81069bd6e35b73f32622f78c73f5f32c3db4e47a"
178177
lazy = true
178+
os = "windows"
179179

180180
[[CUDA112.download]]
181-
sha256 = "6dc0ae6aab8b878864bf926fd9446c71f92f689e6115d6dcedc54ac492d30ea3"
182-
url = "https://github.com/JuliaBinaryWrappers/CUDA_jll.jl/releases/download/CUDA-v11.2.1+1/CUDA.v11.2.1.x86_64-w64-mingw32.tar.gz"
181+
sha256 = "36d22665027bca6eb9a292fd542344e8877b5089b2dacc29d0e334320d49e840"
182+
url = "https://github.com/JuliaBinaryWrappers/CUDA_jll.jl/releases/download/CUDA-v11.2.2+0/CUDA.v11.2.2.x86_64-w64-mingw32.tar.gz"
183183

184184

185185

deps/bindeps.jl

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -126,7 +126,7 @@ lazy_artifact(x) = @artifact_str(x)
126126
# NOTE: we don't use autogenerated JLLs, because we have multiple artifacts and need to
127127
# decide at run time (i.e. not via package dependencies) which one to use.
128128
const cuda_artifacts = Dict(
129-
(release=v"11.2", version=v"11.2.1", preferred=true) => ()->lazy_artifact("CUDA112"),
129+
(release=v"11.2", version=v"11.2.2", preferred=true) => ()->lazy_artifact("CUDA112"),
130130
(release=v"11.1", version=v"11.1.1", preferred=true) => ()->lazy_artifact("CUDA111"),
131131
(release=v"11.0", version=v"11.0.3", preferred=true) => ()->lazy_artifact("CUDA110"),
132132
(release=v"10.2", version=v"10.2.89", preferred=true) => ()->lazy_artifact("CUDA102"),

deps/discovery.jl

Lines changed: 17 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -257,6 +257,20 @@ const cuda_library_versions = Dict(
257257
"npp" => v"11.3.2", #.139
258258
"nvjpeg" => v"11.4.0", #.135
259259
),
260+
v"11.2.2" => Dict(
261+
"cudart" => v"11.2.152",
262+
"cupti" => "2020.3.1", # docs mention 11.2.152
263+
"nvrtc" => v"11.2.152",
264+
"nvtx" => v"11.2.152",
265+
"nvvp" => v"11.2.152",
266+
"cublas" => v"11.4.1", #.1043
267+
"cufft" => v"10.4.1", #.152
268+
"curand" => v"10.2.3", #.152
269+
"cusolver" => v"11.1.0", #.152
270+
"cusparse" => v"11.4.1", #.1152
271+
"npp" => v"11.3.2", #.152
272+
"nvjpeg" => v"11.4.0", #.152
273+
),
260274
)
261275

262276
function cuda_library_version(library, toolkit_version)
@@ -301,6 +315,9 @@ const cuda_binary_versions = Dict(
301315
v"11.2.1" => Dict(
302316
"nvdisasm" => v"11.2.135"
303317
),
318+
v"11.2.2" => Dict(
319+
"nvdisasm" => v"11.2.152"
320+
),
304321
)
305322

306323
# simplified find_library/find_binary entry-points,

lib/cublas/libcublas_common.jl

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@
33
const CUBLAS_VER_MAJOR = 11
44
const CUBLAS_VER_MINOR = 4
55
const CUBLAS_VER_PATCH = 1
6-
const CUBLAS_VER_BUILD = 1026
6+
const CUBLAS_VER_BUILD = 1043
77
const CUBLAS_VERSION = CUBLAS_VER_MAJOR * 1000 + CUBLAS_VER_MINOR * 100 + CUBLAS_VER_PATCH
88

99
@cenum cublasStatus_t::UInt32 begin

lib/cufft/libcufft_common.jl

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -4,9 +4,9 @@
44

55
const CUFFT_VER_MAJOR = 10
66
const CUFFT_VER_MINOR = 4
7-
const CUFFT_VER_PATCH = 0
8-
const CUFFT_VER_BUILD = 135
9-
const CUFFT_VERSION = 10400
7+
const CUFFT_VER_PATCH = 1
8+
const CUFFT_VER_BUILD = 152
9+
const CUFFT_VERSION = 10401
1010
const MAX_CUFFT_ERROR = 0x11
1111
const CUFFT_FORWARD = -1
1212
const CUFFT_INVERSE = 1

lib/curand/libcurand_common.jl

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@
33
const CURAND_VER_MAJOR = 10
44
const CURAND_VER_MINOR = 2
55
const CURAND_VER_PATCH = 3
6-
const CURAND_VER_BUILD = 135
6+
const CURAND_VER_BUILD = 152
77
const CURAND_VERSION = CURAND_VER_MAJOR * 1000 + CURAND_VER_MINOR * 100 + CURAND_VER_PATCH
88

99
@cenum curandStatus::UInt32 begin

lib/cusolver/libcusolver_common.jl

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@
33
const CUSOLVER_VER_MAJOR = 11
44
const CUSOLVER_VER_MINOR = 1
55
const CUSOLVER_VER_PATCH = 0
6-
const CUSOLVER_VER_BUILD = 135
6+
const CUSOLVER_VER_BUILD = 152
77
const CUSOLVER_VERSION = CUSOLVER_VER_MAJOR * 1000 + CUSOLVER_VER_MINOR * 100 + CUSOLVER_VER_PATCH
88
const cusolverDnContext = Cvoid
99
const cusolverDnHandle_t = Ptr{cusolverDnContext}

lib/cusparse/libcusparse_common.jl

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -2,8 +2,8 @@
22

33
const CUSPARSE_VER_MAJOR = 11
44
const CUSPARSE_VER_MINOR = 4
5-
const CUSPARSE_VER_PATCH = 0
6-
const CUSPARSE_VER_BUILD = 135
5+
const CUSPARSE_VER_PATCH = 1
6+
const CUSPARSE_VER_BUILD = 1152
77
const CUSPARSE_VERSION = CUSPARSE_VER_MAJOR * 1000 + CUSPARSE_VER_MINOR * 100 + CUSPARSE_VER_PATCH
88

99
# Skipping MacroDefinition: CUSPARSE_DEPRECATED ( new_func ) __attribute__ ( ( deprecated ( "please use " # new_func " instead" ) ) )

lib/nvml/libnvml.jl

Lines changed: 7 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1601,4 +1601,11 @@ end
16011601
ccall((:nvmlDeviceSetTemperatureThreshold, libnvml()), nvmlReturn_t, (nvmlDevice_t, nvmlTemperatureThresholds_t, Ptr{Cint}), device, thresholdType, temp)
16021602
end
16031603

1604+
## Added in CUDA 11.2 Update 2
1605+
1606+
@checked function nvmlDeviceCreateGpuInstanceWithPlacement(device, profileId, placement, gpuInstance)
1607+
initialize_api()
1608+
ccall((:nvmlDeviceCreateGpuInstanceWithPlacement, libnvml), nvmlReturn_t, (nvmlDevice_t, UInt32, Ptr{nvmlGpuInstancePlacement_t}, Ptr{nvmlGpuInstance_t}), device, profileId, placement, gpuInstance)
1609+
end
1610+
16041611
##

lib/nvml/libnvml_common.jl

Lines changed: 15 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -262,14 +262,16 @@ const NVML_GPU_INSTANCE_PROFILE_3_SLICE = 0x02
262262
const NVML_GPU_INSTANCE_PROFILE_4_SLICE = 0x03
263263
const NVML_GPU_INSTANCE_PROFILE_7_SLICE = 0x04
264264
const NVML_GPU_INSTANCE_PROFILE_8_SLICE = 0x05
265-
const NVML_GPU_INSTANCE_PROFILE_COUNT = 0x06
265+
const NVML_GPU_INSTANCE_PROFILE_6_SLICE = 0x06
266+
const NVML_GPU_INSTANCE_PROFILE_COUNT = 0x07
266267
const NVML_COMPUTE_INSTANCE_PROFILE_1_SLICE = 0x00
267268
const NVML_COMPUTE_INSTANCE_PROFILE_2_SLICE = 0x01
268269
const NVML_COMPUTE_INSTANCE_PROFILE_3_SLICE = 0x02
269270
const NVML_COMPUTE_INSTANCE_PROFILE_4_SLICE = 0x03
270271
const NVML_COMPUTE_INSTANCE_PROFILE_7_SLICE = 0x04
271272
const NVML_COMPUTE_INSTANCE_PROFILE_8_SLICE = 0x05
272-
const NVML_COMPUTE_INSTANCE_PROFILE_COUNT = 0x06
273+
const NVML_COMPUTE_INSTANCE_PROFILE_6_SLICE = 0x06
274+
const NVML_COMPUTE_INSTANCE_PROFILE_COUNT = 0x07
273275
const NVML_COMPUTE_INSTANCE_ENGINE_PROFILE_SHARED = 0x00
274276
const NVML_COMPUTE_INSTANCE_ENGINE_PROFILE_COUNT = 0x01
275277
const nvmlDevice_st = Cvoid
@@ -531,7 +533,17 @@ const nvmlEnableState_t = nvmlEnableState_enum
531533
NVML_BRAND_GRID = 4
532534
NVML_BRAND_GEFORCE = 5
533535
NVML_BRAND_TITAN = 6
534-
NVML_BRAND_COUNT = 7
536+
NVML_BRAND_NVIDIA_VAPPS = 7
537+
NVML_BRAND_NVIDIA_VPC = 8
538+
NVML_BRAND_NVIDIA_VCS = 9
539+
NVML_BRAND_NVIDIA_VWS = 10
540+
NVML_BRAND_NVIDIA_VGAMING = 11
541+
NVML_BRAND_QUADRO_RTX = 12
542+
NVML_BRAND_NVIDIA_RTX = 13
543+
NVML_BRAND_NVIDIA = 14
544+
NVML_BRAND_GEFORCE_RTX = 15
545+
NVML_BRAND_TITAN_RTX = 16
546+
NVML_BRAND_COUNT = 17
535547
end
536548

537549
const nvmlBrandType_t = nvmlBrandType_enum

0 commit comments

Comments
 (0)