79e3f7d54b
When gcc is configured for nvptx offloading with --without-cuda-driver and full CUDA isn't installed, many libgomp.oacc-*/* tests fail, some of them because cuda.h header can't be found, others because the tests can't be linked against -lcuda, -lcudart or -lcublas. I usually only have akmod-nvidia and xorg-x11-drv-nvidia-cuda rpms installed, so libcuda.so.1 can be dlopened and the offloading works, but linking against those libraries isn't possible nor are the headers around (for the plugin itself there is the fallback libgomp/plugin/cuda/cuda.h). The following patch adds 3 new effective targets and uses them in tests that needs those. 2021-05-27 Jakub Jelinek <jakub@redhat.com> * testsuite/lib/libgomp.exp (check_effective_target_openacc_cuda, check_effective_target_openacc_cublas, check_effective_target_openacc_cudart): New. * testsuite/libgomp.oacc-fortran/host_data-4.f90: Require effective target openacc_cublas. * testsuite/libgomp.oacc-fortran/host_data-2.f90: Likewise. * testsuite/libgomp.oacc-fortran/host_data-3.f: Likewise. * testsuite/libgomp.oacc-c-c++-common/lib-91.c: Require effective target openacc_cuda. * testsuite/libgomp.oacc-c-c++-common/lib-70.c: Likewise. * testsuite/libgomp.oacc-c-c++-common/lib-90.c: Likewise. * testsuite/libgomp.oacc-c-c++-common/lib-75.c: Likewise. * testsuite/libgomp.oacc-c-c++-common/lib-69.c: Likewise. * testsuite/libgomp.oacc-c-c++-common/lib-74.c: Likewise. * testsuite/libgomp.oacc-c-c++-common/lib-81.c: Likewise. * testsuite/libgomp.oacc-c-c++-common/lib-72.c: Likewise. * testsuite/libgomp.oacc-c-c++-common/lib-85.c: Likewise. * testsuite/libgomp.oacc-c-c++-common/pr87835.c: Likewise. * testsuite/libgomp.oacc-c-c++-common/lib-82.c: Likewise. * testsuite/libgomp.oacc-c-c++-common/lib-73.c: Likewise. * testsuite/libgomp.oacc-c-c++-common/lib-83.c: Likewise. * testsuite/libgomp.oacc-c-c++-common/lib-78.c: Likewise. * testsuite/libgomp.oacc-c-c++-common/lib-76.c: Likewise. * testsuite/libgomp.oacc-c-c++-common/lib-84.c: Likewise. * testsuite/libgomp.oacc-c-c++-common/lib-79.c: Likewise. * testsuite/libgomp.oacc-c-c++-common/host_data-1.c: Require effective targets openacc_cublas and openacc_cudart. * testsuite/libgomp.oacc-c-c++-common/context-1.c: Likewise. * testsuite/libgomp.oacc-c-c++-common/context-2.c: Likewise. * testsuite/libgomp.oacc-c-c++-common/context-3.c: Likewise. * testsuite/libgomp.oacc-c-c++-common/context-4.c: Likewise. * testsuite/libgomp.oacc-c-c++-common/acc_get_property-nvptx.c: Require effective target openacc_cudart. * testsuite/libgomp.oacc-c-c++-common/asyncwait-1.c: Add -DUSE_CUDA_H for effective target openacc_cuda and add && defined USE_CUDA_H to preprocessor conditionals. Guard -lcuda also on openacc_cuda effective target.
136 lines
2.7 KiB
C
136 lines
2.7 KiB
C
/* { dg-do run { target openacc_nvidia_accel_selected } } */
|
|
/* { dg-additional-options "-lcuda" } */
|
|
/* { dg-require-effective-target openacc_cuda } */
|
|
|
|
#include <stdio.h>
|
|
#include <unistd.h>
|
|
#include <stdlib.h>
|
|
#include <openacc.h>
|
|
#include <cuda.h>
|
|
|
|
int
|
|
main (int argc, char **argv)
|
|
{
|
|
CUdevice dev;
|
|
CUfunction delay;
|
|
CUmodule module;
|
|
CUresult r;
|
|
const int N = 10;
|
|
int i;
|
|
CUstream streams[N];
|
|
unsigned long *a, *d_a, dticks;
|
|
int nbytes;
|
|
float dtime;
|
|
void *kargs[2];
|
|
int clkrate;
|
|
int devnum, nprocs;
|
|
|
|
acc_init (acc_device_nvidia);
|
|
|
|
devnum = acc_get_device_num (acc_device_nvidia);
|
|
|
|
r = cuDeviceGet (&dev, devnum);
|
|
if (r != CUDA_SUCCESS)
|
|
{
|
|
fprintf (stderr, "cuDeviceGet failed: %d\n", r);
|
|
abort ();
|
|
}
|
|
|
|
r =
|
|
cuDeviceGetAttribute (&nprocs, CU_DEVICE_ATTRIBUTE_MULTIPROCESSOR_COUNT,
|
|
dev);
|
|
if (r != CUDA_SUCCESS)
|
|
{
|
|
fprintf (stderr, "cuDeviceGetAttribute failed: %d\n", r);
|
|
abort ();
|
|
}
|
|
|
|
r = cuDeviceGetAttribute (&clkrate, CU_DEVICE_ATTRIBUTE_CLOCK_RATE, dev);
|
|
if (r != CUDA_SUCCESS)
|
|
{
|
|
fprintf (stderr, "cuDeviceGetAttribute failed: %d\n", r);
|
|
abort ();
|
|
}
|
|
|
|
r = cuModuleLoad (&module, "subr.ptx");
|
|
if (r != CUDA_SUCCESS)
|
|
{
|
|
fprintf (stderr, "cuModuleLoad failed: %d\n", r);
|
|
abort ();
|
|
}
|
|
|
|
r = cuModuleGetFunction (&delay, module, "delay");
|
|
if (r != CUDA_SUCCESS)
|
|
{
|
|
fprintf (stderr, "cuModuleGetFunction failed: %d\n", r);
|
|
abort ();
|
|
}
|
|
|
|
nbytes = nprocs * sizeof (unsigned long);
|
|
|
|
dtime = 200.0;
|
|
|
|
dticks = (unsigned long) (dtime * clkrate);
|
|
|
|
a = (unsigned long *) malloc (nbytes);
|
|
d_a = (unsigned long *) acc_malloc (nbytes);
|
|
|
|
acc_map_data (a, d_a, nbytes);
|
|
|
|
kargs[0] = (void *) &d_a;
|
|
kargs[1] = (void *) &dticks;
|
|
|
|
for (i = 0; i < N; i++)
|
|
{
|
|
streams[i] = (CUstream) acc_get_cuda_stream (i);
|
|
if (streams[i] != NULL)
|
|
abort ();
|
|
|
|
r = cuStreamCreate (&streams[i], CU_STREAM_DEFAULT);
|
|
if (r != CUDA_SUCCESS)
|
|
{
|
|
fprintf (stderr, "cuStreamCreate failed: %d\n", r);
|
|
abort ();
|
|
}
|
|
|
|
if (!acc_set_cuda_stream (i, streams[i]))
|
|
abort ();
|
|
}
|
|
|
|
for (i = 0; i < N; i++)
|
|
{
|
|
r = cuLaunchKernel (delay, 1, 1, 1, 1, 1, 1, 0, streams[i], kargs, 0);
|
|
if (r != CUDA_SUCCESS)
|
|
{
|
|
fprintf (stderr, "cuLaunchKernel failed: %d\n", r);
|
|
abort ();
|
|
}
|
|
|
|
}
|
|
|
|
if (acc_async_test_all () != 0)
|
|
{
|
|
fprintf (stderr, "asynchronous operation not running\n");
|
|
abort ();
|
|
}
|
|
|
|
sleep ((int) (dtime / 1000.0f) + 1);
|
|
|
|
if (acc_async_test_all () != 1)
|
|
{
|
|
fprintf (stderr, "asynchronous operation not running\n");
|
|
abort ();
|
|
}
|
|
|
|
acc_unmap_data (a);
|
|
|
|
free (a);
|
|
acc_free (d_a);
|
|
|
|
acc_shutdown (acc_device_nvidia);
|
|
|
|
exit (0);
|
|
}
|
|
|
|
/* { dg-output "" } */
|