All Projects → google → Nvidia_libs_test

google / Nvidia_libs_test

Licence: apache-2.0
Tests and benchmarks for cudnn (and in the future, other nvidia libraries)

Projects that are alternatives of or similar to Nvidia libs test

Mixbench
A GPU benchmark tool for evaluating GPUs on mixed operational intensity kernels (CUDA, OpenCL, HIP, SYCL)
Stars: ✭ 130 (+261.11%)
Mutual labels:  gpu, cuda, benchmark
gpu-monitor
Script to remotely check GPU servers for free GPUs
Stars: ✭ 85 (+136.11%)
Mutual labels:  gpu, cuda, cudnn
HeCBench
software.intel.com/content/www/us/en/develop/articles/repo-evaluating-performance-productivity-oneapi.html
Stars: ✭ 85 (+136.11%)
Mutual labels:  benchmark, cuda, gpu-computing
Pycuda
CUDA integration for Python, plus shiny features
Stars: ✭ 1,112 (+2988.89%)
Mutual labels:  gpu, cuda, gpu-computing
Neanderthal
Fast Clojure Matrix Library
Stars: ✭ 927 (+2475%)
Mutual labels:  gpu, cuda, gpu-computing
Deepnet
Deep.Net machine learning framework for F#
Stars: ✭ 99 (+175%)
Mutual labels:  gpu, cuda, gpu-computing
MatX
An efficient C++17 GPU numerical computing library with Python-like syntax
Stars: ✭ 418 (+1061.11%)
Mutual labels:  gpu, cuda, gpu-computing
cuda memtest
Fork of CUDA GPU memtest 👓
Stars: ✭ 68 (+88.89%)
Mutual labels:  gpu, cuda, gpu-computing
Hipsycl
Implementation of SYCL for CPUs, AMD GPUs, NVIDIA GPUs
Stars: ✭ 377 (+947.22%)
Mutual labels:  gpu, cuda, gpu-computing
Cuda Api Wrappers
Thin C++-flavored wrappers for the CUDA Runtime API
Stars: ✭ 362 (+905.56%)
Mutual labels:  gpu, cuda, gpu-computing
Chainer
A flexible framework of neural networks for deep learning
Stars: ✭ 5,656 (+15611.11%)
Mutual labels:  gpu, cudnn, cuda
Cupy
NumPy & SciPy for GPU
Stars: ✭ 5,625 (+15525%)
Mutual labels:  gpu, cudnn, cuda
Heteroflow
Concurrent CPU-GPU Programming using Task Models
Stars: ✭ 57 (+58.33%)
Mutual labels:  gpu, cuda, gpu-computing
Tensorflow Object Detection Tutorial
The purpose of this tutorial is to learn how to install and prepare TensorFlow framework to train your own convolutional neural network object detection classifier for multiple objects, starting from scratch
Stars: ✭ 113 (+213.89%)
Mutual labels:  gpu, cudnn, cuda
Bayadera
High-performance Bayesian Data Analysis on the GPU in Clojure
Stars: ✭ 342 (+850%)
Mutual labels:  gpu, cuda, gpu-computing
Stdgpu
stdgpu: Efficient STL-like Data Structures on the GPU
Stars: ✭ 531 (+1375%)
Mutual labels:  gpu, cuda, gpu-computing
Arraymancer
A fast, ergonomic and portable tensor library in Nim with a deep learning focus for CPU, GPU and embedded devices via OpenMP, Cuda and OpenCL backends
Stars: ✭ 793 (+2102.78%)
Mutual labels:  cudnn, cuda, gpu-computing
Speedtorch
Library for faster pinned CPU <-> GPU transfer in Pytorch
Stars: ✭ 615 (+1608.33%)
Mutual labels:  gpu, cuda
Cpu X
CPU-X is a Free software that gathers information on CPU, motherboard and more
Stars: ✭ 676 (+1777.78%)
Mutual labels:  gpu, benchmark
Kubernetes Gpu Guide
This guide should help fellow researchers and hobbyists to easily automate and accelerate there deep leaning training with their own Kubernetes GPU cluster.
Stars: ✭ 740 (+1955.56%)
Mutual labels:  gpu, gpu-computing

Tests and Benchmarks for cuDNN

The repository contains a set of convolution tests and benchmarks for NVIDIA's cuDNN library.

This is not an officially supported Google product.

Prerequisites

Install bazel (instructions).

Install the CUDA Toolkit (CUDA 8 is the minimal supported version)

Install the cuDNN SDK (cuDNN 6 is the minimal supported version).

Common parameters

Bazel parameters:

  • -c opt Build with optimizations. Recommended for benchmarks.

  • --action_env=CUDA_PATH=<path>: Path to the CUDA SDK directory. Default is /usr/local/cuda.

  • --action_env=CUDNN_PATH=<path>: Path to the CUDNN SDK directory. Default is CUDA_PATH.

  • --action_env=CC=<compiler>: Name of (or path to) the compiler. Examples: clang, gcc-6.

Executable parameters:

  • --cuda_device=<device_id>: CUDA device to use. Default is 0.

  • --device_memory_limit_mb=<size>: Upper limit of device memory (in megabytes) to use for cuDNN workspace after tensors have been allocated. Negative values specify an offset from the memory available at startup. Default is 4096.

Test instructions

bazel run [bazel parameters] //:cudnn_test -- [test parameters]

Bazel can run tests in a sandbox (which allows reporting crashes as failures). To run in a sandbox, replace 'bazel run ...' with 'bazel test ...' and prefix each test parameter with '--test_arg='

bazel test [bazel parameters] //:cudnn_test --test_arg=[test parameter 1] ...

Test parameters:

  • --gtest_filter=<pattern>: Only run tests that match the given pattern. Example pattern: '*FWD*'.

  • --proto_path=<path>: Path to textproto file that contains additional convolutions to run. Default is 'cudnn_tests.textproto'.

  • --gtest_random_seed=<value>: Seed for random generator. Changing the value produces tests with a different mix of tensor shapes, filter sizes, etc. Default is 0.

  • --gtest_also_run_disabled_tests: Include disabled tests (i.e. tests with names that start with DISABLED_).

  • --help for more options.

Benchmark instructions

bazel run [bazel parameters] //:cudnn_benchmark -- [benchmark parameters]

Benchmark parameters:

  • --benchmark_filter=<regex>: Only run tests that match the given regex pattern. Example: 'BWD_DATA'.

  • --proto_path=<path>: Path to textproto file that contains additional convolutions to run. Default is 'cudnn_benchmarks.textproto'.

  • --timing=<method>: How to measure execution time. One of 'kernel-duration' (default), 'kernel-cycles', or 'host-duration'.

  • --help and --helpfull for more options.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].