All Projects → szagoruyko → openai-gemm.pytorch

szagoruyko / openai-gemm.pytorch

Licence: other
PyTorch bindings for openai-gemm

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to openai-gemm.pytorch

cublasHgemm-P100
Code for testing the native float16 matrix multiplication performance on Tesla P100 and V100 GPU based on cublasHgemm
Stars: ✭ 35 (+75%)
Mutual labels:  gemm
slibs
Single file libraries for C/C++
Stars: ✭ 80 (+300%)
Mutual labels:  gemm
dbcsr
DBCSR: Distributed Block Compressed Sparse Row matrix library
Stars: ✭ 65 (+225%)
Mutual labels:  gemm
Optimizing-SGEMM-on-NVIDIA-Turing-GPUs
Optimizing SGEMM kernel functions on NVIDIA GPUs to a close-to-cuBLAS performance.
Stars: ✭ 78 (+290%)
Mutual labels:  gemm

openai-gemm.pytorch

PyTorch bindings for openai-gemm.

https://github.com/openai/openai-gemm

Installation

Clone original openai-gemm and add it to PYTHONPATH, install pycuda:

pip install pycuda

and follow instructions to install PyTorch on http://pytorch.org

No neon installation needed.

Usage

The library defines matmul function similar to the one that works with neon: https://github.com/openai/openai-gemm/blob/master/openai_gemm.py#L14, which instead of neon matrices takes torch.cuda.FloatTensor or torch.cuda.HalfTensor as A, B and C.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].