عکس PariKhaleghi
Finite Field based on GPUC++
موضوع‌ها
۰
فورک‌ها
۰
ستاره‌ها
۴
تاریخ ایجاد
۳ دی ۱۴۰۰
آخرین بروزرسانی
۸ ماه قبل
لایسنس
BSD 3-Clause "New" or "Revised" License

ff-GPGPU

Finite Field based on GPU

Background

In recent times, I've been interested in Finite Field operations, so I decided to implement few fields in SYCL DPC++, targeting accelerators ( specifically GPGPUs ).

In this repository, currently I keep implementation of two finite field's arithmetic operations, accompanied with relevant benchmarks on both CPU, GPGPU.

  • Binary Extension Field F(2 ** 32)
  • Prime Field F(2 ** 64 - 2 ** 32 + 1)

I've also written following implementations, along with benchmark results on CPU, GPU.

  • Rescue Prime Hash function
  • Six step algorithm based (I)Number Theoretic Transform
  • Cooley-Tukey algorithm based (I)NTT
  • DFT-style (I)NTT

Prerequisites

  • Make sure you've make, clang-format and dpcpp/ clang++ installed
  • You can build DPC++ compiler from source, check here
  • Or you may want to download pre-compiled Intel oneAPI toolkit, includes both compilers, check here [recommended]

Benchmarks

  • I'm running
$ lsb_release -d

Description:    Ubuntu 20.04.3 LTS
  • As compiler, I'm using
$ dpcpp --version

Intel(R) oneAPI DPC++/C++ Compiler 2021.3.0 (2021.3.0.20210619)
Target: x86_64-unknown-linux-gnu
Thread model: posix
InstalledDir: /opt/intel/oneapi/compiler/2021.3.0/linux/bin
  • Compile, link & run
make # JIT kernel compilation on *default* device, for AOT read below
./run
  • But you may be interested in targeting specific device, if you've multiple devices to choose from in runtime. Specify target device (which is to be used in runtime) during compilation phase using
DEVICE=cpu make   # still JIT, but in runtime use CPU
DEVICE=gpu make   # still JIT, but in runtime use GPU
DEVICE=host make  # still JIT, but in runtime use HOST
  • Clean using
make clean
  • Format source, if required
make format
  • Aforementioned steps compile kernels JIT, but if target device is already known it's better to compile them AOT, which saves some time in runtime, though compiled binary is device specific now.

  • I provide AOT kernel compilation recipe for CPUs using avx2 instructions. You can check whether your CPU supports that.

lscpu | grep -i avx
  • AOT Compilation targeting CPU can be invoked using
DEVICE=cpu make aot_cpu
  • I also provide recipe for AOT compiling kernels targeting Intel Iris Xe Max Graphics
DEVICE=gpu make aot_gpu

You may have some other hardware, consider taking a look at AOT compilation guidelines & make necessary changes in Makefile.


Targeting Nvidia GPU with CUDA backend :

For targeting Nvidia GPU, you want to run DEVICE=gpu make cuda, so that benchmark suite is compiled for CUDA backend.


I run benchmark suite on both Intel CPU/ GPU and Nvidia GPU, keeping results 👇

Tests

You can run basic test cases using

# set variable to runtime target device

DEVICE=cpu|gpu|host make test 

There's another set of randomised test cases, which asserts results ( obtained from my prime field implementation ) with another finite field implementation module, written in Python, named galois.

For running those, I suggest you first compile shared object using

# set variable to runtime target device

DEVICE=cpu|gpu|host make genlib

After that you can follow next steps here.