| /*===---- __clang_cuda_device_functions.h - CUDA runtime support -----------=== |
| * |
| * Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. |
| * See https://llvm.org/LICENSE.txt for license information. |
| * SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception |
| * |
| *===-----------------------------------------------------------------------=== |
| */ |
| |
| #ifndef __CLANG_CUDA_DEVICE_FUNCTIONS_H__ |
| #define __CLANG_CUDA_DEVICE_FUNCTIONS_H__ |
| |
| #ifndef _OPENMP |
| #if CUDA_VERSION < 9000 |
| #error This file is intended to be used with CUDA-9+ only. |
| #endif |
| #endif |
| |
| // __DEVICE__ is a helper macro with common set of attributes for the wrappers |
| // we implement in this file. We need static in order to avoid emitting unused |
| // functions and __forceinline__ helps inlining these wrappers at -O1. |
| #pragma push_macro("__DEVICE__") |
| #ifdef _OPENMP |
| #define __DEVICE__ static __attribute__((always_inline)) |
| #else |
| #define __DEVICE__ static __device__ __forceinline__ |
| #endif |
| |
| // libdevice provides fast low precision and slow full-recision implementations |
| // for some functions. Which one gets selected depends on |
| // __CLANG_CUDA_APPROX_TRANSCENDENTALS__ which gets defined by clang if |
| // -ffast-math or -fcuda-approx-transcendentals are in effect. |
| #pragma push_macro("__FAST_OR_SLOW") |
| #if defined(__CLANG_CUDA_APPROX_TRANSCENDENTALS__) |
| #define __FAST_OR_SLOW(fast, slow) fast |
| #else |
| #define __FAST_OR_SLOW(fast, slow) slow |
| #endif |
| |
| // For C++ 17 we need to include noexcept attribute to be compatible |
| // with the header-defined version. This may be removed once |
| // variant is supported. |
| #if defined(_OPENMP) && defined(__cplusplus) && __cplusplus >= 201703L |
| #define __NOEXCEPT noexcept |
| #else |
| #define __NOEXCEPT |
| #endif |
| |
| __DEVICE__ int __all(int __a) { return __nvvm_vote_all(__a); } |
| __DEVICE__ int __any(int __a) { return __nvvm_vote_any(__a); } |
| __DEVICE__ unsigned int __ballot(int __a) { return __nvvm_vote_ballot(__a); } |
| __DEVICE__ unsigned int __brev(unsigned int __a) { return __nv_brev(__a); } |
| __DEVICE__ unsigned long long __brevll(unsigned long long __a) { |
| return __nv_brevll(__a); |
| } |
| #if defined(__cplusplus) |
| __DEVICE__ void __brkpt() { asm volatile("brkpt;"); } |
| __DEVICE__ void __brkpt(int __a) { __brkpt(); } |
| #else |
| __DEVICE__ void __attribute__((overloadable)) __brkpt(void) { asm volatile("brkpt;"); } |
| __DEVICE__ void __attribute__((overloadable)) __brkpt(int __a) { __brkpt(); } |
| #endif |
| __DEVICE__ unsigned int __byte_perm(unsigned int __a, unsigned int __b, |
| unsigned int __c) { |
| return __nv_byte_perm(__a, __b, __c); |
| } |
| __DEVICE__ int __clz(int __a) { return __nv_clz(__a); } |
| __DEVICE__ int __clzll(long long __a) { return __nv_clzll(__a); } |
| __DEVICE__ float __cosf(float __a) { return __nv_fast_cosf(__a); } |
| __DEVICE__ double __dAtomicAdd(double *__p, double __v) { |
| return __nvvm_atom_add_gen_d(__p, __v); |
| } |
| __DEVICE__ double __dAtomicAdd_block(double *__p, double __v) { |
| return __nvvm_atom_cta_add_gen_d(__p, __v); |
| } |
| __DEVICE__ double __dAtomicAdd_system(double *__p, double __v) { |
| return __nvvm_atom_sys_add_gen_d(__p, __v); |
| } |
| __DEVICE__ double __dadd_rd(double __a, double __b) { |
| return __nv_dadd_rd(__a, __b); |
| } |
| __DEVICE__ double __dadd_rn(double __a, double __b) { |
| return __nv_dadd_rn(__a, __b); |
| } |
| __DEVICE__ double __dadd_ru(double __a, double __b) { |
| return __nv_dadd_ru(__a, __b); |
| } |
| __DEVICE__ double __dadd_rz(double __a, double __b) { |
| return __nv_dadd_rz(__a, __b); |
| } |
| __DEVICE__ double __ddiv_rd(double __a, double __b) { |
| return __nv_ddiv_rd(__a, __b); |
| } |
| __DEVICE__ double __ddiv_rn(double __a, double __b) { |
| return __nv_ddiv_rn(__a, __b); |
| } |
| __DEVICE__ double __ddiv_ru(double __a, double __b) { |
| return __nv_ddiv_ru(__a, __b); |
| } |
| __DEVICE__ double __ddiv_rz(double __a, double __b) { |
| return __nv_ddiv_rz(__a, __b); |
| } |
| __DEVICE__ double __dmul_rd(double __a, double __b) { |
| return __nv_dmul_rd(__a, __b); |
| } |
| __DEVICE__ double __dmul_rn(double __a, double __b) { |
| return __nv_dmul_rn(__a, __b); |
| } |
| __DEVICE__ double __dmul_ru(double __a, double __b) { |
| return __nv_dmul_ru(__a, __b); |
| } |
| __DEVICE__ double __dmul_rz(double __a, double __b) { |
| return __nv_dmul_rz(__a, __b); |
| } |
| __DEVICE__ float __double2float_rd(double __a) { |
| return __nv_double2float_rd(__a); |
| } |
| __DEVICE__ float __double2float_rn(double __a) { |
| return __nv_double2float_rn(__a); |
| } |
| __DEVICE__ float __double2float_ru(double __a) { |
| return __nv_double2float_ru(__a); |
| } |
| __DEVICE__ float __double2float_rz(double __a) { |
| return __nv_double2float_rz(__a); |
| } |
| __DEVICE__ int __double2hiint(double __a) { return __nv_double2hiint(__a); } |
| __DEVICE__ int __double2int_rd(double __a) { return __nv_double2int_rd(__a); } |
| __DEVICE__ int __double2int_rn(double __a) { return __nv_double2int_rn(__a); } |
| __DEVICE__ int __double2int_ru(double __a) { return __nv_double2int_ru(__a); } |
| __DEVICE__ int __double2int_rz(double __a) { return __nv_double2int_rz(__a); } |
| __DEVICE__ long long __double2ll_rd(double __a) { |
| return __nv_double2ll_rd(__a); |
| } |
| __DEVICE__ long long __double2ll_rn(double __a) { |
| return __nv_double2ll_rn(__a); |
| } |
| __DEVICE__ long long __double2ll_ru(double __a) { |
| return __nv_double2ll_ru(__a); |
| } |
| __DEVICE__ long long __double2ll_rz(double __a) { |
| return __nv_double2ll_rz(__a); |
| } |
| __DEVICE__ int __double2loint(double __a) { return __nv_double2loint(__a); } |
| __DEVICE__ unsigned int __double2uint_rd(double __a) { |
| return __nv_double2uint_rd(__a); |
| } |
| __DEVICE__ unsigned int __double2uint_rn(double __a) { |
| return __nv_double2uint_rn(__a); |
| } |
| __DEVICE__ unsigned int __double2uint_ru(double __a) { |
| return __nv_double2uint_ru(__a); |
| } |
| __DEVICE__ unsigned int __double2uint_rz(double __a) { |
| return __nv_double2uint_rz(__a); |
| } |
| __DEVICE__ unsigned long long __double2ull_rd(double __a) { |
| return __nv_double2ull_rd(__a); |
| } |
| __DEVICE__ unsigned long long __double2ull_rn(double __a) { |
| return __nv_double2ull_rn(__a); |
| } |
| __DEVICE__ unsigned long long __double2ull_ru(double __a) { |
| return __nv_double2ull_ru(__a); |
| } |
| __DEVICE__ unsigned long long __double2ull_rz(double __a) { |
| return __nv_double2ull_rz(__a); |
| } |
| __DEVICE__ long long __double_as_longlong(double __a) { |
| return __nv_double_as_longlong(__a); |
| } |
| __DEVICE__ double __drcp_rd(double __a) { return __nv_drcp_rd(__a); } |
| __DEVICE__ double __drcp_rn(double __a) { return __nv_drcp_rn(__a); } |
| __DEVICE__ double __drcp_ru(double __a) { return __nv_drcp_ru(__a); } |
| __DEVICE__ double __drcp_rz(double __a) { return __nv_drcp_rz(__a); } |
| __DEVICE__ double __dsqrt_rd(double __a) { return __nv_dsqrt_rd(__a); } |
| __DEVICE__ double __dsqrt_rn(double __a) { return __nv_dsqrt_rn(__a); } |
| __DEVICE__ double __dsqrt_ru(double __a) { return __nv_dsqrt_ru(__a); } |
| __DEVICE__ double __dsqrt_rz(double __a) { return __nv_dsqrt_rz(__a); } |
| __DEVICE__ double __dsub_rd(double __a, double __b) { |
| return __nv_dsub_rd(__a, __b); |
| } |
| __DEVICE__ double __dsub_rn(double __a, double __b) { |
| return __nv_dsub_rn(__a, __b); |
| } |
| __DEVICE__ double __dsub_ru(double __a, double __b) { |
| return __nv_dsub_ru(__a, __b); |
| } |
| __DEVICE__ double __dsub_rz(double __a, double __b) { |
| return __nv_dsub_rz(__a, __b); |
| } |
| __DEVICE__ float __exp10f(float __a) { return __nv_fast_exp10f(__a); } |
| __DEVICE__ float __expf(float __a) { return __nv_fast_expf(__a); } |
| __DEVICE__ float __fAtomicAdd(float *__p, float __v) { |
| return __nvvm_atom_add_gen_f(__p, __v); |
| } |
| __DEVICE__ float __fAtomicAdd_block(float *__p, float __v) { |
| return __nvvm_atom_cta_add_gen_f(__p, __v); |
| } |
| __DEVICE__ float __fAtomicAdd_system(float *__p, float __v) { |
| return __nvvm_atom_sys_add_gen_f(__p, __v); |
| } |
| __DEVICE__ float __fAtomicExch(float *__p, float __v) { |
| return __nv_int_as_float( |
| __nvvm_atom_xchg_gen_i((int *)__p, __nv_float_as_int(__v))); |
| } |
| __DEVICE__ float __fAtomicExch_block(float *__p, float __v) { |
| return __nv_int_as_float( |
| __nvvm_atom_cta_xchg_gen_i((int *)__p, __nv_float_as_int(__v))); |
| } |
| __DEVICE__ float __fAtomicExch_system(float *__p, float __v) { |
| return __nv_int_as_float( |
| __nvvm_atom_sys_xchg_gen_i((int *)__p, __nv_float_as_int(__v))); |
| } |
| __DEVICE__ float __fadd_rd(float __a, float __b) { |
| return __nv_fadd_rd(__a, __b); |
| } |
| __DEVICE__ float __fadd_rn(float __a, float __b) { |
| return __nv_fadd_rn(__a, __b); |
| } |
| __DEVICE__ float __fadd_ru(float __a, float __b) { |
| return __nv_fadd_ru(__a, __b); |
| } |
| __DEVICE__ float __fadd_rz(float __a, float __b) { |
| return __nv_fadd_rz(__a, __b); |
| } |
| __DEVICE__ float __fdiv_rd(float __a, float __b) { |
| return __nv_fdiv_rd(__a, __b); |
| } |
| __DEVICE__ float __fdiv_rn(float __a, float __b) { |
| return __nv_fdiv_rn(__a, __b); |
| } |
| __DEVICE__ float __fdiv_ru(float __a, float __b) { |
| return __nv_fdiv_ru(__a, __b); |
| } |
| __DEVICE__ float __fdiv_rz(float __a, float __b) { |
| return __nv_fdiv_rz(__a, __b); |
| } |
| __DEVICE__ float __fdividef(float __a, float __b) { |
| return __nv_fast_fdividef(__a, __b); |
| } |
| __DEVICE__ int __ffs(int __a) { return __nv_ffs(__a); } |
| __DEVICE__ int __ffsll(long long __a) { return __nv_ffsll(__a); } |
| __DEVICE__ int __finite(double __a) { return __nv_isfinited(__a); } |
| __DEVICE__ int __finitef(float __a) { return __nv_finitef(__a); } |
| #ifdef _MSC_VER |
| __DEVICE__ int __finitel(long double __a); |
| #endif |
| __DEVICE__ int __float2int_rd(float __a) { return __nv_float2int_rd(__a); } |
| __DEVICE__ int __float2int_rn(float __a) { return __nv_float2int_rn(__a); } |
| __DEVICE__ int __float2int_ru(float __a) { return __nv_float2int_ru(__a); } |
| __DEVICE__ int __float2int_rz(float __a) { return __nv_float2int_rz(__a); } |
| __DEVICE__ long long __float2ll_rd(float __a) { return __nv_float2ll_rd(__a); } |
| __DEVICE__ long long __float2ll_rn(float __a) { return __nv_float2ll_rn(__a); } |
| __DEVICE__ long long __float2ll_ru(float __a) { return __nv_float2ll_ru(__a); } |
| __DEVICE__ long long __float2ll_rz(float __a) { return __nv_float2ll_rz(__a); } |
| __DEVICE__ unsigned int __float2uint_rd(float __a) { |
| return __nv_float2uint_rd(__a); |
| } |
| __DEVICE__ unsigned int __float2uint_rn(float __a) { |
| return __nv_float2uint_rn(__a); |
| } |
| __DEVICE__ unsigned int __float2uint_ru(float __a) { |
| return __nv_float2uint_ru(__a); |
| } |
| __DEVICE__ unsigned int __float2uint_rz(float __a) { |
| return __nv_float2uint_rz(__a); |
| } |
| __DEVICE__ unsigned long long __float2ull_rd(float __a) { |
| return __nv_float2ull_rd(__a); |
| } |
| __DEVICE__ unsigned long long __float2ull_rn(float __a) { |
| return __nv_float2ull_rn(__a); |
| } |
| __DEVICE__ unsigned long long __float2ull_ru(float __a) { |
| return __nv_float2ull_ru(__a); |
| } |
| __DEVICE__ unsigned long long __float2ull_rz(float __a) { |
| return __nv_float2ull_rz(__a); |
| } |
| __DEVICE__ int __float_as_int(float __a) { return __nv_float_as_int(__a); } |
| __DEVICE__ unsigned int __float_as_uint(float __a) { |
| return __nv_float_as_uint(__a); |
| } |
| __DEVICE__ double __fma_rd(double __a, double __b, double __c) { |
| return __nv_fma_rd(__a, __b, __c); |
| } |
| __DEVICE__ double __fma_rn(double __a, double __b, double __c) { |
| return __nv_fma_rn(__a, __b, __c); |
| } |
| __DEVICE__ double __fma_ru(double __a, double __b, double __c) { |
| return __nv_fma_ru(__a, __b, __c); |
| } |
| __DEVICE__ double __fma_rz(double __a, double __b, double __c) { |
| return __nv_fma_rz(__a, __b, __c); |
| } |
| __DEVICE__ float __fmaf_ieee_rd(float __a, float __b, float __c) { |
| return __nv_fmaf_ieee_rd(__a, __b, __c); |
| } |
| __DEVICE__ float __fmaf_ieee_rn(float __a, float __b, float __c) { |
| return __nv_fmaf_ieee_rn(__a, __b, __c); |
| } |
| __DEVICE__ float __fmaf_ieee_ru(float __a, float __b, float __c) { |
| return __nv_fmaf_ieee_ru(__a, __b, __c); |
| } |
| __DEVICE__ float __fmaf_ieee_rz(float __a, float __b, float __c) { |
| return __nv_fmaf_ieee_rz(__a, __b, __c); |
| } |
| __DEVICE__ float __fmaf_rd(float __a, float __b, float __c) { |
| return __nv_fmaf_rd(__a, __b, __c); |
| } |
| __DEVICE__ float __fmaf_rn(float __a, float __b, float __c) { |
| return __nv_fmaf_rn(__a, __b, __c); |
| } |
| __DEVICE__ float __fmaf_ru(float __a, float __b, float __c) { |
| return __nv_fmaf_ru(__a, __b, __c); |
| } |
| __DEVICE__ float __fmaf_rz(float __a, float __b, float __c) { |
| return __nv_fmaf_rz(__a, __b, __c); |
| } |
| __DEVICE__ float __fmul_rd(float __a, float __b) { |
| return __nv_fmul_rd(__a, __b); |
| } |
| __DEVICE__ float __fmul_rn(float __a, float __b) { |
| return __nv_fmul_rn(__a, __b); |
| } |
| __DEVICE__ float __fmul_ru(float __a, float __b) { |
| return __nv_fmul_ru(__a, __b); |
| } |
| __DEVICE__ float __fmul_rz(float __a, float __b) { |
| return __nv_fmul_rz(__a, __b); |
| } |
| __DEVICE__ float __frcp_rd(float __a) { return __nv_frcp_rd(__a); } |
| __DEVICE__ float __frcp_rn(float __a) { return __nv_frcp_rn(__a); } |
| __DEVICE__ float __frcp_ru(float __a) { return __nv_frcp_ru(__a); } |
| __DEVICE__ float __frcp_rz(float __a) { return __nv_frcp_rz(__a); } |
| __DEVICE__ float __frsqrt_rn(float __a) { return __nv_frsqrt_rn(__a); } |
| __DEVICE__ float __fsqrt_rd(float __a) { return __nv_fsqrt_rd(__a); } |
| __DEVICE__ float __fsqrt_rn(float __a) { return __nv_fsqrt_rn(__a); } |
| __DEVICE__ float __fsqrt_ru(float __a) { return __nv_fsqrt_ru(__a); } |
| __DEVICE__ float __fsqrt_rz(float __a) { return __nv_fsqrt_rz(__a); } |
| __DEVICE__ float __fsub_rd(float __a, float __b) { |
| return __nv_fsub_rd(__a, __b); |
| } |
| __DEVICE__ float __fsub_rn(float __a, float __b) { |
| return __nv_fsub_rn(__a, __b); |
| } |
| __DEVICE__ float __fsub_ru(float __a, float __b) { |
| return __nv_fsub_ru(__a, __b); |
| } |
| __DEVICE__ float __fsub_rz(float __a, float __b) { |
| return __nv_fsub_rz(__a, __b); |
| } |
| __DEVICE__ int __hadd(int __a, int __b) { return __nv_hadd(__a, __b); } |
| __DEVICE__ double __hiloint2double(int __a, int __b) { |
| return __nv_hiloint2double(__a, __b); |
| } |
| __DEVICE__ int __iAtomicAdd(int *__p, int __v) { |
| return __nvvm_atom_add_gen_i(__p, __v); |
| } |
| __DEVICE__ int __iAtomicAdd_block(int *__p, int __v) { |
| __nvvm_atom_cta_add_gen_i(__p, __v); |
| } |
| __DEVICE__ int __iAtomicAdd_system(int *__p, int __v) { |
| __nvvm_atom_sys_add_gen_i(__p, __v); |
| } |
| __DEVICE__ int __iAtomicAnd(int *__p, int __v) { |
| return __nvvm_atom_and_gen_i(__p, __v); |
| } |
| __DEVICE__ int __iAtomicAnd_block(int *__p, int __v) { |
| return __nvvm_atom_cta_and_gen_i(__p, __v); |
| } |
| __DEVICE__ int __iAtomicAnd_system(int *__p, int __v) { |
| return __nvvm_atom_sys_and_gen_i(__p, __v); |
| } |
| __DEVICE__ int __iAtomicCAS(int *__p, int __cmp, int __v) { |
| return __nvvm_atom_cas_gen_i(__p, __cmp, __v); |
| } |
| __DEVICE__ int __iAtomicCAS_block(int *__p, int __cmp, int __v) { |
| return __nvvm_atom_cta_cas_gen_i(__p, __cmp, __v); |
| } |
| __DEVICE__ int __iAtomicCAS_system(int *__p, int __cmp, int __v) { |
| return __nvvm_atom_sys_cas_gen_i(__p, __cmp, __v); |
| } |
| __DEVICE__ int __iAtomicExch(int *__p, int __v) { |
| return __nvvm_atom_xchg_gen_i(__p, __v); |
| } |
| __DEVICE__ int __iAtomicExch_block(int *__p, int __v) { |
| return __nvvm_atom_cta_xchg_gen_i(__p, __v); |
| } |
| __DEVICE__ int __iAtomicExch_system(int *__p, int __v) { |
| return __nvvm_atom_sys_xchg_gen_i(__p, __v); |
| } |
| __DEVICE__ int __iAtomicMax(int *__p, int __v) { |
| return __nvvm_atom_max_gen_i(__p, __v); |
| } |
| __DEVICE__ int __iAtomicMax_block(int *__p, int __v) { |
| return __nvvm_atom_cta_max_gen_i(__p, __v); |
| } |
| __DEVICE__ int __iAtomicMax_system(int *__p, int __v) { |
| return __nvvm_atom_sys_max_gen_i(__p, __v); |
| } |
| __DEVICE__ int __iAtomicMin(int *__p, int __v) { |
| return __nvvm_atom_min_gen_i(__p, __v); |
| } |
| __DEVICE__ int __iAtomicMin_block(int *__p, int __v) { |
| return __nvvm_atom_cta_min_gen_i(__p, __v); |
| } |
| __DEVICE__ int __iAtomicMin_system(int *__p, int __v) { |
| return __nvvm_atom_sys_min_gen_i(__p, __v); |
| } |
| __DEVICE__ int __iAtomicOr(int *__p, int __v) { |
| return __nvvm_atom_or_gen_i(__p, __v); |
| } |
| __DEVICE__ int __iAtomicOr_block(int *__p, int __v) { |
| return __nvvm_atom_cta_or_gen_i(__p, __v); |
| } |
| __DEVICE__ int __iAtomicOr_system(int *__p, int __v) { |
| return __nvvm_atom_sys_or_gen_i(__p, __v); |
| } |
| __DEVICE__ int __iAtomicXor(int *__p, int __v) { |
| return __nvvm_atom_xor_gen_i(__p, __v); |
| } |
| __DEVICE__ int __iAtomicXor_block(int *__p, int __v) { |
| return __nvvm_atom_cta_xor_gen_i(__p, __v); |
| } |
| __DEVICE__ int __iAtomicXor_system(int *__p, int __v) { |
| return __nvvm_atom_sys_xor_gen_i(__p, __v); |
| } |
| __DEVICE__ long long __illAtomicMax(long long *__p, long long __v) { |
| return __nvvm_atom_max_gen_ll(__p, __v); |
| } |
| __DEVICE__ long long __illAtomicMax_block(long long *__p, long long __v) { |
| return __nvvm_atom_cta_max_gen_ll(__p, __v); |
| } |
| __DEVICE__ long long __illAtomicMax_system(long long *__p, long long __v) { |
| return __nvvm_atom_sys_max_gen_ll(__p, __v); |
| } |
| __DEVICE__ long long __illAtomicMin(long long *__p, long long __v) { |
| return __nvvm_atom_min_gen_ll(__p, __v); |
| } |
| __DEVICE__ long long __illAtomicMin_block(long long *__p, long long __v) { |
| return __nvvm_atom_cta_min_gen_ll(__p, __v); |
| } |
| __DEVICE__ long long __illAtomicMin_system(long long *__p, long long __v) { |
| return __nvvm_atom_sys_min_gen_ll(__p, __v); |
| } |
| __DEVICE__ double __int2double_rn(int __a) { return __nv_int2double_rn(__a); } |
| __DEVICE__ float __int2float_rd(int __a) { return __nv_int2float_rd(__a); } |
| __DEVICE__ float __int2float_rn(int __a) { return __nv_int2float_rn(__a); } |
| __DEVICE__ float __int2float_ru(int __a) { return __nv_int2float_ru(__a); } |
| __DEVICE__ float __int2float_rz(int __a) { return __nv_int2float_rz(__a); } |
| __DEVICE__ float __int_as_float(int __a) { return __nv_int_as_float(__a); } |
| __DEVICE__ int __isfinited(double __a) { return __nv_isfinited(__a); } |
| __DEVICE__ int __isinf(double __a) { return __nv_isinfd(__a); } |
| __DEVICE__ int __isinff(float __a) { return __nv_isinff(__a); } |
| #ifdef _MSC_VER |
| __DEVICE__ int __isinfl(long double __a); |
| #endif |
| __DEVICE__ int __isnan(double __a) { return __nv_isnand(__a); } |
| __DEVICE__ int __isnanf(float __a) { return __nv_isnanf(__a); } |
| #ifdef _MSC_VER |
| __DEVICE__ int __isnanl(long double __a); |
| #endif |
| __DEVICE__ double __ll2double_rd(long long __a) { |
| return __nv_ll2double_rd(__a); |
| } |
| __DEVICE__ double __ll2double_rn(long long __a) { |
| return __nv_ll2double_rn(__a); |
| } |
| __DEVICE__ double __ll2double_ru(long long __a) { |
| return __nv_ll2double_ru(__a); |
| } |
| __DEVICE__ double __ll2double_rz(long long __a) { |
| return __nv_ll2double_rz(__a); |
| } |
| __DEVICE__ float __ll2float_rd(long long __a) { return __nv_ll2float_rd(__a); } |
| __DEVICE__ float __ll2float_rn(long long __a) { return __nv_ll2float_rn(__a); } |
| __DEVICE__ float __ll2float_ru(long long __a) { return __nv_ll2float_ru(__a); } |
| __DEVICE__ float __ll2float_rz(long long __a) { return __nv_ll2float_rz(__a); } |
| __DEVICE__ long long __llAtomicAnd(long long *__p, long long __v) { |
| return __nvvm_atom_and_gen_ll(__p, __v); |
| } |
| __DEVICE__ long long __llAtomicAnd_block(long long *__p, long long __v) { |
| return __nvvm_atom_cta_and_gen_ll(__p, __v); |
| } |
| __DEVICE__ long long __llAtomicAnd_system(long long *__p, long long __v) { |
| return __nvvm_atom_sys_and_gen_ll(__p, __v); |
| } |
| __DEVICE__ long long __llAtomicOr(long long *__p, long long __v) { |
| return __nvvm_atom_or_gen_ll(__p, __v); |
| } |
| __DEVICE__ long long __llAtomicOr_block(long long *__p, long long __v) { |
| return __nvvm_atom_cta_or_gen_ll(__p, __v); |
| } |
| __DEVICE__ long long __llAtomicOr_system(long long *__p, long long __v) { |
| return __nvvm_atom_sys_or_gen_ll(__p, __v); |
| } |
| __DEVICE__ long long __llAtomicXor(long long *__p, long long __v) { |
| return __nvvm_atom_xor_gen_ll(__p, __v); |
| } |
| __DEVICE__ long long __llAtomicXor_block(long long *__p, long long __v) { |
| return __nvvm_atom_cta_xor_gen_ll(__p, __v); |
| } |
| __DEVICE__ long long __llAtomicXor_system(long long *__p, long long __v) { |
| return __nvvm_atom_sys_xor_gen_ll(__p, __v); |
| } |
| __DEVICE__ float __log10f(float __a) { return __nv_fast_log10f(__a); } |
| __DEVICE__ float __log2f(float __a) { return __nv_fast_log2f(__a); } |
| __DEVICE__ float __logf(float __a) { return __nv_fast_logf(__a); } |
| __DEVICE__ double __longlong_as_double(long long __a) { |
| return __nv_longlong_as_double(__a); |
| } |
| __DEVICE__ int __mul24(int __a, int __b) { return __nv_mul24(__a, __b); } |
| __DEVICE__ long long __mul64hi(long long __a, long long __b) { |
| return __nv_mul64hi(__a, __b); |
| } |
| __DEVICE__ int __mulhi(int __a, int __b) { return __nv_mulhi(__a, __b); } |
| __DEVICE__ unsigned int __pm0(void) { return __nvvm_read_ptx_sreg_pm0(); } |
| __DEVICE__ unsigned int __pm1(void) { return __nvvm_read_ptx_sreg_pm1(); } |
| __DEVICE__ unsigned int __pm2(void) { return __nvvm_read_ptx_sreg_pm2(); } |
| __DEVICE__ unsigned int __pm3(void) { return __nvvm_read_ptx_sreg_pm3(); } |
| __DEVICE__ int __popc(int __a) { return __nv_popc(__a); } |
| __DEVICE__ int __popcll(long long __a) { return __nv_popcll(__a); } |
| __DEVICE__ float __powf(float __a, float __b) { |
| return __nv_fast_powf(__a, __b); |
| } |
| |
| // Parameter must have a known integer value. |
| #define __prof_trigger(__a) asm __volatile__("pmevent \t%0;" ::"i"(__a)) |
| __DEVICE__ int __rhadd(int __a, int __b) { return __nv_rhadd(__a, __b); } |
| __DEVICE__ unsigned int __sad(int __a, int __b, unsigned int __c) { |
| return __nv_sad(__a, __b, __c); |
| } |
| __DEVICE__ float __saturatef(float __a) { return __nv_saturatef(__a); } |
| __DEVICE__ int __signbitd(double __a) { return __nv_signbitd(__a); } |
| __DEVICE__ int __signbitf(float __a) { return __nv_signbitf(__a); } |
| __DEVICE__ void __sincosf(float __a, float *__s, float *__c) { |
| return __nv_fast_sincosf(__a, __s, __c); |
| } |
| __DEVICE__ float __sinf(float __a) { return __nv_fast_sinf(__a); } |
| __DEVICE__ int __syncthreads_and(int __a) { return __nvvm_bar0_and(__a); } |
| __DEVICE__ int __syncthreads_count(int __a) { return __nvvm_bar0_popc(__a); } |
| __DEVICE__ int __syncthreads_or(int __a) { return __nvvm_bar0_or(__a); } |
| __DEVICE__ float __tanf(float __a) { return __nv_fast_tanf(__a); } |
| __DEVICE__ void __threadfence(void) { __nvvm_membar_gl(); } |
| __DEVICE__ void __threadfence_block(void) { __nvvm_membar_cta(); }; |
| __DEVICE__ void __threadfence_system(void) { __nvvm_membar_sys(); }; |
| __DEVICE__ void __trap(void) { asm volatile("trap;"); } |
| __DEVICE__ unsigned int __uAtomicAdd(unsigned int *__p, unsigned int __v) { |
| return __nvvm_atom_add_gen_i((int *)__p, __v); |
| } |
| __DEVICE__ unsigned int __uAtomicAdd_block(unsigned int *__p, |
| unsigned int __v) { |
| return __nvvm_atom_cta_add_gen_i((int *)__p, __v); |
| } |
| __DEVICE__ unsigned int __uAtomicAdd_system(unsigned int *__p, |
| unsigned int __v) { |
| return __nvvm_atom_sys_add_gen_i((int *)__p, __v); |
| } |
| __DEVICE__ unsigned int __uAtomicAnd(unsigned int *__p, unsigned int __v) { |
| return __nvvm_atom_and_gen_i((int *)__p, __v); |
| } |
| __DEVICE__ unsigned int __uAtomicAnd_block(unsigned int *__p, |
| unsigned int __v) { |
| return __nvvm_atom_cta_and_gen_i((int *)__p, __v); |
| } |
| __DEVICE__ unsigned int __uAtomicAnd_system(unsigned int *__p, |
| unsigned int __v) { |
| return __nvvm_atom_sys_and_gen_i((int *)__p, __v); |
| } |
| __DEVICE__ unsigned int __uAtomicCAS(unsigned int *__p, unsigned int __cmp, |
| unsigned int __v) { |
| return __nvvm_atom_cas_gen_i((int *)__p, __cmp, __v); |
| } |
| __DEVICE__ unsigned int |
| __uAtomicCAS_block(unsigned int *__p, unsigned int __cmp, unsigned int __v) { |
| return __nvvm_atom_cta_cas_gen_i((int *)__p, __cmp, __v); |
| } |
| __DEVICE__ unsigned int |
| __uAtomicCAS_system(unsigned int *__p, unsigned int __cmp, unsigned int __v) { |
| return __nvvm_atom_sys_cas_gen_i((int *)__p, __cmp, __v); |
| } |
| __DEVICE__ unsigned int __uAtomicDec(unsigned int *__p, unsigned int __v) { |
| return __nvvm_atom_dec_gen_ui(__p, __v); |
| } |
| __DEVICE__ unsigned int __uAtomicDec_block(unsigned int *__p, |
| unsigned int __v) { |
| return __nvvm_atom_cta_dec_gen_ui(__p, __v); |
| } |
| __DEVICE__ unsigned int __uAtomicDec_system(unsigned int *__p, |
| unsigned int __v) { |
| return __nvvm_atom_sys_dec_gen_ui(__p, __v); |
| } |
| __DEVICE__ unsigned int __uAtomicExch(unsigned int *__p, unsigned int __v) { |
| return __nvvm_atom_xchg_gen_i((int *)__p, __v); |
| } |
| __DEVICE__ unsigned int __uAtomicExch_block(unsigned int *__p, |
| unsigned int __v) { |
| return __nvvm_atom_cta_xchg_gen_i((int *)__p, __v); |
| } |
| __DEVICE__ unsigned int __uAtomicExch_system(unsigned int *__p, |
| unsigned int __v) { |
| return __nvvm_atom_sys_xchg_gen_i((int *)__p, __v); |
| } |
| __DEVICE__ unsigned int __uAtomicInc(unsigned int *__p, unsigned int __v) { |
| return __nvvm_atom_inc_gen_ui(__p, __v); |
| } |
| __DEVICE__ unsigned int __uAtomicInc_block(unsigned int *__p, |
| unsigned int __v) { |
| return __nvvm_atom_cta_inc_gen_ui(__p, __v); |
| } |
| __DEVICE__ unsigned int __uAtomicInc_system(unsigned int *__p, |
| unsigned int __v) { |
| return __nvvm_atom_sys_inc_gen_ui(__p, __v); |
| } |
| __DEVICE__ unsigned int __uAtomicMax(unsigned int *__p, unsigned int __v) { |
| return __nvvm_atom_max_gen_ui(__p, __v); |
| } |
| __DEVICE__ unsigned int __uAtomicMax_block(unsigned int *__p, |
| unsigned int __v) { |
| return __nvvm_atom_cta_max_gen_ui(__p, __v); |
| } |
| __DEVICE__ unsigned int __uAtomicMax_system(unsigned int *__p, |
| unsigned int __v) { |
| return __nvvm_atom_sys_max_gen_ui(__p, __v); |
| } |
| __DEVICE__ unsigned int __uAtomicMin(unsigned int *__p, unsigned int __v) { |
| return __nvvm_atom_min_gen_ui(__p, __v); |
| } |
| __DEVICE__ unsigned int __uAtomicMin_block(unsigned int *__p, |
| unsigned int __v) { |
| return __nvvm_atom_cta_min_gen_ui(__p, __v); |
| } |
| __DEVICE__ unsigned int __uAtomicMin_system(unsigned int *__p, |
| unsigned int __v) { |
| return __nvvm_atom_sys_min_gen_ui(__p, __v); |
| } |
| __DEVICE__ unsigned int __uAtomicOr(unsigned int *__p, unsigned int __v) { |
| return __nvvm_atom_or_gen_i((int *)__p, __v); |
| } |
| __DEVICE__ unsigned int __uAtomicOr_block(unsigned int *__p, unsigned int __v) { |
| return __nvvm_atom_cta_or_gen_i((int *)__p, __v); |
| } |
| __DEVICE__ unsigned int __uAtomicOr_system(unsigned int *__p, |
| unsigned int __v) { |
| return __nvvm_atom_sys_or_gen_i((int *)__p, __v); |
| } |
| __DEVICE__ unsigned int __uAtomicXor(unsigned int *__p, unsigned int __v) { |
| return __nvvm_atom_xor_gen_i((int *)__p, __v); |
| } |
| __DEVICE__ unsigned int __uAtomicXor_block(unsigned int *__p, |
| unsigned int __v) { |
| return __nvvm_atom_cta_xor_gen_i((int *)__p, __v); |
| } |
| __DEVICE__ unsigned int __uAtomicXor_system(unsigned int *__p, |
| unsigned int __v) { |
| return __nvvm_atom_sys_xor_gen_i((int *)__p, __v); |
| } |
| __DEVICE__ unsigned int __uhadd(unsigned int __a, unsigned int __b) { |
| return __nv_uhadd(__a, __b); |
| } |
| __DEVICE__ double __uint2double_rn(unsigned int __a) { |
| return __nv_uint2double_rn(__a); |
| } |
| __DEVICE__ float __uint2float_rd(unsigned int __a) { |
| return __nv_uint2float_rd(__a); |
| } |
| __DEVICE__ float __uint2float_rn(unsigned int __a) { |
| return __nv_uint2float_rn(__a); |
| } |
| __DEVICE__ float __uint2float_ru(unsigned int __a) { |
| return __nv_uint2float_ru(__a); |
| } |
| __DEVICE__ float __uint2float_rz(unsigned int __a) { |
| return __nv_uint2float_rz(__a); |
| } |
| __DEVICE__ float __uint_as_float(unsigned int __a) { |
| return __nv_uint_as_float(__a); |
| } // |
| __DEVICE__ double __ull2double_rd(unsigned long long __a) { |
| return __nv_ull2double_rd(__a); |
| } |
| __DEVICE__ double __ull2double_rn(unsigned long long __a) { |
| return __nv_ull2double_rn(__a); |
| } |
| __DEVICE__ double __ull2double_ru(unsigned long long __a) { |
| return __nv_ull2double_ru(__a); |
| } |
| __DEVICE__ double __ull2double_rz(unsigned long long __a) { |
| return __nv_ull2double_rz(__a); |
| } |
| __DEVICE__ float __ull2float_rd(unsigned long long __a) { |
| return __nv_ull2float_rd(__a); |
| } |
| __DEVICE__ float __ull2float_rn(unsigned long long __a) { |
| return __nv_ull2float_rn(__a); |
| } |
| __DEVICE__ float __ull2float_ru(unsigned long long __a) { |
| return __nv_ull2float_ru(__a); |
| } |
| __DEVICE__ float __ull2float_rz(unsigned long long __a) { |
| return __nv_ull2float_rz(__a); |
| } |
| __DEVICE__ unsigned long long __ullAtomicAdd(unsigned long long *__p, |
| unsigned long long __v) { |
| return __nvvm_atom_add_gen_ll((long long *)__p, __v); |
| } |
| __DEVICE__ unsigned long long __ullAtomicAdd_block(unsigned long long *__p, |
| unsigned long long __v) { |
| return __nvvm_atom_cta_add_gen_ll((long long *)__p, __v); |
| } |
| __DEVICE__ unsigned long long __ullAtomicAdd_system(unsigned long long *__p, |
| unsigned long long __v) { |
| return __nvvm_atom_sys_add_gen_ll((long long *)__p, __v); |
| } |
| __DEVICE__ unsigned long long __ullAtomicAnd(unsigned long long *__p, |
| unsigned long long __v) { |
| return __nvvm_atom_and_gen_ll((long long *)__p, __v); |
| } |
| __DEVICE__ unsigned long long __ullAtomicAnd_block(unsigned long long *__p, |
| unsigned long long __v) { |
| return __nvvm_atom_cta_and_gen_ll((long long *)__p, __v); |
| } |
| __DEVICE__ unsigned long long __ullAtomicAnd_system(unsigned long long *__p, |
| unsigned long long __v) { |
| return __nvvm_atom_sys_and_gen_ll((long long *)__p, __v); |
| } |
| __DEVICE__ unsigned long long __ullAtomicCAS(unsigned long long *__p, |
| unsigned long long __cmp, |
| unsigned long long __v) { |
| return __nvvm_atom_cas_gen_ll((long long *)__p, __cmp, __v); |
| } |
| __DEVICE__ unsigned long long __ullAtomicCAS_block(unsigned long long *__p, |
| unsigned long long __cmp, |
| unsigned long long __v) { |
| return __nvvm_atom_cta_cas_gen_ll((long long *)__p, __cmp, __v); |
| } |
| __DEVICE__ unsigned long long __ullAtomicCAS_system(unsigned long long *__p, |
| unsigned long long __cmp, |
| unsigned long long __v) { |
| return __nvvm_atom_sys_cas_gen_ll((long long *)__p, __cmp, __v); |
| } |
| __DEVICE__ unsigned long long __ullAtomicExch(unsigned long long *__p, |
| unsigned long long __v) { |
| return __nvvm_atom_xchg_gen_ll((long long *)__p, __v); |
| } |
| __DEVICE__ unsigned long long __ullAtomicExch_block(unsigned long long *__p, |
| unsigned long long __v) { |
| return __nvvm_atom_cta_xchg_gen_ll((long long *)__p, __v); |
| } |
| __DEVICE__ unsigned long long __ullAtomicExch_system(unsigned long long *__p, |
| unsigned long long __v) { |
| return __nvvm_atom_sys_xchg_gen_ll((long long *)__p, __v); |
| } |
| __DEVICE__ unsigned long long __ullAtomicMax(unsigned long long *__p, |
| unsigned long long __v) { |
| return __nvvm_atom_max_gen_ull(__p, __v); |
| } |
| __DEVICE__ unsigned long long __ullAtomicMax_block(unsigned long long *__p, |
| unsigned long long __v) { |
| return __nvvm_atom_cta_max_gen_ull(__p, __v); |
| } |
| __DEVICE__ unsigned long long __ullAtomicMax_system(unsigned long long *__p, |
| unsigned long long __v) { |
| return __nvvm_atom_sys_max_gen_ull(__p, __v); |
| } |
| __DEVICE__ unsigned long long __ullAtomicMin(unsigned long long *__p, |
| unsigned long long __v) { |
| return __nvvm_atom_min_gen_ull(__p, __v); |
| } |
| __DEVICE__ unsigned long long __ullAtomicMin_block(unsigned long long *__p, |
| unsigned long long __v) { |
| return __nvvm_atom_cta_min_gen_ull(__p, __v); |
| } |
| __DEVICE__ unsigned long long __ullAtomicMin_system(unsigned long long *__p, |
| unsigned long long __v) { |
| return __nvvm_atom_sys_min_gen_ull(__p, __v); |
| } |
| __DEVICE__ unsigned long long __ullAtomicOr(unsigned long long *__p, |
| unsigned long long __v) { |
| return __nvvm_atom_or_gen_ll((long long *)__p, __v); |
| } |
| __DEVICE__ unsigned long long __ullAtomicOr_block(unsigned long long *__p, |
| unsigned long long __v) { |
| return __nvvm_atom_cta_or_gen_ll((long long *)__p, __v); |
| } |
| __DEVICE__ unsigned long long __ullAtomicOr_system(unsigned long long *__p, |
| unsigned long long __v) { |
| return __nvvm_atom_sys_or_gen_ll((long long *)__p, __v); |
| } |
| __DEVICE__ unsigned long long __ullAtomicXor(unsigned long long *__p, |
| unsigned long long __v) { |
| return __nvvm_atom_xor_gen_ll((long long *)__p, __v); |
| } |
| __DEVICE__ unsigned long long __ullAtomicXor_block(unsigned long long *__p, |
| unsigned long long __v) { |
| return __nvvm_atom_cta_xor_gen_ll((long long *)__p, __v); |
| } |
| __DEVICE__ unsigned long long __ullAtomicXor_system(unsigned long long *__p, |
| unsigned long long __v) { |
| return __nvvm_atom_sys_xor_gen_ll((long long *)__p, __v); |
| } |
| __DEVICE__ unsigned int __umul24(unsigned int __a, unsigned int __b) { |
| return __nv_umul24(__a, __b); |
| } |
| __DEVICE__ unsigned long long __umul64hi(unsigned long long __a, |
| unsigned long long __b) { |
| return __nv_umul64hi(__a, __b); |
| } |
| __DEVICE__ unsigned int __umulhi(unsigned int __a, unsigned int __b) { |
| return __nv_umulhi(__a, __b); |
| } |
| __DEVICE__ unsigned int __urhadd(unsigned int __a, unsigned int __b) { |
| return __nv_urhadd(__a, __b); |
| } |
| __DEVICE__ unsigned int __usad(unsigned int __a, unsigned int __b, |
| unsigned int __c) { |
| return __nv_usad(__a, __b, __c); |
| } |
| |
| #if CUDA_VERSION >= 9000 && CUDA_VERSION < 9020 |
| __DEVICE__ unsigned int __vabs2(unsigned int __a) { return __nv_vabs2(__a); } |
| __DEVICE__ unsigned int __vabs4(unsigned int __a) { return __nv_vabs4(__a); } |
| __DEVICE__ unsigned int __vabsdiffs2(unsigned int __a, unsigned int __b) { |
| return __nv_vabsdiffs2(__a, __b); |
| } |
| __DEVICE__ unsigned int __vabsdiffs4(unsigned int __a, unsigned int __b) { |
| return __nv_vabsdiffs4(__a, __b); |
| } |
| __DEVICE__ unsigned int __vabsdiffu2(unsigned int __a, unsigned int __b) { |
| return __nv_vabsdiffu2(__a, __b); |
| } |
| __DEVICE__ unsigned int __vabsdiffu4(unsigned int __a, unsigned int __b) { |
| return __nv_vabsdiffu4(__a, __b); |
| } |
| __DEVICE__ unsigned int __vabsss2(unsigned int __a) { |
| return __nv_vabsss2(__a); |
| } |
| __DEVICE__ unsigned int __vabsss4(unsigned int __a) { |
| return __nv_vabsss4(__a); |
| } |
| __DEVICE__ unsigned int __vadd2(unsigned int __a, unsigned int __b) { |
| return __nv_vadd2(__a, __b); |
| } |
| __DEVICE__ unsigned int __vadd4(unsigned int __a, unsigned int __b) { |
| return __nv_vadd4(__a, __b); |
| } |
| __DEVICE__ unsigned int __vaddss2(unsigned int __a, unsigned int __b) { |
| return __nv_vaddss2(__a, __b); |
| } |
| __DEVICE__ unsigned int __vaddss4(unsigned int __a, unsigned int __b) { |
| return __nv_vaddss4(__a, __b); |
| } |
| __DEVICE__ unsigned int __vaddus2(unsigned int __a, unsigned int __b) { |
| return __nv_vaddus2(__a, __b); |
| } |
| __DEVICE__ unsigned int __vaddus4(unsigned int __a, unsigned int __b) { |
| return __nv_vaddus4(__a, __b); |
| } |
| __DEVICE__ unsigned int __vavgs2(unsigned int __a, unsigned int __b) { |
| return __nv_vavgs2(__a, __b); |
| } |
| __DEVICE__ unsigned int __vavgs4(unsigned int __a, unsigned int __b) { |
| return __nv_vavgs4(__a, __b); |
| } |
| __DEVICE__ unsigned int __vavgu2(unsigned int __a, unsigned int __b) { |
| return __nv_vavgu2(__a, __b); |
| } |
| __DEVICE__ unsigned int __vavgu4(unsigned int __a, unsigned int __b) { |
| return __nv_vavgu4(__a, __b); |
| } |
| __DEVICE__ unsigned int __vcmpeq2(unsigned int __a, unsigned int __b) { |
| return __nv_vcmpeq2(__a, __b); |
| } |
| __DEVICE__ unsigned int __vcmpeq4(unsigned int __a, unsigned int __b) { |
| return __nv_vcmpeq4(__a, __b); |
| } |
| __DEVICE__ unsigned int __vcmpges2(unsigned int __a, unsigned int __b) { |
| return __nv_vcmpges2(__a, __b); |
| } |
| __DEVICE__ unsigned int __vcmpges4(unsigned int __a, unsigned int __b) { |
| return __nv_vcmpges4(__a, __b); |
| } |
| __DEVICE__ unsigned int __vcmpgeu2(unsigned int __a, unsigned int __b) { |
| return __nv_vcmpgeu2(__a, __b); |
| } |
| __DEVICE__ unsigned int __vcmpgeu4(unsigned int __a, unsigned int __b) { |
| return __nv_vcmpgeu4(__a, __b); |
| } |
| __DEVICE__ unsigned int __vcmpgts2(unsigned int __a, unsigned int __b) { |
| return __nv_vcmpgts2(__a, __b); |
| } |
| __DEVICE__ unsigned int __vcmpgts4(unsigned int __a, unsigned int __b) { |
| return __nv_vcmpgts4(__a, __b); |
| } |
| __DEVICE__ unsigned int __vcmpgtu2(unsigned int __a, unsigned int __b) { |
| return __nv_vcmpgtu2(__a, __b); |
| } |
| __DEVICE__ unsigned int __vcmpgtu4(unsigned int __a, unsigned int __b) { |
| return __nv_vcmpgtu4(__a, __b); |
| } |
| __DEVICE__ unsigned int __vcmples2(unsigned int __a, unsigned int __b) { |
| return __nv_vcmples2(__a, __b); |
| } |
| __DEVICE__ unsigned int __vcmples4(unsigned int __a, unsigned int __b) { |
| return __nv_vcmples4(__a, __b); |
| } |
| __DEVICE__ unsigned int __vcmpleu2(unsigned int __a, unsigned int __b) { |
| return __nv_vcmpleu2(__a, __b); |
| } |
| __DEVICE__ unsigned int __vcmpleu4(unsigned int __a, unsigned int __b) { |
| return __nv_vcmpleu4(__a, __b); |
| } |
| __DEVICE__ unsigned int __vcmplts2(unsigned int __a, unsigned int __b) { |
| return __nv_vcmplts2(__a, __b); |
| } |
| __DEVICE__ unsigned int __vcmplts4(unsigned int __a, unsigned int __b) { |
| return __nv_vcmplts4(__a, __b); |
| } |
| __DEVICE__ unsigned int __vcmpltu2(unsigned int __a, unsigned int __b) { |
| return __nv_vcmpltu2(__a, __b); |
| } |
| __DEVICE__ unsigned int __vcmpltu4(unsigned int __a, unsigned int __b) { |
| return __nv_vcmpltu4(__a, __b); |
| } |
| __DEVICE__ unsigned int __vcmpne2(unsigned int __a, unsigned int __b) { |
| return __nv_vcmpne2(__a, __b); |
| } |
| __DEVICE__ unsigned int __vcmpne4(unsigned int __a, unsigned int __b) { |
| return __nv_vcmpne4(__a, __b); |
| } |
| __DEVICE__ unsigned int __vhaddu2(unsigned int __a, unsigned int __b) { |
| return __nv_vhaddu2(__a, __b); |
| } |
| __DEVICE__ unsigned int __vhaddu4(unsigned int __a, unsigned int __b) { |
| return __nv_vhaddu4(__a, __b); |
| } |
| __DEVICE__ unsigned int __vmaxs2(unsigned int __a, unsigned int __b) { |
| return __nv_vmaxs2(__a, __b); |
| } |
| __DEVICE__ unsigned int __vmaxs4(unsigned int __a, unsigned int __b) { |
| return __nv_vmaxs4(__a, __b); |
| } |
| __DEVICE__ unsigned int __vmaxu2(unsigned int __a, unsigned int __b) { |
| return __nv_vmaxu2(__a, __b); |
| } |
| __DEVICE__ unsigned int __vmaxu4(unsigned int __a, unsigned int __b) { |
| return __nv_vmaxu4(__a, __b); |
| } |
| __DEVICE__ unsigned int __vmins2(unsigned int __a, unsigned int __b) { |
| return __nv_vmins2(__a, __b); |
| } |
| __DEVICE__ unsigned int __vmins4(unsigned int __a, unsigned int __b) { |
| return __nv_vmins4(__a, __b); |
| } |
| __DEVICE__ unsigned int __vminu2(unsigned int __a, unsigned int __b) { |
| return __nv_vminu2(__a, __b); |
| } |
| __DEVICE__ unsigned int __vminu4(unsigned int __a, unsigned int __b) { |
| return __nv_vminu4(__a, __b); |
| } |
| __DEVICE__ unsigned int __vneg2(unsigned int __a) { return __nv_vneg2(__a); } |
| __DEVICE__ unsigned int __vneg4(unsigned int __a) { return __nv_vneg4(__a); } |
| __DEVICE__ unsigned int __vnegss2(unsigned int __a) { |
| return __nv_vnegss2(__a); |
| } |
| __DEVICE__ unsigned int __vnegss4(unsigned int __a) { |
| return __nv_vnegss4(__a); |
| } |
| __DEVICE__ unsigned int __vsads2(unsigned int __a, unsigned int __b) { |
| return __nv_vsads2(__a, __b); |
| } |
| __DEVICE__ unsigned int __vsads4(unsigned int __a, unsigned int __b) { |
| return __nv_vsads4(__a, __b); |
| } |
| __DEVICE__ unsigned int __vsadu2(unsigned int __a, unsigned int __b) { |
| return __nv_vsadu2(__a, __b); |
| } |
| __DEVICE__ unsigned int __vsadu4(unsigned int __a, unsigned int __b) { |
| return __nv_vsadu4(__a, __b); |
| } |
| __DEVICE__ unsigned int __vseteq2(unsigned int __a, unsigned int __b) { |
| return __nv_vseteq2(__a, __b); |
| } |
| __DEVICE__ unsigned int __vseteq4(unsigned int __a, unsigned int __b) { |
| return __nv_vseteq4(__a, __b); |
| } |
| __DEVICE__ unsigned int __vsetges2(unsigned int __a, unsigned int __b) { |
| return __nv_vsetges2(__a, __b); |
| } |
| __DEVICE__ unsigned int __vsetges4(unsigned int __a, unsigned int __b) { |
| return __nv_vsetges4(__a, __b); |
| } |
| __DEVICE__ unsigned int __vsetgeu2(unsigned int __a, unsigned int __b) { |
| return __nv_vsetgeu2(__a, __b); |
| } |
| __DEVICE__ unsigned int __vsetgeu4(unsigned int __a, unsigned int __b) { |
| return __nv_vsetgeu4(__a, __b); |
| } |
| __DEVICE__ unsigned int __vsetgts2(unsigned int __a, unsigned int __b) { |
| return __nv_vsetgts2(__a, __b); |
| } |
| __DEVICE__ unsigned int __vsetgts4(unsigned int __a, unsigned int __b) { |
| return __nv_vsetgts4(__a, __b); |
| } |
| __DEVICE__ unsigned int __vsetgtu2(unsigned int __a, unsigned int __b) { |
| return __nv_vsetgtu2(__a, __b); |
| } |
| __DEVICE__ unsigned int __vsetgtu4(unsigned int __a, unsigned int __b) { |
| return __nv_vsetgtu4(__a, __b); |
| } |
| __DEVICE__ unsigned int __vsetles2(unsigned int __a, unsigned int __b) { |
| return __nv_vsetles2(__a, __b); |
| } |
| __DEVICE__ unsigned int __vsetles4(unsigned int __a, unsigned int __b) { |
| return __nv_vsetles4(__a, __b); |
| } |
| __DEVICE__ unsigned int __vsetleu2(unsigned int __a, unsigned int __b) { |
| return __nv_vsetleu2(__a, __b); |
| } |
| __DEVICE__ unsigned int __vsetleu4(unsigned int __a, unsigned int __b) { |
| return __nv_vsetleu4(__a, __b); |
| } |
| __DEVICE__ unsigned int __vsetlts2(unsigned int __a, unsigned int __b) { |
| return __nv_vsetlts2(__a, __b); |
| } |
| __DEVICE__ unsigned int __vsetlts4(unsigned int __a, unsigned int __b) { |
| return __nv_vsetlts4(__a, __b); |
| } |
| __DEVICE__ unsigned int __vsetltu2(unsigned int __a, unsigned int __b) { |
| return __nv_vsetltu2(__a, __b); |
| } |
| __DEVICE__ unsigned int __vsetltu4(unsigned int __a, unsigned int __b) { |
| return __nv_vsetltu4(__a, __b); |
| } |
| __DEVICE__ unsigned int __vsetne2(unsigned int __a, unsigned int __b) { |
| return __nv_vsetne2(__a, __b); |
| } |
| __DEVICE__ unsigned int __vsetne4(unsigned int __a, unsigned int __b) { |
| return __nv_vsetne4(__a, __b); |
| } |
| __DEVICE__ unsigned int __vsub2(unsigned int __a, unsigned int __b) { |
| return __nv_vsub2(__a, __b); |
| } |
| __DEVICE__ unsigned int __vsub4(unsigned int __a, unsigned int __b) { |
| return __nv_vsub4(__a, __b); |
| } |
| __DEVICE__ unsigned int __vsubss2(unsigned int __a, unsigned int __b) { |
| return __nv_vsubss2(__a, __b); |
| } |
| __DEVICE__ unsigned int __vsubss4(unsigned int __a, unsigned int __b) { |
| return __nv_vsubss4(__a, __b); |
| } |
| __DEVICE__ unsigned int __vsubus2(unsigned int __a, unsigned int __b) { |
| return __nv_vsubus2(__a, __b); |
| } |
| __DEVICE__ unsigned int __vsubus4(unsigned int __a, unsigned int __b) { |
| return __nv_vsubus4(__a, __b); |
| } |
| #else // CUDA_VERSION >= 9020 |
| // CUDA no longer provides inline assembly (or bitcode) implementation of these |
| // functions, so we have to reimplment them. The implementation is naive and is |
| // not optimized for performance. |
| |
| // Helper function to convert N-bit boolean subfields into all-0 or all-1. |
| // E.g. __bool2mask(0x01000100,8) -> 0xff00ff00 |
| // __bool2mask(0x00010000,16) -> 0xffff0000 |
| __DEVICE__ unsigned int __bool2mask(unsigned int __a, int shift) { |
| return (__a << shift) - __a; |
| } |
| __DEVICE__ unsigned int __vabs2(unsigned int __a) { |
| unsigned int r; |
| asm("vabsdiff2.s32.s32.s32 %0,%1,%2,%3;" |
| : "=r"(r) |
| : "r"(__a), "r"(0), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vabs4(unsigned int __a) { |
| unsigned int r; |
| asm("vabsdiff4.s32.s32.s32 %0,%1,%2,%3;" |
| : "=r"(r) |
| : "r"(__a), "r"(0), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vabsdiffs2(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vabsdiff2.s32.s32.s32 %0,%1,%2,%3;" |
| : "=r"(r) |
| : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| |
| __DEVICE__ unsigned int __vabsdiffs4(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vabsdiff4.s32.s32.s32 %0,%1,%2,%3;" |
| : "=r"(r) |
| : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vabsdiffu2(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vabsdiff2.u32.u32.u32 %0,%1,%2,%3;" |
| : "=r"(r) |
| : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vabsdiffu4(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vabsdiff4.u32.u32.u32 %0,%1,%2,%3;" |
| : "=r"(r) |
| : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vabsss2(unsigned int __a) { |
| unsigned int r; |
| asm("vabsdiff2.s32.s32.s32.sat %0,%1,%2,%3;" |
| : "=r"(r) |
| : "r"(__a), "r"(0), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vabsss4(unsigned int __a) { |
| unsigned int r; |
| asm("vabsdiff4.s32.s32.s32.sat %0,%1,%2,%3;" |
| : "=r"(r) |
| : "r"(__a), "r"(0), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vadd2(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vadd2.u32.u32.u32 %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vadd4(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vadd4.u32.u32.u32 %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vaddss2(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vadd2.s32.s32.s32.sat %0,%1,%2,%3;" |
| : "=r"(r) |
| : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vaddss4(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vadd4.s32.s32.s32.sat %0,%1,%2,%3;" |
| : "=r"(r) |
| : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vaddus2(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vadd2.u32.u32.u32.sat %0,%1,%2,%3;" |
| : "=r"(r) |
| : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vaddus4(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vadd4.u32.u32.u32.sat %0,%1,%2,%3;" |
| : "=r"(r) |
| : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vavgs2(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vavrg2.s32.s32.s32 %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vavgs4(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vavrg4.s32.s32.s32 %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vavgu2(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vavrg2.u32.u32.u32 %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vavgu4(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vavrg4.u32.u32.u32 %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vseteq2(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vset2.u32.u32.eq %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vcmpeq2(unsigned int __a, unsigned int __b) { |
| return __bool2mask(__vseteq2(__a, __b), 16); |
| } |
| __DEVICE__ unsigned int __vseteq4(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vset4.u32.u32.eq %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vcmpeq4(unsigned int __a, unsigned int __b) { |
| return __bool2mask(__vseteq4(__a, __b), 8); |
| } |
| __DEVICE__ unsigned int __vsetges2(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vset2.s32.s32.ge %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vcmpges2(unsigned int __a, unsigned int __b) { |
| return __bool2mask(__vsetges2(__a, __b), 16); |
| } |
| __DEVICE__ unsigned int __vsetges4(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vset4.s32.s32.ge %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vcmpges4(unsigned int __a, unsigned int __b) { |
| return __bool2mask(__vsetges4(__a, __b), 8); |
| } |
| __DEVICE__ unsigned int __vsetgeu2(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vset2.u32.u32.ge %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vcmpgeu2(unsigned int __a, unsigned int __b) { |
| return __bool2mask(__vsetgeu2(__a, __b), 16); |
| } |
| __DEVICE__ unsigned int __vsetgeu4(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vset4.u32.u32.ge %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vcmpgeu4(unsigned int __a, unsigned int __b) { |
| return __bool2mask(__vsetgeu4(__a, __b), 8); |
| } |
| __DEVICE__ unsigned int __vsetgts2(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vset2.s32.s32.gt %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vcmpgts2(unsigned int __a, unsigned int __b) { |
| return __bool2mask(__vsetgts2(__a, __b), 16); |
| } |
| __DEVICE__ unsigned int __vsetgts4(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vset4.s32.s32.gt %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vcmpgts4(unsigned int __a, unsigned int __b) { |
| return __bool2mask(__vsetgts4(__a, __b), 8); |
| } |
| __DEVICE__ unsigned int __vsetgtu2(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vset2.u32.u32.gt %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vcmpgtu2(unsigned int __a, unsigned int __b) { |
| return __bool2mask(__vsetgtu2(__a, __b), 16); |
| } |
| __DEVICE__ unsigned int __vsetgtu4(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vset4.u32.u32.gt %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vcmpgtu4(unsigned int __a, unsigned int __b) { |
| return __bool2mask(__vsetgtu4(__a, __b), 8); |
| } |
| __DEVICE__ unsigned int __vsetles2(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vset2.s32.s32.le %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vcmples2(unsigned int __a, unsigned int __b) { |
| return __bool2mask(__vsetles2(__a, __b), 16); |
| } |
| __DEVICE__ unsigned int __vsetles4(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vset4.s32.s32.le %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vcmples4(unsigned int __a, unsigned int __b) { |
| return __bool2mask(__vsetles4(__a, __b), 8); |
| } |
| __DEVICE__ unsigned int __vsetleu2(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vset2.u32.u32.le %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vcmpleu2(unsigned int __a, unsigned int __b) { |
| return __bool2mask(__vsetleu2(__a, __b), 16); |
| } |
| __DEVICE__ unsigned int __vsetleu4(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vset4.u32.u32.le %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vcmpleu4(unsigned int __a, unsigned int __b) { |
| return __bool2mask(__vsetleu4(__a, __b), 8); |
| } |
| __DEVICE__ unsigned int __vsetlts2(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vset2.s32.s32.lt %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vcmplts2(unsigned int __a, unsigned int __b) { |
| return __bool2mask(__vsetlts2(__a, __b), 16); |
| } |
| __DEVICE__ unsigned int __vsetlts4(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vset4.s32.s32.lt %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vcmplts4(unsigned int __a, unsigned int __b) { |
| return __bool2mask(__vsetlts4(__a, __b), 8); |
| } |
| __DEVICE__ unsigned int __vsetltu2(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vset2.u32.u32.lt %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vcmpltu2(unsigned int __a, unsigned int __b) { |
| return __bool2mask(__vsetltu2(__a, __b), 16); |
| } |
| __DEVICE__ unsigned int __vsetltu4(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vset4.u32.u32.lt %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vcmpltu4(unsigned int __a, unsigned int __b) { |
| return __bool2mask(__vsetltu4(__a, __b), 8); |
| } |
| __DEVICE__ unsigned int __vsetne2(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vset2.u32.u32.ne %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vcmpne2(unsigned int __a, unsigned int __b) { |
| return __bool2mask(__vsetne2(__a, __b), 16); |
| } |
| __DEVICE__ unsigned int __vsetne4(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vset4.u32.u32.ne %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vcmpne4(unsigned int __a, unsigned int __b) { |
| return __bool2mask(__vsetne4(__a, __b), 8); |
| } |
| |
| // Based on ITEM 23 in AIM-239: http://dspace.mit.edu/handle/1721.1/6086 |
| // (a & b) + (a | b) = a + b = (a ^ b) + 2 * (a & b) => |
| // (a + b) / 2 = ((a ^ b) >> 1) + (a & b) |
| // To operate on multiple sub-elements we need to make sure to mask out bits |
| // that crossed over into adjacent elements during the shift. |
| __DEVICE__ unsigned int __vhaddu2(unsigned int __a, unsigned int __b) { |
| return (((__a ^ __b) >> 1) & ~0x80008000u) + (__a & __b); |
| } |
| __DEVICE__ unsigned int __vhaddu4(unsigned int __a, unsigned int __b) { |
| return (((__a ^ __b) >> 1) & ~0x80808080u) + (__a & __b); |
| } |
| |
| __DEVICE__ unsigned int __vmaxs2(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| if ((__a & 0x8000) && (__b & 0x8000)) { |
| // Work around a bug in ptxas which produces invalid result if low element |
| // is negative. |
| unsigned mask = __vcmpgts2(__a, __b); |
| r = (__a & mask) | (__b & ~mask); |
| } else { |
| asm("vmax2.s32.s32.s32 %0,%1,%2,%3;" |
| : "=r"(r) |
| : "r"(__a), "r"(__b), "r"(0)); |
| } |
| return r; |
| } |
| __DEVICE__ unsigned int __vmaxs4(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vmax4.s32.s32.s32 %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vmaxu2(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vmax2.u32.u32.u32 %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vmaxu4(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vmax4.u32.u32.u32 %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vmins2(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vmin2.s32.s32.s32 %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vmins4(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vmin4.s32.s32.s32 %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vminu2(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vmin2.u32.u32.u32 %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vminu4(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vmin4.u32.u32.u32 %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vsads2(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vabsdiff2.s32.s32.s32.add %0,%1,%2,%3;" |
| : "=r"(r) |
| : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vsads4(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vabsdiff4.s32.s32.s32.add %0,%1,%2,%3;" |
| : "=r"(r) |
| : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vsadu2(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vabsdiff2.u32.u32.u32.add %0,%1,%2,%3;" |
| : "=r"(r) |
| : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vsadu4(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vabsdiff4.u32.u32.u32.add %0,%1,%2,%3;" |
| : "=r"(r) |
| : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| |
| __DEVICE__ unsigned int __vsub2(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vsub2.u32.u32.u32 %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vneg2(unsigned int __a) { return __vsub2(0, __a); } |
| |
| __DEVICE__ unsigned int __vsub4(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vsub4.u32.u32.u32 %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vneg4(unsigned int __a) { return __vsub4(0, __a); } |
| __DEVICE__ unsigned int __vsubss2(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vsub2.s32.s32.s32.sat %0,%1,%2,%3;" |
| : "=r"(r) |
| : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vnegss2(unsigned int __a) { |
| return __vsubss2(0, __a); |
| } |
| __DEVICE__ unsigned int __vsubss4(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vsub4.s32.s32.s32.sat %0,%1,%2,%3;" |
| : "=r"(r) |
| : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vnegss4(unsigned int __a) { |
| return __vsubss4(0, __a); |
| } |
| __DEVICE__ unsigned int __vsubus2(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vsub2.u32.u32.u32.sat %0,%1,%2,%3;" |
| : "=r"(r) |
| : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| __DEVICE__ unsigned int __vsubus4(unsigned int __a, unsigned int __b) { |
| unsigned int r; |
| asm("vsub4.u32.u32.u32.sat %0,%1,%2,%3;" |
| : "=r"(r) |
| : "r"(__a), "r"(__b), "r"(0)); |
| return r; |
| } |
| #endif // CUDA_VERSION >= 9020 |
| __DEVICE__ int abs(int __a) __NOEXCEPT { return __nv_abs(__a); } |
| __DEVICE__ double fabs(double __a) __NOEXCEPT { return __nv_fabs(__a); } |
| __DEVICE__ double acos(double __a) { return __nv_acos(__a); } |
| __DEVICE__ float acosf(float __a) { return __nv_acosf(__a); } |
| __DEVICE__ double acosh(double __a) { return __nv_acosh(__a); } |
| __DEVICE__ float acoshf(float __a) { return __nv_acoshf(__a); } |
| __DEVICE__ double asin(double __a) { return __nv_asin(__a); } |
| __DEVICE__ float asinf(float __a) { return __nv_asinf(__a); } |
| __DEVICE__ double asinh(double __a) { return __nv_asinh(__a); } |
| __DEVICE__ float asinhf(float __a) { return __nv_asinhf(__a); } |
| __DEVICE__ double atan(double __a) { return __nv_atan(__a); } |
| __DEVICE__ double atan2(double __a, double __b) { return __nv_atan2(__a, __b); } |
| __DEVICE__ float atan2f(float __a, float __b) { return __nv_atan2f(__a, __b); } |
| __DEVICE__ float atanf(float __a) { return __nv_atanf(__a); } |
| __DEVICE__ double atanh(double __a) { return __nv_atanh(__a); } |
| __DEVICE__ float atanhf(float __a) { return __nv_atanhf(__a); } |
| __DEVICE__ double cbrt(double __a) { return __nv_cbrt(__a); } |
| __DEVICE__ float cbrtf(float __a) { return __nv_cbrtf(__a); } |
| __DEVICE__ double ceil(double __a) { return __nv_ceil(__a); } |
| __DEVICE__ float ceilf(float __a) { return __nv_ceilf(__a); } |
| #ifndef _OPENMP |
| __DEVICE__ int clock() { return __nvvm_read_ptx_sreg_clock(); } |
| __DEVICE__ long long clock64() { return __nvvm_read_ptx_sreg_clock64(); } |
| #endif |
| __DEVICE__ double copysign(double __a, double __b) { |
| return __nv_copysign(__a, __b); |
| } |
| __DEVICE__ float copysignf(float __a, float __b) { |
| return __nv_copysignf(__a, __b); |
| } |
| __DEVICE__ double cos(double __a) { return __nv_cos(__a); } |
| __DEVICE__ float cosf(float __a) { |
| return __FAST_OR_SLOW(__nv_fast_cosf, __nv_cosf)(__a); |
| } |
| __DEVICE__ double cosh(double __a) { return __nv_cosh(__a); } |
| __DEVICE__ float coshf(float __a) { return __nv_coshf(__a); } |
| __DEVICE__ double cospi(double __a) { return __nv_cospi(__a); } |
| __DEVICE__ float cospif(float __a) { return __nv_cospif(__a); } |
| __DEVICE__ double cyl_bessel_i0(double __a) { return __nv_cyl_bessel_i0(__a); } |
| __DEVICE__ float cyl_bessel_i0f(float __a) { return __nv_cyl_bessel_i0f(__a); } |
| __DEVICE__ double cyl_bessel_i1(double __a) { return __nv_cyl_bessel_i1(__a); } |
| __DEVICE__ float cyl_bessel_i1f(float __a) { return __nv_cyl_bessel_i1f(__a); } |
| __DEVICE__ double erf(double __a) { return __nv_erf(__a); } |
| __DEVICE__ double erfc(double __a) { return __nv_erfc(__a); } |
| __DEVICE__ float erfcf(float __a) { return __nv_erfcf(__a); } |
| __DEVICE__ double erfcinv(double __a) { return __nv_erfcinv(__a); } |
| __DEVICE__ float erfcinvf(float __a) { return __nv_erfcinvf(__a); } |
| __DEVICE__ double erfcx(double __a) { return __nv_erfcx(__a); } |
| __DEVICE__ float erfcxf(float __a) { return __nv_erfcxf(__a); } |
| __DEVICE__ float erff(float __a) { return __nv_erff(__a); } |
| __DEVICE__ double erfinv(double __a) { return __nv_erfinv(__a); } |
| __DEVICE__ float erfinvf(float __a) { return __nv_erfinvf(__a); } |
| __DEVICE__ double exp(double __a) { return __nv_exp(__a); } |
| __DEVICE__ double exp10(double __a) { return __nv_exp10(__a); } |
| __DEVICE__ float exp10f(float __a) { return __nv_exp10f(__a); } |
| __DEVICE__ double exp2(double __a) { return __nv_exp2(__a); } |
| __DEVICE__ float exp2f(float __a) { return __nv_exp2f(__a); } |
| __DEVICE__ float expf(float __a) { return __nv_expf(__a); } |
| __DEVICE__ double expm1(double __a) { return __nv_expm1(__a); } |
| __DEVICE__ float expm1f(float __a) { return __nv_expm1f(__a); } |
| __DEVICE__ float fabsf(float __a) { return __nv_fabsf(__a); } |
| __DEVICE__ double fdim(double __a, double __b) { return __nv_fdim(__a, __b); } |
| __DEVICE__ float fdimf(float __a, float __b) { return __nv_fdimf(__a, __b); } |
| __DEVICE__ double fdivide(double __a, double __b) { return __a / __b; } |
| __DEVICE__ float fdividef(float __a, float __b) { |
| #if __FAST_MATH__ && !__CUDA_PREC_DIV |
| return __nv_fast_fdividef(__a, __b); |
| #else |
| return __a / __b; |
| #endif |
| } |
| __DEVICE__ double floor(double __f) { return __nv_floor(__f); } |
| __DEVICE__ float floorf(float __f) { return __nv_floorf(__f); } |
| __DEVICE__ double fma(double __a, double __b, double __c) { |
| return __nv_fma(__a, __b, __c); |
| } |
| __DEVICE__ float fmaf(float __a, float __b, float __c) { |
| return __nv_fmaf(__a, __b, __c); |
| } |
| __DEVICE__ double fmax(double __a, double __b) { return __nv_fmax(__a, __b); } |
| __DEVICE__ float fmaxf(float __a, float __b) { return __nv_fmaxf(__a, __b); } |
| __DEVICE__ double fmin(double __a, double __b) { return __nv_fmin(__a, __b); } |
| __DEVICE__ float fminf(float __a, float __b) { return __nv_fminf(__a, __b); } |
| __DEVICE__ double fmod(double __a, double __b) { return __nv_fmod(__a, __b); } |
| __DEVICE__ float fmodf(float __a, float __b) { return __nv_fmodf(__a, __b); } |
| __DEVICE__ double frexp(double __a, int *__b) { return __nv_frexp(__a, __b); } |
| __DEVICE__ float frexpf(float __a, int *__b) { return __nv_frexpf(__a, __b); } |
| __DEVICE__ double hypot(double __a, double __b) { return __nv_hypot(__a, __b); } |
| __DEVICE__ float hypotf(float __a, float __b) { return __nv_hypotf(__a, __b); } |
| __DEVICE__ int ilogb(double __a) { return __nv_ilogb(__a); } |
| __DEVICE__ int ilogbf(float __a) { return __nv_ilogbf(__a); } |
| __DEVICE__ double j0(double __a) { return __nv_j0(__a); } |
| __DEVICE__ float j0f(float __a) { return __nv_j0f(__a); } |
| __DEVICE__ double j1(double __a) { return __nv_j1(__a); } |
| __DEVICE__ float j1f(float __a) { return __nv_j1f(__a); } |
| __DEVICE__ double jn(int __n, double __a) { return __nv_jn(__n, __a); } |
| __DEVICE__ float jnf(int __n, float __a) { return __nv_jnf(__n, __a); } |
| #if defined(__LP64__) || defined(_WIN64) |
| __DEVICE__ long labs(long __a) __NOEXCEPT { return __nv_llabs(__a); }; |
| #else |
| __DEVICE__ long labs(long __a) __NOEXCEPT { return __nv_abs(__a); }; |
| #endif |
| __DEVICE__ double ldexp(double __a, int __b) { return __nv_ldexp(__a, __b); } |
| __DEVICE__ float ldexpf(float __a, int __b) { return __nv_ldexpf(__a, __b); } |
| __DEVICE__ double lgamma(double __a) { return __nv_lgamma(__a); } |
| __DEVICE__ float lgammaf(float __a) { return __nv_lgammaf(__a); } |
| __DEVICE__ long long llabs(long long __a) __NOEXCEPT { return __nv_llabs(__a); } |
| __DEVICE__ long long llmax(long long __a, long long __b) { |
| return __nv_llmax(__a, __b); |
| } |
| __DEVICE__ long long llmin(long long __a, long long __b) { |
| return __nv_llmin(__a, __b); |
| } |
| __DEVICE__ long long llrint(double __a) { return __nv_llrint(__a); } |
| __DEVICE__ long long llrintf(float __a) { return __nv_llrintf(__a); } |
| __DEVICE__ long long llround(double __a) { return __nv_llround(__a); } |
| __DEVICE__ long long llroundf(float __a) { return __nv_llroundf(__a); } |
| __DEVICE__ double log(double __a) { return __nv_log(__a); } |
| __DEVICE__ double log10(double __a) { return __nv_log10(__a); } |
| __DEVICE__ float log10f(float __a) { return __nv_log10f(__a); } |
| __DEVICE__ double log1p(double __a) { return __nv_log1p(__a); } |
| __DEVICE__ float log1pf(float __a) { return __nv_log1pf(__a); } |
| __DEVICE__ double log2(double __a) { return __nv_log2(__a); } |
| __DEVICE__ float log2f(float __a) { |
| return __FAST_OR_SLOW(__nv_fast_log2f, __nv_log2f)(__a); |
| } |
| __DEVICE__ double logb(double __a) { return __nv_logb(__a); } |
| __DEVICE__ float logbf(float __a) { return __nv_logbf(__a); } |
| __DEVICE__ float logf(float __a) { |
| return __FAST_OR_SLOW(__nv_fast_logf, __nv_logf)(__a); |
| } |
| #if defined(__LP64__) || defined(_WIN64) |
| __DEVICE__ long lrint(double __a) { return llrint(__a); } |
| __DEVICE__ long lrintf(float __a) { return __float2ll_rn(__a); } |
| __DEVICE__ long lround(double __a) { return llround(__a); } |
| __DEVICE__ long lroundf(float __a) { return llroundf(__a); } |
| #else |
| __DEVICE__ long lrint(double __a) { return (long)rint(__a); } |
| __DEVICE__ long lrintf(float __a) { return __float2int_rn(__a); } |
| __DEVICE__ long lround(double __a) { return round(__a); } |
| __DEVICE__ long lroundf(float __a) { return roundf(__a); } |
| #endif |
| __DEVICE__ int max(int __a, int __b) { return __nv_max(__a, __b); } |
| // These functions shouldn't be declared when including this header |
| // for math function resolution purposes. |
| #ifndef _OPENMP |
| __DEVICE__ void *memcpy(void *__a, const void *__b, size_t __c) { |
| return __builtin_memcpy(__a, __b, __c); |
| } |
| __DEVICE__ void *memset(void *__a, int __b, size_t __c) { |
| return __builtin_memset(__a, __b, __c); |
| } |
| #endif |
| __DEVICE__ int min(int __a, int __b) { return __nv_min(__a, __b); } |
| __DEVICE__ double modf(double __a, double *__b) { return __nv_modf(__a, __b); } |
| __DEVICE__ float modff(float __a, float *__b) { return __nv_modff(__a, __b); } |
| __DEVICE__ double nearbyint(double __a) { return __nv_nearbyint(__a); } |
| __DEVICE__ float nearbyintf(float __a) { return __nv_nearbyintf(__a); } |
| __DEVICE__ double nextafter(double __a, double __b) { |
| return __nv_nextafter(__a, __b); |
| } |
| __DEVICE__ float nextafterf(float __a, float __b) { |
| return __nv_nextafterf(__a, __b); |
| } |
| __DEVICE__ double norm(int __dim, const double *__t) { |
| return __nv_norm(__dim, __t); |
| } |
| __DEVICE__ double norm3d(double __a, double __b, double __c) { |
| return __nv_norm3d(__a, __b, __c); |
| } |
| __DEVICE__ float norm3df(float __a, float __b, float __c) { |
| return __nv_norm3df(__a, __b, __c); |
| } |
| __DEVICE__ double norm4d(double __a, double __b, double __c, double __d) { |
| return __nv_norm4d(__a, __b, __c, __d); |
| } |
| __DEVICE__ float norm4df(float __a, float __b, float __c, float __d) { |
| return __nv_norm4df(__a, __b, __c, __d); |
| } |
| __DEVICE__ double normcdf(double __a) { return __nv_normcdf(__a); } |
| __DEVICE__ float normcdff(float __a) { return __nv_normcdff(__a); } |
| __DEVICE__ double normcdfinv(double __a) { return __nv_normcdfinv(__a); } |
| __DEVICE__ float normcdfinvf(float __a) { return __nv_normcdfinvf(__a); } |
| __DEVICE__ float normf(int __dim, const float *__t) { |
| return __nv_normf(__dim, __t); |
| } |
| __DEVICE__ double pow(double __a, double __b) { return __nv_pow(__a, __b); } |
| __DEVICE__ float powf(float __a, float __b) { return __nv_powf(__a, __b); } |
| __DEVICE__ double powi(double __a, int __b) { return __nv_powi(__a, __b); } |
| __DEVICE__ float powif(float __a, int __b) { return __nv_powif(__a, __b); } |
| __DEVICE__ double rcbrt(double __a) { return __nv_rcbrt(__a); } |
| __DEVICE__ float rcbrtf(float __a) { return __nv_rcbrtf(__a); } |
| __DEVICE__ double remainder(double __a, double __b) { |
| return __nv_remainder(__a, __b); |
| } |
| __DEVICE__ float remainderf(float __a, float __b) { |
| return __nv_remainderf(__a, __b); |
| } |
| __DEVICE__ double remquo(double __a, double __b, int *__c) { |
| return __nv_remquo(__a, __b, __c); |
| } |
| __DEVICE__ float remquof(float __a, float __b, int *__c) { |
| return __nv_remquof(__a, __b, __c); |
| } |
| __DEVICE__ double rhypot(double __a, double __b) { |
| return __nv_rhypot(__a, __b); |
| } |
| __DEVICE__ float rhypotf(float __a, float __b) { |
| return __nv_rhypotf(__a, __b); |
| } |
| __DEVICE__ double rint(double __a) { return __nv_rint(__a); } |
| __DEVICE__ float rintf(float __a) { return __nv_rintf(__a); } |
| __DEVICE__ double rnorm(int __a, const double *__b) { |
| return __nv_rnorm(__a, __b); |
| } |
| __DEVICE__ double rnorm3d(double __a, double __b, double __c) { |
| return __nv_rnorm3d(__a, __b, __c); |
| } |
| __DEVICE__ float rnorm3df(float __a, float __b, float __c) { |
| return __nv_rnorm3df(__a, __b, __c); |
| } |
| __DEVICE__ double rnorm4d(double __a, double __b, double __c, double __d) { |
| return __nv_rnorm4d(__a, __b, __c, __d); |
| } |
| __DEVICE__ float rnorm4df(float __a, float __b, float __c, float __d) { |
| return __nv_rnorm4df(__a, __b, __c, __d); |
| } |
| __DEVICE__ float rnormf(int __dim, const float *__t) { |
| return __nv_rnormf(__dim, __t); |
| } |
| __DEVICE__ double round(double __a) { return __nv_round(__a); } |
| __DEVICE__ float roundf(float __a) { return __nv_roundf(__a); } |
| __DEVICE__ double rsqrt(double __a) { return __nv_rsqrt(__a); } |
| __DEVICE__ float rsqrtf(float __a) { return __nv_rsqrtf(__a); } |
| __DEVICE__ double scalbn(double __a, int __b) { return __nv_scalbn(__a, __b); } |
| __DEVICE__ float scalbnf(float __a, int __b) { return __nv_scalbnf(__a, __b); } |
| // TODO: remove once variant is supported |
| #ifndef _OPENMP |
| __DEVICE__ double scalbln(double __a, long __b) { |
| if (__b > INT_MAX) |
| return __a > 0 ? HUGE_VAL : -HUGE_VAL; |
| if (__b < INT_MIN) |
| return __a > 0 ? 0.0 : -0.0; |
| return scalbn(__a, (int)__b); |
| } |
| __DEVICE__ float scalblnf(float __a, long __b) { |
| if (__b > INT_MAX) |
| return __a > 0 ? HUGE_VALF : -HUGE_VALF; |
| if (__b < INT_MIN) |
| return __a > 0 ? 0.f : -0.f; |
| return scalbnf(__a, (int)__b); |
| } |
| #endif |
| __DEVICE__ double sin(double __a) { return __nv_sin(__a); } |
| __DEVICE__ void sincos(double __a, double *__s, double *__c) { |
| return __nv_sincos(__a, __s, __c); |
| } |
| __DEVICE__ void sincosf(float __a, float *__s, float *__c) { |
| return __FAST_OR_SLOW(__nv_fast_sincosf, __nv_sincosf)(__a, __s, __c); |
| } |
| __DEVICE__ void sincospi(double __a, double *__s, double *__c) { |
| return __nv_sincospi(__a, __s, __c); |
| } |
| __DEVICE__ void sincospif(float __a, float *__s, float *__c) { |
| return __nv_sincospif(__a, __s, __c); |
| } |
| __DEVICE__ float sinf(float __a) { |
| return __FAST_OR_SLOW(__nv_fast_sinf, __nv_sinf)(__a); |
| } |
| __DEVICE__ double sinh(double __a) { return __nv_sinh(__a); } |
| __DEVICE__ float sinhf(float __a) { return __nv_sinhf(__a); } |
| __DEVICE__ double sinpi(double __a) { return __nv_sinpi(__a); } |
| __DEVICE__ float sinpif(float __a) { return __nv_sinpif(__a); } |
| __DEVICE__ double sqrt(double __a) { return __nv_sqrt(__a); } |
| __DEVICE__ float sqrtf(float __a) { return __nv_sqrtf(__a); } |
| __DEVICE__ double tan(double __a) { return __nv_tan(__a); } |
| __DEVICE__ float tanf(float __a) { return __nv_tanf(__a); } |
| __DEVICE__ double tanh(double __a) { return __nv_tanh(__a); } |
| __DEVICE__ float tanhf(float __a) { return __nv_tanhf(__a); } |
| __DEVICE__ double tgamma(double __a) { return __nv_tgamma(__a); } |
| __DEVICE__ float tgammaf(float __a) { return __nv_tgammaf(__a); } |
| __DEVICE__ double trunc(double __a) { return __nv_trunc(__a); } |
| __DEVICE__ float truncf(float __a) { return __nv_truncf(__a); } |
| __DEVICE__ unsigned long long ullmax(unsigned long long __a, |
| unsigned long long __b) { |
| return __nv_ullmax(__a, __b); |
| } |
| __DEVICE__ unsigned long long ullmin(unsigned long long __a, |
| unsigned long long __b) { |
| return __nv_ullmin(__a, __b); |
| } |
| __DEVICE__ unsigned int umax(unsigned int __a, unsigned int __b) { |
| return __nv_umax(__a, __b); |
| } |
| __DEVICE__ unsigned int umin(unsigned int __a, unsigned int __b) { |
| return __nv_umin(__a, __b); |
| } |
| __DEVICE__ double y0(double __a) { return __nv_y0(__a); } |
| __DEVICE__ float y0f(float __a) { return __nv_y0f(__a); } |
| __DEVICE__ double y1(double __a) { return __nv_y1(__a); } |
| __DEVICE__ float y1f(float __a) { return __nv_y1f(__a); } |
| __DEVICE__ double yn(int __a, double __b) { return __nv_yn(__a, __b); } |
| __DEVICE__ float ynf(int __a, float __b) { return __nv_ynf(__a, __b); } |
| |
| #undef __NOEXCEPT |
| #pragma pop_macro("__DEVICE__") |
| #pragma pop_macro("__FAST_OR_SLOW") |
| #endif // __CLANG_CUDA_DEVICE_FUNCTIONS_H__ |