/*===---- riscv_vector.h - RISC-V V-extension RVVIntrinsics -------------------=== * * * Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. * See https://llvm.org/LICENSE.txt for license information. * SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception * *===-----------------------------------------------------------------------=== */ #ifndef __RISCV_VECTOR_H #define __RISCV_VECTOR_H #include #include #ifndef __riscv_vector #error "Vector intrinsics require the vector extension." #endif #ifdef __cplusplus extern "C" { #endif #pragma clang riscv intrinsic vector #define vlenb() __builtin_rvv_vlenb() enum RVV_CSR { RVV_VSTART = 0, RVV_VXSAT, RVV_VXRM, RVV_VCSR, }; static __inline__ __attribute__((__always_inline__, __nodebug__)) unsigned long vread_csr(enum RVV_CSR __csr) { unsigned long __rv = 0; switch (__csr) { case RVV_VSTART: __asm__ __volatile__ ("csrr\t%0, vstart" : "=r"(__rv) : : "memory"); break; case RVV_VXSAT: __asm__ __volatile__ ("csrr\t%0, vxsat" : "=r"(__rv) : : "memory"); break; case RVV_VXRM: __asm__ __volatile__ ("csrr\t%0, vxrm" : "=r"(__rv) : : "memory"); break; case RVV_VCSR: __asm__ __volatile__ ("csrr\t%0, vcsr" : "=r"(__rv) : : "memory"); break; } return __rv; } static __inline__ __attribute__((__always_inline__, __nodebug__)) void vwrite_csr(enum RVV_CSR __csr, unsigned long __value) { switch (__csr) { case RVV_VSTART: __asm__ __volatile__ ("csrw\tvstart, %z0" : : "rJ"(__value) : "memory"); break; case RVV_VXSAT: __asm__ __volatile__ ("csrw\tvxsat, %z0" : : "rJ"(__value) : "memory"); break; case RVV_VXRM: __asm__ __volatile__ ("csrw\tvxrm, %z0" : : "rJ"(__value) : "memory"); break; case RVV_VCSR: __asm__ __volatile__ ("csrw\tvcsr, %z0" : : "rJ"(__value) : "memory"); break; } } #define vsetvl_e8mf4(avl) __builtin_rvv_vsetvli((size_t)(avl), 0, 6) #define vsetvl_e8mf2(avl) __builtin_rvv_vsetvli((size_t)(avl), 0, 7) #define vsetvl_e8m1(avl) __builtin_rvv_vsetvli((size_t)(avl), 0, 0) #define vsetvl_e8m2(avl) __builtin_rvv_vsetvli((size_t)(avl), 0, 1) #define vsetvl_e8m4(avl) __builtin_rvv_vsetvli((size_t)(avl), 0, 2) #define vsetvl_e8m8(avl) __builtin_rvv_vsetvli((size_t)(avl), 0, 3) #define vsetvl_e16mf2(avl) __builtin_rvv_vsetvli((size_t)(avl), 1, 7) #define vsetvl_e16m1(avl) __builtin_rvv_vsetvli((size_t)(avl), 1, 0) #define vsetvl_e16m2(avl) __builtin_rvv_vsetvli((size_t)(avl), 1, 1) #define vsetvl_e16m4(avl) __builtin_rvv_vsetvli((size_t)(avl), 1, 2) #define vsetvl_e16m8(avl) __builtin_rvv_vsetvli((size_t)(avl), 1, 3) #define vsetvl_e32m1(avl) __builtin_rvv_vsetvli((size_t)(avl), 2, 0) #define vsetvl_e32m2(avl) __builtin_rvv_vsetvli((size_t)(avl), 2, 1) #define vsetvl_e32m4(avl) __builtin_rvv_vsetvli((size_t)(avl), 2, 2) #define vsetvl_e32m8(avl) __builtin_rvv_vsetvli((size_t)(avl), 2, 3) #if __riscv_v_elen >= 64 #define vsetvl_e8mf8(avl) __builtin_rvv_vsetvli((size_t)(avl), 0, 5) #define vsetvl_e16mf4(avl) __builtin_rvv_vsetvli((size_t)(avl), 1, 6) #define vsetvl_e32mf2(avl) __builtin_rvv_vsetvli((size_t)(avl), 2, 7) #define vsetvl_e64m1(avl) __builtin_rvv_vsetvli((size_t)(avl), 3, 0) #define vsetvl_e64m2(avl) __builtin_rvv_vsetvli((size_t)(avl), 3, 1) #define vsetvl_e64m4(avl) __builtin_rvv_vsetvli((size_t)(avl), 3, 2) #define vsetvl_e64m8(avl) __builtin_rvv_vsetvli((size_t)(avl), 3, 3) #endif #define vsetvlmax_e8mf4() __builtin_rvv_vsetvlimax(0, 6) #define vsetvlmax_e8mf2() __builtin_rvv_vsetvlimax(0, 7) #define vsetvlmax_e8m1() __builtin_rvv_vsetvlimax(0, 0) #define vsetvlmax_e8m2() __builtin_rvv_vsetvlimax(0, 1) #define vsetvlmax_e8m4() __builtin_rvv_vsetvlimax(0, 2) #define vsetvlmax_e8m8() __builtin_rvv_vsetvlimax(0, 3) #define vsetvlmax_e16mf2() __builtin_rvv_vsetvlimax(1, 7) #define vsetvlmax_e16m1() __builtin_rvv_vsetvlimax(1, 0) #define vsetvlmax_e16m2() __builtin_rvv_vsetvlimax(1, 1) #define vsetvlmax_e16m4() __builtin_rvv_vsetvlimax(1, 2) #define vsetvlmax_e16m8() __builtin_rvv_vsetvlimax(1, 3) #define vsetvlmax_e32m1() __builtin_rvv_vsetvlimax(2, 0) #define vsetvlmax_e32m2() __builtin_rvv_vsetvlimax(2, 1) #define vsetvlmax_e32m4() __builtin_rvv_vsetvlimax(2, 2) #define vsetvlmax_e32m8() __builtin_rvv_vsetvlimax(2, 3) #if __riscv_v_elen >= 64 #define vsetvlmax_e8mf8() __builtin_rvv_vsetvlimax(0, 5) #define vsetvlmax_e16mf4() __builtin_rvv_vsetvlimax(1, 6) #define vsetvlmax_e32mf2() __builtin_rvv_vsetvlimax(2, 7) #define vsetvlmax_e64m1() __builtin_rvv_vsetvlimax(3, 0) #define vsetvlmax_e64m2() __builtin_rvv_vsetvlimax(3, 1) #define vsetvlmax_e64m4() __builtin_rvv_vsetvlimax(3, 2) #define vsetvlmax_e64m8() __builtin_rvv_vsetvlimax(3, 3) #endif typedef __rvv_bool64_t vbool64_t; typedef __rvv_bool32_t vbool32_t; typedef __rvv_bool16_t vbool16_t; typedef __rvv_bool8_t vbool8_t; typedef __rvv_bool4_t vbool4_t; typedef __rvv_bool2_t vbool2_t; typedef __rvv_bool1_t vbool1_t; typedef __rvv_int8mf8_t vint8mf8_t; typedef __rvv_uint8mf8_t vuint8mf8_t; typedef __rvv_int8mf4_t vint8mf4_t; typedef __rvv_uint8mf4_t vuint8mf4_t; typedef __rvv_int8mf2_t vint8mf2_t; typedef __rvv_uint8mf2_t vuint8mf2_t; typedef __rvv_int8m1_t vint8m1_t; typedef __rvv_uint8m1_t vuint8m1_t; typedef __rvv_int8m2_t vint8m2_t; typedef __rvv_uint8m2_t vuint8m2_t; typedef __rvv_int8m4_t vint8m4_t; typedef __rvv_uint8m4_t vuint8m4_t; typedef __rvv_int8m8_t vint8m8_t; typedef __rvv_uint8m8_t vuint8m8_t; typedef __rvv_int16mf4_t vint16mf4_t; typedef __rvv_uint16mf4_t vuint16mf4_t; typedef __rvv_int16mf2_t vint16mf2_t; typedef __rvv_uint16mf2_t vuint16mf2_t; typedef __rvv_int16m1_t vint16m1_t; typedef __rvv_uint16m1_t vuint16m1_t; typedef __rvv_int16m2_t vint16m2_t; typedef __rvv_uint16m2_t vuint16m2_t; typedef __rvv_int16m4_t vint16m4_t; typedef __rvv_uint16m4_t vuint16m4_t; typedef __rvv_int16m8_t vint16m8_t; typedef __rvv_uint16m8_t vuint16m8_t; typedef __rvv_int32mf2_t vint32mf2_t; typedef __rvv_uint32mf2_t vuint32mf2_t; typedef __rvv_int32m1_t vint32m1_t; typedef __rvv_uint32m1_t vuint32m1_t; typedef __rvv_int32m2_t vint32m2_t; typedef __rvv_uint32m2_t vuint32m2_t; typedef __rvv_int32m4_t vint32m4_t; typedef __rvv_uint32m4_t vuint32m4_t; typedef __rvv_int32m8_t vint32m8_t; typedef __rvv_uint32m8_t vuint32m8_t; typedef __rvv_int64m1_t vint64m1_t; typedef __rvv_uint64m1_t vuint64m1_t; typedef __rvv_int64m2_t vint64m2_t; typedef __rvv_uint64m2_t vuint64m2_t; typedef __rvv_int64m4_t vint64m4_t; typedef __rvv_uint64m4_t vuint64m4_t; typedef __rvv_int64m8_t vint64m8_t; typedef __rvv_uint64m8_t vuint64m8_t; #if defined(__riscv_zvfh) typedef __rvv_float16mf4_t vfloat16mf4_t; typedef __rvv_float16mf2_t vfloat16mf2_t; typedef __rvv_float16m1_t vfloat16m1_t; typedef __rvv_float16m2_t vfloat16m2_t; typedef __rvv_float16m4_t vfloat16m4_t; typedef __rvv_float16m8_t vfloat16m8_t; #endif #if (__riscv_v_elen_fp >= 32) typedef __rvv_float32mf2_t vfloat32mf2_t; typedef __rvv_float32m1_t vfloat32m1_t; typedef __rvv_float32m2_t vfloat32m2_t; typedef __rvv_float32m4_t vfloat32m4_t; typedef __rvv_float32m8_t vfloat32m8_t; #endif #if (__riscv_v_elen_fp >= 64) typedef __rvv_float64m1_t vfloat64m1_t; typedef __rvv_float64m2_t vfloat64m2_t; typedef __rvv_float64m4_t vfloat64m4_t; typedef __rvv_float64m8_t vfloat64m8_t; #endif #define __riscv_v_intrinsic_overloading 1 #ifdef __cplusplus } #endif // __cplusplus #endif // __RISCV_VECTOR_H