Mercurial > games > semicongine
diff fuhtark_test/include/fvec.h @ 1500:91c8c3b7cbf0
add: futhark tests for generating vulkan api
| author | sam <sam@basx.dev> |
|---|---|
| date | Wed, 26 Nov 2025 21:36:48 +0700 |
| parents | |
| children |
line wrap: on
line diff
--- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/fuhtark_test/include/fvec.h Wed Nov 26 21:36:48 2025 +0700 @@ -0,0 +1,249 @@ +/** + * This file has no copyright assigned and is placed in the Public Domain. + * This file is part of the w64 mingw-runtime package. + * No warranty is given; refer to the file DISCLAIMER within this package. + */ +#ifndef _FVEC_H_INCLUDED +#define _FVEC_H_INCLUDED + +#ifndef RC_INVOKED +#ifndef __cplusplus +#error ERROR: This file is only supported in C++ compilations! +#endif + +#include <intrin.h> +#include <assert.h> +#include <ivec.h> +#include <_mingw.h> + +#if defined(_ENABLE_VEC_DEBUG) +#include <iostream> +#endif + +#pragma pack(push,_CRT_PACKING) + +#ifdef __SSE__ + +#pragma pack(push,16) + +#define EXPLICIT explicit + +class F32vec4 { +protected: + __m128 vec; +public: + F32vec4() {} + F32vec4(__m128 m) { vec = m;} + F32vec4(float f3,float f2,float f1,float f0) { vec= _mm_set_ps(f3,f2,f1,f0); } + EXPLICIT F32vec4(float f) { vec = _mm_set_ps1(f); } + EXPLICIT F32vec4(double d) { vec = _mm_set_ps1((float) d); } + F32vec4& operator =(float f) { vec = _mm_set_ps1(f); return *this; } + F32vec4& operator =(double d) { vec = _mm_set_ps1((float) d); return *this; } + operator __m128() const { return vec; } + friend F32vec4 operator &(const F32vec4 &a,const F32vec4 &b) { return _mm_and_ps(a,b); } + friend F32vec4 operator |(const F32vec4 &a,const F32vec4 &b) { return _mm_or_ps(a,b); } + friend F32vec4 operator ^(const F32vec4 &a,const F32vec4 &b) { return _mm_xor_ps(a,b); } + friend F32vec4 operator +(const F32vec4 &a,const F32vec4 &b) { return _mm_add_ps(a,b); } + friend F32vec4 operator -(const F32vec4 &a,const F32vec4 &b) { return _mm_sub_ps(a,b); } + friend F32vec4 operator *(const F32vec4 &a,const F32vec4 &b) { return _mm_mul_ps(a,b); } + friend F32vec4 operator /(const F32vec4 &a,const F32vec4 &b) { return _mm_div_ps(a,b); } + F32vec4& operator =(const F32vec4 &a) { vec = a.vec; return *this; } + F32vec4& operator =(const __m128 &avec) { vec = avec; return *this; } + F32vec4& operator +=(F32vec4 &a) { return *this = _mm_add_ps(vec,a); } + F32vec4& operator -=(F32vec4 &a) { return *this = _mm_sub_ps(vec,a); } + F32vec4& operator *=(F32vec4 &a) { return *this = _mm_mul_ps(vec,a); } + F32vec4& operator /=(F32vec4 &a) { return *this = _mm_div_ps(vec,a); } + F32vec4& operator &=(F32vec4 &a) { return *this = _mm_and_ps(vec,a); } + F32vec4& operator |=(F32vec4 &a) { return *this = _mm_or_ps(vec,a); } + F32vec4& operator ^=(F32vec4 &a) { return *this = _mm_xor_ps(vec,a); } + friend float add_horizontal(F32vec4 &a) { + F32vec4 ftemp = _mm_add_ss(a,_mm_add_ss(_mm_shuffle_ps(a,a,1),_mm_add_ss(_mm_shuffle_ps(a,a,2),_mm_shuffle_ps(a,a,3)))); + return ftemp[0]; + } + friend F32vec4 sqrt(const F32vec4 &a) { return _mm_sqrt_ps(a); } + friend F32vec4 rcp(const F32vec4 &a) { return _mm_rcp_ps(a); } + friend F32vec4 rsqrt(const F32vec4 &a) { return _mm_rsqrt_ps(a); } + friend F32vec4 rcp_nr(const F32vec4 &a) { + F32vec4 Ra0 = _mm_rcp_ps(a); + return _mm_sub_ps(_mm_add_ps(Ra0,Ra0),_mm_mul_ps(_mm_mul_ps(Ra0,a),Ra0)); + } + friend F32vec4 rsqrt_nr(const F32vec4 &a) { + static const F32vec4 fvecf0pt5(0.5f); + static const F32vec4 fvecf3pt0(3.0f); + F32vec4 Ra0 = _mm_rsqrt_ps(a); + return (fvecf0pt5 *Ra0) *(fvecf3pt0 - (a *Ra0) *Ra0); + + } +#define Fvec32s4_COMP(op) friend F32vec4 cmp##op (const F32vec4 &a,const F32vec4 &b) { return _mm_cmp##op##_ps(a,b); } + Fvec32s4_COMP(eq) + Fvec32s4_COMP(lt) + Fvec32s4_COMP(le) + Fvec32s4_COMP(gt) + Fvec32s4_COMP(ge) + Fvec32s4_COMP(neq) + Fvec32s4_COMP(nlt) + Fvec32s4_COMP(nle) + Fvec32s4_COMP(ngt) + Fvec32s4_COMP(nge) +#undef Fvec32s4_COMP + + friend F32vec4 simd_min(const F32vec4 &a,const F32vec4 &b) { return _mm_min_ps(a,b); } + friend F32vec4 simd_max(const F32vec4 &a,const F32vec4 &b) { return _mm_max_ps(a,b); } + +#if defined(_ENABLE_VEC_DEBUG) + friend std::ostream & operator<<(std::ostream & os,const F32vec4 &a) { + float *fp = (float*)&a; + os << "[3]:" << *(fp+3) + << " [2]:" << *(fp+2) + << " [1]:" << *(fp+1) + << " [0]:" << *fp; + return os; + } +#endif + const float& operator[](int i) const { + assert((0 <= i) && (i <= 3)); + float *fp = (float*)&vec; + return *(fp+i); + } + float& operator[](int i) { + assert((0 <= i) && (i <= 3)); + float *fp = (float*)&vec; + return *(fp+i); + } +}; + +inline F32vec4 unpack_low(const F32vec4 &a,const F32vec4 &b) { return _mm_unpacklo_ps(a,b); } +inline F32vec4 unpack_high(const F32vec4 &a,const F32vec4 &b) { return _mm_unpackhi_ps(a,b); } +inline int move_mask(const F32vec4 &a) { return _mm_movemask_ps(a); } +inline void loadu(F32vec4 &a,float *p) { a = _mm_loadu_ps(p); } +inline void storeu(float *p,const F32vec4 &a) { _mm_storeu_ps(p,a); } +inline void store_nta(float *p,F32vec4 &a) { _mm_stream_ps(p,a); } + +#define Fvec32s4_SELECT(op) inline F32vec4 select_##op (const F32vec4 &a,const F32vec4 &b,const F32vec4 &c,const F32vec4 &d) { F32vec4 mask = _mm_cmp##op##_ps(a,b); return((mask & c) | F32vec4((_mm_andnot_ps(mask,d)))); } +Fvec32s4_SELECT(eq) +Fvec32s4_SELECT(lt) +Fvec32s4_SELECT(le) +Fvec32s4_SELECT(gt) +Fvec32s4_SELECT(ge) +Fvec32s4_SELECT(neq) +Fvec32s4_SELECT(nlt) +Fvec32s4_SELECT(nle) +Fvec32s4_SELECT(ngt) +Fvec32s4_SELECT(nge) +#undef Fvec32s4_SELECT + +inline Is16vec4 simd_max(const Is16vec4 &a,const Is16vec4 &b) { return _m_pmaxsw(a,b); } +inline Is16vec4 simd_min(const Is16vec4 &a,const Is16vec4 &b) { return _m_pminsw(a,b); } +inline Iu8vec8 simd_max(const Iu8vec8 &a,const Iu8vec8 &b) { return _m_pmaxub(a,b); } +inline Iu8vec8 simd_min(const Iu8vec8 &a,const Iu8vec8 &b) { return _m_pminub(a,b); } +inline Iu16vec4 simd_avg(const Iu16vec4 &a,const Iu16vec4 &b) { return _m_pavgw(a,b); } +inline Iu8vec8 simd_avg(const Iu8vec8 &a,const Iu8vec8 &b) { return _m_pavgb(a,b); } +inline int move_mask(const I8vec8 &a) { return _m_pmovmskb(a); } +inline Iu16vec4 mul_high(const Iu16vec4 &a,const Iu16vec4 &b) { return _m_pmulhuw(a,b); } +inline void mask_move(const I8vec8 &a,const I8vec8 &b,char *addr) { _m_maskmovq(a,b,addr); } +inline void store_nta(__m64 *p,M64 &a) { _mm_stream_pi(p,a); } +inline int F32vec4ToInt(const F32vec4 &a) { return _mm_cvtt_ss2si(a); } +inline Is32vec2 F32vec4ToIs32vec2 (const F32vec4 &a) { + __m64 result; + result = _mm_cvtt_ps2pi(a); + return Is32vec2(result); +} + +inline F32vec4 IntToF32vec4(const F32vec4 &a,int i) { + __m128 result; + result = _mm_cvt_si2ss(a,i); + return F32vec4(result); +} + +inline F32vec4 Is32vec2ToF32vec4(const F32vec4 &a,const Is32vec2 &b) { + __m128 result; + result = _mm_cvt_pi2ps(a,b); + return F32vec4(result); +} + +class F32vec1 { +protected: + __m128 vec; +public: + F32vec1() {} + F32vec1(int i) { vec = _mm_cvt_si2ss(vec,i);}; + EXPLICIT F32vec1(float f) { vec = _mm_set_ss(f); } + EXPLICIT F32vec1(double d) { vec = _mm_set_ss((float) d); } + F32vec1(__m128 m) { vec = m; } + operator __m128() const { return vec; } + friend F32vec1 operator &(const F32vec1 &a,const F32vec1 &b) { return _mm_and_ps(a,b); } + friend F32vec1 operator |(const F32vec1 &a,const F32vec1 &b) { return _mm_or_ps(a,b); } + friend F32vec1 operator ^(const F32vec1 &a,const F32vec1 &b) { return _mm_xor_ps(a,b); } + friend F32vec1 operator +(const F32vec1 &a,const F32vec1 &b) { return _mm_add_ss(a,b); } + friend F32vec1 operator -(const F32vec1 &a,const F32vec1 &b) { return _mm_sub_ss(a,b); } + friend F32vec1 operator *(const F32vec1 &a,const F32vec1 &b) { return _mm_mul_ss(a,b); } + friend F32vec1 operator /(const F32vec1 &a,const F32vec1 &b) { return _mm_div_ss(a,b); } + F32vec1& operator +=(F32vec1 &a) { return *this = _mm_add_ss(vec,a); } + F32vec1& operator -=(F32vec1 &a) { return *this = _mm_sub_ss(vec,a); } + F32vec1& operator *=(F32vec1 &a) { return *this = _mm_mul_ss(vec,a); } + F32vec1& operator /=(F32vec1 &a) { return *this = _mm_div_ss(vec,a); } + F32vec1& operator &=(F32vec1 &a) { return *this = _mm_and_ps(vec,a); } + F32vec1& operator |=(F32vec1 &a) { return *this = _mm_or_ps(vec,a); } + F32vec1& operator ^=(F32vec1 &a) { return *this = _mm_xor_ps(vec,a); } + friend F32vec1 sqrt(const F32vec1 &a) { return _mm_sqrt_ss(a); } + friend F32vec1 rcp(const F32vec1 &a) { return _mm_rcp_ss(a); } + friend F32vec1 rsqrt(const F32vec1 &a) { return _mm_rsqrt_ss(a); } + friend F32vec1 rcp_nr(const F32vec1 &a) { + F32vec1 Ra0 = _mm_rcp_ss(a); + return _mm_sub_ss(_mm_add_ss(Ra0,Ra0),_mm_mul_ss(_mm_mul_ss(Ra0,a),Ra0)); + } + friend F32vec1 rsqrt_nr(const F32vec1 &a) { + static const F32vec1 fvecf0pt5(0.5f); + static const F32vec1 fvecf3pt0(3.0f); + F32vec1 Ra0 = _mm_rsqrt_ss(a); + return (fvecf0pt5 *Ra0) *(fvecf3pt0 - (a *Ra0) *Ra0); + } +#define Fvec32s1_COMP(op) friend F32vec1 cmp##op (const F32vec1 &a,const F32vec1 &b) { return _mm_cmp##op##_ss(a,b); } + Fvec32s1_COMP(eq) + Fvec32s1_COMP(lt) + Fvec32s1_COMP(le) + Fvec32s1_COMP(gt) + Fvec32s1_COMP(ge) + Fvec32s1_COMP(neq) + Fvec32s1_COMP(nlt) + Fvec32s1_COMP(nle) + Fvec32s1_COMP(ngt) + Fvec32s1_COMP(nge) +#undef Fvec32s1_COMP + + friend F32vec1 simd_min(const F32vec1 &a,const F32vec1 &b) { return _mm_min_ss(a,b); } + friend F32vec1 simd_max(const F32vec1 &a,const F32vec1 &b) { return _mm_max_ss(a,b); } + +#if defined(_ENABLE_VEC_DEBUG) + friend std::ostream & operator<<(std::ostream & os,const F32vec1 &a) { + float *fp = (float*)&a; + os << "float:" << *fp; + return os; + } +#endif +}; + +#define Fvec32s1_SELECT(op) inline F32vec1 select_##op (const F32vec1 &a,const F32vec1 &b,const F32vec1 &c,const F32vec1 &d) { F32vec1 mask = _mm_cmp##op##_ss(a,b); return((mask & c) | F32vec1((_mm_andnot_ps(mask,d)))); } +Fvec32s1_SELECT(eq) +Fvec32s1_SELECT(lt) +Fvec32s1_SELECT(le) +Fvec32s1_SELECT(gt) +Fvec32s1_SELECT(ge) +Fvec32s1_SELECT(neq) +Fvec32s1_SELECT(nlt) +Fvec32s1_SELECT(nle) +Fvec32s1_SELECT(ngt) +Fvec32s1_SELECT(nge) +#undef Fvec32s1_SELECT + +inline int F32vec1ToInt(const F32vec1 &a) +{ + return _mm_cvtt_ss2si(a); +} + +#pragma pack(pop) + +#endif /* #ifdef __SSE__ */ +#pragma pack(pop) +#endif +#endif
