mirror of
https://github.com/CTCaer/RetroArch.git
synced 2024-12-14 06:18:34 +00:00
2f08e760ff
depends on glm for now and is C++-based, no way around that for now
147 lines
4.7 KiB
C++
147 lines
4.7 KiB
C++
///////////////////////////////////////////////////////////////////////////////////
|
|
/// OpenGL Mathematics (glm.g-truc.net)
|
|
///
|
|
/// Copyright (c) 2005 - 2014 G-Truc Creation (www.g-truc.net)
|
|
/// Permission is hereby granted, free of charge, to any person obtaining a copy
|
|
/// of this software and associated documentation files (the "Software"), to deal
|
|
/// in the Software without restriction, including without limitation the rights
|
|
/// to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
|
|
/// copies of the Software, and to permit persons to whom the Software is
|
|
/// furnished to do so, subject to the following conditions:
|
|
///
|
|
/// The above copyright notice and this permission notice shall be included in
|
|
/// all copies or substantial portions of the Software.
|
|
///
|
|
/// THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
/// IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
/// FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
|
|
/// AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
/// LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
|
/// OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
|
|
/// THE SOFTWARE.
|
|
///
|
|
/// @ref core
|
|
/// @file glm/core/intrinsic_geometric.inl
|
|
/// @date 2009-05-08 / 2011-06-15
|
|
/// @author Christophe Riccio
|
|
///////////////////////////////////////////////////////////////////////////////////
|
|
|
|
namespace glm{
|
|
namespace detail{
|
|
|
|
//length
|
|
GLM_FUNC_QUALIFIER __m128 sse_len_ps(__m128 x)
|
|
{
|
|
__m128 dot0 = sse_dot_ps(x, x);
|
|
__m128 sqt0 = _mm_sqrt_ps(dot0);
|
|
return sqt0;
|
|
}
|
|
|
|
//distance
|
|
GLM_FUNC_QUALIFIER __m128 sse_dst_ps(__m128 p0, __m128 p1)
|
|
{
|
|
__m128 sub0 = _mm_sub_ps(p0, p1);
|
|
__m128 len0 = sse_len_ps(sub0);
|
|
return len0;
|
|
}
|
|
|
|
//dot
|
|
GLM_FUNC_QUALIFIER __m128 sse_dot_ps(__m128 v1, __m128 v2)
|
|
{
|
|
# if((GLM_ARCH & GLM_ARCH_AVX) == GLM_ARCH_AVX)
|
|
return _mm_dp_ps(v1, v2, 0xff);
|
|
# else
|
|
__m128 mul0 = _mm_mul_ps(v1, v2);
|
|
__m128 swp0 = _mm_shuffle_ps(mul0, mul0, _MM_SHUFFLE(2, 3, 0, 1));
|
|
__m128 add0 = _mm_add_ps(mul0, swp0);
|
|
__m128 swp1 = _mm_shuffle_ps(add0, add0, _MM_SHUFFLE(0, 1, 2, 3));
|
|
__m128 add1 = _mm_add_ps(add0, swp1);
|
|
return add1;
|
|
# endif
|
|
}
|
|
|
|
// SSE1
|
|
GLM_FUNC_QUALIFIER __m128 sse_dot_ss(__m128 v1, __m128 v2)
|
|
{
|
|
__m128 mul0 = _mm_mul_ps(v1, v2);
|
|
__m128 mov0 = _mm_movehl_ps(mul0, mul0);
|
|
__m128 add0 = _mm_add_ps(mov0, mul0);
|
|
__m128 swp1 = _mm_shuffle_ps(add0, add0, 1);
|
|
__m128 add1 = _mm_add_ss(add0, swp1);
|
|
return add1;
|
|
}
|
|
|
|
//cross
|
|
GLM_FUNC_QUALIFIER __m128 sse_xpd_ps(__m128 v1, __m128 v2)
|
|
{
|
|
__m128 swp0 = _mm_shuffle_ps(v1, v1, _MM_SHUFFLE(3, 0, 2, 1));
|
|
__m128 swp1 = _mm_shuffle_ps(v1, v1, _MM_SHUFFLE(3, 1, 0, 2));
|
|
__m128 swp2 = _mm_shuffle_ps(v2, v2, _MM_SHUFFLE(3, 0, 2, 1));
|
|
__m128 swp3 = _mm_shuffle_ps(v2, v2, _MM_SHUFFLE(3, 1, 0, 2));
|
|
__m128 mul0 = _mm_mul_ps(swp0, swp3);
|
|
__m128 mul1 = _mm_mul_ps(swp1, swp2);
|
|
__m128 sub0 = _mm_sub_ps(mul0, mul1);
|
|
return sub0;
|
|
}
|
|
|
|
//normalize
|
|
GLM_FUNC_QUALIFIER __m128 sse_nrm_ps(__m128 v)
|
|
{
|
|
__m128 dot0 = sse_dot_ps(v, v);
|
|
__m128 isr0 = _mm_rsqrt_ps(dot0);
|
|
__m128 mul0 = _mm_mul_ps(v, isr0);
|
|
return mul0;
|
|
}
|
|
|
|
//faceforward
|
|
GLM_FUNC_QUALIFIER __m128 sse_ffd_ps(__m128 N, __m128 I, __m128 Nref)
|
|
{
|
|
//__m128 dot0 = _mm_dot_ps(v, v);
|
|
//__m128 neg0 = _mm_neg_ps(N);
|
|
//__m128 sgn0 = _mm_sgn_ps(dot0);
|
|
//__m128 mix0 = _mm_mix_ps(N, neg0, sgn0);
|
|
//return mix0;
|
|
|
|
__m128 dot0 = sse_dot_ps(Nref, I);
|
|
__m128 sgn0 = sse_sgn_ps(dot0);
|
|
__m128 mul0 = _mm_mul_ps(sgn0, glm::detail::minus_one);
|
|
__m128 mul1 = _mm_mul_ps(N, mul0);
|
|
return mul1;
|
|
}
|
|
|
|
//reflect
|
|
GLM_FUNC_QUALIFIER __m128 sse_rfe_ps(__m128 I, __m128 N)
|
|
{
|
|
__m128 dot0 = sse_dot_ps(N, I);
|
|
__m128 mul0 = _mm_mul_ps(N, dot0);
|
|
__m128 mul1 = _mm_mul_ps(mul0, glm::detail::two);
|
|
__m128 sub0 = _mm_sub_ps(I, mul1);
|
|
return sub0;
|
|
}
|
|
|
|
//refract
|
|
GLM_FUNC_QUALIFIER __m128 sse_rfa_ps(__m128 I, __m128 N, __m128 eta)
|
|
{
|
|
__m128 dot0 = sse_dot_ps(N, I);
|
|
__m128 mul0 = _mm_mul_ps(eta, eta);
|
|
__m128 mul1 = _mm_mul_ps(dot0, dot0);
|
|
__m128 sub0 = _mm_sub_ps(glm::detail::one, mul0);
|
|
__m128 sub1 = _mm_sub_ps(glm::detail::one, mul1);
|
|
__m128 mul2 = _mm_mul_ps(sub0, sub1);
|
|
|
|
if(_mm_movemask_ps(_mm_cmplt_ss(mul2, glm::detail::zero)) == 0)
|
|
return glm::detail::zero;
|
|
|
|
__m128 sqt0 = _mm_sqrt_ps(mul2);
|
|
__m128 mul3 = _mm_mul_ps(eta, dot0);
|
|
__m128 add0 = _mm_add_ps(mul3, sqt0);
|
|
__m128 mul4 = _mm_mul_ps(add0, N);
|
|
__m128 mul5 = _mm_mul_ps(eta, I);
|
|
__m128 sub2 = _mm_sub_ps(mul5, mul4);
|
|
|
|
return sub2;
|
|
}
|
|
|
|
}//namespace detail
|
|
}//namespace glm
|