125 lines
		
	
	
		
			3.8 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			125 lines
		
	
	
		
			3.8 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /// @ref simd
 | |
| /// @file glm/simd/geometric.h
 | |
| 
 | |
| #pragma once
 | |
| 
 | |
| #include "common.h"
 | |
| 
 | |
| #if GLM_ARCH & GLM_ARCH_SSE2_BIT
 | |
| 
 | |
| GLM_FUNC_DECL glm_vec4 glm_vec4_dot(glm_vec4 v1, glm_vec4 v2);
 | |
| GLM_FUNC_DECL glm_vec4 glm_vec1_dot(glm_vec4 v1, glm_vec4 v2);
 | |
| 
 | |
| GLM_FUNC_QUALIFIER glm_vec4 glm_vec4_length(glm_vec4 x)
 | |
| {
 | |
| 	glm_vec4 const dot0 = glm_vec4_dot(x, x);
 | |
| 	glm_vec4 const sqt0 = _mm_sqrt_ps(dot0);
 | |
| 	return sqt0;
 | |
| }
 | |
| 
 | |
| GLM_FUNC_QUALIFIER glm_vec4 glm_vec4_distance(glm_vec4 p0, glm_vec4 p1)
 | |
| {
 | |
| 	glm_vec4 const sub0 = _mm_sub_ps(p0, p1);
 | |
| 	glm_vec4 const len0 = glm_vec4_length(sub0);
 | |
| 	return len0;
 | |
| }
 | |
| 
 | |
| GLM_FUNC_QUALIFIER glm_vec4 glm_vec4_dot(glm_vec4 v1, glm_vec4 v2)
 | |
| {
 | |
| #	if GLM_ARCH & GLM_ARCH_AVX_BIT
 | |
| 		return _mm_dp_ps(v1, v2, 0xff);
 | |
| #	elif GLM_ARCH & GLM_ARCH_SSE3_BIT
 | |
| 		glm_vec4 const mul0 = _mm_mul_ps(v1, v2);
 | |
| 		glm_vec4 const hadd0 = _mm_hadd_ps(mul0, mul0);
 | |
| 		glm_vec4 const hadd1 = _mm_hadd_ps(hadd0, hadd0);
 | |
| 		return hadd1;
 | |
| #	else
 | |
| 		glm_vec4 const mul0 = _mm_mul_ps(v1, v2);
 | |
| 		glm_vec4 const swp0 = _mm_shuffle_ps(mul0, mul0, _MM_SHUFFLE(2, 3, 0, 1));
 | |
| 		glm_vec4 const add0 = _mm_add_ps(mul0, swp0);
 | |
| 		glm_vec4 const swp1 = _mm_shuffle_ps(add0, add0, _MM_SHUFFLE(0, 1, 2, 3));
 | |
| 		glm_vec4 const add1 = _mm_add_ps(add0, swp1);
 | |
| 		return add1;
 | |
| #	endif
 | |
| }
 | |
| 
 | |
| GLM_FUNC_QUALIFIER glm_vec4 glm_vec1_dot(glm_vec4 v1, glm_vec4 v2)
 | |
| {
 | |
| #	if GLM_ARCH & GLM_ARCH_AVX_BIT
 | |
| 		return _mm_dp_ps(v1, v2, 0xff);
 | |
| #	elif GLM_ARCH & GLM_ARCH_SSE3_BIT
 | |
| 		glm_vec4 const mul0 = _mm_mul_ps(v1, v2);
 | |
| 		glm_vec4 const had0 = _mm_hadd_ps(mul0, mul0);
 | |
| 		glm_vec4 const had1 = _mm_hadd_ps(had0, had0);
 | |
| 		return had1;
 | |
| #	else
 | |
| 		glm_vec4 const mul0 = _mm_mul_ps(v1, v2);
 | |
| 		glm_vec4 const mov0 = _mm_movehl_ps(mul0, mul0);
 | |
| 		glm_vec4 const add0 = _mm_add_ps(mov0, mul0);
 | |
| 		glm_vec4 const swp1 = _mm_shuffle_ps(add0, add0, 1);
 | |
| 		glm_vec4 const add1 = _mm_add_ss(add0, swp1);
 | |
| 		return add1;
 | |
| #	endif
 | |
| }
 | |
| 
 | |
| GLM_FUNC_QUALIFIER glm_vec4 glm_vec4_cross(glm_vec4 v1, glm_vec4 v2)
 | |
| {
 | |
| 	glm_vec4 const swp0 = _mm_shuffle_ps(v1, v1, _MM_SHUFFLE(3, 0, 2, 1));
 | |
| 	glm_vec4 const swp1 = _mm_shuffle_ps(v1, v1, _MM_SHUFFLE(3, 1, 0, 2));
 | |
| 	glm_vec4 const swp2 = _mm_shuffle_ps(v2, v2, _MM_SHUFFLE(3, 0, 2, 1));
 | |
| 	glm_vec4 const swp3 = _mm_shuffle_ps(v2, v2, _MM_SHUFFLE(3, 1, 0, 2));
 | |
| 	glm_vec4 const mul0 = _mm_mul_ps(swp0, swp3);
 | |
| 	glm_vec4 const mul1 = _mm_mul_ps(swp1, swp2);
 | |
| 	glm_vec4 const sub0 = _mm_sub_ps(mul0, mul1);
 | |
| 	return sub0;
 | |
| }
 | |
| 
 | |
| GLM_FUNC_QUALIFIER glm_vec4 glm_vec4_normalize(glm_vec4 v)
 | |
| {
 | |
| 	glm_vec4 const dot0 = glm_vec4_dot(v, v);
 | |
| 	glm_vec4 const isr0 = _mm_rsqrt_ps(dot0);
 | |
| 	glm_vec4 const mul0 = _mm_mul_ps(v, isr0);
 | |
| 	return mul0;
 | |
| }
 | |
| 
 | |
| GLM_FUNC_QUALIFIER glm_vec4 glm_vec4_faceforward(glm_vec4 N, glm_vec4 I, glm_vec4 Nref)
 | |
| {
 | |
| 	glm_vec4 const dot0 = glm_vec4_dot(Nref, I);
 | |
| 	glm_vec4 const sgn0 = glm_vec4_sign(dot0);
 | |
| 	glm_vec4 const mul0 = _mm_mul_ps(sgn0, _mm_set1_ps(-1.0f));
 | |
| 	glm_vec4 const mul1 = _mm_mul_ps(N, mul0);
 | |
| 	return mul1;
 | |
| }
 | |
| 
 | |
| GLM_FUNC_QUALIFIER glm_vec4 glm_vec4_reflect(glm_vec4 I, glm_vec4 N)
 | |
| {
 | |
| 	glm_vec4 const dot0 = glm_vec4_dot(N, I);
 | |
| 	glm_vec4 const mul0 = _mm_mul_ps(N, dot0);
 | |
| 	glm_vec4 const mul1 = _mm_mul_ps(mul0, _mm_set1_ps(2.0f));
 | |
| 	glm_vec4 const sub0 = _mm_sub_ps(I, mul1);
 | |
| 	return sub0;
 | |
| }
 | |
| 
 | |
| GLM_FUNC_QUALIFIER __m128 glm_vec4_refract(glm_vec4 I, glm_vec4 N, glm_vec4 eta)
 | |
| {
 | |
| 	glm_vec4 const dot0 = glm_vec4_dot(N, I);
 | |
| 	glm_vec4 const mul0 = _mm_mul_ps(eta, eta);
 | |
| 	glm_vec4 const mul1 = _mm_mul_ps(dot0, dot0);
 | |
| 	glm_vec4 const sub0 = _mm_sub_ps(_mm_set1_ps(1.0f), mul0);
 | |
| 	glm_vec4 const sub1 = _mm_sub_ps(_mm_set1_ps(1.0f), mul1);
 | |
| 	glm_vec4 const mul2 = _mm_mul_ps(sub0, sub1);
 | |
| 	
 | |
| 	if(_mm_movemask_ps(_mm_cmplt_ss(mul2, _mm_set1_ps(0.0f))) == 0)
 | |
| 		return _mm_set1_ps(0.0f);
 | |
| 
 | |
| 	glm_vec4 const sqt0 = _mm_sqrt_ps(mul2);
 | |
| 	glm_vec4 const mad0 = glm_vec4_fma(eta, dot0, sqt0);
 | |
| 	glm_vec4 const mul4 = _mm_mul_ps(mad0, N);
 | |
| 	glm_vec4 const mul5 = _mm_mul_ps(eta, I);
 | |
| 	glm_vec4 const sub2 = _mm_sub_ps(mul5, mul4);
 | |
| 
 | |
| 	return sub2;
 | |
| }
 | |
| 
 | |
| #endif//GLM_ARCH & GLM_ARCH_SSE2_BIT
 | 
