ref: cfe6fa98f72cae9587152591ffd52035e71b9562
parent: aa86438a0d5f7e24506b3bc66603eefe3cf26096
 parent: 849b63ffe1723e4e3c021c821cbd50db7cb5211d
	author: James Zern <jzern@google.com>
	date: Thu Oct 24 02:36:38 EDT 2019
	
Merge "vpx_int_pro_col_sse2: use unaligned loads"
--- a/vpx_dsp/x86/avg_intrin_sse2.c
+++ b/vpx_dsp/x86/avg_intrin_sse2.c
@@ -515,7 +515,7 @@
 int16_t vpx_int_pro_col_sse2(const uint8_t *ref, const int width) {__m128i zero = _mm_setzero_si128();
- __m128i src_line = _mm_load_si128((const __m128i *)ref);
+ __m128i src_line = _mm_loadu_si128((const __m128i *)ref);
__m128i s0 = _mm_sad_epu8(src_line, zero);
__m128i s1;
int i;
@@ -522,7 +522,7 @@
   for (i = 16; i < width; i += 16) {ref += 16;
- src_line = _mm_load_si128((const __m128i *)ref);
+ src_line = _mm_loadu_si128((const __m128i *)ref);
s1 = _mm_sad_epu8(src_line, zero);
s0 = _mm_adds_epu16(s0, s1);
}
--
⑨