/external/libvpx/libvpx/vpx_util/ |
D | vpx_debug_util.c | 120 const uint16_t *src16 = highbd ? CONVERT_TO_SHORTPTR(src) : NULL; in mismatch_record_block_pre() local 132 src16 ? src16[r * src_stride + c] : src[r * src_stride + c]; in mismatch_record_block_pre() 155 const uint16_t *src16 = highbd ? CONVERT_TO_SHORTPTR(src) : NULL; in mismatch_record_block_tx() local 166 src16 ? src16[r * src_stride + c] : src[r * src_stride + c]; in mismatch_record_block_tx() 189 const uint16_t *src16 = highbd ? CONVERT_TO_SHORTPTR(src) : NULL; in mismatch_check_block_pre() local 201 (uint16_t)(src16 ? src16[r * src_stride + c] in mismatch_check_block_pre() 227 src16 ? src16[rr * src_stride + cc] : src[rr * src_stride + cc]); in mismatch_check_block_pre() 237 const uint16_t *src16 = highbd ? CONVERT_TO_SHORTPTR(src) : NULL; in mismatch_check_block_tx() local 249 (uint16_t)(src16 ? src16[r * src_stride + c] in mismatch_check_block_tx() 275 src16 ? src16[rr * src_stride + cc] : src[rr * src_stride + cc]); in mismatch_check_block_tx()
|
/external/libaom/libaom/aom_util/ |
D | debug_util.c | 130 const uint16_t *src16 = highbd ? CONVERT_TO_SHORTPTR(src) : NULL; in mismatch_record_block_pre() local 135 src16 ? src16[r * src_stride + c] : src[r * src_stride + c]; in mismatch_record_block_pre() 163 const uint16_t *src16 = highbd ? CONVERT_TO_SHORTPTR(src) : NULL; in mismatch_record_block_tx() local 168 src16 ? src16[r * src_stride + c] : src[r * src_stride + c]; in mismatch_record_block_tx() 195 const uint16_t *src16 = highbd ? CONVERT_TO_SHORTPTR(src) : NULL; in mismatch_check_block_pre() local 201 (uint16_t)(src16 ? src16[r * src_stride + c] in mismatch_check_block_pre() 226 src16 ? src16[rr * src_stride + cc] : src[rr * src_stride + cc]); in mismatch_check_block_pre() 241 const uint16_t *src16 = highbd ? CONVERT_TO_SHORTPTR(src) : NULL; in mismatch_check_block_tx() local 247 (uint16_t)(src16 ? src16[r * src_stride + c] in mismatch_check_block_tx() 272 src16 ? src16[rr * src_stride + cc] : src[rr * src_stride + cc]); in mismatch_check_block_tx()
|
/external/llvm-project/lldb/source/Host/common/ |
D | SocketAddress.cpp | 56 const uint16_t *src16 = static_cast<const uint16_t *>(src); in inet_ntop() local 58 tmp, sizeof(tmp), "%x:%x:%x:%x:%x:%x:%x:%x", ntohs(src16[0]), in inet_ntop() 59 ntohs(src16[1]), ntohs(src16[2]), ntohs(src16[3]), ntohs(src16[4]), in inet_ntop() 60 ntohs(src16[5]), ntohs(src16[6]), ntohs(src16[7])); in inet_ntop()
|
/external/libaom/libaom/aom_scale/generic/ |
D | yv12extend.c | 273 const uint16_t *src16 = CONVERT_TO_SHORTPTR(src); in aom_yv12_copy_y_c() local 276 memcpy(dst16, src16, src_ybc->y_width * sizeof(uint16_t)); in aom_yv12_copy_y_c() 277 src16 += src_ybc->y_stride; in aom_yv12_copy_y_c() 298 const uint16_t *src16 = CONVERT_TO_SHORTPTR(src); in aom_yv12_copy_u_c() local 301 memcpy(dst16, src16, src_bc->uv_width * sizeof(uint16_t)); in aom_yv12_copy_u_c() 302 src16 += src_bc->uv_stride; in aom_yv12_copy_u_c() 322 const uint16_t *src16 = CONVERT_TO_SHORTPTR(src); in aom_yv12_copy_v_c() local 325 memcpy(dst16, src16, src_bc->uv_width * sizeof(uint16_t)); in aom_yv12_copy_v_c() 326 src16 += src_bc->uv_stride; in aom_yv12_copy_v_c() 348 const uint16_t *src16 = in aom_yv12_partial_copy_y_c() local [all …]
|
/external/skia/src/core/ |
D | SkConvertPixels.cpp | 88 auto src16 = (const uint16_t*) src; in convert_to_alpha8() local 91 dst[x] = src16[x] >> 8; in convert_to_alpha8() 94 src16 = SkTAddOffset<const uint16_t>(src16, srcRB); in convert_to_alpha8() 115 auto src16 = (const uint16_t*) src; in convert_to_alpha8() local 118 dst[x] = SkPacked4444ToA32(src16[x]); in convert_to_alpha8() 121 src16 = SkTAddOffset<const uint16_t>(src16, srcRB); in convert_to_alpha8()
|
/external/llvm/test/Transforms/SLPVectorizer/X86/ |
D | bswap.ll | 12 @src16 = common global [16 x i16] zeroinitializer, align 32 148 ; CHECK-NEXT: [[TMP1:%.*]] = load <8 x i16>, <8 x i16>* bitcast ([16 x i16]* @src16 to <8 x i16>… 153 …%ld0 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 0), align… 154 …%ld1 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 1), align… 155 …%ld2 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 2), align… 156 …%ld3 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 3), align… 157 …%ld4 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 4), align… 158 …%ld5 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 5), align… 159 …%ld6 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 6), align… 160 …%ld7 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 7), align… [all …]
|
D | ctpop.ll | 12 @src16 = common global [16 x i16] zeroinitializer, align 32 136 ; CHECK-NEXT: [[TMP1:%.*]] = load <8 x i16>, <8 x i16>* bitcast ([16 x i16]* @src16 to <8 x i16>… 141 …%ld0 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 0), align… 142 …%ld1 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 1), align… 143 …%ld2 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 2), align… 144 …%ld3 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 3), align… 145 …%ld4 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 4), align… 146 …%ld5 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 5), align… 147 …%ld6 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 6), align… 148 …%ld7 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 7), align… [all …]
|
D | ctlz.ll | 12 @src16 = common global [16 x i16] zeroinitializer, align 32 164 …0:%.*]] = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 0), al… 165 …1:%.*]] = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 1), al… 166 …2:%.*]] = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 2), al… 167 …3:%.*]] = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 3), al… 168 …4:%.*]] = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 4), al… 169 …5:%.*]] = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 5), al… 170 …6:%.*]] = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 6), al… 171 …7:%.*]] = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 7), al… 190 …%ld0 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 0), align… [all …]
|
D | cttz.ll | 12 @src16 = common global [16 x i16] zeroinitializer, align 32 164 …0:%.*]] = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 0), al… 165 …1:%.*]] = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 1), al… 166 …2:%.*]] = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 2), al… 167 …3:%.*]] = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 3), al… 168 …4:%.*]] = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 4), al… 169 …5:%.*]] = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 5), al… 170 …6:%.*]] = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 6), al… 171 …7:%.*]] = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 7), al… 190 …%ld0 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 0), align… [all …]
|
D | bitreverse.ll | 14 @src16 = common global [16 x i16] zeroinitializer, align 32 228 ; SSE-NEXT: [[LD0:%.*]] = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16,… 229 ; SSE-NEXT: [[LD1:%.*]] = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16,… 230 ; SSE-NEXT: [[LD2:%.*]] = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16,… 231 ; SSE-NEXT: [[LD3:%.*]] = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16,… 232 ; SSE-NEXT: [[LD4:%.*]] = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16,… 233 ; SSE-NEXT: [[LD5:%.*]] = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16,… 234 ; SSE-NEXT: [[LD6:%.*]] = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16,… 235 ; SSE-NEXT: [[LD7:%.*]] = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16,… 255 ; AVX-NEXT: [[TMP1:%.*]] = load <8 x i16>, <8 x i16>* bitcast ([16 x i16]* @src16 to <8 x i16>*)… [all …]
|
/external/llvm-project/llvm/test/Transforms/SLPVectorizer/X86/ |
D | bswap.ll | 12 @src16 = common global [16 x i16] zeroinitializer, align 32 148 ; CHECK-NEXT: [[TMP1:%.*]] = load <8 x i16>, <8 x i16>* bitcast ([16 x i16]* @src16 to <8 x i16>… 153 …%ld0 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 0), align… 154 …%ld1 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 1), align… 155 …%ld2 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 2), align… 156 …%ld3 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 3), align… 157 …%ld4 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 4), align… 158 …%ld5 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 5), align… 159 …%ld6 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 6), align… 160 …%ld7 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 7), align… [all …]
|
D | uitofp.ll | 13 @src16 = common global [32 x i16] zeroinitializer, align 64 277 …0:%.*]] = load i16, i16* getelementptr inbounds ([32 x i16], [32 x i16]* @src16, i32 0, i64 0), al… 278 …1:%.*]] = load i16, i16* getelementptr inbounds ([32 x i16], [32 x i16]* @src16, i32 0, i64 1), al… 285 …%ld0 = load i16, i16* getelementptr inbounds ([32 x i16], [32 x i16]* @src16, i32 0, i64 0), align… 286 …%ld1 = load i16, i16* getelementptr inbounds ([32 x i16], [32 x i16]* @src16, i32 0, i64 1), align… 296 ; SSE-NEXT: [[LD0:%.*]] = load i16, i16* getelementptr inbounds ([32 x i16], [32 x i16]* @src16,… 297 ; SSE-NEXT: [[LD1:%.*]] = load i16, i16* getelementptr inbounds ([32 x i16], [32 x i16]* @src16,… 298 ; SSE-NEXT: [[LD2:%.*]] = load i16, i16* getelementptr inbounds ([32 x i16], [32 x i16]* @src16,… 299 ; SSE-NEXT: [[LD3:%.*]] = load i16, i16* getelementptr inbounds ([32 x i16], [32 x i16]* @src16,… 311 ; AVX-NEXT: [[TMP1:%.*]] = load <4 x i16>, <4 x i16>* bitcast ([32 x i16]* @src16 to <4 x i16>*)… [all …]
|
D | sitofp.ll | 13 @src16 = common global [32 x i16] zeroinitializer, align 64 346 …0:%.*]] = load i16, i16* getelementptr inbounds ([32 x i16], [32 x i16]* @src16, i32 0, i64 0), al… 347 …1:%.*]] = load i16, i16* getelementptr inbounds ([32 x i16], [32 x i16]* @src16, i32 0, i64 1), al… 354 …%ld0 = load i16, i16* getelementptr inbounds ([32 x i16], [32 x i16]* @src16, i32 0, i64 0), align… 355 …%ld1 = load i16, i16* getelementptr inbounds ([32 x i16], [32 x i16]* @src16, i32 0, i64 1), align… 365 ; SSE-NEXT: [[LD0:%.*]] = load i16, i16* getelementptr inbounds ([32 x i16], [32 x i16]* @src16,… 366 ; SSE-NEXT: [[LD1:%.*]] = load i16, i16* getelementptr inbounds ([32 x i16], [32 x i16]* @src16,… 367 ; SSE-NEXT: [[LD2:%.*]] = load i16, i16* getelementptr inbounds ([32 x i16], [32 x i16]* @src16,… 368 ; SSE-NEXT: [[LD3:%.*]] = load i16, i16* getelementptr inbounds ([32 x i16], [32 x i16]* @src16,… 380 ; AVX-NEXT: [[TMP1:%.*]] = load <4 x i16>, <4 x i16>* bitcast ([32 x i16]* @src16 to <4 x i16>*)… [all …]
|
D | bitreverse.ll | 14 @src16 = common global [16 x i16] zeroinitializer, align 32 150 ; CHECK-NEXT: [[TMP1:%.*]] = load <8 x i16>, <8 x i16>* bitcast ([16 x i16]* @src16 to <8 x i16>… 155 …%ld0 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 0), align… 156 …%ld1 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 1), align… 157 …%ld2 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 2), align… 158 …%ld3 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 3), align… 159 …%ld4 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 4), align… 160 …%ld5 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 5), align… 161 …%ld6 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 6), align… 162 …%ld7 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 7), align… [all …]
|
D | ctlz.ll | 13 @src16 = common global [16 x i16] zeroinitializer, align 32 207 ; CHECK-NEXT: [[TMP1:%.*]] = load <8 x i16>, <8 x i16>* bitcast ([16 x i16]* @src16 to <8 x i16>… 212 …%ld0 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 0), align… 213 …%ld1 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 1), align… 214 …%ld2 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 2), align… 215 …%ld3 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 3), align… 216 …%ld4 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 4), align… 217 …%ld5 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 5), align… 218 …%ld6 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 6), align… 219 …%ld7 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 7), align… [all …]
|
D | cttz.ll | 13 @src16 = common global [16 x i16] zeroinitializer, align 32 207 ; CHECK-NEXT: [[TMP1:%.*]] = load <8 x i16>, <8 x i16>* bitcast ([16 x i16]* @src16 to <8 x i16>… 212 …%ld0 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 0), align… 213 …%ld1 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 1), align… 214 …%ld2 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 2), align… 215 …%ld3 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 3), align… 216 …%ld4 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 4), align… 217 …%ld5 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 5), align… 218 …%ld6 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 6), align… 219 …%ld7 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 7), align… [all …]
|
D | ctpop.ll | 13 @src16 = common global [16 x i16] zeroinitializer, align 32 269 ; CHECK-NEXT: [[TMP1:%.*]] = load <8 x i16>, <8 x i16>* bitcast ([16 x i16]* @src16 to <8 x i16>… 274 …%ld0 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 0), align… 275 …%ld1 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 1), align… 276 …%ld2 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 2), align… 277 …%ld3 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 3), align… 278 …%ld4 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 4), align… 279 …%ld5 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 5), align… 280 …%ld6 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 6), align… 281 …%ld7 = load i16, i16* getelementptr inbounds ([16 x i16], [16 x i16]* @src16, i16 0, i64 7), align… [all …]
|
/external/libaom/libaom/test/ |
D | transform_test_base.h | 66 uint16_t *src16 = reinterpret_cast<uint16_t *>( in RunAccuracyCheck() local 77 src16[j] = rnd.Rand16() & mask_; in RunAccuracyCheck() 79 test_input_block[j] = src16[j] - dst16[j]; in RunAccuracyCheck() 94 bit_depth_ == AOM_BITS_8 ? dst[j] - src[j] : dst16[j] - src16[j]; in RunAccuracyCheck() 116 aom_free(src16); in RunAccuracyCheck() 275 uint16_t *src16 = reinterpret_cast<uint16_t *>( in RunInvAccuracyCheck() local 286 src16[j] = rnd.Rand16() & mask_; in RunInvAccuracyCheck() 288 in[j] = src16[j] - dst16[j]; in RunInvAccuracyCheck() 303 bit_depth_ == AOM_BITS_8 ? dst[j] - src[j] : dst16[j] - src16[j]; in RunInvAccuracyCheck() 313 aom_free(src16); in RunInvAccuracyCheck()
|
/external/libvpx/libvpx/vpx_dsp/x86/ |
D | avg_intrin_avx2.c | 97 __m128i src16[8]; in vpx_highbd_hadamard_8x8_avx2() local 100 src16[0] = _mm_loadu_si128((const __m128i *)src_diff); in vpx_highbd_hadamard_8x8_avx2() 101 src16[1] = _mm_loadu_si128((const __m128i *)(src_diff += src_stride)); in vpx_highbd_hadamard_8x8_avx2() 102 src16[2] = _mm_loadu_si128((const __m128i *)(src_diff += src_stride)); in vpx_highbd_hadamard_8x8_avx2() 103 src16[3] = _mm_loadu_si128((const __m128i *)(src_diff += src_stride)); in vpx_highbd_hadamard_8x8_avx2() 104 src16[4] = _mm_loadu_si128((const __m128i *)(src_diff += src_stride)); in vpx_highbd_hadamard_8x8_avx2() 105 src16[5] = _mm_loadu_si128((const __m128i *)(src_diff += src_stride)); in vpx_highbd_hadamard_8x8_avx2() 106 src16[6] = _mm_loadu_si128((const __m128i *)(src_diff += src_stride)); in vpx_highbd_hadamard_8x8_avx2() 107 src16[7] = _mm_loadu_si128((const __m128i *)(src_diff += src_stride)); in vpx_highbd_hadamard_8x8_avx2() 109 src32[0] = _mm256_cvtepi16_epi32(src16[0]); in vpx_highbd_hadamard_8x8_avx2() [all …]
|
/external/libaom/libaom/aom_dsp/x86/ |
D | avg_intrin_avx2.c | 338 __m128i src16[8]; in aom_highbd_hadamard_8x8_avx2() local 341 src16[0] = _mm_loadu_si128((const __m128i *)src_diff); in aom_highbd_hadamard_8x8_avx2() 342 src16[1] = _mm_loadu_si128((const __m128i *)(src_diff += src_stride)); in aom_highbd_hadamard_8x8_avx2() 343 src16[2] = _mm_loadu_si128((const __m128i *)(src_diff += src_stride)); in aom_highbd_hadamard_8x8_avx2() 344 src16[3] = _mm_loadu_si128((const __m128i *)(src_diff += src_stride)); in aom_highbd_hadamard_8x8_avx2() 345 src16[4] = _mm_loadu_si128((const __m128i *)(src_diff += src_stride)); in aom_highbd_hadamard_8x8_avx2() 346 src16[5] = _mm_loadu_si128((const __m128i *)(src_diff += src_stride)); in aom_highbd_hadamard_8x8_avx2() 347 src16[6] = _mm_loadu_si128((const __m128i *)(src_diff += src_stride)); in aom_highbd_hadamard_8x8_avx2() 348 src16[7] = _mm_loadu_si128((const __m128i *)(src_diff += src_stride)); in aom_highbd_hadamard_8x8_avx2() 350 src32[0] = _mm256_cvtepi16_epi32(src16[0]); in aom_highbd_hadamard_8x8_avx2() [all …]
|
/external/libvpx/libvpx/test/ |
D | fdct8x8_test.cc | 224 DECLARE_ALIGNED(16, uint16_t, src16[64]); in RunRoundTripErrorCheck() 236 src16[j] = rnd.Rand16() & mask_; in RunRoundTripErrorCheck() 238 test_input_block[j] = src16[j] - dst16[j]; in RunRoundTripErrorCheck() 268 bit_depth_ == VPX_BITS_8 ? dst[j] - src[j] : dst16[j] - src16[j]; in RunRoundTripErrorCheck() 300 DECLARE_ALIGNED(16, uint16_t, src16[64]); in RunExtremalCheck() 321 src16[j] = mask_; in RunExtremalCheck() 324 src16[j] = 0; in RunExtremalCheck() 327 src16[j] = rnd.Rand8() % 2 ? mask_ : 0; in RunExtremalCheck() 330 test_input_block[j] = src16[j] - dst16[j]; in RunExtremalCheck() 351 bit_depth_ == VPX_BITS_8 ? dst[j] - src[j] : dst16[j] - src16[j]; in RunExtremalCheck() [all …]
|
D | dct32x32_test.cc | 124 DECLARE_ALIGNED(16, uint16_t, src16[kNumCoeffs]); in TEST_P() 136 src16[j] = rnd.Rand16() & mask_; in TEST_P() 138 test_input_block[j] = src16[j] - dst16[j]; in TEST_P() 156 bit_depth_ == VPX_BITS_8 ? dst[j] - src[j] : dst16[j] - src16[j]; in TEST_P() 271 DECLARE_ALIGNED(16, uint16_t, src16[kNumCoeffs]); in TEST_P() 285 src16[j] = rnd.Rand16() & mask_; in TEST_P() 287 in[j] = src16[j] - dst16[j]; in TEST_P() 306 bit_depth_ == VPX_BITS_8 ? dst[j] - src[j] : dst16[j] - src16[j]; in TEST_P()
|
/external/icu/icu4c/source/test/cintltst/ |
D | custrtrn.c | 103 static const UChar src16[] = { variable 152 u_strToUTF32(u32Target, 0, &u32DestLen, src16, UPRV_LENGTHOF(src16),&err); in Test_strToUTF32() 161 u_strToUTF32(u32Target, UPRV_LENGTHOF(src32)+1, &u32DestLen, src16, UPRV_LENGTHOF(src16),&err); in Test_strToUTF32() 185 u_strToUTF32(NULL,0, &u32DestLen, src16, -1,&err); in Test_strToUTF32() 194 u_strToUTF32(u32Target, UPRV_LENGTHOF(src32), &u32DestLen, src16, -1,&err); in Test_strToUTF32() 332 if(err != U_BUFFER_OVERFLOW_ERROR || uDestLen != UPRV_LENGTHOF(src16)) { in Test_strFromUTF32() 335 (long)uDestLen, (long)UPRV_LENGTHOF(src16), u_errorName(err)); in Test_strFromUTF32() 340 u_strFromUTF32(uTarget, UPRV_LENGTHOF(src16)+1,&uDestLen,src32,UPRV_LENGTHOF(src32),&err); in Test_strFromUTF32() 341 if(err != U_ZERO_ERROR || uDestLen != UPRV_LENGTHOF(src16)) { in Test_strFromUTF32() 344 (long)uDestLen, (long)UPRV_LENGTHOF(src16), u_errorName(err)); in Test_strFromUTF32() [all …]
|
/external/skqp/src/core/ |
D | SkConvertPixels.cpp | 99 auto src16 = (const uint16_t*) src; in convert_to_alpha8() local 102 dst[x] = SkPacked4444ToA32(src16[x]); in convert_to_alpha8() 105 src16 = SkTAddOffset<const uint16_t>(src16, srcRB); in convert_to_alpha8()
|
/external/llvm-project/llvm/test/Transforms/LICM/ |
D | pr36228.ll | 15 %src16 = bitcast i8* %src to i16* 16 %srcval = load i16, i16* %src16
|