/aosp_15_r20/external/libgav1/src/dsp/x86/ |
H A D | inverse_transform_sse4.cc | 229 const __m128i v_src = in DctDcOnly() local 273 const __m128i v_src = LoadLo8(dst); in DctDcOnlyColumn() local 279 const __m128i v_src = LoadUnaligned16(&dst[i]); in DctDcOnlyColumn() local 1040 const __m128i v_src = in Adst4DcOnly() local 1073 const __m128i v_src = _mm_cvtepi16_epi32(LoadLo8(&dst[i])); in Adst4DcOnlyColumn() local 1197 const __m128i v_src = _mm_shufflelo_epi16(_mm_cvtsi32_si128(dst[0]), 0); in Adst8DcOnly() local 1266 const __m128i v_src = LoadLo8(dst); in Adst8DcOnlyColumn() local 1523 const __m128i v_src = _mm_shufflelo_epi16(_mm_cvtsi32_si128(dst[0]), 0); in Adst16DcOnly() local 1565 const __m128i v_src = LoadUnaligned16(dst); in Adst16DcOnlyColumn() local 1594 const __m128i v_src = LoadUnaligned16(&dst[i * step]); in Identity4_SSE4_1() local [all …]
|
H A D | convolve_sse4.cc | 43 __m128i v_src[4]; in SumHorizontalTaps() local 484 const __m128i v_src = LoadUnaligned16(&src[x]); in ConvolveCompoundCopy_SSE4_1() local 502 const __m128i v_src = LoadLo8(&src[0]); in ConvolveCompoundCopy_SSE4_1() local 514 const __m128i v_src = _mm_unpacklo_epi32(v_src0, v_src1); in ConvolveCompoundCopy_SSE4_1() local
|
H A D | cdef_sse4.cc | 206 LIBGAV1_ALWAYS_INLINE void AddPartial_D5_D7(__m128i* v_src, __m128i* partial_lo, in AddPartial_D5_D7() 256 __m128i v_src[8]; in AddPartial() local
|
H A D | cdef_avx2.cc | 228 LIBGAV1_ALWAYS_INLINE void AddPartial_D7_D5(__m256i* v_src, __m256i* partial_lo, in AddPartial_D7_D5() 283 __m256i v_src[8]; in AddPartial() local
|
H A D | convolve_avx2.cc | 77 __m256i v_src[4]; in SumHorizontalTaps() local 724 __m256i v_src[4]; in SumVerticalTaps() local
|
/aosp_15_r20/external/libgav1/src/dsp/arm/ |
H A D | inverse_transform_neon.cc | 465 const int16x8_t v_src = vdupq_n_s16(dst[0]); in DctDcOnly() local 496 const int16x4_t v_src = vld1_s16(dst); in DctDcOnlyColumn() local 502 const int16x8_t v_src = vld1q_s16(&dst[i]); in DctDcOnlyColumn() local 1252 const int16x4_t v_src = vbsl_s16(v_mask, v_src_round, v_src0); in Adst4DcOnly() local 1279 const int16x4_t v_src = vld1_s16(&dst[i]); in Adst4DcOnlyColumn() local 1399 const int16x8_t v_src = vdupq_n_s16(dst[0]); in Adst8DcOnly() local 1455 const int16x8_t v_src = vld1q_s16(dst); in Adst8DcOnlyColumn() local 1719 const int16x8_t v_src = vdupq_n_s16(dst[0]); in Adst16DcOnly() local 1747 const int16x8_t v_src = vld1q_s16(dst); in Adst16DcOnlyColumn() local 1776 const int16x8_t v_src = vld1q_s16(&dst[i * step]); in Identity4_NEON() local [all …]
|
H A D | inverse_transform_10bit_neon.cc | 198 const int32x4_t v_src = vdupq_n_s32(dst[0]); in DctDcOnly() local 230 const int32x4_t v_src = vld1q_s32(dst); in DctDcOnlyColumn() local 236 const int32x4_t v_src = vld1q_s32(&dst[i]); in DctDcOnlyColumn() local 1016 const int32x4_t v_src = vbslq_s32(v_mask, v_src0_round, v_src0); in Adst4DcOnly() local 1044 const int32x4_t v_src = vld1q_s32(&dst[i]); in Adst4DcOnlyColumn() local 1155 const int32x4_t v_src = vdupq_n_s32(dst[0]); in Adst8DcOnly() local 1211 const int32x4_t v_src = vld1q_s32(dst); in Adst8DcOnlyColumn() local 1449 const int32x4_t v_src = vdupq_n_s32(dst[0]); in Adst16DcOnly() local 1477 const int32x4_t v_src = vld1q_s32(dst); in Adst16DcOnlyColumn() local 1502 const int32x4_t v_src = vld1q_s32(&dst[i * step]); in Identity4_NEON() local [all …]
|
H A D | cdef_neon.cc | 86 LIBGAV1_ALWAYS_INLINE void AddPartial_D0_D4(uint8x8_t* v_src, in AddPartial_D0_D4() 142 LIBGAV1_ALWAYS_INLINE void AddPartial_D1_D3(uint8x8_t* v_src, in AddPartial_D1_D3() 196 LIBGAV1_ALWAYS_INLINE void AddPartial_D5_D7(uint8x8_t* v_src, in AddPartial_D5_D7() 251 uint8x8_t v_src[8]; in AddPartial() local
|
H A D | convolve_neon.cc | 120 uint8x8_t v_src[8]; in FilterHorizontalWidth8AndUp() local 181 uint8x8_t v_src[8]; in FilterHorizontalWidth8AndUp() local 234 uint8x8_t v_src[4]; in FilterHorizontalWidth4() local 2377 const uint8x16_t v_src = vld1q_u8(&src[x]); in ConvolveCompoundCopy_NEON() local 2393 const uint8x8_t v_src = vld1_u8(&src[0]); in ConvolveCompoundCopy_NEON() local 2400 uint8x8_t v_src = vdup_n_u8(0); in ConvolveCompoundCopy_NEON() local
|
H A D | convolve_10bit_neon.cc | 163 uint16x8_t v_src[8]; in FilterHorizontalWidth8AndUp() local 214 uint16x8_t v_src[8]; in FilterHorizontalWidth8AndUp() local 290 uint16x4_t v_src[4]; in FilterHorizontalWidth4() local
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/vulkan/ops/ |
H A D | Copy.cpp | 85 void transfer_vulkan_to_cpu(vTensor& v_src, Tensor& dst) { in transfer_vulkan_to_cpu() 247 vTensor& v_src = convert(src); in copy_() local 257 vTensor& v_src = convert(src); in copy_() local 293 at::Tensor from_vulkan(vTensor& v_src) { in from_vulkan()
|
H A D | Utils.cpp | 167 vTensor& v_src, in copy_vtensor_to_buffer()
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/vulkan/impl/ |
H A D | Packing.cpp | 69 api::ShaderInfo get_image_to_nchw_shader(const vTensor& v_src) { in get_image_to_nchw_shader() 167 vTensor& v_src, in record_image_to_nchw_op() 266 vTensor& v_src, in record_buffer_to_nchw_op()
|
/aosp_15_r20/external/executorch/backends/vulkan/test/utils/ |
H A D | test_utils.cpp | 50 api::vTensor& v_src, in record_buffer_to_nchw_op() 94 api::vTensor& v_src, in record_image_to_nchw_op() 114 api::vTensor& v_src, in record_bitw8_image_to_nchw_nobitw8buffer_op()
|
/aosp_15_r20/external/libaom/aom_dsp/arm/ |
H A D | subtract_neon.c | 54 const uint8x16_t v_src = vld1q_u8(&src[0]); in aom_subtract_block_neon() local 69 const uint8x8_t v_src = vld1_u8(&src[0]); in aom_subtract_block_neon() local
|
/aosp_15_r20/hardware/libhardware/modules/camera/3_4/arc/ |
H A D | image_processor.cpp | 319 const uint8_t* v_src = src + width * height * 5 / 4; in YU12ToYV12() local 337 const uint8_t* v_src = src + width * height * 5 / 4; in YU12ToNV21() local
|
/aosp_15_r20/external/libvpx/test/ |
H A D | yuv_temporal_filter_test.cc | 119 const Buffer<PixelType> &u_src, const Buffer<PixelType> &v_src, in ApplyReferenceFilter() 345 Buffer<PixelType> v_src = Buffer<PixelType>(uv_width, uv_height, 0); in CompareTestWithParam() local 466 Buffer<PixelType> v_src = Buffer<PixelType>(uv_width, uv_height, 0); in RunTestFilterWithParam() local
|
/aosp_15_r20/external/libaom/av1/common/arm/ |
H A D | warp_plane_neon.h | 246 const int16x8_t *v_src = tmp + (k + 4); in warp_affine_vertical() local 296 const int16x8_t *v_src = tmp + (k + 4); in warp_affine_vertical() local
|
/aosp_15_r20/external/executorch/backends/vulkan/runtime/graph/ops/utils/ |
H A D | StagingUtils.cpp | 51 const api::vTensor& v_src, in get_tensor_to_nchw_shader()
|
/aosp_15_r20/external/mesa3d/src/gallium/auxiliary/util/ |
H A D | u_video.h | 148 uint8_t *v_src = (uint8_t *)source_data[1] + source_pitches[1] * dst_field; in u_copy_nv12_from_yv12() local
|
/aosp_15_r20/external/libvpx/vp8/common/ |
H A D | mfqe.c | 67 unsigned char *u_src, unsigned char *v_src, in apply_ifactor()
|
/aosp_15_r20/external/libvpx/vp8/decoder/ |
H A D | decodeframe.c | 360 unsigned char *v_src) { in yv12_extend_frame_left_right_c()
|
/aosp_15_r20/external/libyuv/source/ |
H A D | scale_rvv.cc | 36 vuint8m4_t v_src = __riscv_vle8_v_u8m4(src_ptr, vl); in ScaleAddRow_RVV() local 204 vuint16m8_t v_src = __riscv_vle16_v_u16m8(src, vl); in ScaleRowDown2_RVV() local
|
/aosp_15_r20/external/libvpx/vp9/encoder/arm/neon/ |
H A D | vp9_highbd_temporal_filter_neon.c | 794 int y_pre_stride, const uint16_t *u_src, const uint16_t *v_src, in vp9_highbd_apply_temporal_filter_neon()
|
H A D | vp9_temporal_filter_neon.c | 771 int y_pre_stride, const uint8_t *u_src, const uint8_t *v_src, in vp9_apply_temporal_filter_neon()
|