123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892 |
- /*
- * Copyright 2020 The LibYuv Project Authors. All rights reserved.
- *
- * Use of this source code is governed by a BSD-style license
- * that can be found in the LICENSE file in the root of the source
- * tree. An additional intellectual property rights grant can be found
- * in the file PATENTS. All contributing project authors may
- * be found in the AUTHORS file in the root of the source tree.
- */
- #include "libyuv/scale.h"
- #include <assert.h>
- #include <string.h>
- #include "libyuv/cpu_id.h"
- #include "libyuv/planar_functions.h" // For CopyUV
- #include "libyuv/row.h"
- #include "libyuv/scale_row.h"
- #ifdef __cplusplus
- namespace libyuv {
- extern "C" {
- #endif
- // Macros to enable specialized scalers
- #ifndef HAS_SCALEUVDOWN2
- #define HAS_SCALEUVDOWN2 1
- #endif
- #ifndef HAS_SCALEUVDOWN4BOX
- #define HAS_SCALEUVDOWN4BOX 1
- #endif
- #ifndef HAS_SCALEUVDOWNEVEN
- #define HAS_SCALEUVDOWNEVEN 1
- #endif
- #ifndef HAS_SCALEUVBILINEARDOWN
- #define HAS_SCALEUVBILINEARDOWN 1
- #endif
- #ifndef HAS_SCALEUVBILINEARUP
- #define HAS_SCALEUVBILINEARUP 1
- #endif
- #ifndef HAS_UVCOPY
- #define HAS_UVCOPY 1
- #endif
- #ifndef HAS_SCALEPLANEVERTICAL
- #define HAS_SCALEPLANEVERTICAL 1
- #endif
- static __inline int Abs(int v) {
- return v >= 0 ? v : -v;
- }
- // ScaleUV, 1/2
- // This is an optimized version for scaling down a UV to 1/2 of
- // its original size.
- #if HAS_SCALEUVDOWN2
- static void ScaleUVDown2(int src_width,
- int src_height,
- int dst_width,
- int dst_height,
- int src_stride,
- int dst_stride,
- const uint8_t* src_uv,
- uint8_t* dst_uv,
- int x,
- int dx,
- int y,
- int dy,
- enum FilterMode filtering) {
- int j;
- int row_stride = src_stride * (dy >> 16);
- void (*ScaleUVRowDown2)(const uint8_t* src_uv, ptrdiff_t src_stride,
- uint8_t* dst_uv, int dst_width) =
- filtering == kFilterNone
- ? ScaleUVRowDown2_C
- : (filtering == kFilterLinear ? ScaleUVRowDown2Linear_C
- : ScaleUVRowDown2Box_C);
- (void)src_width;
- (void)src_height;
- (void)dx;
- assert(dx == 65536 * 2); // Test scale factor of 2.
- assert((dy & 0x1ffff) == 0); // Test vertical scale is multiple of 2.
- // Advance to odd row, even column.
- if (filtering == kFilterBilinear) {
- src_uv += (y >> 16) * src_stride + (x >> 16) * 2;
- } else {
- src_uv += (y >> 16) * src_stride + ((x >> 16) - 1) * 2;
- }
- #if defined(HAS_SCALEUVROWDOWN2BOX_SSSE3)
- if (TestCpuFlag(kCpuHasSSSE3) && filtering) {
- ScaleUVRowDown2 = ScaleUVRowDown2Box_Any_SSSE3;
- if (IS_ALIGNED(dst_width, 4)) {
- ScaleUVRowDown2 = ScaleUVRowDown2Box_SSSE3;
- }
- }
- #endif
- #if defined(HAS_SCALEUVROWDOWN2BOX_AVX2)
- if (TestCpuFlag(kCpuHasAVX2) && filtering) {
- ScaleUVRowDown2 = ScaleUVRowDown2Box_Any_AVX2;
- if (IS_ALIGNED(dst_width, 8)) {
- ScaleUVRowDown2 = ScaleUVRowDown2Box_AVX2;
- }
- }
- #endif
- #if defined(HAS_SCALEUVROWDOWN2BOX_NEON)
- if (TestCpuFlag(kCpuHasNEON) && filtering) {
- ScaleUVRowDown2 = ScaleUVRowDown2Box_Any_NEON;
- if (IS_ALIGNED(dst_width, 8)) {
- ScaleUVRowDown2 = ScaleUVRowDown2Box_NEON;
- }
- }
- #endif
- // This code is not enabled. Only box filter is available at this time.
- #if defined(HAS_SCALEUVROWDOWN2_SSSE3)
- if (TestCpuFlag(kCpuHasSSSE3)) {
- ScaleUVRowDown2 =
- filtering == kFilterNone
- ? ScaleUVRowDown2_Any_SSSE3
- : (filtering == kFilterLinear ? ScaleUVRowDown2Linear_Any_SSSE3
- : ScaleUVRowDown2Box_Any_SSSE3);
- if (IS_ALIGNED(dst_width, 2)) {
- ScaleUVRowDown2 =
- filtering == kFilterNone
- ? ScaleUVRowDown2_SSSE3
- : (filtering == kFilterLinear ? ScaleUVRowDown2Linear_SSSE3
- : ScaleUVRowDown2Box_SSSE3);
- }
- }
- #endif
- // This code is not enabled. Only box filter is available at this time.
- #if defined(HAS_SCALEUVROWDOWN2_NEON)
- if (TestCpuFlag(kCpuHasNEON)) {
- ScaleUVRowDown2 =
- filtering == kFilterNone
- ? ScaleUVRowDown2_Any_NEON
- : (filtering == kFilterLinear ? ScaleUVRowDown2Linear_Any_NEON
- : ScaleUVRowDown2Box_Any_NEON);
- if (IS_ALIGNED(dst_width, 8)) {
- ScaleUVRowDown2 =
- filtering == kFilterNone
- ? ScaleUVRowDown2_NEON
- : (filtering == kFilterLinear ? ScaleUVRowDown2Linear_NEON
- : ScaleUVRowDown2Box_NEON);
- }
- }
- #endif
- #if defined(HAS_SCALEUVROWDOWN2_MMI)
- if (TestCpuFlag(kCpuHasMMI)) {
- ScaleUVRowDown2 =
- filtering == kFilterNone
- ? ScaleUVRowDown2_Any_MMI
- : (filtering == kFilterLinear ? ScaleUVRowDown2Linear_Any_MMI
- : ScaleUVRowDown2Box_Any_MMI);
- if (IS_ALIGNED(dst_width, 2)) {
- ScaleUVRowDown2 =
- filtering == kFilterNone
- ? ScaleUVRowDown2_MMI
- : (filtering == kFilterLinear ? ScaleUVRowDown2Linear_MMI
- : ScaleUVRowDown2Box_MMI);
- }
- }
- #endif
- #if defined(HAS_SCALEUVROWDOWN2_MSA)
- if (TestCpuFlag(kCpuHasMSA)) {
- ScaleUVRowDown2 =
- filtering == kFilterNone
- ? ScaleUVRowDown2_Any_MSA
- : (filtering == kFilterLinear ? ScaleUVRowDown2Linear_Any_MSA
- : ScaleUVRowDown2Box_Any_MSA);
- if (IS_ALIGNED(dst_width, 2)) {
- ScaleUVRowDown2 =
- filtering == kFilterNone
- ? ScaleUVRowDown2_MSA
- : (filtering == kFilterLinear ? ScaleUVRowDown2Linear_MSA
- : ScaleUVRowDown2Box_MSA);
- }
- }
- #endif
- if (filtering == kFilterLinear) {
- src_stride = 0;
- }
- for (j = 0; j < dst_height; ++j) {
- ScaleUVRowDown2(src_uv, src_stride, dst_uv, dst_width);
- src_uv += row_stride;
- dst_uv += dst_stride;
- }
- }
- #endif // HAS_SCALEUVDOWN2
- // ScaleUV, 1/4
- // This is an optimized version for scaling down a UV to 1/4 of
- // its original size.
- #if HAS_SCALEUVDOWN4BOX
- static void ScaleUVDown4Box(int src_width,
- int src_height,
- int dst_width,
- int dst_height,
- int src_stride,
- int dst_stride,
- const uint8_t* src_uv,
- uint8_t* dst_uv,
- int x,
- int dx,
- int y,
- int dy) {
- int j;
- // Allocate 2 rows of UV.
- const int kRowSize = (dst_width * 2 * 2 + 15) & ~15;
- align_buffer_64(row, kRowSize * 2);
- int row_stride = src_stride * (dy >> 16);
- void (*ScaleUVRowDown2)(const uint8_t* src_uv, ptrdiff_t src_stride,
- uint8_t* dst_uv, int dst_width) =
- ScaleUVRowDown2Box_C;
- // Advance to odd row, even column.
- src_uv += (y >> 16) * src_stride + (x >> 16) * 2;
- (void)src_width;
- (void)src_height;
- (void)dx;
- assert(dx == 65536 * 4); // Test scale factor of 4.
- assert((dy & 0x3ffff) == 0); // Test vertical scale is multiple of 4.
- #if defined(HAS_SCALEUVROWDOWN2BOX_SSSE3)
- if (TestCpuFlag(kCpuHasSSSE3)) {
- ScaleUVRowDown2 = ScaleUVRowDown2Box_Any_SSSE3;
- if (IS_ALIGNED(dst_width, 4)) {
- ScaleUVRowDown2 = ScaleUVRowDown2Box_SSSE3;
- }
- }
- #endif
- #if defined(HAS_SCALEUVROWDOWN2BOX_AVX2)
- if (TestCpuFlag(kCpuHasAVX2)) {
- ScaleUVRowDown2 = ScaleUVRowDown2Box_Any_AVX2;
- if (IS_ALIGNED(dst_width, 8)) {
- ScaleUVRowDown2 = ScaleUVRowDown2Box_AVX2;
- }
- }
- #endif
- #if defined(HAS_SCALEUVROWDOWN2BOX_NEON)
- if (TestCpuFlag(kCpuHasNEON)) {
- ScaleUVRowDown2 = ScaleUVRowDown2Box_Any_NEON;
- if (IS_ALIGNED(dst_width, 8)) {
- ScaleUVRowDown2 = ScaleUVRowDown2Box_NEON;
- }
- }
- #endif
- for (j = 0; j < dst_height; ++j) {
- ScaleUVRowDown2(src_uv, src_stride, row, dst_width * 2);
- ScaleUVRowDown2(src_uv + src_stride * 2, src_stride, row + kRowSize,
- dst_width * 2);
- ScaleUVRowDown2(row, kRowSize, dst_uv, dst_width);
- src_uv += row_stride;
- dst_uv += dst_stride;
- }
- free_aligned_buffer_64(row);
- }
- #endif // HAS_SCALEUVDOWN4BOX
- // ScaleUV Even
- // This is an optimized version for scaling down a UV to even
- // multiple of its original size.
- #if HAS_SCALEUVDOWNEVEN
- static void ScaleUVDownEven(int src_width,
- int src_height,
- int dst_width,
- int dst_height,
- int src_stride,
- int dst_stride,
- const uint8_t* src_uv,
- uint8_t* dst_uv,
- int x,
- int dx,
- int y,
- int dy,
- enum FilterMode filtering) {
- int j;
- int col_step = dx >> 16;
- int row_stride = (dy >> 16) * src_stride;
- void (*ScaleUVRowDownEven)(const uint8_t* src_uv, ptrdiff_t src_stride,
- int src_step, uint8_t* dst_uv, int dst_width) =
- filtering ? ScaleUVRowDownEvenBox_C : ScaleUVRowDownEven_C;
- (void)src_width;
- (void)src_height;
- assert(IS_ALIGNED(src_width, 2));
- assert(IS_ALIGNED(src_height, 2));
- src_uv += (y >> 16) * src_stride + (x >> 16) * 2;
- #if defined(HAS_SCALEUVROWDOWNEVEN_SSSE3)
- if (TestCpuFlag(kCpuHasSSSE3)) {
- ScaleUVRowDownEven = filtering ? ScaleUVRowDownEvenBox_Any_SSSE3
- : ScaleUVRowDownEven_Any_SSSE3;
- if (IS_ALIGNED(dst_width, 4)) {
- ScaleUVRowDownEven =
- filtering ? ScaleUVRowDownEvenBox_SSE2 : ScaleUVRowDownEven_SSSE3;
- }
- }
- #endif
- #if defined(HAS_SCALEUVROWDOWNEVEN_NEON)
- if (TestCpuFlag(kCpuHasNEON) && !filtering) {
- ScaleUVRowDownEven = ScaleUVRowDownEven_Any_NEON;
- if (IS_ALIGNED(dst_width, 4)) {
- ScaleUVRowDownEven = ScaleUVRowDownEven_NEON;
- }
- }
- #endif// TODO(fbarchard): Enable Box filter
- #if defined(HAS_SCALEUVROWDOWNEVENBOX_NEON)
- if (TestCpuFlag(kCpuHasNEON)) {
- ScaleUVRowDownEven = filtering ? ScaleUVRowDownEvenBox_Any_NEON
- : ScaleUVRowDownEven_Any_NEON;
- if (IS_ALIGNED(dst_width, 4)) {
- ScaleUVRowDownEven =
- filtering ? ScaleUVRowDownEvenBox_NEON : ScaleUVRowDownEven_NEON;
- }
- }
- #endif
- #if defined(HAS_SCALEUVROWDOWNEVEN_MMI)
- if (TestCpuFlag(kCpuHasMMI)) {
- ScaleUVRowDownEven =
- filtering ? ScaleUVRowDownEvenBox_Any_MMI : ScaleUVRowDownEven_Any_MMI;
- if (IS_ALIGNED(dst_width, 2)) {
- ScaleUVRowDownEven =
- filtering ? ScaleUVRowDownEvenBox_MMI : ScaleUVRowDownEven_MMI;
- }
- }
- #endif
- #if defined(HAS_SCALEUVROWDOWNEVEN_MSA)
- if (TestCpuFlag(kCpuHasMSA)) {
- ScaleUVRowDownEven =
- filtering ? ScaleUVRowDownEvenBox_Any_MSA : ScaleUVRowDownEven_Any_MSA;
- if (IS_ALIGNED(dst_width, 4)) {
- ScaleUVRowDownEven =
- filtering ? ScaleUVRowDownEvenBox_MSA : ScaleUVRowDownEven_MSA;
- }
- }
- #endif
- if (filtering == kFilterLinear) {
- src_stride = 0;
- }
- for (j = 0; j < dst_height; ++j) {
- ScaleUVRowDownEven(src_uv, src_stride, col_step, dst_uv, dst_width);
- src_uv += row_stride;
- dst_uv += dst_stride;
- }
- }
- #endif
- // Scale UV down with bilinear interpolation.
- #if HAS_SCALEUVBILINEARDOWN
- static void ScaleUVBilinearDown(int src_width,
- int src_height,
- int dst_width,
- int dst_height,
- int src_stride,
- int dst_stride,
- const uint8_t* src_uv,
- uint8_t* dst_uv,
- int x,
- int dx,
- int y,
- int dy,
- enum FilterMode filtering) {
- int j;
- void (*InterpolateRow)(uint8_t * dst_uv, const uint8_t* src_uv,
- ptrdiff_t src_stride, int dst_width,
- int source_y_fraction) = InterpolateRow_C;
- void (*ScaleUVFilterCols)(uint8_t * dst_uv, const uint8_t* src_uv,
- int dst_width, int x, int dx) =
- (src_width >= 32768) ? ScaleUVFilterCols64_C : ScaleUVFilterCols_C;
- int64_t xlast = x + (int64_t)(dst_width - 1) * dx;
- int64_t xl = (dx >= 0) ? x : xlast;
- int64_t xr = (dx >= 0) ? xlast : x;
- int clip_src_width;
- xl = (xl >> 16) & ~3; // Left edge aligned.
- xr = (xr >> 16) + 1; // Right most pixel used. Bilinear uses 2 pixels.
- xr = (xr + 1 + 3) & ~3; // 1 beyond 4 pixel aligned right most pixel.
- if (xr > src_width) {
- xr = src_width;
- }
- clip_src_width = (int)(xr - xl) * 2; // Width aligned to 2.
- src_uv += xl * 2;
- x -= (int)(xl << 16);
- #if defined(HAS_INTERPOLATEROW_SSSE3)
- if (TestCpuFlag(kCpuHasSSSE3)) {
- InterpolateRow = InterpolateRow_Any_SSSE3;
- if (IS_ALIGNED(clip_src_width, 16)) {
- InterpolateRow = InterpolateRow_SSSE3;
- }
- }
- #endif
- #if defined(HAS_INTERPOLATEROW_AVX2)
- if (TestCpuFlag(kCpuHasAVX2)) {
- InterpolateRow = InterpolateRow_Any_AVX2;
- if (IS_ALIGNED(clip_src_width, 32)) {
- InterpolateRow = InterpolateRow_AVX2;
- }
- }
- #endif
- #if defined(HAS_INTERPOLATEROW_NEON)
- if (TestCpuFlag(kCpuHasNEON)) {
- InterpolateRow = InterpolateRow_Any_NEON;
- if (IS_ALIGNED(clip_src_width, 16)) {
- InterpolateRow = InterpolateRow_NEON;
- }
- }
- #endif
- #if defined(HAS_INTERPOLATEROW_MSA)
- if (TestCpuFlag(kCpuHasMSA)) {
- InterpolateRow = InterpolateRow_Any_MSA;
- if (IS_ALIGNED(clip_src_width, 32)) {
- InterpolateRow = InterpolateRow_MSA;
- }
- }
- #endif
- #if defined(HAS_SCALEUVFILTERCOLS_SSSE3)
- if (TestCpuFlag(kCpuHasSSSE3) && src_width < 32768) {
- ScaleUVFilterCols = ScaleUVFilterCols_SSSE3;
- }
- #endif
- #if defined(HAS_SCALEUVFILTERCOLS_NEON)
- if (TestCpuFlag(kCpuHasNEON)) {
- ScaleUVFilterCols = ScaleUVFilterCols_Any_NEON;
- if (IS_ALIGNED(dst_width, 4)) {
- ScaleUVFilterCols = ScaleUVFilterCols_NEON;
- }
- }
- #endif
- #if defined(HAS_SCALEUVFILTERCOLS_MSA)
- if (TestCpuFlag(kCpuHasMSA)) {
- ScaleUVFilterCols = ScaleUVFilterCols_Any_MSA;
- if (IS_ALIGNED(dst_width, 8)) {
- ScaleUVFilterCols = ScaleUVFilterCols_MSA;
- }
- }
- #endif
- // TODO(fbarchard): Consider not allocating row buffer for kFilterLinear.
- // Allocate a row of UV.
- {
- align_buffer_64(row, clip_src_width * 2);
- const int max_y = (src_height - 1) << 16;
- if (y > max_y) {
- y = max_y;
- }
- for (j = 0; j < dst_height; ++j) {
- int yi = y >> 16;
- const uint8_t* src = src_uv + yi * src_stride;
- if (filtering == kFilterLinear) {
- ScaleUVFilterCols(dst_uv, src, dst_width, x, dx);
- } else {
- int yf = (y >> 8) & 255;
- InterpolateRow(row, src, src_stride, clip_src_width, yf);
- ScaleUVFilterCols(dst_uv, row, dst_width, x, dx);
- }
- dst_uv += dst_stride;
- y += dy;
- if (y > max_y) {
- y = max_y;
- }
- }
- free_aligned_buffer_64(row);
- }
- }
- #endif
- // Scale UV up with bilinear interpolation.
- #if HAS_SCALEUVBILINEARUP
- static void ScaleUVBilinearUp(int src_width,
- int src_height,
- int dst_width,
- int dst_height,
- int src_stride,
- int dst_stride,
- const uint8_t* src_uv,
- uint8_t* dst_uv,
- int x,
- int dx,
- int y,
- int dy,
- enum FilterMode filtering) {
- int j;
- void (*InterpolateRow)(uint8_t * dst_uv, const uint8_t* src_uv,
- ptrdiff_t src_stride, int dst_width,
- int source_y_fraction) = InterpolateRow_C;
- void (*ScaleUVFilterCols)(uint8_t * dst_uv, const uint8_t* src_uv,
- int dst_width, int x, int dx) =
- filtering ? ScaleUVFilterCols_C : ScaleUVCols_C;
- const int max_y = (src_height - 1) << 16;
- #if defined(HAS_INTERPOLATEROW_SSSE3)
- if (TestCpuFlag(kCpuHasSSSE3)) {
- InterpolateRow = InterpolateRow_Any_SSSE3;
- if (IS_ALIGNED(dst_width, 4)) {
- InterpolateRow = InterpolateRow_SSSE3;
- }
- }
- #endif
- #if defined(HAS_INTERPOLATEROW_AVX2)
- if (TestCpuFlag(kCpuHasAVX2)) {
- InterpolateRow = InterpolateRow_Any_AVX2;
- if (IS_ALIGNED(dst_width, 8)) {
- InterpolateRow = InterpolateRow_AVX2;
- }
- }
- #endif
- #if defined(HAS_INTERPOLATEROW_NEON)
- if (TestCpuFlag(kCpuHasNEON)) {
- InterpolateRow = InterpolateRow_Any_NEON;
- if (IS_ALIGNED(dst_width, 4)) {
- InterpolateRow = InterpolateRow_NEON;
- }
- }
- #endif
- #if defined(HAS_INTERPOLATEROW_MMI)
- if (TestCpuFlag(kCpuHasMMI)) {
- InterpolateRow = InterpolateRow_Any_MMI;
- if (IS_ALIGNED(dst_width, 2)) {
- InterpolateRow = InterpolateRow_MMI;
- }
- }
- #endif
- #if defined(HAS_INTERPOLATEROW_MSA)
- if (TestCpuFlag(kCpuHasMSA)) {
- InterpolateRow = InterpolateRow_Any_MSA;
- if (IS_ALIGNED(dst_width, 8)) {
- InterpolateRow = InterpolateRow_MSA;
- }
- }
- #endif
- if (src_width >= 32768) {
- ScaleUVFilterCols = filtering ? ScaleUVFilterCols64_C : ScaleUVCols64_C;
- }
- #if defined(HAS_SCALEUVFILTERCOLS_SSSE3)
- if (filtering && TestCpuFlag(kCpuHasSSSE3) && src_width < 32768) {
- ScaleUVFilterCols = ScaleUVFilterCols_SSSE3;
- }
- #endif
- #if defined(HAS_SCALEUVFILTERCOLS_NEON)
- if (filtering && TestCpuFlag(kCpuHasNEON)) {
- ScaleUVFilterCols = ScaleUVFilterCols_Any_NEON;
- if (IS_ALIGNED(dst_width, 4)) {
- ScaleUVFilterCols = ScaleUVFilterCols_NEON;
- }
- }
- #endif
- #if defined(HAS_SCALEUVFILTERCOLS_MSA)
- if (filtering && TestCpuFlag(kCpuHasMSA)) {
- ScaleUVFilterCols = ScaleUVFilterCols_Any_MSA;
- if (IS_ALIGNED(dst_width, 8)) {
- ScaleUVFilterCols = ScaleUVFilterCols_MSA;
- }
- }
- #endif
- #if defined(HAS_SCALEUVCOLS_SSSE3)
- if (!filtering && TestCpuFlag(kCpuHasSSSE3) && src_width < 32768) {
- ScaleUVFilterCols = ScaleUVCols_SSSE3;
- }
- #endif
- #if defined(HAS_SCALEUVCOLS_NEON)
- if (!filtering && TestCpuFlag(kCpuHasNEON)) {
- ScaleUVFilterCols = ScaleUVCols_Any_NEON;
- if (IS_ALIGNED(dst_width, 8)) {
- ScaleUVFilterCols = ScaleUVCols_NEON;
- }
- }
- #endif
- #if defined(HAS_SCALEUVCOLS_MMI)
- if (!filtering && TestCpuFlag(kCpuHasMMI)) {
- ScaleUVFilterCols = ScaleUVCols_Any_MMI;
- if (IS_ALIGNED(dst_width, 1)) {
- ScaleUVFilterCols = ScaleUVCols_MMI;
- }
- }
- #endif
- #if defined(HAS_SCALEUVCOLS_MSA)
- if (!filtering && TestCpuFlag(kCpuHasMSA)) {
- ScaleUVFilterCols = ScaleUVCols_Any_MSA;
- if (IS_ALIGNED(dst_width, 4)) {
- ScaleUVFilterCols = ScaleUVCols_MSA;
- }
- }
- #endif
- if (!filtering && src_width * 2 == dst_width && x < 0x8000) {
- ScaleUVFilterCols = ScaleUVColsUp2_C;
- #if defined(HAS_SCALEUVCOLSUP2_SSSE3)
- if (TestCpuFlag(kCpuHasSSSE3) && IS_ALIGNED(dst_width, 8)) {
- ScaleUVFilterCols = ScaleUVColsUp2_SSSE3;
- }
- #endif
- #if defined(HAS_SCALEUVCOLSUP2_MMI)
- if (TestCpuFlag(kCpuHasMMI) && IS_ALIGNED(dst_width, 4)) {
- ScaleUVFilterCols = ScaleUVColsUp2_MMI;
- }
- #endif
- }
- if (y > max_y) {
- y = max_y;
- }
- {
- int yi = y >> 16;
- const uint8_t* src = src_uv + yi * src_stride;
- // Allocate 2 rows of UV.
- const int kRowSize = (dst_width * 2 + 15) & ~15;
- align_buffer_64(row, kRowSize * 2);
- uint8_t* rowptr = row;
- int rowstride = kRowSize;
- int lasty = yi;
- ScaleUVFilterCols(rowptr, src, dst_width, x, dx);
- if (src_height > 1) {
- src += src_stride;
- }
- ScaleUVFilterCols(rowptr + rowstride, src, dst_width, x, dx);
- src += src_stride;
- for (j = 0; j < dst_height; ++j) {
- yi = y >> 16;
- if (yi != lasty) {
- if (y > max_y) {
- y = max_y;
- yi = y >> 16;
- src = src_uv + yi * src_stride;
- }
- if (yi != lasty) {
- ScaleUVFilterCols(rowptr, src, dst_width, x, dx);
- rowptr += rowstride;
- rowstride = -rowstride;
- lasty = yi;
- src += src_stride;
- }
- }
- if (filtering == kFilterLinear) {
- InterpolateRow(dst_uv, rowptr, 0, dst_width * 2, 0);
- } else {
- int yf = (y >> 8) & 255;
- InterpolateRow(dst_uv, rowptr, rowstride, dst_width * 2, yf);
- }
- dst_uv += dst_stride;
- y += dy;
- }
- free_aligned_buffer_64(row);
- }
- }
- #endif // HAS_SCALEUVBILINEARUP
- // Scale UV to/from any dimensions, without interpolation.
- // Fixed point math is used for performance: The upper 16 bits
- // of x and dx is the integer part of the source position and
- // the lower 16 bits are the fixed decimal part.
- static void ScaleUVSimple(int src_width,
- int src_height,
- int dst_width,
- int dst_height,
- int src_stride,
- int dst_stride,
- const uint8_t* src_uv,
- uint8_t* dst_uv,
- int x,
- int dx,
- int y,
- int dy) {
- int j;
- void (*ScaleUVCols)(uint8_t * dst_uv, const uint8_t* src_uv, int dst_width,
- int x, int dx) =
- (src_width >= 32768) ? ScaleUVCols64_C : ScaleUVCols_C;
- (void)src_height;
- #if defined(HAS_SCALEUVCOLS_SSSE3)
- if (TestCpuFlag(kCpuHasSSSE3) && src_width < 32768) {
- ScaleUVCols = ScaleUVCols_SSSE3;
- }
- #endif
- #if defined(HAS_SCALEUVCOLS_NEON)
- if (TestCpuFlag(kCpuHasNEON)) {
- ScaleUVCols = ScaleUVCols_Any_NEON;
- if (IS_ALIGNED(dst_width, 8)) {
- ScaleUVCols = ScaleUVCols_NEON;
- }
- }
- #endif
- #if defined(HAS_SCALEUVCOLS_MMI)
- if (TestCpuFlag(kCpuHasMMI)) {
- ScaleUVCols = ScaleUVCols_Any_MMI;
- if (IS_ALIGNED(dst_width, 1)) {
- ScaleUVCols = ScaleUVCols_MMI;
- }
- }
- #endif
- #if defined(HAS_SCALEUVCOLS_MSA)
- if (TestCpuFlag(kCpuHasMSA)) {
- ScaleUVCols = ScaleUVCols_Any_MSA;
- if (IS_ALIGNED(dst_width, 4)) {
- ScaleUVCols = ScaleUVCols_MSA;
- }
- }
- #endif
- if (src_width * 2 == dst_width && x < 0x8000) {
- ScaleUVCols = ScaleUVColsUp2_C;
- #if defined(HAS_SCALEUVCOLSUP2_SSSE3)
- if (TestCpuFlag(kCpuHasSSSE3) && IS_ALIGNED(dst_width, 8)) {
- ScaleUVCols = ScaleUVColsUp2_SSSE3;
- }
- #endif
- #if defined(HAS_SCALEUVCOLSUP2_MMI)
- if (TestCpuFlag(kCpuHasMMI) && IS_ALIGNED(dst_width, 4)) {
- ScaleUVCols = ScaleUVColsUp2_MMI;
- }
- #endif
- }
- for (j = 0; j < dst_height; ++j) {
- ScaleUVCols(dst_uv, src_uv + (y >> 16) * src_stride, dst_width, x, dx);
- dst_uv += dst_stride;
- y += dy;
- }
- }
- // Copy UV with optional flipping
- #if HAS_UVCOPY
- static int UVCopy(const uint8_t* src_UV,
- int src_stride_UV,
- uint8_t* dst_UV,
- int dst_stride_UV,
- int width,
- int height) {
- if (!src_UV || !dst_UV || width <= 0 || height == 0) {
- return -1;
- }
- // Negative height means invert the image.
- if (height < 0) {
- height = -height;
- src_UV = src_UV + (height - 1) * src_stride_UV;
- src_stride_UV = -src_stride_UV;
- }
- CopyPlane(src_UV, src_stride_UV, dst_UV, dst_stride_UV, width * 2, height);
- return 0;
- }
- #endif // HAS_UVCOPY
- // Scale a UV plane (from NV12)
- // This function in turn calls a scaling function
- // suitable for handling the desired resolutions.
- static void ScaleUV(const uint8_t* src,
- int src_stride,
- int src_width,
- int src_height,
- uint8_t* dst,
- int dst_stride,
- int dst_width,
- int dst_height,
- int clip_x,
- int clip_y,
- int clip_width,
- int clip_height,
- enum FilterMode filtering) {
- // Initial source x/y coordinate and step values as 16.16 fixed point.
- int x = 0;
- int y = 0;
- int dx = 0;
- int dy = 0;
- // UV does not support box filter yet, but allow the user to pass it.
- // Simplify filtering when possible.
- filtering = ScaleFilterReduce(src_width, src_height, dst_width, dst_height,
- filtering);
- // Negative src_height means invert the image.
- if (src_height < 0) {
- src_height = -src_height;
- src = src + (src_height - 1) * src_stride;
- src_stride = -src_stride;
- }
- ScaleSlope(src_width, src_height, dst_width, dst_height, filtering, &x, &y,
- &dx, &dy);
- src_width = Abs(src_width);
- if (clip_x) {
- int64_t clipf = (int64_t)(clip_x)*dx;
- x += (clipf & 0xffff);
- src += (clipf >> 16) * 2;
- dst += clip_x * 2;
- }
- if (clip_y) {
- int64_t clipf = (int64_t)(clip_y)*dy;
- y += (clipf & 0xffff);
- src += (clipf >> 16) * src_stride;
- dst += clip_y * dst_stride;
- }
- // Special case for integer step values.
- if (((dx | dy) & 0xffff) == 0) {
- if (!dx || !dy) { // 1 pixel wide and/or tall.
- filtering = kFilterNone;
- } else {
- // Optimized even scale down. ie 2, 4, 6, 8, 10x.
- if (!(dx & 0x10000) && !(dy & 0x10000)) {
- #if HAS_SCALEUVDOWN2
- if (dx == 0x20000) {
- // Optimized 1/2 downsample.
- ScaleUVDown2(src_width, src_height, clip_width, clip_height,
- src_stride, dst_stride, src, dst, x, dx, y, dy,
- filtering);
- return;
- }
- #endif
- #if HAS_SCALEUVDOWN4BOX
- if (dx == 0x40000 && filtering == kFilterBox) {
- // Optimized 1/4 box downsample.
- ScaleUVDown4Box(src_width, src_height, clip_width, clip_height,
- src_stride, dst_stride, src, dst, x, dx, y, dy);
- return;
- }
- #endif
- #if HAS_SCALEUVDOWNEVEN
- ScaleUVDownEven(src_width, src_height, clip_width, clip_height,
- src_stride, dst_stride, src, dst, x, dx, y, dy,
- filtering);
- return;
- #endif
- }
- // Optimized odd scale down. ie 3, 5, 7, 9x.
- if ((dx & 0x10000) && (dy & 0x10000)) {
- filtering = kFilterNone;
- #ifdef HAS_UVCOPY
- if (dx == 0x10000 && dy == 0x10000) {
- // Straight copy.
- UVCopy(src + (y >> 16) * src_stride + (x >> 16) * 2, src_stride, dst,
- dst_stride, clip_width, clip_height);
- return;
- }
- #endif
- }
- }
- }
- // HAS_SCALEPLANEVERTICAL
- if (dx == 0x10000 && (x & 0xffff) == 0) {
- // Arbitrary scale vertically, but unscaled horizontally.
- ScalePlaneVertical(src_height, clip_width, clip_height, src_stride,
- dst_stride, src, dst, x, y, dy, 4, filtering);
- return;
- }
- #if HAS_SCALEUVBILINEARUP
- if (filtering && dy < 65536) {
- ScaleUVBilinearUp(src_width, src_height, clip_width, clip_height,
- src_stride, dst_stride, src, dst, x, dx, y, dy,
- filtering);
- return;
- }
- #endif
- #if HAS_SCALEUVBILINEARDOWN
- if (filtering) {
- ScaleUVBilinearDown(src_width, src_height, clip_width, clip_height,
- src_stride, dst_stride, src, dst, x, dx, y, dy,
- filtering);
- return;
- }
- #endif
- ScaleUVSimple(src_width, src_height, clip_width, clip_height, src_stride,
- dst_stride, src, dst, x, dx, y, dy);
- }
- // Scale an UV image.
- LIBYUV_API
- int UVScale(const uint8_t* src_uv,
- int src_stride_uv,
- int src_width,
- int src_height,
- uint8_t* dst_uv,
- int dst_stride_uv,
- int dst_width,
- int dst_height,
- enum FilterMode filtering) {
- if (!src_uv || src_width == 0 || src_height == 0 || src_width > 32768 ||
- src_height > 32768 || !dst_uv || dst_width <= 0 || dst_height <= 0) {
- return -1;
- }
- ScaleUV(src_uv, src_stride_uv, src_width, src_height, dst_uv, dst_stride_uv,
- dst_width, dst_height, 0, 0, dst_width, dst_height, filtering);
- return 0;
- }
- #ifdef __cplusplus
- } // extern "C"
- } // namespace libyuv
- #endif
|