forked from KolibriOS/kolibrios
a4b787f4b8
git-svn-id: svn://kolibrios.org@6147 a494cfbc-eb01-0410-851d-a64ba20cac60
251 lines
8.9 KiB
C
251 lines
8.9 KiB
C
/*
|
|
* Copyright (c) 2015 Manojkumar Bhosale (Manojkumar.Bhosale@imgtec.com)
|
|
*
|
|
* This file is part of FFmpeg.
|
|
*
|
|
* FFmpeg is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* FFmpeg is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with FFmpeg; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
|
|
#include "libavutil/mips/generic_macros_msa.h"
|
|
#include "h263dsp_mips.h"
|
|
|
|
static void h263_dct_unquantize_msa(int16_t *block, int16_t qmul,
|
|
int16_t qadd, int8_t n_coeffs,
|
|
uint8_t loop_start)
|
|
{
|
|
int16_t *block_dup = block;
|
|
int32_t level, cnt;
|
|
v8i16 block_vec, qmul_vec, qadd_vec, sub;
|
|
v8i16 add, mask, mul, zero_mask;
|
|
|
|
qmul_vec = __msa_fill_h(qmul);
|
|
qadd_vec = __msa_fill_h(qadd);
|
|
for (cnt = 0; cnt < (n_coeffs >> 3); cnt++) {
|
|
block_vec = LD_SH(block_dup + loop_start);
|
|
mask = __msa_clti_s_h(block_vec, 0);
|
|
zero_mask = __msa_ceqi_h(block_vec, 0);
|
|
mul = block_vec * qmul_vec;
|
|
sub = mul - qadd_vec;
|
|
add = mul + qadd_vec;
|
|
add = (v8i16) __msa_bmnz_v((v16u8) add, (v16u8) sub, (v16u8) mask);
|
|
block_vec = (v8i16) __msa_bmnz_v((v16u8) add, (v16u8) block_vec,
|
|
(v16u8) zero_mask);
|
|
ST_SH(block_vec, block_dup + loop_start);
|
|
block_dup += 8;
|
|
}
|
|
|
|
cnt = ((n_coeffs >> 3) * 8) + loop_start;
|
|
|
|
for (; cnt <= n_coeffs; cnt++) {
|
|
level = block[cnt];
|
|
if (level) {
|
|
if (level < 0) {
|
|
level = level * qmul - qadd;
|
|
} else {
|
|
level = level * qmul + qadd;
|
|
}
|
|
block[cnt] = level;
|
|
}
|
|
}
|
|
}
|
|
|
|
static int32_t mpeg2_dct_unquantize_inter_msa(int16_t *block,
|
|
int32_t qscale,
|
|
const int16_t *quant_matrix)
|
|
{
|
|
int32_t cnt, sum_res = -1;
|
|
v8i16 block_vec, block_neg, qscale_vec, mask;
|
|
v8i16 block_org0, block_org1, block_org2, block_org3;
|
|
v8i16 quant_m0, quant_m1, quant_m2, quant_m3;
|
|
v8i16 sum, mul, zero_mask;
|
|
v4i32 mul_vec, qscale_l, qscale_r, quant_m_r, quant_m_l;
|
|
v4i32 block_l, block_r, sad;
|
|
|
|
qscale_vec = __msa_fill_h(qscale);
|
|
for (cnt = 0; cnt < 2; cnt++) {
|
|
LD_SH4(block, 8, block_org0, block_org1, block_org2, block_org3);
|
|
LD_SH4(quant_matrix, 8, quant_m0, quant_m1, quant_m2, quant_m3);
|
|
mask = __msa_clti_s_h(block_org0, 0);
|
|
zero_mask = __msa_ceqi_h(block_org0, 0);
|
|
block_neg = -block_org0;
|
|
block_vec = (v8i16) __msa_bmnz_v((v16u8) block_org0, (v16u8) block_neg,
|
|
(v16u8) mask);
|
|
block_vec <<= 1;
|
|
block_vec += 1;
|
|
UNPCK_SH_SW(block_vec, block_r, block_l);
|
|
UNPCK_SH_SW(qscale_vec, qscale_r, qscale_l);
|
|
UNPCK_SH_SW(quant_m0, quant_m_r, quant_m_l);
|
|
mul_vec = block_l * qscale_l;
|
|
mul_vec *= quant_m_l;
|
|
block_l = mul_vec >> 4;
|
|
mul_vec = block_r * qscale_r;
|
|
mul_vec *= quant_m_r;
|
|
block_r = mul_vec >> 4;
|
|
mul = (v8i16) __msa_pckev_h((v8i16) block_l, (v8i16) block_r);
|
|
block_neg = - mul;
|
|
sum = (v8i16) __msa_bmnz_v((v16u8) mul, (v16u8) block_neg,
|
|
(v16u8) mask);
|
|
sum = (v8i16) __msa_bmnz_v((v16u8) sum, (v16u8) block_org0,
|
|
(v16u8) zero_mask);
|
|
ST_SH(sum, block);
|
|
block += 8;
|
|
quant_matrix += 8;
|
|
sad = __msa_hadd_s_w(sum, sum);
|
|
sum_res += HADD_SW_S32(sad);
|
|
mask = __msa_clti_s_h(block_org1, 0);
|
|
zero_mask = __msa_ceqi_h(block_org1, 0);
|
|
block_neg = - block_org1;
|
|
block_vec = (v8i16) __msa_bmnz_v((v16u8) block_org1, (v16u8) block_neg,
|
|
(v16u8) mask);
|
|
block_vec <<= 1;
|
|
block_vec += 1;
|
|
UNPCK_SH_SW(block_vec, block_r, block_l);
|
|
UNPCK_SH_SW(qscale_vec, qscale_r, qscale_l);
|
|
UNPCK_SH_SW(quant_m1, quant_m_r, quant_m_l);
|
|
mul_vec = block_l * qscale_l;
|
|
mul_vec *= quant_m_l;
|
|
block_l = mul_vec >> 4;
|
|
mul_vec = block_r * qscale_r;
|
|
mul_vec *= quant_m_r;
|
|
block_r = mul_vec >> 4;
|
|
mul = __msa_pckev_h((v8i16) block_l, (v8i16) block_r);
|
|
block_neg = - mul;
|
|
sum = (v8i16) __msa_bmnz_v((v16u8) mul, (v16u8) block_neg,
|
|
(v16u8) mask);
|
|
sum = (v8i16) __msa_bmnz_v((v16u8) sum, (v16u8) block_org1,
|
|
(v16u8) zero_mask);
|
|
ST_SH(sum, block);
|
|
|
|
block += 8;
|
|
quant_matrix += 8;
|
|
sad = __msa_hadd_s_w(sum, sum);
|
|
sum_res += HADD_SW_S32(sad);
|
|
mask = __msa_clti_s_h(block_org2, 0);
|
|
zero_mask = __msa_ceqi_h(block_org2, 0);
|
|
block_neg = - block_org2;
|
|
block_vec = (v8i16) __msa_bmnz_v((v16u8) block_org2, (v16u8) block_neg,
|
|
(v16u8) mask);
|
|
block_vec <<= 1;
|
|
block_vec += 1;
|
|
UNPCK_SH_SW(block_vec, block_r, block_l);
|
|
UNPCK_SH_SW(qscale_vec, qscale_r, qscale_l);
|
|
UNPCK_SH_SW(quant_m2, quant_m_r, quant_m_l);
|
|
mul_vec = block_l * qscale_l;
|
|
mul_vec *= quant_m_l;
|
|
block_l = mul_vec >> 4;
|
|
mul_vec = block_r * qscale_r;
|
|
mul_vec *= quant_m_r;
|
|
block_r = mul_vec >> 4;
|
|
mul = __msa_pckev_h((v8i16) block_l, (v8i16) block_r);
|
|
block_neg = - mul;
|
|
sum = (v8i16) __msa_bmnz_v((v16u8) mul, (v16u8) block_neg,
|
|
(v16u8) mask);
|
|
sum = (v8i16) __msa_bmnz_v((v16u8) sum, (v16u8) block_org2,
|
|
(v16u8) zero_mask);
|
|
ST_SH(sum, block);
|
|
|
|
block += 8;
|
|
quant_matrix += 8;
|
|
sad = __msa_hadd_s_w(sum, sum);
|
|
sum_res += HADD_SW_S32(sad);
|
|
mask = __msa_clti_s_h(block_org3, 0);
|
|
zero_mask = __msa_ceqi_h(block_org3, 0);
|
|
block_neg = - block_org3;
|
|
block_vec = (v8i16) __msa_bmnz_v((v16u8) block_org3, (v16u8) block_neg,
|
|
(v16u8) mask);
|
|
block_vec <<= 1;
|
|
block_vec += 1;
|
|
UNPCK_SH_SW(block_vec, block_r, block_l);
|
|
UNPCK_SH_SW(qscale_vec, qscale_r, qscale_l);
|
|
UNPCK_SH_SW(quant_m3, quant_m_r, quant_m_l);
|
|
mul_vec = block_l * qscale_l;
|
|
mul_vec *= quant_m_l;
|
|
block_l = mul_vec >> 4;
|
|
mul_vec = block_r * qscale_r;
|
|
mul_vec *= quant_m_r;
|
|
block_r = mul_vec >> 4;
|
|
mul = __msa_pckev_h((v8i16) block_l, (v8i16) block_r);
|
|
block_neg = - mul;
|
|
sum = (v8i16) __msa_bmnz_v((v16u8) mul, (v16u8) block_neg,
|
|
(v16u8) mask);
|
|
sum = (v8i16) __msa_bmnz_v((v16u8) sum, (v16u8) block_org3,
|
|
(v16u8) zero_mask);
|
|
ST_SH(sum, block);
|
|
|
|
block += 8;
|
|
quant_matrix += 8;
|
|
sad = __msa_hadd_s_w(sum, sum);
|
|
sum_res += HADD_SW_S32(sad);
|
|
}
|
|
|
|
return sum_res;
|
|
}
|
|
|
|
void ff_dct_unquantize_h263_intra_msa(MpegEncContext *s,
|
|
int16_t *block, int32_t index,
|
|
int32_t qscale)
|
|
{
|
|
int32_t qmul, qadd;
|
|
int32_t nCoeffs;
|
|
|
|
av_assert2(s->block_last_index[index] >= 0 || s->h263_aic);
|
|
|
|
qmul = qscale << 1;
|
|
|
|
if (!s->h263_aic) {
|
|
block[0] *= index < 4 ? s->y_dc_scale : s->c_dc_scale;
|
|
qadd = (qscale - 1) | 1;
|
|
} else {
|
|
qadd = 0;
|
|
}
|
|
if (s->ac_pred)
|
|
nCoeffs = 63;
|
|
else
|
|
nCoeffs = s->inter_scantable.raster_end[s->block_last_index[index]];
|
|
|
|
h263_dct_unquantize_msa(block, qmul, qadd, nCoeffs, 1);
|
|
}
|
|
|
|
void ff_dct_unquantize_h263_inter_msa(MpegEncContext *s,
|
|
int16_t *block, int32_t index,
|
|
int32_t qscale)
|
|
{
|
|
int32_t qmul, qadd;
|
|
int32_t nCoeffs;
|
|
|
|
av_assert2(s->block_last_index[index] >= 0);
|
|
|
|
qadd = (qscale - 1) | 1;
|
|
qmul = qscale << 1;
|
|
|
|
nCoeffs = s->inter_scantable.raster_end[s->block_last_index[index]];
|
|
|
|
h263_dct_unquantize_msa(block, qmul, qadd, nCoeffs, 0);
|
|
}
|
|
|
|
void ff_dct_unquantize_mpeg2_inter_msa(MpegEncContext *s,
|
|
int16_t *block, int32_t index,
|
|
int32_t qscale)
|
|
{
|
|
const uint16_t *quant_matrix;
|
|
int32_t sum = -1;
|
|
|
|
quant_matrix = s->inter_matrix;
|
|
|
|
sum = mpeg2_dct_unquantize_inter_msa(block, qscale, quant_matrix);
|
|
|
|
block[63] ^= sum & 1;
|
|
}
|