2 * Copyright (C) 2009 David Conrad
4 * This file is part of FFmpeg.
6 * FFmpeg is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU Lesser General Public
8 * License as published by the Free Software Foundation; either
9 * version 2.1 of the License, or (at your option) any later version.
11 * FFmpeg is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
14 * Lesser General Public License for more details.
16 * You should have received a copy of the GNU Lesser General Public
17 * License along with FFmpeg; if not, write to the Free Software
18 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
21 #include "libavcodec/dsputil.h"
22 #include "util_altivec.h"
23 #include "types_altivec.h"
25 static const vec_s16 constants
=
26 {0, 64277, 60547, 54491, 46341, 36410, 25080, 12785};
27 static const vec_u8 interleave_high
=
28 {0, 1, 16, 17, 4, 5, 20, 21, 8, 9, 24, 25, 12, 13, 28, 29};
31 vec_s16 A, B, C, D, Ad, Bd, Cd, Dd, E, F, G, H;\
32 vec_s16 Ed, Gd, Add, Bdd, Fd, Hd;\
33 vec_s16 eight = vec_splat_s16(8);\
34 vec_u16 four = vec_splat_u16(4);\
36 vec_s16 C1 = vec_splat(constants, 1);\
37 vec_s16 C2 = vec_splat(constants, 2);\
38 vec_s16 C3 = vec_splat(constants, 3);\
39 vec_s16 C4 = vec_splat(constants, 4);\
40 vec_s16 C5 = vec_splat(constants, 5);\
41 vec_s16 C6 = vec_splat(constants, 6);\
42 vec_s16 C7 = vec_splat(constants, 7);\
44 vec_s16 b0 = vec_ld(0x00, block);\
45 vec_s16 b1 = vec_ld(0x10, block);\
46 vec_s16 b2 = vec_ld(0x20, block);\
47 vec_s16 b3 = vec_ld(0x30, block);\
48 vec_s16 b4 = vec_ld(0x40, block);\
49 vec_s16 b5 = vec_ld(0x50, block);\
50 vec_s16 b6 = vec_ld(0x60, block);\
51 vec_s16 b7 = vec_ld(0x70, block);
53 // these functions do (a*C)>>16
54 // things are tricky because a is signed, but C unsigned.
55 // M15 is used if C fits in 15 bit unsigned (C6,C7)
56 // M16 is used if C requires 16 bits unsigned
57 static inline vec_s16
M15(vec_s16 a
, vec_s16 C
)
59 return (vec_s16
)vec_perm(vec_mule(a
,C
), vec_mulo(a
,C
), interleave_high
);
61 static inline vec_s16
M16(vec_s16 a
, vec_s16 C
)
63 return vec_add(a
, M15(a
, C
));
66 #define IDCT_1D(ADD, SHIFT)\
67 A = vec_add(M16(b1, C1), M15(b7, C7));\
68 B = vec_sub(M15(b1, C7), M16(b7, C1));\
69 C = vec_add(M16(b3, C3), M16(b5, C5));\
70 D = vec_sub(M16(b5, C3), M16(b3, C5));\
72 Ad = M16(vec_sub(A, C), C4);\
73 Bd = M16(vec_sub(B, D), C4);\
78 E = ADD(M16(vec_add(b0, b4), C4));\
79 F = ADD(M16(vec_sub(b0, b4), C4));\
81 G = vec_add(M16(b2, C2), M15(b6, C6));\
82 H = vec_sub(M15(b2, C6), M16(b6, C2));\
87 Add = vec_add(F, Ad);\
88 Bdd = vec_sub(Bd, H);\
93 b0 = SHIFT(vec_add(Gd, Cd));\
94 b7 = SHIFT(vec_sub(Gd, Cd));\
96 b1 = SHIFT(vec_add(Add, Hd));\
97 b2 = SHIFT(vec_sub(Add, Hd));\
99 b3 = SHIFT(vec_add(Ed, Dd));\
100 b4 = SHIFT(vec_sub(Ed, Dd));\
102 b5 = SHIFT(vec_add(Fd, Bdd));\
103 b6 = SHIFT(vec_sub(Fd, Bdd));
106 #define ADD8(a) vec_add(a, eight)
107 #define SHIFT4(a) vec_sra(a, four)
109 void ff_vp3_idct_altivec(DCTELEM block
[64])
114 TRANSPOSE8(b0
, b1
, b2
, b3
, b4
, b5
, b6
, b7
);
115 IDCT_1D(ADD8
, SHIFT4
)
117 vec_st(b0
, 0x00, block
);
118 vec_st(b1
, 0x10, block
);
119 vec_st(b2
, 0x20, block
);
120 vec_st(b3
, 0x30, block
);
121 vec_st(b4
, 0x40, block
);
122 vec_st(b5
, 0x50, block
);
123 vec_st(b6
, 0x60, block
);
124 vec_st(b7
, 0x70, block
);
127 void ff_vp3_idct_put_altivec(uint8_t *dst
, int stride
, DCTELEM block
[64])
132 // pixels are signed; so add 128*16 in addition to the normal 8
133 vec_s16 v2048
= vec_sl(vec_splat_s16(1), vec_splat_u16(11));
134 eight
= vec_add(eight
, v2048
);
137 TRANSPOSE8(b0
, b1
, b2
, b3
, b4
, b5
, b6
, b7
);
138 IDCT_1D(ADD8
, SHIFT4
)
141 t = vec_packsu(a, a);\
142 vec_ste((vec_u32)t, 0, (unsigned int *)dst);\
143 vec_ste((vec_u32)t, 4, (unsigned int *)dst);
145 PUT(b0
) dst
+= stride
;
146 PUT(b1
) dst
+= stride
;
147 PUT(b2
) dst
+= stride
;
148 PUT(b3
) dst
+= stride
;
149 PUT(b4
) dst
+= stride
;
150 PUT(b5
) dst
+= stride
;
151 PUT(b6
) dst
+= stride
;
155 void ff_vp3_idct_add_altivec(uint8_t *dst
, int stride
, DCTELEM block
[64])
160 vec_u8 vdst_mask
= vec_mergeh(vec_splat_u8(-1), vec_lvsl(0, dst
));
165 TRANSPOSE8(b0
, b1
, b2
, b3
, b4
, b5
, b6
, b7
);
166 IDCT_1D(ADD8
, SHIFT4
)
169 vdst = vec_ld(0, dst);\
170 vdst_16 = (vec_s16)vec_perm(vdst, zero_u8v, vdst_mask);\
171 vdst_16 = vec_adds(a, vdst_16);\
172 t = vec_packsu(vdst_16, vdst_16);\
173 vec_ste((vec_u32)t, 0, (unsigned int *)dst);\
174 vec_ste((vec_u32)t, 4, (unsigned int *)dst);
176 ADD(b0
) dst
+= stride
;
177 ADD(b1
) dst
+= stride
;
178 ADD(b2
) dst
+= stride
;
179 ADD(b3
) dst
+= stride
;
180 ADD(b4
) dst
+= stride
;
181 ADD(b5
) dst
+= stride
;
182 ADD(b6
) dst
+= stride
;