sync with en/mplayer.1 r30075
[mplayer/glamo.git] / tremor / misc.h
blob541019a5bd7db1ef0eafbc8c29357d6b532864af
1 /********************************************************************
2 * *
3 * THIS FILE IS PART OF THE OggVorbis 'TREMOR' CODEC SOURCE CODE. *
4 * *
5 * USE, DISTRIBUTION AND REPRODUCTION OF THIS LIBRARY SOURCE IS *
6 * GOVERNED BY A BSD-STYLE SOURCE LICENSE INCLUDED WITH THIS SOURCE *
7 * IN 'COPYING'. PLEASE READ THESE TERMS BEFORE DISTRIBUTING. *
8 * *
9 * THE OggVorbis 'TREMOR' SOURCE CODE IS (C) COPYRIGHT 1994-2002 *
10 * BY THE Xiph.Org FOUNDATION http://www.xiph.org/ *
11 * *
12 ********************************************************************
14 function: miscellaneous math and prototypes
16 ********************************************************************/
18 #ifndef _V_RANDOM_H_
19 #define _V_RANDOM_H_
20 #include "ivorbiscodec.h"
21 #include "codec_internal.h"
22 #include "os_types.h"
24 #include "asm_arm.h"
26 #ifndef _V_WIDE_MATH
27 #define _V_WIDE_MATH
29 #ifndef _LOW_ACCURACY_
30 /* 64 bit multiply */
32 #include <sys/types.h>
33 #include <stdlib.h>
34 #include "config.h"
36 #if !HAVE_BIGENDIAN
37 union magic {
38 struct {
39 ogg_int32_t lo;
40 ogg_int32_t hi;
41 } halves;
42 ogg_int64_t whole;
44 #else
45 union magic {
46 struct {
47 ogg_int32_t hi;
48 ogg_int32_t lo;
49 } halves;
50 ogg_int64_t whole;
52 #endif
54 static inline ogg_int32_t MULT32(ogg_int32_t x, ogg_int32_t y) {
55 union magic magic;
56 magic.whole = (ogg_int64_t)x * y;
57 return magic.halves.hi;
60 static inline ogg_int32_t MULT31(ogg_int32_t x, ogg_int32_t y) {
61 return MULT32(x,y)<<1;
64 static inline ogg_int32_t MULT31_SHIFT15(ogg_int32_t x, ogg_int32_t y) {
65 union magic magic;
66 magic.whole = (ogg_int64_t)x * y;
67 return ((ogg_uint32_t)(magic.halves.lo)>>15) | ((magic.halves.hi)<<17);
70 #else
71 /* 32 bit multiply, more portable but less accurate */
74 * Note: Precision is biased towards the first argument therefore ordering
75 * is important. Shift values were chosen for the best sound quality after
76 * many listening tests.
80 * For MULT32 and MULT31: The second argument is always a lookup table
81 * value already preshifted from 31 to 8 bits. We therefore take the
82 * opportunity to save on text space and use unsigned char for those
83 * tables in this case.
86 static inline ogg_int32_t MULT32(ogg_int32_t x, ogg_int32_t y) {
87 return (x >> 9) * y; /* y preshifted >>23 */
90 static inline ogg_int32_t MULT31(ogg_int32_t x, ogg_int32_t y) {
91 return (x >> 8) * y; /* y preshifted >>23 */
94 static inline ogg_int32_t MULT31_SHIFT15(ogg_int32_t x, ogg_int32_t y) {
95 return (x >> 6) * y; /* y preshifted >>9 */
98 #endif
101 * This should be used as a memory barrier, forcing all cached values in
102 * registers to wr writen back to memory. Might or might not be beneficial
103 * depending on the architecture and compiler.
105 #define MB()
108 * The XPROD functions are meant to optimize the cross products found all
109 * over the place in mdct.c by forcing memory operation ordering to avoid
110 * unnecessary register reloads as soon as memory is being written to.
111 * However this is only beneficial on CPUs with a sane number of general
112 * purpose registers which exclude the Intel x86. On Intel, better let the
113 * compiler actually reload registers directly from original memory by using
114 * macros.
117 #ifdef __i386__
119 #define XPROD32(_a, _b, _t, _v, _x, _y) \
120 { *(_x)=MULT32(_a,_t)+MULT32(_b,_v); \
121 *(_y)=MULT32(_b,_t)-MULT32(_a,_v); }
122 #define XPROD31(_a, _b, _t, _v, _x, _y) \
123 { *(_x)=MULT31(_a,_t)+MULT31(_b,_v); \
124 *(_y)=MULT31(_b,_t)-MULT31(_a,_v); }
125 #define XNPROD31(_a, _b, _t, _v, _x, _y) \
126 { *(_x)=MULT31(_a,_t)-MULT31(_b,_v); \
127 *(_y)=MULT31(_b,_t)+MULT31(_a,_v); }
129 #else
131 static inline void XPROD32(ogg_int32_t a, ogg_int32_t b,
132 ogg_int32_t t, ogg_int32_t v,
133 ogg_int32_t *x, ogg_int32_t *y)
135 *x = MULT32(a, t) + MULT32(b, v);
136 *y = MULT32(b, t) - MULT32(a, v);
139 static inline void XPROD31(ogg_int32_t a, ogg_int32_t b,
140 ogg_int32_t t, ogg_int32_t v,
141 ogg_int32_t *x, ogg_int32_t *y)
143 *x = MULT31(a, t) + MULT31(b, v);
144 *y = MULT31(b, t) - MULT31(a, v);
147 static inline void XNPROD31(ogg_int32_t a, ogg_int32_t b,
148 ogg_int32_t t, ogg_int32_t v,
149 ogg_int32_t *x, ogg_int32_t *y)
151 *x = MULT31(a, t) - MULT31(b, v);
152 *y = MULT31(b, t) + MULT31(a, v);
155 #endif
157 #endif
159 #ifndef _V_CLIP_MATH
160 #define _V_CLIP_MATH
162 static inline ogg_int32_t CLIP_TO_15(ogg_int32_t x) {
163 int ret=x;
164 ret-= ((x<=32767)-1)&(x-32767);
165 ret-= ((x>=-32768)-1)&(x+32768);
166 return(ret);
169 #endif
171 static inline ogg_int32_t VFLOAT_MULT(ogg_int32_t a,ogg_int32_t ap,
172 ogg_int32_t b,ogg_int32_t bp,
173 ogg_int32_t *p){
174 if(a && b){
175 #ifndef _LOW_ACCURACY_
176 *p=ap+bp+32;
177 return MULT32(a,b);
178 #else
179 *p=ap+bp+31;
180 return (a>>15)*(b>>16);
181 #endif
182 }else
183 return 0;
186 static inline ogg_int32_t VFLOAT_MULTI(ogg_int32_t a,ogg_int32_t ap,
187 ogg_int32_t i,
188 ogg_int32_t *p){
190 int ip=_ilog(abs(i))-31;
191 return VFLOAT_MULT(a,ap,i<<-ip,ip,p);
194 static inline ogg_int32_t VFLOAT_ADD(ogg_int32_t a,ogg_int32_t ap,
195 ogg_int32_t b,ogg_int32_t bp,
196 ogg_int32_t *p){
198 if(!a){
199 *p=bp;
200 return b;
201 }else if(!b){
202 *p=ap;
203 return a;
206 /* yes, this can leak a bit. */
207 if(ap>bp){
208 int shift=ap-bp+1;
209 *p=ap+1;
210 a>>=1;
211 if(shift<32){
212 b=(b+(1<<(shift-1)))>>shift;
213 }else{
214 b=0;
216 }else{
217 int shift=bp-ap+1;
218 *p=bp+1;
219 b>>=1;
220 if(shift<32){
221 a=(a+(1<<(shift-1)))>>shift;
222 }else{
223 a=0;
227 a+=b;
228 if((a&0xc0000000)==0xc0000000 ||
229 (a&0xc0000000)==0){
230 a<<=1;
231 (*p)--;
233 return(a);
236 #endif