3 * This source code is part of
7 * GROningen MAchine for Chemical Simulations
10 * Written by David van der Spoel, Erik Lindahl, Berk Hess, and others.
11 * Copyright (c) 1991-2000, University of Groningen, The Netherlands.
12 * Copyright (c) 2001-2004, The GROMACS development team,
13 * check out http://www.gromacs.org for more information.
15 * This program is free software; you can redistribute it and/or
16 * modify it under the terms of the GNU General Public License
17 * as published by the Free Software Foundation; either version 2
18 * of the License, or (at your option) any later version.
20 * If you want to redistribute modifications, please consider that
21 * scientific software is very special. Version control is crucial -
22 * bugs must be traceable. We will be happy to consider code for
23 * inclusion in the official distribution, but derived work must not
24 * be called official GROMACS. Details are found in the README & COPYING
25 * files - if they are missing, get the official version at www.gromacs.org.
27 * To help us fund GROMACS development, we humbly ask that you cite
28 * the papers on the package - you can find them in the top README file.
30 * For more info, check our website at http://www.gromacs.org
33 * Gromacs Runs On Most of All Computer Systems
41 #define gmx_inline inline
44 #define gmx_inline __inline
52 collection of in-line ready operations:
54 lookup-table optimized scalar operations:
55 real gmx_invsqrt(real x)
56 void vecinvsqrt(real in[],real out[],int n)
57 void vecrecip(real in[],real out[],int n)
62 void rvec_add(const rvec a,const rvec b,rvec c) c = a + b
63 void dvec_add(const dvec a,const dvec b,dvec c) c = a + b
64 void ivec_add(const ivec a,const ivec b,ivec c) c = a + b
65 void rvec_inc(rvec a,const rvec b) a += b
66 void dvec_inc(dvec a,const dvec b) a += b
67 void ivec_inc(ivec a,const ivec b) a += b
68 void rvec_sub(const rvec a,const rvec b,rvec c) c = a - b
69 void dvec_sub(const dvec a,const dvec b,dvec c) c = a - b
70 void rvec_dec(rvec a,rvec b) a -= b
71 void copy_rvec(const rvec a,rvec b) b = a (reals)
72 void copy_dvec(const dvec a,dvec b) b = a (reals)
73 void copy_ivec(const ivec a,ivec b) b = a (integers)
74 void ivec_sub(const ivec a,const ivec b,ivec c) c = a - b
75 void svmul(real a,rvec v1,rvec v2) v2 = a * v1
76 void dsvmul(double a,dvec v1,dvec v2) v2 = a * v1
77 void clear_rvec(rvec a) a = 0
78 void clear_dvec(dvec a) a = 0
79 void clear_ivec(rvec a) a = 0
80 void clear_rvecs(int n,rvec v[])
81 real iprod(rvec a,rvec b) = a . b (inner product)
82 double diprod(dvec a,dvec b) = a . b (inner product)
83 real iiprod(ivec a,ivec b) = a . b (integers)
84 real norm2(rvec a) = | a |^2 ( = x*y*z )
85 double dnorm2(dvec a) = | a |^2 ( = x*y*z )
86 real norm(rvec a) = | a |
87 double dnorm(dvec a) = | a |
88 void cprod(rvec a,rvec b,rvec c) c = a x b (cross product)
89 void dprod(rvec a,rvec b,rvec c) c = a x b (cross product)
90 void dprod(rvec a,rvec b,rvec c) c = a * b (direct product)
91 real cos_angle(rvec a,rvec b)
92 real cos_angle_no_table(rvec a,rvec b)
93 real distance2(rvec v1, rvec v2) = | v2 - v1 |^2
94 void unitv(rvec src,rvec dest) dest = src / |src|
95 void unitv_no_table(rvec src,rvec dest) dest = src / |src|
97 matrix (3x3) operations:
98 ! indicates that dest should not be the same as a, b or src
99 the _ur0 varieties work on matrices that have only zeros
100 in the upper right part, such as box matrices, these varieties
101 could produce less rounding errors, not due to the operations themselves,
102 but because the compiler can easier recombine the operations
103 void copy_mat(matrix a,matrix b) b = a
104 void clear_mat(matrix a) a = 0
105 void mmul(matrix a,matrix b,matrix dest) ! dest = a . b
106 void mmul_ur0(matrix a,matrix b,matrix dest) dest = a . b
107 void transpose(matrix src,matrix dest) ! dest = src*
108 void tmmul(matrix a,matrix b,matrix dest) ! dest = a* . b
109 void mtmul(matrix a,matrix b,matrix dest) ! dest = a . b*
110 real det(matrix a) = det(a)
111 void m_add(matrix a,matrix b,matrix dest) dest = a + b
112 void m_sub(matrix a,matrix b,matrix dest) dest = a - b
113 void msmul(matrix m1,real r1,matrix dest) dest = r1 * m1
114 void m_inv_ur0(matrix src,matrix dest) dest = src^-1
115 void m_inv(matrix src,matrix dest) ! dest = src^-1
116 void mvmul(matrix a,rvec src,rvec dest) ! dest = a . src
117 void mvmul_ur0(matrix a,rvec src,rvec dest) dest = a . src
118 void tmvmul_ur0(matrix a,rvec src,rvec dest) dest = a* . src
119 real trace(matrix m) = trace(m)
122 #include "types/simple.h"
124 #include "typedefs.h"
125 #include "sysstuff.h"
127 #include "gmx_fatal.h"
128 #include "mpelogging.h"
134 } /* avoid screwing up indentation */
138 #define EXP_LSB 0x00800000
139 #define EXP_MASK 0x7f800000
141 #define FRACT_MASK 0x007fffff
142 #define FRACT_SIZE 11 /* significant part of fraction */
143 #define FRACT_SHIFT (EXP_SHIFT-FRACT_SIZE)
144 #define EXP_ADDR(val) (((val)&EXP_MASK)>>EXP_SHIFT)
145 #define FRACT_ADDR(val) (((val)&(FRACT_MASK|EXP_LSB))>>FRACT_SHIFT)
147 #define PR_VEC(a) a[XX],a[YY],a[ZZ]
149 #ifdef GMX_SOFTWARE_INVSQRT
150 extern const unsigned int * gmx_invsqrt_exptab
;
151 extern const unsigned int * gmx_invsqrt_fracttab
;
162 #ifdef GMX_SOFTWARE_INVSQRT
163 static real
gmx_invsqrt(real x
)
166 const real three
=3.0;
167 t_convert result
,bit_pattern
;
168 unsigned int exp
,fract
;
176 exp
= EXP_ADDR(bit_pattern
.bval
);
177 fract
= FRACT_ADDR(bit_pattern
.bval
);
178 result
.bval
=gmx_invsqrt_exptab
[exp
] | gmx_invsqrt_fracttab
[fract
];
181 y
=(half
*lu
*(three
-((x
*lu
)*lu
)));
183 y2
=(half
*y
*(three
-((x
*y
)*y
)));
185 return y2
; /* 10 Flops */
187 return y
; /* 5 Flops */
191 #endif /* gmx_invsqrt */
193 #ifdef GMX_POWERPC_SQRT
194 static real
gmx_invsqrt(real x
)
197 const real three
=3.0;
198 t_convert result
,bit_pattern
;
199 unsigned int exp
,fract
;
206 lu
= __frsqrte((double)x
);
208 y
=(half
*lu
*(three
-((x
*lu
)*lu
)));
210 #if (GMX_POWERPC_SQRT==2)
211 /* Extra iteration required */
212 y
=(half
*y
*(three
-((x
*y
)*y
)));
216 y2
=(half
*y
*(three
-((x
*y
)*y
)));
218 return y2
; /* 10 Flops */
220 return y
; /* 5 Flops */
224 #endif /* powerpc_invsqrt */
228 #define gmx_invsqrt(x) (1.0f/sqrt(x))
235 static real
sqr(real x
)
240 static inline double dsqr(double x
)
245 /* Maclaurin series for sinh(x)/x, useful for NH chains and MTTK pressure control
246 Here, we compute it to 10th order, which might be overkill, 8th is probably enough,
247 but it's not very much more expensive. */
249 static inline real
series_sinhx(real x
)
252 return (1 + (x2
/6.0)*(1 + (x2
/20.0)*(1 + (x2
/42.0)*(1 + (x2
/72.0)*(1 + (x2
/110.0))))));
255 extern void vecinvsqrt(real in
[],real out
[],int n
);
256 /* Perform out[i]=1.0/sqrt(in[i]) for n elements */
259 extern void vecrecip(real in
[],real out
[],int n
);
260 /* Perform out[i]=1.0/(in[i]) for n elements */
262 /* Note: If you need a fast version of vecinvsqrt
263 * and/or vecrecip, call detectcpu() and run the SSE/3DNow/SSE2/Altivec
264 * versions if your hardware supports it.
266 * To use those routines, your memory HAS TO BE CACHE-ALIGNED.
267 * Use snew_aligned(ptr,size,32) to allocate and sfree_aligned to free.
271 static inline void rvec_add(const rvec a
,const rvec b
,rvec c
)
284 static inline void dvec_add(const dvec a
,const dvec b
,dvec c
)
297 static inline void ivec_add(const ivec a
,const ivec b
,ivec c
)
310 static inline void rvec_inc(rvec a
,const rvec b
)
323 static inline void dvec_inc(dvec a
,const dvec b
)
336 static inline void rvec_sub(const rvec a
,const rvec b
,rvec c
)
349 static inline void dvec_sub(const dvec a
,const dvec b
,dvec c
)
362 static inline void rvec_dec(rvec a
,const rvec b
)
375 static inline void copy_rvec(const rvec a
,rvec b
)
382 static inline void copy_rvecn(rvec
*a
,rvec
*b
,int startn
, int endn
)
385 for (i
=startn
;i
<endn
;i
++) {
392 static inline void copy_dvec(const dvec a
,dvec b
)
399 static inline void copy_ivec(const ivec a
,ivec b
)
406 static inline void ivec_sub(const ivec a
,const ivec b
,ivec c
)
419 static inline void copy_mat(matrix a
,matrix b
)
421 copy_rvec(a
[XX
],b
[XX
]);
422 copy_rvec(a
[YY
],b
[YY
]);
423 copy_rvec(a
[ZZ
],b
[ZZ
]);
426 static inline void svmul(real a
,const rvec v1
,rvec v2
)
433 static inline void dsvmul(double a
,const dvec v1
,dvec v2
)
440 static inline real
distance2(const rvec v1
,const rvec v2
)
442 return sqr(v2
[XX
]-v1
[XX
]) + sqr(v2
[YY
]-v1
[YY
]) + sqr(v2
[ZZ
]-v1
[ZZ
]);
445 static inline void clear_rvec(rvec a
)
447 /* The ibm compiler has problems with inlining this
448 * when we use a const real variable
455 static inline void clear_dvec(dvec a
)
457 /* The ibm compiler has problems with inlining this
458 * when we use a const real variable
465 static inline void clear_ivec(ivec a
)
472 static inline void clear_rvecs(int n
,rvec v
[])
474 /* memset(v[0],0,DIM*n*sizeof(v[0][0])); */
477 GMX_MPE_LOG(ev_clear_rvecs_start
);
482 GMX_MPE_LOG(ev_clear_rvecs_finish
);
485 static inline void clear_mat(matrix a
)
487 /* memset(a[0],0,DIM*DIM*sizeof(a[0][0])); */
491 a
[XX
][XX
]=a
[XX
][YY
]=a
[XX
][ZZ
]=nul
;
492 a
[YY
][XX
]=a
[YY
][YY
]=a
[YY
][ZZ
]=nul
;
493 a
[ZZ
][XX
]=a
[ZZ
][YY
]=a
[ZZ
][ZZ
]=nul
;
496 static inline real
iprod(const rvec a
,const rvec b
)
498 return (a
[XX
]*b
[XX
]+a
[YY
]*b
[YY
]+a
[ZZ
]*b
[ZZ
]);
501 static inline double diprod(const dvec a
,const dvec b
)
503 return (a
[XX
]*b
[XX
]+a
[YY
]*b
[YY
]+a
[ZZ
]*b
[ZZ
]);
506 static inline int iiprod(const ivec a
,const ivec b
)
508 return (a
[XX
]*b
[XX
]+a
[YY
]*b
[YY
]+a
[ZZ
]*b
[ZZ
]);
511 static inline real
norm2(const rvec a
)
513 return a
[XX
]*a
[XX
]+a
[YY
]*a
[YY
]+a
[ZZ
]*a
[ZZ
];
516 static inline double dnorm2(const dvec a
)
518 return a
[XX
]*a
[XX
]+a
[YY
]*a
[YY
]+a
[ZZ
]*a
[ZZ
];
521 static inline real
norm(const rvec a
)
523 return (real
)sqrt(a
[XX
]*a
[XX
]+a
[YY
]*a
[YY
]+a
[ZZ
]*a
[ZZ
]);
526 static inline double dnorm(const dvec a
)
528 return sqrt(a
[XX
]*a
[XX
]+a
[YY
]*a
[YY
]+a
[ZZ
]*a
[ZZ
]);
532 * Do _not_ use these routines to calculate the angle between two vectors
533 * as acos(cos_angle(u,v)). While it might seem obvious, the acos function
534 * is very flat close to -1 and 1, which will lead to accuracy-loss.
535 * Instead, use the new gmx_angle() function directly.
538 cos_angle(const rvec a
,const rvec b
)
541 * ax*bx + ay*by + az*bz
542 * cos-vec (a,b) = ---------------------
547 double aa
,bb
,ip
,ipa
,ipb
,ipab
; /* For accuracy these must be double! */
550 for(m
=0; (m
<DIM
); m
++) { /* 18 */
559 cosval
= ip
*gmx_invsqrt(ipab
); /* 7 */
572 * Do _not_ use these routines to calculate the angle between two vectors
573 * as acos(cos_angle(u,v)). While it might seem obvious, the acos function
574 * is very flat close to -1 and 1, which will lead to accuracy-loss.
575 * Instead, use the new gmx_angle() function directly.
578 cos_angle_no_table(const rvec a
,const rvec b
)
580 /* This version does not need the invsqrt lookup table */
583 double aa
,bb
,ip
,ipa
,ipb
; /* For accuracy these must be double! */
586 for(m
=0; (m
<DIM
); m
++) { /* 18 */
593 cosval
=ip
/sqrt(ipa
*ipb
); /* 12 */
604 static inline void cprod(const rvec a
,const rvec b
,rvec c
)
606 c
[XX
]=a
[YY
]*b
[ZZ
]-a
[ZZ
]*b
[YY
];
607 c
[YY
]=a
[ZZ
]*b
[XX
]-a
[XX
]*b
[ZZ
];
608 c
[ZZ
]=a
[XX
]*b
[YY
]-a
[YY
]*b
[XX
];
611 static inline void dcprod(const dvec a
,const dvec b
,dvec c
)
613 c
[XX
]=a
[YY
]*b
[ZZ
]-a
[ZZ
]*b
[YY
];
614 c
[YY
]=a
[ZZ
]*b
[XX
]-a
[XX
]*b
[ZZ
];
615 c
[ZZ
]=a
[XX
]*b
[YY
]-a
[YY
]*b
[XX
];
618 /* This routine calculates the angle between a & b without any loss of accuracy close to 0/PI.
619 * If you only need cos(theta), use the cos_angle() routines to save a few cycles.
620 * This routine is faster than it might appear, since atan2 is accelerated on many CPUs (e.g. x86).
623 gmx_angle(const rvec a
, const rvec b
)
633 return atan2(wlen
,s
);
636 static inline void mmul_ur0(matrix a
,matrix b
,matrix dest
)
638 dest
[XX
][XX
]=a
[XX
][XX
]*b
[XX
][XX
];
641 dest
[YY
][XX
]=a
[YY
][XX
]*b
[XX
][XX
]+a
[YY
][YY
]*b
[YY
][XX
];
642 dest
[YY
][YY
]= a
[YY
][YY
]*b
[YY
][YY
];
644 dest
[ZZ
][XX
]=a
[ZZ
][XX
]*b
[XX
][XX
]+a
[ZZ
][YY
]*b
[YY
][XX
]+a
[ZZ
][ZZ
]*b
[ZZ
][XX
];
645 dest
[ZZ
][YY
]= a
[ZZ
][YY
]*b
[YY
][YY
]+a
[ZZ
][ZZ
]*b
[ZZ
][YY
];
646 dest
[ZZ
][ZZ
]= a
[ZZ
][ZZ
]*b
[ZZ
][ZZ
];
649 static inline void mmul(matrix a
,matrix b
,matrix dest
)
651 dest
[XX
][XX
]=a
[XX
][XX
]*b
[XX
][XX
]+a
[XX
][YY
]*b
[YY
][XX
]+a
[XX
][ZZ
]*b
[ZZ
][XX
];
652 dest
[YY
][XX
]=a
[YY
][XX
]*b
[XX
][XX
]+a
[YY
][YY
]*b
[YY
][XX
]+a
[YY
][ZZ
]*b
[ZZ
][XX
];
653 dest
[ZZ
][XX
]=a
[ZZ
][XX
]*b
[XX
][XX
]+a
[ZZ
][YY
]*b
[YY
][XX
]+a
[ZZ
][ZZ
]*b
[ZZ
][XX
];
654 dest
[XX
][YY
]=a
[XX
][XX
]*b
[XX
][YY
]+a
[XX
][YY
]*b
[YY
][YY
]+a
[XX
][ZZ
]*b
[ZZ
][YY
];
655 dest
[YY
][YY
]=a
[YY
][XX
]*b
[XX
][YY
]+a
[YY
][YY
]*b
[YY
][YY
]+a
[YY
][ZZ
]*b
[ZZ
][YY
];
656 dest
[ZZ
][YY
]=a
[ZZ
][XX
]*b
[XX
][YY
]+a
[ZZ
][YY
]*b
[YY
][YY
]+a
[ZZ
][ZZ
]*b
[ZZ
][YY
];
657 dest
[XX
][ZZ
]=a
[XX
][XX
]*b
[XX
][ZZ
]+a
[XX
][YY
]*b
[YY
][ZZ
]+a
[XX
][ZZ
]*b
[ZZ
][ZZ
];
658 dest
[YY
][ZZ
]=a
[YY
][XX
]*b
[XX
][ZZ
]+a
[YY
][YY
]*b
[YY
][ZZ
]+a
[YY
][ZZ
]*b
[ZZ
][ZZ
];
659 dest
[ZZ
][ZZ
]=a
[ZZ
][XX
]*b
[XX
][ZZ
]+a
[ZZ
][YY
]*b
[YY
][ZZ
]+a
[ZZ
][ZZ
]*b
[ZZ
][ZZ
];
662 static inline void transpose(matrix src
,matrix dest
)
664 dest
[XX
][XX
]=src
[XX
][XX
];
665 dest
[YY
][XX
]=src
[XX
][YY
];
666 dest
[ZZ
][XX
]=src
[XX
][ZZ
];
667 dest
[XX
][YY
]=src
[YY
][XX
];
668 dest
[YY
][YY
]=src
[YY
][YY
];
669 dest
[ZZ
][YY
]=src
[YY
][ZZ
];
670 dest
[XX
][ZZ
]=src
[ZZ
][XX
];
671 dest
[YY
][ZZ
]=src
[ZZ
][YY
];
672 dest
[ZZ
][ZZ
]=src
[ZZ
][ZZ
];
675 static inline void tmmul(matrix a
,matrix b
,matrix dest
)
677 /* Computes dest=mmul(transpose(a),b,dest) - used in do_pr_pcoupl */
678 dest
[XX
][XX
]=a
[XX
][XX
]*b
[XX
][XX
]+a
[YY
][XX
]*b
[YY
][XX
]+a
[ZZ
][XX
]*b
[ZZ
][XX
];
679 dest
[XX
][YY
]=a
[XX
][XX
]*b
[XX
][YY
]+a
[YY
][XX
]*b
[YY
][YY
]+a
[ZZ
][XX
]*b
[ZZ
][YY
];
680 dest
[XX
][ZZ
]=a
[XX
][XX
]*b
[XX
][ZZ
]+a
[YY
][XX
]*b
[YY
][ZZ
]+a
[ZZ
][XX
]*b
[ZZ
][ZZ
];
681 dest
[YY
][XX
]=a
[XX
][YY
]*b
[XX
][XX
]+a
[YY
][YY
]*b
[YY
][XX
]+a
[ZZ
][YY
]*b
[ZZ
][XX
];
682 dest
[YY
][YY
]=a
[XX
][YY
]*b
[XX
][YY
]+a
[YY
][YY
]*b
[YY
][YY
]+a
[ZZ
][YY
]*b
[ZZ
][YY
];
683 dest
[YY
][ZZ
]=a
[XX
][YY
]*b
[XX
][ZZ
]+a
[YY
][YY
]*b
[YY
][ZZ
]+a
[ZZ
][YY
]*b
[ZZ
][ZZ
];
684 dest
[ZZ
][XX
]=a
[XX
][ZZ
]*b
[XX
][XX
]+a
[YY
][ZZ
]*b
[YY
][XX
]+a
[ZZ
][ZZ
]*b
[ZZ
][XX
];
685 dest
[ZZ
][YY
]=a
[XX
][ZZ
]*b
[XX
][YY
]+a
[YY
][ZZ
]*b
[YY
][YY
]+a
[ZZ
][ZZ
]*b
[ZZ
][YY
];
686 dest
[ZZ
][ZZ
]=a
[XX
][ZZ
]*b
[XX
][ZZ
]+a
[YY
][ZZ
]*b
[YY
][ZZ
]+a
[ZZ
][ZZ
]*b
[ZZ
][ZZ
];
689 static inline void mtmul(matrix a
,matrix b
,matrix dest
)
691 /* Computes dest=mmul(a,transpose(b),dest) - used in do_pr_pcoupl */
692 dest
[XX
][XX
]=a
[XX
][XX
]*b
[XX
][XX
]+a
[XX
][YY
]*b
[XX
][YY
]+a
[XX
][ZZ
]*b
[XX
][ZZ
];
693 dest
[XX
][YY
]=a
[XX
][XX
]*b
[YY
][XX
]+a
[XX
][YY
]*b
[YY
][YY
]+a
[XX
][ZZ
]*b
[YY
][ZZ
];
694 dest
[XX
][ZZ
]=a
[XX
][XX
]*b
[ZZ
][XX
]+a
[XX
][YY
]*b
[ZZ
][YY
]+a
[XX
][ZZ
]*b
[ZZ
][ZZ
];
695 dest
[YY
][XX
]=a
[YY
][XX
]*b
[XX
][XX
]+a
[YY
][YY
]*b
[XX
][YY
]+a
[YY
][ZZ
]*b
[XX
][ZZ
];
696 dest
[YY
][YY
]=a
[YY
][XX
]*b
[YY
][XX
]+a
[YY
][YY
]*b
[YY
][YY
]+a
[YY
][ZZ
]*b
[YY
][ZZ
];
697 dest
[YY
][ZZ
]=a
[YY
][XX
]*b
[ZZ
][XX
]+a
[YY
][YY
]*b
[ZZ
][YY
]+a
[YY
][ZZ
]*b
[ZZ
][ZZ
];
698 dest
[ZZ
][XX
]=a
[ZZ
][XX
]*b
[XX
][XX
]+a
[ZZ
][YY
]*b
[XX
][YY
]+a
[ZZ
][ZZ
]*b
[XX
][ZZ
];
699 dest
[ZZ
][YY
]=a
[ZZ
][XX
]*b
[YY
][XX
]+a
[ZZ
][YY
]*b
[YY
][YY
]+a
[ZZ
][ZZ
]*b
[YY
][ZZ
];
700 dest
[ZZ
][ZZ
]=a
[ZZ
][XX
]*b
[ZZ
][XX
]+a
[ZZ
][YY
]*b
[ZZ
][YY
]+a
[ZZ
][ZZ
]*b
[ZZ
][ZZ
];
703 static inline real
det(matrix a
)
705 return ( a
[XX
][XX
]*(a
[YY
][YY
]*a
[ZZ
][ZZ
]-a
[ZZ
][YY
]*a
[YY
][ZZ
])
706 -a
[YY
][XX
]*(a
[XX
][YY
]*a
[ZZ
][ZZ
]-a
[ZZ
][YY
]*a
[XX
][ZZ
])
707 +a
[ZZ
][XX
]*(a
[XX
][YY
]*a
[YY
][ZZ
]-a
[YY
][YY
]*a
[XX
][ZZ
]));
710 static inline void m_add(matrix a
,matrix b
,matrix dest
)
712 dest
[XX
][XX
]=a
[XX
][XX
]+b
[XX
][XX
];
713 dest
[XX
][YY
]=a
[XX
][YY
]+b
[XX
][YY
];
714 dest
[XX
][ZZ
]=a
[XX
][ZZ
]+b
[XX
][ZZ
];
715 dest
[YY
][XX
]=a
[YY
][XX
]+b
[YY
][XX
];
716 dest
[YY
][YY
]=a
[YY
][YY
]+b
[YY
][YY
];
717 dest
[YY
][ZZ
]=a
[YY
][ZZ
]+b
[YY
][ZZ
];
718 dest
[ZZ
][XX
]=a
[ZZ
][XX
]+b
[ZZ
][XX
];
719 dest
[ZZ
][YY
]=a
[ZZ
][YY
]+b
[ZZ
][YY
];
720 dest
[ZZ
][ZZ
]=a
[ZZ
][ZZ
]+b
[ZZ
][ZZ
];
723 static inline void m_sub(matrix a
,matrix b
,matrix dest
)
725 dest
[XX
][XX
]=a
[XX
][XX
]-b
[XX
][XX
];
726 dest
[XX
][YY
]=a
[XX
][YY
]-b
[XX
][YY
];
727 dest
[XX
][ZZ
]=a
[XX
][ZZ
]-b
[XX
][ZZ
];
728 dest
[YY
][XX
]=a
[YY
][XX
]-b
[YY
][XX
];
729 dest
[YY
][YY
]=a
[YY
][YY
]-b
[YY
][YY
];
730 dest
[YY
][ZZ
]=a
[YY
][ZZ
]-b
[YY
][ZZ
];
731 dest
[ZZ
][XX
]=a
[ZZ
][XX
]-b
[ZZ
][XX
];
732 dest
[ZZ
][YY
]=a
[ZZ
][YY
]-b
[ZZ
][YY
];
733 dest
[ZZ
][ZZ
]=a
[ZZ
][ZZ
]-b
[ZZ
][ZZ
];
736 static inline void msmul(matrix m1
,real r1
,matrix dest
)
738 dest
[XX
][XX
]=r1
*m1
[XX
][XX
];
739 dest
[XX
][YY
]=r1
*m1
[XX
][YY
];
740 dest
[XX
][ZZ
]=r1
*m1
[XX
][ZZ
];
741 dest
[YY
][XX
]=r1
*m1
[YY
][XX
];
742 dest
[YY
][YY
]=r1
*m1
[YY
][YY
];
743 dest
[YY
][ZZ
]=r1
*m1
[YY
][ZZ
];
744 dest
[ZZ
][XX
]=r1
*m1
[ZZ
][XX
];
745 dest
[ZZ
][YY
]=r1
*m1
[ZZ
][YY
];
746 dest
[ZZ
][ZZ
]=r1
*m1
[ZZ
][ZZ
];
749 static inline void m_inv_ur0(matrix src
,matrix dest
)
751 double tmp
= src
[XX
][XX
]*src
[YY
][YY
]*src
[ZZ
][ZZ
];
752 if (fabs(tmp
) <= 100*GMX_REAL_MIN
)
753 gmx_fatal(FARGS
,"Can not invert matrix, determinant is zero");
755 dest
[XX
][XX
] = 1/src
[XX
][XX
];
756 dest
[YY
][YY
] = 1/src
[YY
][YY
];
757 dest
[ZZ
][ZZ
] = 1/src
[ZZ
][ZZ
];
758 dest
[ZZ
][XX
] = (src
[YY
][XX
]*src
[ZZ
][YY
]*dest
[YY
][YY
]
759 - src
[ZZ
][XX
])*dest
[XX
][XX
]*dest
[ZZ
][ZZ
];
760 dest
[YY
][XX
] = -src
[YY
][XX
]*dest
[XX
][XX
]*dest
[YY
][YY
];
761 dest
[ZZ
][YY
] = -src
[ZZ
][YY
]*dest
[YY
][YY
]*dest
[ZZ
][ZZ
];
767 static inline void m_inv(matrix src
,matrix dest
)
769 const real smallreal
= (real
)1.0e-24;
770 const real largereal
= (real
)1.0e24
;
777 if ((fc
<= smallreal
) || (fc
>= largereal
))
778 gmx_fatal(FARGS
,"Can not invert matrix, determinant = %e",deter
);
780 dest
[XX
][XX
]= c
*(src
[YY
][YY
]*src
[ZZ
][ZZ
]-src
[ZZ
][YY
]*src
[YY
][ZZ
]);
781 dest
[XX
][YY
]=-c
*(src
[XX
][YY
]*src
[ZZ
][ZZ
]-src
[ZZ
][YY
]*src
[XX
][ZZ
]);
782 dest
[XX
][ZZ
]= c
*(src
[XX
][YY
]*src
[YY
][ZZ
]-src
[YY
][YY
]*src
[XX
][ZZ
]);
783 dest
[YY
][XX
]=-c
*(src
[YY
][XX
]*src
[ZZ
][ZZ
]-src
[ZZ
][XX
]*src
[YY
][ZZ
]);
784 dest
[YY
][YY
]= c
*(src
[XX
][XX
]*src
[ZZ
][ZZ
]-src
[ZZ
][XX
]*src
[XX
][ZZ
]);
785 dest
[YY
][ZZ
]=-c
*(src
[XX
][XX
]*src
[YY
][ZZ
]-src
[YY
][XX
]*src
[XX
][ZZ
]);
786 dest
[ZZ
][XX
]= c
*(src
[YY
][XX
]*src
[ZZ
][YY
]-src
[ZZ
][XX
]*src
[YY
][YY
]);
787 dest
[ZZ
][YY
]=-c
*(src
[XX
][XX
]*src
[ZZ
][YY
]-src
[ZZ
][XX
]*src
[XX
][YY
]);
788 dest
[ZZ
][ZZ
]= c
*(src
[XX
][XX
]*src
[YY
][YY
]-src
[YY
][XX
]*src
[XX
][YY
]);
791 static inline void mvmul(matrix a
,const rvec src
,rvec dest
)
793 dest
[XX
]=a
[XX
][XX
]*src
[XX
]+a
[XX
][YY
]*src
[YY
]+a
[XX
][ZZ
]*src
[ZZ
];
794 dest
[YY
]=a
[YY
][XX
]*src
[XX
]+a
[YY
][YY
]*src
[YY
]+a
[YY
][ZZ
]*src
[ZZ
];
795 dest
[ZZ
]=a
[ZZ
][XX
]*src
[XX
]+a
[ZZ
][YY
]*src
[YY
]+a
[ZZ
][ZZ
]*src
[ZZ
];
798 static inline void mvmul_ur0(matrix a
,const rvec src
,rvec dest
)
800 dest
[ZZ
]=a
[ZZ
][XX
]*src
[XX
]+a
[ZZ
][YY
]*src
[YY
]+a
[ZZ
][ZZ
]*src
[ZZ
];
801 dest
[YY
]=a
[YY
][XX
]*src
[XX
]+a
[YY
][YY
];
802 dest
[XX
]=a
[XX
][XX
]*src
[XX
];
805 static inline void tmvmul_ur0(matrix a
,const rvec src
,rvec dest
)
807 dest
[XX
]=a
[XX
][XX
]*src
[XX
]+a
[YY
][XX
]*src
[YY
]+a
[ZZ
][XX
]*src
[ZZ
];
808 dest
[YY
]= a
[YY
][YY
]*src
[YY
]+a
[ZZ
][YY
]*src
[ZZ
];
809 dest
[ZZ
]= a
[ZZ
][ZZ
]*src
[ZZ
];
812 static inline void unitv(const rvec src
,rvec dest
)
816 linv
=gmx_invsqrt(norm2(src
));
817 dest
[XX
]=linv
*src
[XX
];
818 dest
[YY
]=linv
*src
[YY
];
819 dest
[ZZ
]=linv
*src
[ZZ
];
822 static inline void unitv_no_table(const rvec src
,rvec dest
)
826 linv
=1.0/sqrt(norm2(src
));
827 dest
[XX
]=linv
*src
[XX
];
828 dest
[YY
]=linv
*src
[YY
];
829 dest
[ZZ
]=linv
*src
[ZZ
];
832 static void calc_lll(rvec box
,rvec lll
)
834 lll
[XX
] = 2.0*M_PI
/box
[XX
];
835 lll
[YY
] = 2.0*M_PI
/box
[YY
];
836 lll
[ZZ
] = 2.0*M_PI
/box
[ZZ
];
839 static inline real
trace(matrix m
)
841 return (m
[XX
][XX
]+m
[YY
][YY
]+m
[ZZ
][ZZ
]);
844 static inline real
_divide(real a
,real b
,const char *file
,int line
)
846 if (fabs(b
) <= GMX_REAL_MIN
)
847 gmx_fatal(FARGS
,"Dividing by zero, file %s, line %d",file
,line
);
851 static inline int _mod(int a
,int b
,char *file
,int line
)
854 gmx_fatal(FARGS
,"Modulo zero, file %s, line %d",file
,line
);
858 /* Operations on multidimensional rvecs, used e.g. in edsam.c */
859 static void m_rveccopy(int dim
, rvec
*a
, rvec
*b
)
864 for (i
=0; i
<dim
; i
++)
865 copy_rvec(a
[i
],b
[i
]);
868 /*computer matrix vectors from base vectors and angles */
869 static void matrix_convert(matrix box
, rvec vec
, rvec angle
)
871 svmul(DEG2RAD
,angle
,angle
);
872 box
[XX
][XX
] = vec
[XX
];
873 box
[YY
][XX
] = vec
[YY
]*cos(angle
[ZZ
]);
874 box
[YY
][YY
] = vec
[YY
]*sin(angle
[ZZ
]);
875 box
[ZZ
][XX
] = vec
[ZZ
]*cos(angle
[YY
]);
876 box
[ZZ
][YY
] = vec
[ZZ
]
877 *(cos(angle
[XX
])-cos(angle
[YY
])*cos(angle
[ZZ
]))/sin(angle
[ZZ
]);
878 box
[ZZ
][ZZ
] = sqrt(sqr(vec
[ZZ
])
879 -box
[ZZ
][XX
]*box
[ZZ
][XX
]-box
[ZZ
][YY
]*box
[ZZ
][YY
]);
882 #define divide(a,b) _divide((a),(b),__FILE__,__LINE__)
883 #define mod(a,b) _mod((a),(b),__FILE__,__LINE__)