/* libs/opengles/matrix.h ** ** Copyright 2006, The Android Open Source Project ** ** Licensed under the Apache License, Version 2.0 (the "License"); ** you may not use this file except in compliance with the License. ** You may obtain a copy of the License at ** ** http://www.apache.org/licenses/LICENSE-2.0 ** ** Unless required by applicable law or agreed to in writing, software ** distributed under the License is distributed on an "AS IS" BASIS, ** WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. ** See the License for the specific language governing permissions and ** limitations under the License. */ #ifndef ANDROID_OPENGLES_MATRIX_H #define ANDROID_OPENGLES_MATRIX_H #include #include #include #include #include #include namespace android { const int OGLES_MODELVIEW_STACK_DEPTH = 16; const int OGLES_PROJECTION_STACK_DEPTH = 2; const int OGLES_TEXTURE_STACK_DEPTH = 2; void ogles_init_matrix(ogles_context_t*); void ogles_uninit_matrix(ogles_context_t*); void ogles_invalidate_perspective(ogles_context_t* c); void ogles_validate_transform_impl(ogles_context_t* c, uint32_t want); int ogles_surfaceport(ogles_context_t* c, GLint x, GLint y); void ogles_scissor(ogles_context_t* c, GLint x, GLint y, GLsizei w, GLsizei h); void ogles_viewport(ogles_context_t* c, GLint x, GLint y, GLsizei w, GLsizei h); inline void ogles_validate_transform( ogles_context_t* c, uint32_t want) { if (c->transforms.dirty & want) ogles_validate_transform_impl(c, want); } // ---------------------------------------------------------------------------- inline GLfixed vsquare3(GLfixed a, GLfixed b, GLfixed c) { #if defined(__arm__) && !defined(__thumb__) GLfixed r; int32_t t; asm( "smull %0, %1, %2, %2 \n" "smlal %0, %1, %3, %3 \n" "smlal %0, %1, %4, %4 \n" "movs %0, %0, lsr #16 \n" "adc %0, %0, %1, lsl #16 \n" : "=&r"(r), "=&r"(t) : "%r"(a), "r"(b), "r"(c) : "cc" ); return r; #elif defined(__mips__) GLfixed res; int32_t t1,t2,t3; asm( "mult %[a], %[a] \r\n" "li %[res],0x8000 \r\n" "madd %[b],%[b] \r\n" "move %[t3],$zero \r\n" "madd %[c],%[c] \r\n" "mflo %[t1]\r\n" "mfhi %[t2]\r\n" "addu %[t1],%[res],%[t1]\r\n" /*add 0x8000*/ "sltu %[t3],%[t1],%[res]\r\n" "addu %[t2],%[t2],%[t3]\r\n" "srl %[res],%[t1],16\r\n" "sll %[t2],%[t2],16\r\n" "or %[res],%[res],%[t2]\r\n" : [res]"=&r"(res),[t1]"=&r"(t1),[t2]"=&r"(t2),[t3]"=&r"(t3) : [a] "r" (a),[b] "r" (b),[c] "r" (c) : "%hi","%lo" ); return res; #else return (( int64_t(a)*a + int64_t(b)*b + int64_t(c)*c + 0x8000)>>16); #endif } static inline GLfixed mla2a( GLfixed a0, GLfixed b0, GLfixed a1, GLfixed b1, GLfixed c) { #if defined(__arm__) && !defined(__thumb__) GLfixed r; int32_t t; asm( "smull %0, %1, %2, %3 \n" "smlal %0, %1, %4, %5 \n" "add %0, %6, %0, lsr #16 \n" "add %0, %0, %1, lsl #16 \n" : "=&r"(r), "=&r"(t) : "%r"(a0), "r"(b0), "%r"(a1), "r"(b1), "r"(c) : ); return r; #else return (( int64_t(a0)*b0 + int64_t(a1)*b1)>>16) + c; #endif } static inline GLfixed mla3a( GLfixed a0, GLfixed b0, GLfixed a1, GLfixed b1, GLfixed a2, GLfixed b2, GLfixed c) { #if defined(__arm__) && !defined(__thumb__) GLfixed r; int32_t t; asm( "smull %0, %1, %2, %3 \n" "smlal %0, %1, %4, %5 \n" "smlal %0, %1, %6, %7 \n" "add %0, %8, %0, lsr #16 \n" "add %0, %0, %1, lsl #16 \n" : "=&r"(r), "=&r"(t) : "%r"(a0), "r"(b0), "%r"(a1), "r"(b1), "%r"(a2), "r"(b2), "r"(c) : ); return r; #elif defined(__mips__) GLfixed res; int32_t t1,t2; asm( "mult %[a0],%[b0] \r\n" "madd %[a1],%[b1] \r\n" "madd %[a2],%[b2] \r\n" "mflo %[t2]\r\n" "mfhi %[t1]\r\n" "srl %[t2],%[t2],16\r\n" "sll %[t1],%[t1],16\r\n" "or %[t2],%[t2],%[t1]\r\n" "addu %[res],%[t2],%[c]" : [res]"=&r"(res),[t1]"=&r"(t1),[t2]"=&r"(t2) : [a0] "r" (a0),[b0] "r" (b0),[a1] "r" (a1),[b1] "r" (b1),[a2] "r" (a2),[b2] "r" (b2),[c] "r" (c) : "%hi","%lo" ); return res; #else return (( int64_t(a0)*b0 + int64_t(a1)*b1 + int64_t(a2)*b2)>>16) + c; #endif } // b0, b1, b2 are signed 16-bit quanities // that have been shifted right by 'shift' bits relative to normal // S16.16 fixed point static inline GLfixed mla3a16( GLfixed a0, int32_t b1b0, GLfixed a1, GLfixed a2, int32_t b2, GLint shift, GLfixed c) { #if defined(__arm__) && !defined(__thumb__) GLfixed r; asm( "smulwb %0, %1, %2 \n" "smlawt %0, %3, %2, %0 \n" "smlawb %0, %4, %5, %0 \n" "add %0, %7, %0, lsl %6 \n" : "=&r"(r) : "r"(a0), "r"(b1b0), "r"(a1), "r"(a2), "r"(b2), "r"(shift), "r"(c) : ); return r; #else int32_t accum; int16_t b0 = b1b0 & 0xffff; int16_t b1 = (b1b0 >> 16) & 0xffff; accum = int64_t(a0)*int16_t(b0) >> 16; accum += int64_t(a1)*int16_t(b1) >> 16; accum += int64_t(a2)*int16_t(b2) >> 16; accum = (accum << shift) + c; return accum; #endif } static inline GLfixed mla3a16_btb( GLfixed a0, GLfixed a1, GLfixed a2, int32_t b1b0, int32_t xxb2, GLint shift, GLfixed c) { #if defined(__arm__) && !defined(__thumb__) GLfixed r; asm( "smulwb %0, %1, %4 \n" "smlawt %0, %2, %4, %0 \n" "smlawb %0, %3, %5, %0 \n" "add %0, %7, %0, lsl %6 \n" : "=&r"(r) : "r"(a0), "r"(a1), "r"(a2), "r"(b1b0), "r"(xxb2), "r"(shift), "r"(c) : ); return r; #else int32_t accum; int16_t b0 = b1b0 & 0xffff; int16_t b1 = (b1b0 >> 16) & 0xffff; int16_t b2 = xxb2 & 0xffff; accum = int64_t(a0)*int16_t(b0) >> 16; accum += int64_t(a1)*int16_t(b1) >> 16; accum += int64_t(a2)*int16_t(b2) >> 16; accum = (accum << shift) + c; return accum; #endif } static inline GLfixed mla3a16_btt( GLfixed a0, GLfixed a1, GLfixed a2, int32_t b1b0, int32_t b2xx, GLint shift, GLfixed c) { #if defined(__arm__) && !defined(__thumb__) GLfixed r; asm( "smulwb %0, %1, %4 \n" "smlawt %0, %2, %4, %0 \n" "smlawt %0, %3, %5, %0 \n" "add %0, %7, %0, lsl %6 \n" : "=&r"(r) : "r"(a0), "r"(a1), "r"(a2), "r"(b1b0), "r"(b2xx), "r"(shift), "r"(c) : ); return r; #else int32_t accum; int16_t b0 = b1b0 & 0xffff; int16_t b1 = (b1b0 >> 16) & 0xffff; int16_t b2 = (b2xx >> 16) & 0xffff; accum = int64_t(a0)*int16_t(b0) >> 16; accum += int64_t(a1)*int16_t(b1) >> 16; accum += int64_t(a2)*int16_t(b2) >> 16; accum = (accum << shift) + c; return accum; #endif } static inline GLfixed mla3( GLfixed a0, GLfixed b0, GLfixed a1, GLfixed b1, GLfixed a2, GLfixed b2) { #if defined(__arm__) && !defined(__thumb__) GLfixed r; int32_t t; asm( "smull %0, %1, %2, %3 \n" "smlal %0, %1, %4, %5 \n" "smlal %0, %1, %6, %7 \n" "movs %0, %0, lsr #16 \n" "adc %0, %0, %1, lsl #16 \n" : "=&r"(r), "=&r"(t) : "%r"(a0), "r"(b0), "%r"(a1), "r"(b1), "%r"(a2), "r"(b2) : "cc" ); return r; #else return (( int64_t(a0)*b0 + int64_t(a1)*b1 + int64_t(a2)*b2 + 0x8000)>>16); #endif } static inline GLfixed mla4( GLfixed a0, GLfixed b0, GLfixed a1, GLfixed b1, GLfixed a2, GLfixed b2, GLfixed a3, GLfixed b3) { #if defined(__arm__) && !defined(__thumb__) GLfixed r; int32_t t; asm( "smull %0, %1, %2, %3 \n" "smlal %0, %1, %4, %5 \n" "smlal %0, %1, %6, %7 \n" "smlal %0, %1, %8, %9 \n" "movs %0, %0, lsr #16 \n" "adc %0, %0, %1, lsl #16 \n" : "=&r"(r), "=&r"(t) : "%r"(a0), "r"(b0), "%r"(a1), "r"(b1), "%r"(a2), "r"(b2), "%r"(a3), "r"(b3) : "cc" ); return r; #else return (( int64_t(a0)*b0 + int64_t(a1)*b1 + int64_t(a2)*b2 + int64_t(a3)*b3 + 0x8000)>>16); #endif } inline GLfixed dot4(const GLfixed* a, const GLfixed* b) { return mla4(a[0], b[0], a[1], b[1], a[2], b[2], a[3], b[3]); } inline GLfixed dot3(const GLfixed* a, const GLfixed* b) { return mla3(a[0], b[0], a[1], b[1], a[2], b[2]); } }; // namespace android #endif // ANDROID_OPENGLES_MATRIX_H