Home | History | Annotate | Download | only in math
      1 /*
      2  * Mesa 3-D graphics library
      3  * Version:  6.1
      4  *
      5  * Copyright (C) 1999-2004  Brian Paul   All Rights Reserved.
      6  *
      7  * Permission is hereby granted, free of charge, to any person obtaining a
      8  * copy of this software and associated documentation files (the "Software"),
      9  * to deal in the Software without restriction, including without limitation
     10  * the rights to use, copy, modify, merge, publish, distribute, sublicense,
     11  * and/or sell copies of the Software, and to permit persons to whom the
     12  * Software is furnished to do so, subject to the following conditions:
     13  *
     14  * The above copyright notice and this permission notice shall be included
     15  * in all copies or substantial portions of the Software.
     16  *
     17  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
     18  * OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
     19  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
     20  * BRIAN PAUL BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN
     21  * AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
     22  * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
     23  *
     24  * Authors:
     25  *    Gareth Hughes
     26  */
     27 
     28 #ifndef __M_DEBUG_UTIL_H__
     29 #define __M_DEBUG_UTIL_H__
     30 
     31 
     32 #ifdef DEBUG_MATH  /* This code only used for debugging */
     33 
     34 
     35 /* Comment this out to deactivate the cycle counter.
     36  * NOTE: it works only on CPUs which know the 'rdtsc' command (586 or higher)
     37  * (hope, you don't try to debug Mesa on a 386 ;)
     38  */
     39 #if defined(__GNUC__) && \
     40     ((defined(__i386__) && defined(USE_X86_ASM)) || \
     41      (defined(__sparc__) && defined(USE_SPARC_ASM)))
     42 #define  RUN_DEBUG_BENCHMARK
     43 #endif
     44 
     45 #define TEST_COUNT		128	/* size of the tested vector array   */
     46 
     47 #define REQUIRED_PRECISION	10	/* allow 4 bits to miss              */
     48 #define MAX_PRECISION		24	/* max. precision possible           */
     49 
     50 
     51 #ifdef  RUN_DEBUG_BENCHMARK
     52 /* Overhead of profiling counter in cycles.  Automatically adjusted to
     53  * your machine at run time - counter initialization should give very
     54  * consistent results.
     55  */
     56 extern long counter_overhead;
     57 
     58 /* This is the value of the environment variable MESA_PROFILE, and is
     59  * used to determine if we should benchmark the functions as well as
     60  * verify their correctness.
     61  */
     62 extern char *mesa_profile;
     63 
     64 /* Modify the number of tests if you like.
     65  * We take the minimum of all results, because every error should be
     66  * positive (time used by other processes, task switches etc).
     67  * It is assumed that all calculations are done in the cache.
     68  */
     69 
     70 #if defined(__i386__)
     71 
     72 #if 1 /* PPro, PII, PIII version */
     73 
     74 /* Profiling on the P6 architecture requires a little more work, due to
     75  * the internal out-of-order execution.  We must perform a serializing
     76  * 'cpuid' instruction before and after the 'rdtsc' instructions to make
     77  * sure no other uops are executed when we sample the timestamp counter.
     78  */
     79 #define  INIT_COUNTER()							\
     80    do {									\
     81       int cycle_i;							\
     82       counter_overhead = LONG_MAX;					\
     83       for ( cycle_i = 0 ; cycle_i < 8 ; cycle_i++ ) {			\
     84 	 long cycle_tmp1 = 0, cycle_tmp2 = 0;				\
     85 	 __asm__ __volatile__ ( "push %%ebx       \n"			\
     86 				"xor %%eax, %%eax \n"			\
     87 				"cpuid            \n"			\
     88 				"rdtsc            \n"			\
     89 				"mov %%eax, %0    \n"			\
     90 				"xor %%eax, %%eax \n"			\
     91 				"cpuid            \n"			\
     92 				"pop %%ebx        \n"			\
     93 				"push %%ebx       \n"			\
     94 				"xor %%eax, %%eax \n"			\
     95 				"cpuid            \n"			\
     96 				"rdtsc            \n"			\
     97 				"mov %%eax, %1    \n"			\
     98 				"xor %%eax, %%eax \n"			\
     99 				"cpuid            \n"			\
    100 				"pop %%ebx        \n"			\
    101 				: "=m" (cycle_tmp1), "=m" (cycle_tmp2)	\
    102 				: : "eax", "ecx", "edx" );		\
    103 	 if ( counter_overhead > (cycle_tmp2 - cycle_tmp1) ) {		\
    104 	    counter_overhead = cycle_tmp2 - cycle_tmp1;			\
    105 	 }								\
    106       }									\
    107    } while (0)
    108 
    109 #define  BEGIN_RACE(x)							\
    110    x = LONG_MAX;							\
    111    for ( cycle_i = 0 ; cycle_i < 10 ; cycle_i++ ) {			\
    112       long cycle_tmp1 = 0, cycle_tmp2 = 0;				\
    113       __asm__ __volatile__ ( "push %%ebx       \n"			\
    114 			     "xor %%eax, %%eax \n"			\
    115 			     "cpuid            \n"			\
    116 			     "rdtsc            \n"			\
    117 			     "mov %%eax, %0    \n"			\
    118 			     "xor %%eax, %%eax \n"			\
    119 			     "cpuid            \n"			\
    120 			     "pop %%ebx        \n"			\
    121 			     : "=m" (cycle_tmp1)			\
    122 			     : : "eax", "ecx", "edx" );
    123 
    124 #define END_RACE(x)							\
    125       __asm__ __volatile__ ( "push %%ebx       \n"			\
    126 			     "xor %%eax, %%eax \n"			\
    127 			     "cpuid            \n"			\
    128 			     "rdtsc            \n"			\
    129 			     "mov %%eax, %0    \n"			\
    130 			     "xor %%eax, %%eax \n"			\
    131 			     "cpuid            \n"			\
    132 			     "pop %%ebx        \n"			\
    133 			     : "=m" (cycle_tmp2)			\
    134 			     : : "eax", "ecx", "edx" );			\
    135       if ( x > (cycle_tmp2 - cycle_tmp1) ) {				\
    136 	 x = cycle_tmp2 - cycle_tmp1;					\
    137       }									\
    138    }									\
    139    x -= counter_overhead;
    140 
    141 #else /* PPlain, PMMX version */
    142 
    143 /* To ensure accurate results, we stall the pipelines with the
    144  * non-pairable 'cdq' instruction.  This ensures all the code being
    145  * profiled is complete when the 'rdtsc' instruction executes.
    146  */
    147 #define  INIT_COUNTER(x)						\
    148    do {									\
    149       int cycle_i;							\
    150       x = LONG_MAX;							\
    151       for ( cycle_i = 0 ; cycle_i < 32 ; cycle_i++ ) {			\
    152 	 long cycle_tmp1, cycle_tmp2, dummy;				\
    153 	 __asm__ ( "mov %%eax, %0" : "=a" (cycle_tmp1) );		\
    154 	 __asm__ ( "mov %%eax, %0" : "=a" (cycle_tmp2) );		\
    155 	 __asm__ ( "cdq" );						\
    156 	 __asm__ ( "cdq" );						\
    157 	 __asm__ ( "rdtsc" : "=a" (cycle_tmp1), "=d" (dummy) );		\
    158 	 __asm__ ( "cdq" );						\
    159 	 __asm__ ( "cdq" );						\
    160 	 __asm__ ( "rdtsc" : "=a" (cycle_tmp2), "=d" (dummy) );		\
    161 	 if ( x > (cycle_tmp2 - cycle_tmp1) )				\
    162 	    x = cycle_tmp2 - cycle_tmp1;				\
    163       }									\
    164    } while (0)
    165 
    166 #define  BEGIN_RACE(x)							\
    167    x = LONG_MAX;							\
    168    for ( cycle_i = 0 ; cycle_i < 16 ; cycle_i++ ) {			\
    169       long cycle_tmp1, cycle_tmp2, dummy;				\
    170       __asm__ ( "mov %%eax, %0" : "=a" (cycle_tmp1) );			\
    171       __asm__ ( "mov %%eax, %0" : "=a" (cycle_tmp2) );			\
    172       __asm__ ( "cdq" );						\
    173       __asm__ ( "cdq" );						\
    174       __asm__ ( "rdtsc" : "=a" (cycle_tmp1), "=d" (dummy) );
    175 
    176 
    177 #define END_RACE(x)							\
    178       __asm__ ( "cdq" );						\
    179       __asm__ ( "cdq" );						\
    180       __asm__ ( "rdtsc" : "=a" (cycle_tmp2), "=d" (dummy) );		\
    181       if ( x > (cycle_tmp2 - cycle_tmp1) )				\
    182 	 x = cycle_tmp2 - cycle_tmp1;					\
    183    }									\
    184    x -= counter_overhead;
    185 
    186 #endif
    187 
    188 #elif defined(__x86_64__)
    189 
    190 #define rdtscll(val) do { \
    191      unsigned int a,d; \
    192      __asm__ volatile("rdtsc" : "=a" (a), "=d" (d)); \
    193      (val) = ((unsigned long)a) | (((unsigned long)d)<<32); \
    194 } while(0)
    195 
    196 /* Copied from i386 PIII version */
    197 #define  INIT_COUNTER()							\
    198    do {									\
    199       int cycle_i;							\
    200       counter_overhead = LONG_MAX;					\
    201       for ( cycle_i = 0 ; cycle_i < 16 ; cycle_i++ ) {			\
    202 	 unsigned long cycle_tmp1, cycle_tmp2;        			\
    203 	 rdtscll(cycle_tmp1);						\
    204 	 rdtscll(cycle_tmp2);						\
    205 	 if ( counter_overhead > (cycle_tmp2 - cycle_tmp1) ) {		\
    206 	    counter_overhead = cycle_tmp2 - cycle_tmp1;			\
    207 	 }								\
    208       }									\
    209    } while (0)
    210 
    211 
    212 #define  BEGIN_RACE(x)							\
    213    x = LONG_MAX;							\
    214    for ( cycle_i = 0 ; cycle_i < 10 ; cycle_i++ ) {			\
    215       unsigned long cycle_tmp1, cycle_tmp2;				\
    216       rdtscll(cycle_tmp1);						\
    217 
    218 #define END_RACE(x)							\
    219       rdtscll(cycle_tmp2);						\
    220       if ( x > (cycle_tmp2 - cycle_tmp1) ) {				\
    221 	 x = cycle_tmp2 - cycle_tmp1;					\
    222       }									\
    223    }									\
    224    x -= counter_overhead;
    225 
    226 #elif defined(__sparc__)
    227 
    228 #define  INIT_COUNTER()	\
    229 	 do { counter_overhead = 5; } while(0)
    230 
    231 #define  BEGIN_RACE(x)                                                        \
    232 x = LONG_MAX;                                                                 \
    233 for (cycle_i = 0; cycle_i <10; cycle_i++) {                                   \
    234    register long cycle_tmp1 __asm__("l0");				      \
    235    register long cycle_tmp2 __asm__("l1");				      \
    236    /* rd %tick, %l0 */							      \
    237    __asm__ __volatile__ (".word 0xa1410000" : "=r" (cycle_tmp1));  /*  save timestamp   */
    238 
    239 #define END_RACE(x)                                                           \
    240    /* rd %tick, %l1 */							      \
    241    __asm__ __volatile__ (".word 0xa3410000" : "=r" (cycle_tmp2));	      \
    242    if (x > (cycle_tmp2-cycle_tmp1)) x = cycle_tmp2 - cycle_tmp1;              \
    243 }                                                                             \
    244 x -= counter_overhead;
    245 
    246 #else
    247 #error Your processor is not supported for RUN_XFORM_BENCHMARK
    248 #endif
    249 
    250 #else
    251 
    252 #define BEGIN_RACE(x)
    253 #define END_RACE(x)
    254 
    255 #endif
    256 
    257 
    258 /* =============================================================
    259  * Helper functions
    260  */
    261 
    262 static GLfloat rnd( void )
    263 {
    264    GLfloat f = (GLfloat)rand() / (GLfloat)RAND_MAX;
    265    GLfloat gran = (GLfloat)(1 << 13);
    266 
    267    f = (GLfloat)(GLint)(f * gran) / gran;
    268 
    269    return f * 2.0 - 1.0;
    270 }
    271 
    272 static int significand_match( GLfloat a, GLfloat b )
    273 {
    274    GLfloat d = a - b;
    275    int a_ex, b_ex, d_ex;
    276 
    277    if ( d == 0.0F ) {
    278       return MAX_PRECISION;   /* Exact match */
    279    }
    280 
    281    if ( a == 0.0F || b == 0.0F ) {
    282       /* It would probably be better to check if the
    283        * non-zero number is denormalized and return
    284        * the index of the highest set bit here.
    285        */
    286       return 0;
    287    }
    288 
    289    FREXPF( a, &a_ex );
    290    FREXPF( b, &b_ex );
    291    FREXPF( d, &d_ex );
    292 
    293    if ( a_ex < b_ex ) {
    294       return a_ex - d_ex;
    295    } else {
    296       return b_ex - d_ex;
    297    }
    298 }
    299 
    300 enum { NIL = 0, ONE = 1, NEG = -1, VAR = 2 };
    301 
    302 /* Ensure our arrays are correctly aligned.
    303  */
    304 #if defined(__GNUC__)
    305 #  define ALIGN16(type, array)	type array __attribute__ ((aligned (16)))
    306 #elif defined(_MSC_VER)
    307 #  define ALIGN16(type, array)	type array __declspec(align(16)) /* GH: Does this work? */
    308 #elif defined(__WATCOMC__)
    309 #  define ALIGN16(type, array)	                    /* Watcom does not support this */
    310 #elif defined(__xlC__)
    311 #  define ALIGN16(type, array)       type __align (16) array
    312 #else
    313 #  warning "ALIGN16 will not 16-byte align!\n"
    314 #  define ALIGN16
    315 #endif
    316 
    317 
    318 #endif /* DEBUG_MATH */
    319 
    320 #endif /* __M_DEBUG_UTIL_H__ */
    321