diff options
Diffstat (limited to 'cachegrind/cg-x86.c')
-rw-r--r-- | cachegrind/cg-x86.c | 352 |
1 files changed, 352 insertions, 0 deletions
diff --git a/cachegrind/cg-x86.c b/cachegrind/cg-x86.c new file mode 100644 index 0000000..be5eb82 --- /dev/null +++ b/cachegrind/cg-x86.c @@ -0,0 +1,352 @@ + +/*--------------------------------------------------------------------*/ +/*--- x86-specific (and AMD64-specific) definitions. cg-x86.c ---*/ +/*--------------------------------------------------------------------*/ + +/* + This file is part of Cachegrind, a Valgrind tool for cache + profiling programs. + + Copyright (C) 2002-2009 Nicholas Nethercote + njn@valgrind.org + + This program is free software; you can redistribute it and/or + modify it under the terms of the GNU General Public License as + published by the Free Software Foundation; either version 2 of the + License, or (at your option) any later version. + + This program is distributed in the hope that it will be useful, but + WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + General Public License for more details. + + You should have received a copy of the GNU General Public License + along with this program; if not, write to the Free Software + Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA + 02111-1307, USA. + + The GNU General Public License is contained in the file COPYING. +*/ + +#include "pub_tool_basics.h" +#include "pub_tool_cpuid.h" +#include "pub_tool_libcbase.h" +#include "pub_tool_libcassert.h" +#include "pub_tool_libcprint.h" + +#include "cg_arch.h" + +// All CPUID info taken from sandpile.org/a32/cpuid.htm */ +// Probably only works for Intel and AMD chips, and probably only for some of +// them. + +static void micro_ops_warn(Int actual_size, Int used_size, Int line_size) +{ + VG_DMSG("warning: Pentium 4 with %d KB micro-op instruction trace cache", + actual_size); + VG_DMSG(" Simulating a %d KB I-cache with %d B lines", + used_size, line_size); +} + +/* Intel method is truly wretched. We have to do an insane indexing into an + * array of pre-defined configurations for various parts of the memory + * hierarchy. + * According to Intel Processor Identification, App Note 485. + */ +static +Int Intel_cache_info(Int level, cache_t* I1c, cache_t* D1c, cache_t* L2c) +{ + Int cpuid1_eax; + Int cpuid1_ignore; + Int family; + Int model; + UChar info[16]; + Int i, trials; + Bool L2_found = False; + + if (level < 2) { + VG_DMSG("warning: CPUID level < 2 for Intel processor (%d)", level); + return -1; + } + + /* family/model needed to distinguish code reuse (currently 0x49) */ + VG_(cpuid)(1, &cpuid1_eax, &cpuid1_ignore, + &cpuid1_ignore, &cpuid1_ignore); + family = (((cpuid1_eax >> 20) & 0xff) << 4) + ((cpuid1_eax >> 8) & 0xf); + model = (((cpuid1_eax >> 16) & 0xf) << 4) + ((cpuid1_eax >> 4) & 0xf); + + VG_(cpuid)(2, (Int*)&info[0], (Int*)&info[4], + (Int*)&info[8], (Int*)&info[12]); + trials = info[0] - 1; /* AL register - bits 0..7 of %eax */ + info[0] = 0x0; /* reset AL */ + + if (0 != trials) { + VG_DMSG("warning: non-zero CPUID trials for Intel processor (%d)", + trials); + return -1; + } + + for (i = 0; i < 16; i++) { + + switch (info[i]) { + + case 0x0: /* ignore zeros */ + break; + + /* TLB info, ignore */ + case 0x01: case 0x02: case 0x03: case 0x04: case 0x05: + case 0x4f: case 0x50: case 0x51: case 0x52: + case 0x56: case 0x57: case 0x59: + case 0x5b: case 0x5c: case 0x5d: + case 0xb0: case 0xb1: + case 0xb3: case 0xb4: case 0xba: case 0xc0: + break; + + case 0x06: *I1c = (cache_t) { 8, 4, 32 }; break; + case 0x08: *I1c = (cache_t) { 16, 4, 32 }; break; + case 0x30: *I1c = (cache_t) { 32, 8, 64 }; break; + + case 0x0a: *D1c = (cache_t) { 8, 2, 32 }; break; + case 0x0c: *D1c = (cache_t) { 16, 4, 32 }; break; + case 0x0e: *D1c = (cache_t) { 24, 6, 64 }; break; + case 0x2c: *D1c = (cache_t) { 32, 8, 64 }; break; + + /* IA-64 info -- panic! */ + case 0x10: case 0x15: case 0x1a: + case 0x88: case 0x89: case 0x8a: case 0x8d: + case 0x90: case 0x96: case 0x9b: + VG_(tool_panic)("IA-64 cache detected?!"); + + case 0x22: case 0x23: case 0x25: case 0x29: + case 0x46: case 0x47: case 0x4a: case 0x4b: case 0x4c: case 0x4d: + VG_DMSG("warning: L3 cache detected but ignored"); + break; + + /* These are sectored, whatever that means */ + case 0x39: *L2c = (cache_t) { 128, 4, 64 }; L2_found = True; break; + case 0x3c: *L2c = (cache_t) { 256, 4, 64 }; L2_found = True; break; + + /* If a P6 core, this means "no L2 cache". + If a P4 core, this means "no L3 cache". + We don't know what core it is, so don't issue a warning. To detect + a missing L2 cache, we use 'L2_found'. */ + case 0x40: + break; + + case 0x41: *L2c = (cache_t) { 128, 4, 32 }; L2_found = True; break; + case 0x42: *L2c = (cache_t) { 256, 4, 32 }; L2_found = True; break; + case 0x43: *L2c = (cache_t) { 512, 4, 32 }; L2_found = True; break; + case 0x44: *L2c = (cache_t) { 1024, 4, 32 }; L2_found = True; break; + case 0x45: *L2c = (cache_t) { 2048, 4, 32 }; L2_found = True; break; + case 0x48: *L2c = (cache_t) { 3072,12, 64 }; L2_found = True; break; + case 0x49: + if ((family == 15) && (model == 6)) + /* On Xeon MP (family F, model 6), this is for L3 */ + VG_DMSG("warning: L3 cache detected but ignored"); + else + *L2c = (cache_t) { 4096, 16, 64 }; L2_found = True; + break; + case 0x4e: *L2c = (cache_t) { 6144, 24, 64 }; L2_found = True; break; + + /* These are sectored, whatever that means */ + case 0x60: *D1c = (cache_t) { 16, 8, 64 }; break; /* sectored */ + case 0x66: *D1c = (cache_t) { 8, 4, 64 }; break; /* sectored */ + case 0x67: *D1c = (cache_t) { 16, 4, 64 }; break; /* sectored */ + case 0x68: *D1c = (cache_t) { 32, 4, 64 }; break; /* sectored */ + + /* HACK ALERT: Instruction trace cache -- capacity is micro-ops based. + * conversion to byte size is a total guess; treat the 12K and 16K + * cases the same since the cache byte size must be a power of two for + * everything to work!. Also guessing 32 bytes for the line size... + */ + case 0x70: /* 12K micro-ops, 8-way */ + *I1c = (cache_t) { 16, 8, 32 }; + micro_ops_warn(12, 16, 32); + break; + case 0x71: /* 16K micro-ops, 8-way */ + *I1c = (cache_t) { 16, 8, 32 }; + micro_ops_warn(16, 16, 32); + break; + case 0x72: /* 32K micro-ops, 8-way */ + *I1c = (cache_t) { 32, 8, 32 }; + micro_ops_warn(32, 32, 32); + break; + + /* These are sectored, whatever that means */ + case 0x79: *L2c = (cache_t) { 128, 8, 64 }; L2_found = True; break; + case 0x7a: *L2c = (cache_t) { 256, 8, 64 }; L2_found = True; break; + case 0x7b: *L2c = (cache_t) { 512, 8, 64 }; L2_found = True; break; + case 0x7c: *L2c = (cache_t) { 1024, 8, 64 }; L2_found = True; break; + case 0x7d: *L2c = (cache_t) { 2048, 8, 64 }; L2_found = True; break; + case 0x7e: *L2c = (cache_t) { 256, 8, 128 }; L2_found = True; break; + + case 0x7f: *L2c = (cache_t) { 512, 2, 64 }; L2_found = True; break; + case 0x80: *L2c = (cache_t) { 512, 8, 64 }; L2_found = True; break; + + case 0x81: *L2c = (cache_t) { 128, 8, 32 }; L2_found = True; break; + case 0x82: *L2c = (cache_t) { 256, 8, 32 }; L2_found = True; break; + case 0x83: *L2c = (cache_t) { 512, 8, 32 }; L2_found = True; break; + case 0x84: *L2c = (cache_t) { 1024, 8, 32 }; L2_found = True; break; + case 0x85: *L2c = (cache_t) { 2048, 8, 32 }; L2_found = True; break; + case 0x86: *L2c = (cache_t) { 512, 4, 64 }; L2_found = True; break; + case 0x87: *L2c = (cache_t) { 1024, 8, 64 }; L2_found = True; break; + + /* Ignore prefetch information */ + case 0xf0: case 0xf1: + break; + + default: + VG_DMSG("warning: Unknown Intel cache config value (0x%x), ignoring", + info[i]); + break; + } + } + + if (!L2_found) + VG_DMSG("warning: L2 cache not installed, ignore L2 results."); + + return 0; +} + +/* AMD method is straightforward, just extract appropriate bits from the + * result registers. + * + * Bits, for D1 and I1: + * 31..24 data L1 cache size in KBs + * 23..16 data L1 cache associativity (FFh=full) + * 15.. 8 data L1 cache lines per tag + * 7.. 0 data L1 cache line size in bytes + * + * Bits, for L2: + * 31..16 unified L2 cache size in KBs + * 15..12 unified L2 cache associativity (0=off, FFh=full) + * 11.. 8 unified L2 cache lines per tag + * 7.. 0 unified L2 cache line size in bytes + * + * #3 The AMD K7 processor's L2 cache must be configured prior to relying + * upon this information. (Whatever that means -- njn) + * + * Also, according to Cyrille Chepelov, Duron stepping A0 processors (model + * 0x630) have a bug and misreport their L2 size as 1KB (it's really 64KB), + * so we detect that. + * + * Returns 0 on success, non-zero on failure. + */ +static +Int AMD_cache_info(cache_t* I1c, cache_t* D1c, cache_t* L2c) +{ + UInt ext_level; + UInt dummy, model; + UInt I1i, D1i, L2i; + + VG_(cpuid)(0x80000000, &ext_level, &dummy, &dummy, &dummy); + + if (0 == (ext_level & 0x80000000) || ext_level < 0x80000006) { + VG_DMSG("warning: ext_level < 0x80000006 for AMD processor (0x%x)", + ext_level); + return -1; + } + + VG_(cpuid)(0x80000005, &dummy, &dummy, &D1i, &I1i); + VG_(cpuid)(0x80000006, &dummy, &dummy, &L2i, &dummy); + + VG_(cpuid)(0x1, &model, &dummy, &dummy, &dummy); + + /* Check for Duron bug */ + if (model == 0x630) { + VG_DMSG("warning: Buggy Duron stepping A0. Assuming L2 size=65536 bytes"); + L2i = (64 << 16) | (L2i & 0xffff); + } + + D1c->size = (D1i >> 24) & 0xff; + D1c->assoc = (D1i >> 16) & 0xff; + D1c->line_size = (D1i >> 0) & 0xff; + + I1c->size = (I1i >> 24) & 0xff; + I1c->assoc = (I1i >> 16) & 0xff; + I1c->line_size = (I1i >> 0) & 0xff; + + L2c->size = (L2i >> 16) & 0xffff; /* Nb: different bits used for L2 */ + L2c->assoc = (L2i >> 12) & 0xf; + L2c->line_size = (L2i >> 0) & 0xff; + + return 0; +} + +static +Int get_caches_from_CPUID(cache_t* I1c, cache_t* D1c, cache_t* L2c) +{ + Int level, ret; + Char vendor_id[13]; + + if (!VG_(has_cpuid)()) { + VG_DMSG("CPUID instruction not supported"); + return -1; + } + + VG_(cpuid)(0, &level, (int*)&vendor_id[0], + (int*)&vendor_id[8], (int*)&vendor_id[4]); + vendor_id[12] = '\0'; + + if (0 == level) { + VG_DMSG("CPUID level is 0, early Pentium?"); + return -1; + } + + /* Only handling Intel and AMD chips... no Cyrix, Transmeta, etc */ + if (0 == VG_(strcmp)(vendor_id, "GenuineIntel")) { + ret = Intel_cache_info(level, I1c, D1c, L2c); + + } else if (0 == VG_(strcmp)(vendor_id, "AuthenticAMD")) { + ret = AMD_cache_info(I1c, D1c, L2c); + + } else if (0 == VG_(strcmp)(vendor_id, "CentaurHauls")) { + /* Total kludge. Pretend to be a VIA Nehemiah. */ + D1c->size = 64; + D1c->assoc = 16; + D1c->line_size = 16; + I1c->size = 64; + I1c->assoc = 4; + I1c->line_size = 16; + L2c->size = 64; + L2c->assoc = 16; + L2c->line_size = 16; + ret = 0; + + } else { + VG_DMSG("CPU vendor ID not recognised (%s)", vendor_id); + return -1; + } + + /* Successful! Convert sizes from KB to bytes */ + I1c->size *= 1024; + D1c->size *= 1024; + L2c->size *= 1024; + + return ret; +} + + +void VG_(configure_caches)(cache_t* I1c, cache_t* D1c, cache_t* L2c, + Bool all_caches_clo_defined) +{ + Int res; + + // Set caches to default. + *I1c = (cache_t) { 65536, 2, 64 }; + *D1c = (cache_t) { 65536, 2, 64 }; + *L2c = (cache_t) { 262144, 8, 64 }; + + // Then replace with any info we can get from CPUID. + res = get_caches_from_CPUID(I1c, D1c, L2c); + + // Warn if CPUID failed and config not completely specified from cmd line. + if (res != 0 && !all_caches_clo_defined) { + VG_DMSG("Warning: Couldn't auto-detect cache config, using one " + "or more defaults "); + } +} + +/*--------------------------------------------------------------------*/ +/*--- end ---*/ +/*--------------------------------------------------------------------*/ |