提交 42cbd8ef 编写于 作者: L Linus Torvalds

Merge branch 'x86-amd-nb-for-linus' of...

Merge branch 'x86-amd-nb-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/linux-2.6-tip

* 'x86-amd-nb-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/linux-2.6-tip:
  x86, cacheinfo: Cleanup L3 cache index disable support
  x86, amd-nb: Cleanup AMD northbridge caching code
  x86, amd-nb: Complete the rename of AMD NB and related code
...@@ -1141,16 +1141,16 @@ config NUMA ...@@ -1141,16 +1141,16 @@ config NUMA
comment "NUMA (Summit) requires SMP, 64GB highmem support, ACPI" comment "NUMA (Summit) requires SMP, 64GB highmem support, ACPI"
depends on X86_32 && X86_SUMMIT && (!HIGHMEM64G || !ACPI) depends on X86_32 && X86_SUMMIT && (!HIGHMEM64G || !ACPI)
config K8_NUMA config AMD_NUMA
def_bool y def_bool y
prompt "Old style AMD Opteron NUMA detection" prompt "Old style AMD Opteron NUMA detection"
depends on X86_64 && NUMA && PCI depends on X86_64 && NUMA && PCI
---help--- ---help---
Enable K8 NUMA node topology detection. You should say Y here if Enable AMD NUMA node topology detection. You should say Y here if
you have a multi processor AMD K8 system. This uses an old you have a multi processor AMD system. This uses an old method to
method to read the NUMA configuration directly from the builtin read the NUMA configuration directly from the builtin Northbridge
Northbridge of Opteron. It is recommended to use X86_64_ACPI_NUMA of Opteron. It is recommended to use X86_64_ACPI_NUMA instead,
instead, which also takes priority if both are compiled in. which also takes priority if both are compiled in.
config X86_64_ACPI_NUMA config X86_64_ACPI_NUMA
def_bool y def_bool y
......
...@@ -3,36 +3,53 @@ ...@@ -3,36 +3,53 @@
#include <linux/pci.h> #include <linux/pci.h>
extern struct pci_device_id k8_nb_ids[]; extern struct pci_device_id amd_nb_misc_ids[];
struct bootnode; struct bootnode;
extern int early_is_k8_nb(u32 value); extern int early_is_amd_nb(u32 value);
extern int cache_k8_northbridges(void); extern int amd_cache_northbridges(void);
extern void k8_flush_garts(void); extern void amd_flush_garts(void);
extern int k8_get_nodes(struct bootnode *nodes); extern int amd_get_nodes(struct bootnode *nodes);
extern int k8_numa_init(unsigned long start_pfn, unsigned long end_pfn); extern int amd_numa_init(unsigned long start_pfn, unsigned long end_pfn);
extern int k8_scan_nodes(void); extern int amd_scan_nodes(void);
struct k8_northbridge_info { struct amd_northbridge {
struct pci_dev *misc;
};
struct amd_northbridge_info {
u16 num; u16 num;
u8 gart_supported; u64 flags;
struct pci_dev **nb_misc; struct amd_northbridge *nb;
}; };
extern struct k8_northbridge_info k8_northbridges; extern struct amd_northbridge_info amd_northbridges;
#define AMD_NB_GART 0x1
#define AMD_NB_L3_INDEX_DISABLE 0x2
#ifdef CONFIG_AMD_NB #ifdef CONFIG_AMD_NB
static inline struct pci_dev *node_to_k8_nb_misc(int node) static inline int amd_nb_num(void)
{ {
return (node < k8_northbridges.num) ? k8_northbridges.nb_misc[node] : NULL; return amd_northbridges.num;
} }
#else static inline int amd_nb_has_feature(int feature)
{
return ((amd_northbridges.flags & feature) == feature);
}
static inline struct pci_dev *node_to_k8_nb_misc(int node) static inline struct amd_northbridge *node_to_amd_nb(int node)
{ {
return NULL; return (node < amd_northbridges.num) ? &amd_northbridges.nb[node] : NULL;
} }
#else
#define amd_nb_num(x) 0
#define amd_nb_has_feature(x) false
#define node_to_amd_nb(x) NULL
#endif #endif
......
...@@ -12,95 +12,116 @@ ...@@ -12,95 +12,116 @@
static u32 *flush_words; static u32 *flush_words;
struct pci_device_id k8_nb_ids[] = { struct pci_device_id amd_nb_misc_ids[] = {
{ PCI_DEVICE(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_K8_NB_MISC) }, { PCI_DEVICE(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_K8_NB_MISC) },
{ PCI_DEVICE(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_10H_NB_MISC) }, { PCI_DEVICE(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_10H_NB_MISC) },
{ PCI_DEVICE(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_15H_NB_MISC) }, { PCI_DEVICE(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_15H_NB_MISC) },
{} {}
}; };
EXPORT_SYMBOL(k8_nb_ids); EXPORT_SYMBOL(amd_nb_misc_ids);
struct k8_northbridge_info k8_northbridges; struct amd_northbridge_info amd_northbridges;
EXPORT_SYMBOL(k8_northbridges); EXPORT_SYMBOL(amd_northbridges);
static struct pci_dev *next_k8_northbridge(struct pci_dev *dev) static struct pci_dev *next_northbridge(struct pci_dev *dev,
struct pci_device_id *ids)
{ {
do { do {
dev = pci_get_device(PCI_ANY_ID, PCI_ANY_ID, dev); dev = pci_get_device(PCI_ANY_ID, PCI_ANY_ID, dev);
if (!dev) if (!dev)
break; break;
} while (!pci_match_id(&k8_nb_ids[0], dev)); } while (!pci_match_id(ids, dev));
return dev; return dev;
} }
int cache_k8_northbridges(void) int amd_cache_northbridges(void)
{ {
int i; int i = 0;
struct pci_dev *dev; struct amd_northbridge *nb;
struct pci_dev *misc;
if (k8_northbridges.num) if (amd_nb_num())
return 0; return 0;
dev = NULL; misc = NULL;
while ((dev = next_k8_northbridge(dev)) != NULL) while ((misc = next_northbridge(misc, amd_nb_misc_ids)) != NULL)
k8_northbridges.num++; i++;
/* some CPU families (e.g. family 0x11) do not support GART */ if (i == 0)
if (boot_cpu_data.x86 == 0xf || boot_cpu_data.x86 == 0x10 || return 0;
boot_cpu_data.x86 == 0x15)
k8_northbridges.gart_supported = 1;
k8_northbridges.nb_misc = kmalloc((k8_northbridges.num + 1) * nb = kzalloc(i * sizeof(struct amd_northbridge), GFP_KERNEL);
sizeof(void *), GFP_KERNEL); if (!nb)
if (!k8_northbridges.nb_misc)
return -ENOMEM; return -ENOMEM;
if (!k8_northbridges.num) { amd_northbridges.nb = nb;
k8_northbridges.nb_misc[0] = NULL; amd_northbridges.num = i;
return 0;
}
if (k8_northbridges.gart_supported) { misc = NULL;
flush_words = kmalloc(k8_northbridges.num * sizeof(u32), for (i = 0; i != amd_nb_num(); i++) {
GFP_KERNEL); node_to_amd_nb(i)->misc = misc =
if (!flush_words) { next_northbridge(misc, amd_nb_misc_ids);
kfree(k8_northbridges.nb_misc);
return -ENOMEM;
}
} }
dev = NULL; /* some CPU families (e.g. family 0x11) do not support GART */
i = 0; if (boot_cpu_data.x86 == 0xf || boot_cpu_data.x86 == 0x10 ||
while ((dev = next_k8_northbridge(dev)) != NULL) { boot_cpu_data.x86 == 0x15)
k8_northbridges.nb_misc[i] = dev; amd_northbridges.flags |= AMD_NB_GART;
if (k8_northbridges.gart_supported)
pci_read_config_dword(dev, 0x9c, &flush_words[i++]); /*
} * Some CPU families support L3 Cache Index Disable. There are some
k8_northbridges.nb_misc[i] = NULL; * limitations because of E382 and E388 on family 0x10.
*/
if (boot_cpu_data.x86 == 0x10 &&
boot_cpu_data.x86_model >= 0x8 &&
(boot_cpu_data.x86_model > 0x9 ||
boot_cpu_data.x86_mask >= 0x1))
amd_northbridges.flags |= AMD_NB_L3_INDEX_DISABLE;
return 0; return 0;
} }
EXPORT_SYMBOL_GPL(cache_k8_northbridges); EXPORT_SYMBOL_GPL(amd_cache_northbridges);
/* Ignores subdevice/subvendor but as far as I can figure out /* Ignores subdevice/subvendor but as far as I can figure out
they're useless anyways */ they're useless anyways */
int __init early_is_k8_nb(u32 device) int __init early_is_amd_nb(u32 device)
{ {
struct pci_device_id *id; struct pci_device_id *id;
u32 vendor = device & 0xffff; u32 vendor = device & 0xffff;
device >>= 16; device >>= 16;
for (id = k8_nb_ids; id->vendor; id++) for (id = amd_nb_misc_ids; id->vendor; id++)
if (vendor == id->vendor && device == id->device) if (vendor == id->vendor && device == id->device)
return 1; return 1;
return 0; return 0;
} }
void k8_flush_garts(void) int amd_cache_gart(void)
{
int i;
if (!amd_nb_has_feature(AMD_NB_GART))
return 0;
flush_words = kmalloc(amd_nb_num() * sizeof(u32), GFP_KERNEL);
if (!flush_words) {
amd_northbridges.flags &= ~AMD_NB_GART;
return -ENOMEM;
}
for (i = 0; i != amd_nb_num(); i++)
pci_read_config_dword(node_to_amd_nb(i)->misc, 0x9c,
&flush_words[i]);
return 0;
}
void amd_flush_garts(void)
{ {
int flushed, i; int flushed, i;
unsigned long flags; unsigned long flags;
static DEFINE_SPINLOCK(gart_lock); static DEFINE_SPINLOCK(gart_lock);
if (!k8_northbridges.gart_supported) if (!amd_nb_has_feature(AMD_NB_GART))
return; return;
/* Avoid races between AGP and IOMMU. In theory it's not needed /* Avoid races between AGP and IOMMU. In theory it's not needed
...@@ -109,16 +130,16 @@ void k8_flush_garts(void) ...@@ -109,16 +130,16 @@ void k8_flush_garts(void)
that it doesn't matter to serialize more. -AK */ that it doesn't matter to serialize more. -AK */
spin_lock_irqsave(&gart_lock, flags); spin_lock_irqsave(&gart_lock, flags);
flushed = 0; flushed = 0;
for (i = 0; i < k8_northbridges.num; i++) { for (i = 0; i < amd_nb_num(); i++) {
pci_write_config_dword(k8_northbridges.nb_misc[i], 0x9c, pci_write_config_dword(node_to_amd_nb(i)->misc, 0x9c,
flush_words[i]|1); flush_words[i] | 1);
flushed++; flushed++;
} }
for (i = 0; i < k8_northbridges.num; i++) { for (i = 0; i < amd_nb_num(); i++) {
u32 w; u32 w;
/* Make sure the hardware actually executed the flush*/ /* Make sure the hardware actually executed the flush*/
for (;;) { for (;;) {
pci_read_config_dword(k8_northbridges.nb_misc[i], pci_read_config_dword(node_to_amd_nb(i)->misc,
0x9c, &w); 0x9c, &w);
if (!(w & 1)) if (!(w & 1))
break; break;
...@@ -129,19 +150,23 @@ void k8_flush_garts(void) ...@@ -129,19 +150,23 @@ void k8_flush_garts(void)
if (!flushed) if (!flushed)
printk("nothing to flush?\n"); printk("nothing to flush?\n");
} }
EXPORT_SYMBOL_GPL(k8_flush_garts); EXPORT_SYMBOL_GPL(amd_flush_garts);
static __init int init_k8_nbs(void) static __init int init_amd_nbs(void)
{ {
int err = 0; int err = 0;
err = cache_k8_northbridges(); err = amd_cache_northbridges();
if (err < 0) if (err < 0)
printk(KERN_NOTICE "K8 NB: Cannot enumerate AMD northbridges.\n"); printk(KERN_NOTICE "AMD NB: Cannot enumerate AMD northbridges.\n");
if (amd_cache_gart() < 0)
printk(KERN_NOTICE "AMD NB: Cannot initialize GART flush words, "
"GART support disabled.\n");
return err; return err;
} }
/* This has to go after the PCI subsystem */ /* This has to go after the PCI subsystem */
fs_initcall(init_k8_nbs); fs_initcall(init_amd_nbs);
...@@ -206,7 +206,7 @@ static u32 __init read_agp(int bus, int slot, int func, int cap, u32 *order) ...@@ -206,7 +206,7 @@ static u32 __init read_agp(int bus, int slot, int func, int cap, u32 *order)
* Do an PCI bus scan by hand because we're running before the PCI * Do an PCI bus scan by hand because we're running before the PCI
* subsystem. * subsystem.
* *
* All K8 AGP bridges are AGPv3 compliant, so we can do this scan * All AMD AGP bridges are AGPv3 compliant, so we can do this scan
* generically. It's probably overkill to always scan all slots because * generically. It's probably overkill to always scan all slots because
* the AGP bridges should be always an own bus on the HT hierarchy, * the AGP bridges should be always an own bus on the HT hierarchy,
* but do it here for future safety. * but do it here for future safety.
...@@ -303,7 +303,7 @@ void __init early_gart_iommu_check(void) ...@@ -303,7 +303,7 @@ void __init early_gart_iommu_check(void)
dev_limit = bus_dev_ranges[i].dev_limit; dev_limit = bus_dev_ranges[i].dev_limit;
for (slot = dev_base; slot < dev_limit; slot++) { for (slot = dev_base; slot < dev_limit; slot++) {
if (!early_is_k8_nb(read_pci_config(bus, slot, 3, 0x00))) if (!early_is_amd_nb(read_pci_config(bus, slot, 3, 0x00)))
continue; continue;
ctl = read_pci_config(bus, slot, 3, AMD64_GARTAPERTURECTL); ctl = read_pci_config(bus, slot, 3, AMD64_GARTAPERTURECTL);
...@@ -358,7 +358,7 @@ void __init early_gart_iommu_check(void) ...@@ -358,7 +358,7 @@ void __init early_gart_iommu_check(void)
dev_limit = bus_dev_ranges[i].dev_limit; dev_limit = bus_dev_ranges[i].dev_limit;
for (slot = dev_base; slot < dev_limit; slot++) { for (slot = dev_base; slot < dev_limit; slot++) {
if (!early_is_k8_nb(read_pci_config(bus, slot, 3, 0x00))) if (!early_is_amd_nb(read_pci_config(bus, slot, 3, 0x00)))
continue; continue;
ctl = read_pci_config(bus, slot, 3, AMD64_GARTAPERTURECTL); ctl = read_pci_config(bus, slot, 3, AMD64_GARTAPERTURECTL);
...@@ -400,7 +400,7 @@ int __init gart_iommu_hole_init(void) ...@@ -400,7 +400,7 @@ int __init gart_iommu_hole_init(void)
dev_limit = bus_dev_ranges[i].dev_limit; dev_limit = bus_dev_ranges[i].dev_limit;
for (slot = dev_base; slot < dev_limit; slot++) { for (slot = dev_base; slot < dev_limit; slot++) {
if (!early_is_k8_nb(read_pci_config(bus, slot, 3, 0x00))) if (!early_is_amd_nb(read_pci_config(bus, slot, 3, 0x00)))
continue; continue;
iommu_detected = 1; iommu_detected = 1;
...@@ -518,7 +518,7 @@ int __init gart_iommu_hole_init(void) ...@@ -518,7 +518,7 @@ int __init gart_iommu_hole_init(void)
dev_base = bus_dev_ranges[i].dev_base; dev_base = bus_dev_ranges[i].dev_base;
dev_limit = bus_dev_ranges[i].dev_limit; dev_limit = bus_dev_ranges[i].dev_limit;
for (slot = dev_base; slot < dev_limit; slot++) { for (slot = dev_base; slot < dev_limit; slot++) {
if (!early_is_k8_nb(read_pci_config(bus, slot, 3, 0x00))) if (!early_is_amd_nb(read_pci_config(bus, slot, 3, 0x00)))
continue; continue;
write_pci_config(bus, slot, 3, AMD64_GARTAPERTURECTL, ctl); write_pci_config(bus, slot, 3, AMD64_GARTAPERTURECTL, ctl);
......
...@@ -149,8 +149,7 @@ union _cpuid4_leaf_ecx { ...@@ -149,8 +149,7 @@ union _cpuid4_leaf_ecx {
}; };
struct amd_l3_cache { struct amd_l3_cache {
struct pci_dev *dev; struct amd_northbridge *nb;
bool can_disable;
unsigned indices; unsigned indices;
u8 subcaches[4]; u8 subcaches[4];
}; };
...@@ -311,14 +310,12 @@ struct _cache_attr { ...@@ -311,14 +310,12 @@ struct _cache_attr {
/* /*
* L3 cache descriptors * L3 cache descriptors
*/ */
static struct amd_l3_cache **__cpuinitdata l3_caches;
static void __cpuinit amd_calc_l3_indices(struct amd_l3_cache *l3) static void __cpuinit amd_calc_l3_indices(struct amd_l3_cache *l3)
{ {
unsigned int sc0, sc1, sc2, sc3; unsigned int sc0, sc1, sc2, sc3;
u32 val = 0; u32 val = 0;
pci_read_config_dword(l3->dev, 0x1C4, &val); pci_read_config_dword(l3->nb->misc, 0x1C4, &val);
/* calculate subcache sizes */ /* calculate subcache sizes */
l3->subcaches[0] = sc0 = !(val & BIT(0)); l3->subcaches[0] = sc0 = !(val & BIT(0));
...@@ -330,47 +327,14 @@ static void __cpuinit amd_calc_l3_indices(struct amd_l3_cache *l3) ...@@ -330,47 +327,14 @@ static void __cpuinit amd_calc_l3_indices(struct amd_l3_cache *l3)
l3->indices = (max(max3(sc0, sc1, sc2), sc3) << 10) - 1; l3->indices = (max(max3(sc0, sc1, sc2), sc3) << 10) - 1;
} }
static struct amd_l3_cache * __cpuinit amd_init_l3_cache(int node) static void __cpuinit amd_init_l3_cache(struct _cpuid4_info_regs *this_leaf,
{
struct amd_l3_cache *l3;
struct pci_dev *dev = node_to_k8_nb_misc(node);
l3 = kzalloc(sizeof(struct amd_l3_cache), GFP_ATOMIC);
if (!l3) {
printk(KERN_WARNING "Error allocating L3 struct\n");
return NULL;
}
l3->dev = dev;
amd_calc_l3_indices(l3);
return l3;
}
static void __cpuinit amd_check_l3_disable(struct _cpuid4_info_regs *this_leaf,
int index) int index)
{ {
static struct amd_l3_cache *__cpuinitdata l3_caches;
int node; int node;
if (boot_cpu_data.x86 != 0x10) /* only for L3, and not in virtualized environments */
return; if (index < 3 || amd_nb_num() == 0)
if (index < 3)
return;
/* see errata #382 and #388 */
if (boot_cpu_data.x86_model < 0x8)
return;
if ((boot_cpu_data.x86_model == 0x8 ||
boot_cpu_data.x86_model == 0x9)
&&
boot_cpu_data.x86_mask < 0x1)
return;
/* not in virtualized environments */
if (k8_northbridges.num == 0)
return; return;
/* /*
...@@ -378,7 +342,7 @@ static void __cpuinit amd_check_l3_disable(struct _cpuid4_info_regs *this_leaf, ...@@ -378,7 +342,7 @@ static void __cpuinit amd_check_l3_disable(struct _cpuid4_info_regs *this_leaf,
* never freed but this is done only on shutdown so it doesn't matter. * never freed but this is done only on shutdown so it doesn't matter.
*/ */
if (!l3_caches) { if (!l3_caches) {
int size = k8_northbridges.num * sizeof(struct amd_l3_cache *); int size = amd_nb_num() * sizeof(struct amd_l3_cache);
l3_caches = kzalloc(size, GFP_ATOMIC); l3_caches = kzalloc(size, GFP_ATOMIC);
if (!l3_caches) if (!l3_caches)
...@@ -387,14 +351,12 @@ static void __cpuinit amd_check_l3_disable(struct _cpuid4_info_regs *this_leaf, ...@@ -387,14 +351,12 @@ static void __cpuinit amd_check_l3_disable(struct _cpuid4_info_regs *this_leaf,
node = amd_get_nb_id(smp_processor_id()); node = amd_get_nb_id(smp_processor_id());
if (!l3_caches[node]) { if (!l3_caches[node].nb) {
l3_caches[node] = amd_init_l3_cache(node); l3_caches[node].nb = node_to_amd_nb(node);
l3_caches[node]->can_disable = true; amd_calc_l3_indices(&l3_caches[node]);
} }
WARN_ON(!l3_caches[node]); this_leaf->l3 = &l3_caches[node];
this_leaf->l3 = l3_caches[node];
} }
/* /*
...@@ -408,7 +370,7 @@ int amd_get_l3_disable_slot(struct amd_l3_cache *l3, unsigned slot) ...@@ -408,7 +370,7 @@ int amd_get_l3_disable_slot(struct amd_l3_cache *l3, unsigned slot)
{ {
unsigned int reg = 0; unsigned int reg = 0;
pci_read_config_dword(l3->dev, 0x1BC + slot * 4, &reg); pci_read_config_dword(l3->nb->misc, 0x1BC + slot * 4, &reg);
/* check whether this slot is activated already */ /* check whether this slot is activated already */
if (reg & (3UL << 30)) if (reg & (3UL << 30))
...@@ -422,7 +384,8 @@ static ssize_t show_cache_disable(struct _cpuid4_info *this_leaf, char *buf, ...@@ -422,7 +384,8 @@ static ssize_t show_cache_disable(struct _cpuid4_info *this_leaf, char *buf,
{ {
int index; int index;
if (!this_leaf->l3 || !this_leaf->l3->can_disable) if (!this_leaf->l3 ||
!amd_nb_has_feature(AMD_NB_L3_INDEX_DISABLE))
return -EINVAL; return -EINVAL;
index = amd_get_l3_disable_slot(this_leaf->l3, slot); index = amd_get_l3_disable_slot(this_leaf->l3, slot);
...@@ -457,7 +420,7 @@ static void amd_l3_disable_index(struct amd_l3_cache *l3, int cpu, ...@@ -457,7 +420,7 @@ static void amd_l3_disable_index(struct amd_l3_cache *l3, int cpu,
if (!l3->subcaches[i]) if (!l3->subcaches[i])
continue; continue;
pci_write_config_dword(l3->dev, 0x1BC + slot * 4, reg); pci_write_config_dword(l3->nb->misc, 0x1BC + slot * 4, reg);
/* /*
* We need to WBINVD on a core on the node containing the L3 * We need to WBINVD on a core on the node containing the L3
...@@ -467,7 +430,7 @@ static void amd_l3_disable_index(struct amd_l3_cache *l3, int cpu, ...@@ -467,7 +430,7 @@ static void amd_l3_disable_index(struct amd_l3_cache *l3, int cpu,
wbinvd_on_cpu(cpu); wbinvd_on_cpu(cpu);
reg |= BIT(31); reg |= BIT(31);
pci_write_config_dword(l3->dev, 0x1BC + slot * 4, reg); pci_write_config_dword(l3->nb->misc, 0x1BC + slot * 4, reg);
} }
} }
...@@ -524,7 +487,8 @@ static ssize_t store_cache_disable(struct _cpuid4_info *this_leaf, ...@@ -524,7 +487,8 @@ static ssize_t store_cache_disable(struct _cpuid4_info *this_leaf,
if (!capable(CAP_SYS_ADMIN)) if (!capable(CAP_SYS_ADMIN))
return -EPERM; return -EPERM;
if (!this_leaf->l3 || !this_leaf->l3->can_disable) if (!this_leaf->l3 ||
!amd_nb_has_feature(AMD_NB_L3_INDEX_DISABLE))
return -EINVAL; return -EINVAL;
cpu = cpumask_first(to_cpumask(this_leaf->shared_cpu_map)); cpu = cpumask_first(to_cpumask(this_leaf->shared_cpu_map));
...@@ -558,10 +522,7 @@ static struct _cache_attr cache_disable_1 = __ATTR(cache_disable_1, 0644, ...@@ -558,10 +522,7 @@ static struct _cache_attr cache_disable_1 = __ATTR(cache_disable_1, 0644,
show_cache_disable_1, store_cache_disable_1); show_cache_disable_1, store_cache_disable_1);
#else /* CONFIG_AMD_NB */ #else /* CONFIG_AMD_NB */
static void __cpuinit #define amd_init_l3_cache(x, y)
amd_check_l3_disable(struct _cpuid4_info_regs *this_leaf, int index)
{
};
#endif /* CONFIG_AMD_NB */ #endif /* CONFIG_AMD_NB */
static int static int
...@@ -575,7 +536,7 @@ __cpuinit cpuid4_cache_lookup_regs(int index, ...@@ -575,7 +536,7 @@ __cpuinit cpuid4_cache_lookup_regs(int index,
if (boot_cpu_data.x86_vendor == X86_VENDOR_AMD) { if (boot_cpu_data.x86_vendor == X86_VENDOR_AMD) {
amd_cpuid4(index, &eax, &ebx, &ecx); amd_cpuid4(index, &eax, &ebx, &ecx);
amd_check_l3_disable(this_leaf, index); amd_init_l3_cache(this_leaf, index);
} else { } else {
cpuid_count(4, index, &eax.full, &ebx.full, &ecx.full, &edx); cpuid_count(4, index, &eax.full, &ebx.full, &ecx.full, &edx);
} }
...@@ -983,30 +944,48 @@ define_one_ro(size); ...@@ -983,30 +944,48 @@ define_one_ro(size);
define_one_ro(shared_cpu_map); define_one_ro(shared_cpu_map);
define_one_ro(shared_cpu_list); define_one_ro(shared_cpu_list);
#define DEFAULT_SYSFS_CACHE_ATTRS \
&type.attr, \
&level.attr, \
&coherency_line_size.attr, \
&physical_line_partition.attr, \
&ways_of_associativity.attr, \
&number_of_sets.attr, \
&size.attr, \
&shared_cpu_map.attr, \
&shared_cpu_list.attr
static struct attribute *default_attrs[] = { static struct attribute *default_attrs[] = {
DEFAULT_SYSFS_CACHE_ATTRS, &type.attr,
&level.attr,
&coherency_line_size.attr,
&physical_line_partition.attr,
&ways_of_associativity.attr,
&number_of_sets.attr,
&size.attr,
&shared_cpu_map.attr,
&shared_cpu_list.attr,
NULL NULL
}; };
static struct attribute *default_l3_attrs[] = {
DEFAULT_SYSFS_CACHE_ATTRS,
#ifdef CONFIG_AMD_NB #ifdef CONFIG_AMD_NB
&cache_disable_0.attr, static struct attribute ** __cpuinit amd_l3_attrs(void)
&cache_disable_1.attr, {
static struct attribute **attrs;
int n;
if (attrs)
return attrs;
n = sizeof (default_attrs) / sizeof (struct attribute *);
if (amd_nb_has_feature(AMD_NB_L3_INDEX_DISABLE))
n += 2;
attrs = kzalloc(n * sizeof (struct attribute *), GFP_KERNEL);
if (attrs == NULL)
return attrs = default_attrs;
for (n = 0; default_attrs[n]; n++)
attrs[n] = default_attrs[n];
if (amd_nb_has_feature(AMD_NB_L3_INDEX_DISABLE)) {
attrs[n++] = &cache_disable_0.attr;
attrs[n++] = &cache_disable_1.attr;
}
return attrs;
}
#endif #endif
NULL
};
static ssize_t show(struct kobject *kobj, struct attribute *attr, char *buf) static ssize_t show(struct kobject *kobj, struct attribute *attr, char *buf)
{ {
...@@ -1117,11 +1096,11 @@ static int __cpuinit cache_add_dev(struct sys_device * sys_dev) ...@@ -1117,11 +1096,11 @@ static int __cpuinit cache_add_dev(struct sys_device * sys_dev)
this_leaf = CPUID4_INFO_IDX(cpu, i); this_leaf = CPUID4_INFO_IDX(cpu, i);
if (this_leaf->l3 && this_leaf->l3->can_disable)
ktype_cache.default_attrs = default_l3_attrs;
else
ktype_cache.default_attrs = default_attrs; ktype_cache.default_attrs = default_attrs;
#ifdef CONFIG_AMD_NB
if (this_leaf->l3)
ktype_cache.default_attrs = amd_l3_attrs();
#endif
retval = kobject_init_and_add(&(this_object->kobj), retval = kobject_init_and_add(&(this_object->kobj),
&ktype_cache, &ktype_cache,
per_cpu(ici_cache_kobject, cpu), per_cpu(ici_cache_kobject, cpu),
......
...@@ -143,7 +143,7 @@ static void flush_gart(void) ...@@ -143,7 +143,7 @@ static void flush_gart(void)
spin_lock_irqsave(&iommu_bitmap_lock, flags); spin_lock_irqsave(&iommu_bitmap_lock, flags);
if (need_flush) { if (need_flush) {
k8_flush_garts(); amd_flush_garts();
need_flush = false; need_flush = false;
} }
spin_unlock_irqrestore(&iommu_bitmap_lock, flags); spin_unlock_irqrestore(&iommu_bitmap_lock, flags);
...@@ -561,17 +561,17 @@ static void enable_gart_translations(void) ...@@ -561,17 +561,17 @@ static void enable_gart_translations(void)
{ {
int i; int i;
if (!k8_northbridges.gart_supported) if (!amd_nb_has_feature(AMD_NB_GART))
return; return;
for (i = 0; i < k8_northbridges.num; i++) { for (i = 0; i < amd_nb_num(); i++) {
struct pci_dev *dev = k8_northbridges.nb_misc[i]; struct pci_dev *dev = node_to_amd_nb(i)->misc;
enable_gart_translation(dev, __pa(agp_gatt_table)); enable_gart_translation(dev, __pa(agp_gatt_table));
} }
/* Flush the GART-TLB to remove stale entries */ /* Flush the GART-TLB to remove stale entries */
k8_flush_garts(); amd_flush_garts();
} }
/* /*
...@@ -596,13 +596,13 @@ static void gart_fixup_northbridges(struct sys_device *dev) ...@@ -596,13 +596,13 @@ static void gart_fixup_northbridges(struct sys_device *dev)
if (!fix_up_north_bridges) if (!fix_up_north_bridges)
return; return;
if (!k8_northbridges.gart_supported) if (!amd_nb_has_feature(AMD_NB_GART))
return; return;
pr_info("PCI-DMA: Restoring GART aperture settings\n"); pr_info("PCI-DMA: Restoring GART aperture settings\n");
for (i = 0; i < k8_northbridges.num; i++) { for (i = 0; i < amd_nb_num(); i++) {
struct pci_dev *dev = k8_northbridges.nb_misc[i]; struct pci_dev *dev = node_to_amd_nb(i)->misc;
/* /*
* Don't enable translations just yet. That is the next * Don't enable translations just yet. That is the next
...@@ -644,7 +644,7 @@ static struct sys_device device_gart = { ...@@ -644,7 +644,7 @@ static struct sys_device device_gart = {
* Private Northbridge GATT initialization in case we cannot use the * Private Northbridge GATT initialization in case we cannot use the
* AGP driver for some reason. * AGP driver for some reason.
*/ */
static __init int init_k8_gatt(struct agp_kern_info *info) static __init int init_amd_gatt(struct agp_kern_info *info)
{ {
unsigned aper_size, gatt_size, new_aper_size; unsigned aper_size, gatt_size, new_aper_size;
unsigned aper_base, new_aper_base; unsigned aper_base, new_aper_base;
...@@ -656,8 +656,8 @@ static __init int init_k8_gatt(struct agp_kern_info *info) ...@@ -656,8 +656,8 @@ static __init int init_k8_gatt(struct agp_kern_info *info)
aper_size = aper_base = info->aper_size = 0; aper_size = aper_base = info->aper_size = 0;
dev = NULL; dev = NULL;
for (i = 0; i < k8_northbridges.num; i++) { for (i = 0; i < amd_nb_num(); i++) {
dev = k8_northbridges.nb_misc[i]; dev = node_to_amd_nb(i)->misc;
new_aper_base = read_aperture(dev, &new_aper_size); new_aper_base = read_aperture(dev, &new_aper_size);
if (!new_aper_base) if (!new_aper_base)
goto nommu; goto nommu;
...@@ -725,13 +725,13 @@ static void gart_iommu_shutdown(void) ...@@ -725,13 +725,13 @@ static void gart_iommu_shutdown(void)
if (!no_agp) if (!no_agp)
return; return;
if (!k8_northbridges.gart_supported) if (!amd_nb_has_feature(AMD_NB_GART))
return; return;
for (i = 0; i < k8_northbridges.num; i++) { for (i = 0; i < amd_nb_num(); i++) {
u32 ctl; u32 ctl;
dev = k8_northbridges.nb_misc[i]; dev = node_to_amd_nb(i)->misc;
pci_read_config_dword(dev, AMD64_GARTAPERTURECTL, &ctl); pci_read_config_dword(dev, AMD64_GARTAPERTURECTL, &ctl);
ctl &= ~GARTEN; ctl &= ~GARTEN;
...@@ -749,14 +749,14 @@ int __init gart_iommu_init(void) ...@@ -749,14 +749,14 @@ int __init gart_iommu_init(void)
unsigned long scratch; unsigned long scratch;
long i; long i;
if (!k8_northbridges.gart_supported) if (!amd_nb_has_feature(AMD_NB_GART))
return 0; return 0;
#ifndef CONFIG_AGP_AMD64 #ifndef CONFIG_AGP_AMD64
no_agp = 1; no_agp = 1;
#else #else
/* Makefile puts PCI initialization via subsys_initcall first. */ /* Makefile puts PCI initialization via subsys_initcall first. */
/* Add other K8 AGP bridge drivers here */ /* Add other AMD AGP bridge drivers here */
no_agp = no_agp || no_agp = no_agp ||
(agp_amd64_init() < 0) || (agp_amd64_init() < 0) ||
(agp_copy_info(agp_bridge, &info) < 0); (agp_copy_info(agp_bridge, &info) < 0);
...@@ -765,7 +765,7 @@ int __init gart_iommu_init(void) ...@@ -765,7 +765,7 @@ int __init gart_iommu_init(void)
if (no_iommu || if (no_iommu ||
(!force_iommu && max_pfn <= MAX_DMA32_PFN) || (!force_iommu && max_pfn <= MAX_DMA32_PFN) ||
!gart_iommu_aperture || !gart_iommu_aperture ||
(no_agp && init_k8_gatt(&info) < 0)) { (no_agp && init_amd_gatt(&info) < 0)) {
if (max_pfn > MAX_DMA32_PFN) { if (max_pfn > MAX_DMA32_PFN) {
pr_warning("More than 4GB of memory but GART IOMMU not available.\n"); pr_warning("More than 4GB of memory but GART IOMMU not available.\n");
pr_warning("falling back to iommu=soft.\n"); pr_warning("falling back to iommu=soft.\n");
......
...@@ -705,7 +705,7 @@ static u64 __init get_max_mapped(void) ...@@ -705,7 +705,7 @@ static u64 __init get_max_mapped(void)
void __init setup_arch(char **cmdline_p) void __init setup_arch(char **cmdline_p)
{ {
int acpi = 0; int acpi = 0;
int k8 = 0; int amd = 0;
unsigned long flags; unsigned long flags;
#ifdef CONFIG_X86_32 #ifdef CONFIG_X86_32
...@@ -991,12 +991,12 @@ void __init setup_arch(char **cmdline_p) ...@@ -991,12 +991,12 @@ void __init setup_arch(char **cmdline_p)
acpi = acpi_numa_init(); acpi = acpi_numa_init();
#endif #endif
#ifdef CONFIG_K8_NUMA #ifdef CONFIG_AMD_NUMA
if (!acpi) if (!acpi)
k8 = !k8_numa_init(0, max_pfn); amd = !amd_numa_init(0, max_pfn);
#endif #endif
initmem_init(0, max_pfn, acpi, k8); initmem_init(0, max_pfn, acpi, amd);
memblock_find_dma_reserve(); memblock_find_dma_reserve();
dma32_reserve_bootmem(); dma32_reserve_bootmem();
......
...@@ -23,7 +23,7 @@ mmiotrace-y := kmmio.o pf_in.o mmio-mod.o ...@@ -23,7 +23,7 @@ mmiotrace-y := kmmio.o pf_in.o mmio-mod.o
obj-$(CONFIG_MMIOTRACE_TEST) += testmmiotrace.o obj-$(CONFIG_MMIOTRACE_TEST) += testmmiotrace.o
obj-$(CONFIG_NUMA) += numa.o numa_$(BITS).o obj-$(CONFIG_NUMA) += numa.o numa_$(BITS).o
obj-$(CONFIG_K8_NUMA) += k8topology_64.o obj-$(CONFIG_AMD_NUMA) += amdtopology_64.o
obj-$(CONFIG_ACPI_NUMA) += srat_$(BITS).o obj-$(CONFIG_ACPI_NUMA) += srat_$(BITS).o
obj-$(CONFIG_HAVE_MEMBLOCK) += memblock.o obj-$(CONFIG_HAVE_MEMBLOCK) += memblock.o
......
/* /*
* AMD K8 NUMA support. * AMD NUMA support.
* Discover the memory map and associated nodes. * Discover the memory map and associated nodes.
* *
* This version reads it directly from the K8 northbridge. * This version reads it directly from the AMD northbridge.
* *
* Copyright 2002,2003 Andi Kleen, SuSE Labs. * Copyright 2002,2003 Andi Kleen, SuSE Labs.
*/ */
...@@ -57,7 +57,7 @@ static __init void early_get_boot_cpu_id(void) ...@@ -57,7 +57,7 @@ static __init void early_get_boot_cpu_id(void)
{ {
/* /*
* need to get the APIC ID of the BSP so can use that to * need to get the APIC ID of the BSP so can use that to
* create apicid_to_node in k8_scan_nodes() * create apicid_to_node in amd_scan_nodes()
*/ */
#ifdef CONFIG_X86_MPPARSE #ifdef CONFIG_X86_MPPARSE
/* /*
...@@ -69,7 +69,7 @@ static __init void early_get_boot_cpu_id(void) ...@@ -69,7 +69,7 @@ static __init void early_get_boot_cpu_id(void)
early_init_lapic_mapping(); early_init_lapic_mapping();
} }
int __init k8_get_nodes(struct bootnode *physnodes) int __init amd_get_nodes(struct bootnode *physnodes)
{ {
int i; int i;
int ret = 0; int ret = 0;
...@@ -82,7 +82,7 @@ int __init k8_get_nodes(struct bootnode *physnodes) ...@@ -82,7 +82,7 @@ int __init k8_get_nodes(struct bootnode *physnodes)
return ret; return ret;
} }
int __init k8_numa_init(unsigned long start_pfn, unsigned long end_pfn) int __init amd_numa_init(unsigned long start_pfn, unsigned long end_pfn)
{ {
unsigned long start = PFN_PHYS(start_pfn); unsigned long start = PFN_PHYS(start_pfn);
unsigned long end = PFN_PHYS(end_pfn); unsigned long end = PFN_PHYS(end_pfn);
...@@ -194,7 +194,7 @@ int __init k8_numa_init(unsigned long start_pfn, unsigned long end_pfn) ...@@ -194,7 +194,7 @@ int __init k8_numa_init(unsigned long start_pfn, unsigned long end_pfn)
return 0; return 0;
} }
int __init k8_scan_nodes(void) int __init amd_scan_nodes(void)
{ {
unsigned int bits; unsigned int bits;
unsigned int cores; unsigned int cores;
......
...@@ -264,7 +264,7 @@ static struct bootnode physnodes[MAX_NUMNODES] __initdata; ...@@ -264,7 +264,7 @@ static struct bootnode physnodes[MAX_NUMNODES] __initdata;
static char *cmdline __initdata; static char *cmdline __initdata;
static int __init setup_physnodes(unsigned long start, unsigned long end, static int __init setup_physnodes(unsigned long start, unsigned long end,
int acpi, int k8) int acpi, int amd)
{ {
int nr_nodes = 0; int nr_nodes = 0;
int ret = 0; int ret = 0;
...@@ -274,13 +274,13 @@ static int __init setup_physnodes(unsigned long start, unsigned long end, ...@@ -274,13 +274,13 @@ static int __init setup_physnodes(unsigned long start, unsigned long end,
if (acpi) if (acpi)
nr_nodes = acpi_get_nodes(physnodes); nr_nodes = acpi_get_nodes(physnodes);
#endif #endif
#ifdef CONFIG_K8_NUMA #ifdef CONFIG_AMD_NUMA
if (k8) if (amd)
nr_nodes = k8_get_nodes(physnodes); nr_nodes = amd_get_nodes(physnodes);
#endif #endif
/* /*
* Basic sanity checking on the physical node map: there may be errors * Basic sanity checking on the physical node map: there may be errors
* if the SRAT or K8 incorrectly reported the topology or the mem= * if the SRAT or AMD code incorrectly reported the topology or the mem=
* kernel parameter is used. * kernel parameter is used.
*/ */
for (i = 0; i < nr_nodes; i++) { for (i = 0; i < nr_nodes; i++) {
...@@ -549,7 +549,7 @@ static int __init split_nodes_size_interleave(u64 addr, u64 max_addr, u64 size) ...@@ -549,7 +549,7 @@ static int __init split_nodes_size_interleave(u64 addr, u64 max_addr, u64 size)
* numa=fake command-line option. * numa=fake command-line option.
*/ */
static int __init numa_emulation(unsigned long start_pfn, static int __init numa_emulation(unsigned long start_pfn,
unsigned long last_pfn, int acpi, int k8) unsigned long last_pfn, int acpi, int amd)
{ {
u64 addr = start_pfn << PAGE_SHIFT; u64 addr = start_pfn << PAGE_SHIFT;
u64 max_addr = last_pfn << PAGE_SHIFT; u64 max_addr = last_pfn << PAGE_SHIFT;
...@@ -557,7 +557,7 @@ static int __init numa_emulation(unsigned long start_pfn, ...@@ -557,7 +557,7 @@ static int __init numa_emulation(unsigned long start_pfn,
int num_nodes; int num_nodes;
int i; int i;
num_phys_nodes = setup_physnodes(addr, max_addr, acpi, k8); num_phys_nodes = setup_physnodes(addr, max_addr, acpi, amd);
/* /*
* If the numa=fake command-line contains a 'M' or 'G', it represents * If the numa=fake command-line contains a 'M' or 'G', it represents
* the fixed node size. Otherwise, if it is just a single number N, * the fixed node size. Otherwise, if it is just a single number N,
...@@ -602,7 +602,7 @@ static int __init numa_emulation(unsigned long start_pfn, ...@@ -602,7 +602,7 @@ static int __init numa_emulation(unsigned long start_pfn,
#endif /* CONFIG_NUMA_EMU */ #endif /* CONFIG_NUMA_EMU */
void __init initmem_init(unsigned long start_pfn, unsigned long last_pfn, void __init initmem_init(unsigned long start_pfn, unsigned long last_pfn,
int acpi, int k8) int acpi, int amd)
{ {
int i; int i;
...@@ -610,7 +610,7 @@ void __init initmem_init(unsigned long start_pfn, unsigned long last_pfn, ...@@ -610,7 +610,7 @@ void __init initmem_init(unsigned long start_pfn, unsigned long last_pfn,
nodes_clear(node_online_map); nodes_clear(node_online_map);
#ifdef CONFIG_NUMA_EMU #ifdef CONFIG_NUMA_EMU
if (cmdline && !numa_emulation(start_pfn, last_pfn, acpi, k8)) if (cmdline && !numa_emulation(start_pfn, last_pfn, acpi, amd))
return; return;
nodes_clear(node_possible_map); nodes_clear(node_possible_map);
nodes_clear(node_online_map); nodes_clear(node_online_map);
...@@ -624,8 +624,8 @@ void __init initmem_init(unsigned long start_pfn, unsigned long last_pfn, ...@@ -624,8 +624,8 @@ void __init initmem_init(unsigned long start_pfn, unsigned long last_pfn,
nodes_clear(node_online_map); nodes_clear(node_online_map);
#endif #endif
#ifdef CONFIG_K8_NUMA #ifdef CONFIG_AMD_NUMA
if (!numa_off && k8 && !k8_scan_nodes()) if (!numa_off && amd && !amd_scan_nodes())
return; return;
nodes_clear(node_possible_map); nodes_clear(node_possible_map);
nodes_clear(node_online_map); nodes_clear(node_online_map);
......
...@@ -38,7 +38,7 @@ static int agp_bridges_found; ...@@ -38,7 +38,7 @@ static int agp_bridges_found;
static void amd64_tlbflush(struct agp_memory *temp) static void amd64_tlbflush(struct agp_memory *temp)
{ {
k8_flush_garts(); amd_flush_garts();
} }
static int amd64_insert_memory(struct agp_memory *mem, off_t pg_start, int type) static int amd64_insert_memory(struct agp_memory *mem, off_t pg_start, int type)
...@@ -124,7 +124,7 @@ static int amd64_fetch_size(void) ...@@ -124,7 +124,7 @@ static int amd64_fetch_size(void)
u32 temp; u32 temp;
struct aper_size_info_32 *values; struct aper_size_info_32 *values;
dev = k8_northbridges.nb_misc[0]; dev = node_to_amd_nb(0)->misc;
if (dev==NULL) if (dev==NULL)
return 0; return 0;
...@@ -181,16 +181,15 @@ static int amd_8151_configure(void) ...@@ -181,16 +181,15 @@ static int amd_8151_configure(void)
unsigned long gatt_bus = virt_to_phys(agp_bridge->gatt_table_real); unsigned long gatt_bus = virt_to_phys(agp_bridge->gatt_table_real);
int i; int i;
if (!k8_northbridges.gart_supported) if (!amd_nb_has_feature(AMD_NB_GART))
return 0; return 0;
/* Configure AGP regs in each x86-64 host bridge. */ /* Configure AGP regs in each x86-64 host bridge. */
for (i = 0; i < k8_northbridges.num; i++) { for (i = 0; i < amd_nb_num(); i++) {
agp_bridge->gart_bus_addr = agp_bridge->gart_bus_addr =
amd64_configure(k8_northbridges.nb_misc[i], amd64_configure(node_to_amd_nb(i)->misc, gatt_bus);
gatt_bus);
} }
k8_flush_garts(); amd_flush_garts();
return 0; return 0;
} }
...@@ -200,11 +199,11 @@ static void amd64_cleanup(void) ...@@ -200,11 +199,11 @@ static void amd64_cleanup(void)
u32 tmp; u32 tmp;
int i; int i;
if (!k8_northbridges.gart_supported) if (!amd_nb_has_feature(AMD_NB_GART))
return; return;
for (i = 0; i < k8_northbridges.num; i++) { for (i = 0; i < amd_nb_num(); i++) {
struct pci_dev *dev = k8_northbridges.nb_misc[i]; struct pci_dev *dev = node_to_amd_nb(i)->misc;
/* disable gart translation */ /* disable gart translation */
pci_read_config_dword(dev, AMD64_GARTAPERTURECTL, &tmp); pci_read_config_dword(dev, AMD64_GARTAPERTURECTL, &tmp);
tmp &= ~GARTEN; tmp &= ~GARTEN;
...@@ -331,15 +330,15 @@ static __devinit int cache_nbs(struct pci_dev *pdev, u32 cap_ptr) ...@@ -331,15 +330,15 @@ static __devinit int cache_nbs(struct pci_dev *pdev, u32 cap_ptr)
{ {
int i; int i;
if (cache_k8_northbridges() < 0) if (amd_cache_northbridges() < 0)
return -ENODEV; return -ENODEV;
if (!k8_northbridges.gart_supported) if (!amd_nb_has_feature(AMD_NB_GART))
return -ENODEV; return -ENODEV;
i = 0; i = 0;
for (i = 0; i < k8_northbridges.num; i++) { for (i = 0; i < amd_nb_num(); i++) {
struct pci_dev *dev = k8_northbridges.nb_misc[i]; struct pci_dev *dev = node_to_amd_nb(i)->misc;
if (fix_northbridge(dev, pdev, cap_ptr) < 0) { if (fix_northbridge(dev, pdev, cap_ptr) < 0) {
dev_err(&dev->dev, "no usable aperture found\n"); dev_err(&dev->dev, "no usable aperture found\n");
#ifdef __x86_64__ #ifdef __x86_64__
...@@ -416,7 +415,7 @@ static int __devinit uli_agp_init(struct pci_dev *pdev) ...@@ -416,7 +415,7 @@ static int __devinit uli_agp_init(struct pci_dev *pdev)
} }
/* shadow x86-64 registers into ULi registers */ /* shadow x86-64 registers into ULi registers */
pci_read_config_dword (k8_northbridges.nb_misc[0], AMD64_GARTAPERTUREBASE, pci_read_config_dword (node_to_amd_nb(0)->misc, AMD64_GARTAPERTUREBASE,
&httfea); &httfea);
/* if x86-64 aperture base is beyond 4G, exit here */ /* if x86-64 aperture base is beyond 4G, exit here */
...@@ -484,7 +483,7 @@ static int nforce3_agp_init(struct pci_dev *pdev) ...@@ -484,7 +483,7 @@ static int nforce3_agp_init(struct pci_dev *pdev)
pci_write_config_dword(dev1, NVIDIA_X86_64_1_APSIZE, tmp); pci_write_config_dword(dev1, NVIDIA_X86_64_1_APSIZE, tmp);
/* shadow x86-64 registers into NVIDIA registers */ /* shadow x86-64 registers into NVIDIA registers */
pci_read_config_dword (k8_northbridges.nb_misc[0], AMD64_GARTAPERTUREBASE, pci_read_config_dword (node_to_amd_nb(0)->misc, AMD64_GARTAPERTUREBASE,
&apbase); &apbase);
/* if x86-64 aperture base is beyond 4G, exit here */ /* if x86-64 aperture base is beyond 4G, exit here */
...@@ -778,7 +777,7 @@ int __init agp_amd64_init(void) ...@@ -778,7 +777,7 @@ int __init agp_amd64_init(void)
} }
/* First check that we have at least one AMD64 NB */ /* First check that we have at least one AMD64 NB */
if (!pci_dev_present(k8_nb_ids)) if (!pci_dev_present(amd_nb_misc_ids))
return -ENODEV; return -ENODEV;
/* Look for any AGP bridge */ /* Look for any AGP bridge */
......
...@@ -2917,7 +2917,7 @@ static int __init amd64_edac_init(void) ...@@ -2917,7 +2917,7 @@ static int __init amd64_edac_init(void)
opstate_init(); opstate_init();
if (cache_k8_northbridges() < 0) if (amd_cache_northbridges() < 0)
goto err_ret; goto err_ret;
msrs = msrs_alloc(); msrs = msrs_alloc();
...@@ -2934,7 +2934,7 @@ static int __init amd64_edac_init(void) ...@@ -2934,7 +2934,7 @@ static int __init amd64_edac_init(void)
* to finish initialization of the MC instances. * to finish initialization of the MC instances.
*/ */
err = -ENODEV; err = -ENODEV;
for (nb = 0; nb < k8_northbridges.num; nb++) { for (nb = 0; nb < amd_nb_num(); nb++) {
if (!pvt_lookup[nb]) if (!pvt_lookup[nb])
continue; continue;
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册