]> git.karo-electronics.de Git - karo-tx-linux.git/commitdiff
Merge commit 'v2.6.28-rc8' into x86/uv
authorIngo Molnar <mingo@elte.hu>
Tue, 16 Dec 2008 21:53:53 +0000 (22:53 +0100)
committerIngo Molnar <mingo@elte.hu>
Tue, 16 Dec 2008 21:53:53 +0000 (22:53 +0100)
13 files changed:
arch/x86/include/asm/uv/bios.h
arch/x86/include/asm/uv/uv_hub.h
arch/x86/kernel/bios_uv.c
arch/x86/kernel/genx2apic_uv_x.c
arch/x86/kernel/tlb_uv.c
drivers/misc/sgi-gru/gruprocfs.c
drivers/misc/sgi-xp/xp.h
drivers/misc/sgi-xp/xp_main.c
drivers/misc/sgi-xp/xp_sn2.c
drivers/misc/sgi-xp/xp_uv.c
drivers/misc/sgi-xp/xpc.h
drivers/misc/sgi-xp/xpc_sn2.c
drivers/misc/sgi-xp/xpc_uv.c

index d931d3b7e6f71b394d87f5766835dd8ef892dfec..da1c4e8e78fcc9b39b78675f5f999c37626f59db 100644 (file)
 enum uv_bios_cmd {
        UV_BIOS_COMMON,
        UV_BIOS_GET_SN_INFO,
-       UV_BIOS_FREQ_BASE
+       UV_BIOS_FREQ_BASE,
+       UV_BIOS_WATCHLIST_ALLOC,
+       UV_BIOS_WATCHLIST_FREE,
+       UV_BIOS_MEMPROTECT,
+       UV_BIOS_GET_PARTITION_ADDR
 };
 
 /*
  * Status values returned from a BIOS call.
  */
 enum {
+       BIOS_STATUS_MORE_PASSES         =  1,
        BIOS_STATUS_SUCCESS             =  0,
        BIOS_STATUS_UNIMPLEMENTED       = -ENOSYS,
        BIOS_STATUS_EINVAL              = -EINVAL,
@@ -71,6 +76,21 @@ union partition_info_u {
        };
 };
 
+union uv_watchlist_u {
+       u64     val;
+       struct {
+               u64     blade   : 16,
+                       size    : 32,
+                       filler  : 16;
+       };
+};
+
+enum uv_memprotect {
+       UV_MEMPROT_RESTRICT_ACCESS,
+       UV_MEMPROT_ALLOW_AMO,
+       UV_MEMPROT_ALLOW_RW
+};
+
 /*
  * bios calls have 6 parameters
  */
@@ -80,14 +100,20 @@ extern s64 uv_bios_call_reentrant(enum uv_bios_cmd, u64, u64, u64, u64, u64);
 
 extern s64 uv_bios_get_sn_info(int, int *, long *, long *, long *);
 extern s64 uv_bios_freq_base(u64, u64 *);
+extern int uv_bios_mq_watchlist_alloc(int, void *, unsigned int,
+                                       unsigned long *);
+extern int uv_bios_mq_watchlist_free(int, int);
+extern s64 uv_bios_change_memprotect(u64, u64, enum uv_memprotect);
+extern s64 uv_bios_reserved_page_pa(u64, u64 *, u64 *, u64 *);
 
 extern void uv_bios_init(void);
 
+extern unsigned long sn_rtc_cycles_per_second;
 extern int uv_type;
 extern long sn_partition_id;
-extern long uv_coherency_id;
-extern long uv_region_size;
-#define partition_coherence_id()       (uv_coherency_id)
+extern long sn_coherency_id;
+extern long sn_region_size;
+#define partition_coherence_id()       (sn_coherency_id)
 
 extern struct kobject *sgi_uv_kobj;    /* /sys/firmware/sgi_uv */
 
index 7a5782610b2bcb0af69d99be174d0815a6722663..52aa943c634fe1f8d4d87a05335ae9979f534590 100644 (file)
  */
 #define UV_MAX_NASID_VALUE     (UV_MAX_NUMALINK_NODES * 2)
 
+struct uv_scir_s {
+       struct timer_list timer;
+       unsigned long   offset;
+       unsigned long   last;
+       unsigned long   idle_on;
+       unsigned long   idle_off;
+       unsigned char   state;
+       unsigned char   enabled;
+};
+
 /*
  * The following defines attributes of the HUB chip. These attributes are
  * frequently referenced and are kept in the per-cpu data areas of each cpu.
  * They are kept together in a struct to minimize cache misses.
  */
 struct uv_hub_info_s {
-       unsigned long   global_mmr_base;
-       unsigned long   gpa_mask;
-       unsigned long   gnode_upper;
-       unsigned long   lowmem_remap_top;
-       unsigned long   lowmem_remap_base;
-       unsigned short  pnode;
-       unsigned short  pnode_mask;
-       unsigned short  coherency_domain_number;
-       unsigned short  numa_blade_id;
-       unsigned char   blade_processor_id;
-       unsigned char   m_val;
-       unsigned char   n_val;
+       unsigned long           global_mmr_base;
+       unsigned long           gpa_mask;
+       unsigned long           gnode_upper;
+       unsigned long           lowmem_remap_top;
+       unsigned long           lowmem_remap_base;
+       unsigned short          pnode;
+       unsigned short          pnode_mask;
+       unsigned short          coherency_domain_number;
+       unsigned short          numa_blade_id;
+       unsigned char           blade_processor_id;
+       unsigned char           m_val;
+       unsigned char           n_val;
+       struct uv_scir_s        scir;
 };
+
 DECLARE_PER_CPU(struct uv_hub_info_s, __uv_hub_info);
 #define uv_hub_info            (&__get_cpu_var(__uv_hub_info))
 #define uv_cpu_hub_info(cpu)   (&per_cpu(__uv_hub_info, cpu))
@@ -163,6 +175,30 @@ DECLARE_PER_CPU(struct uv_hub_info_s, __uv_hub_info);
 
 #define UV_APIC_PNODE_SHIFT    6
 
+/* Local Bus from cpu's perspective */
+#define LOCAL_BUS_BASE         0x1c00000
+#define LOCAL_BUS_SIZE         (4 * 1024 * 1024)
+
+/*
+ * System Controller Interface Reg
+ *
+ * Note there are NO leds on a UV system.  This register is only
+ * used by the system controller to monitor system-wide operation.
+ * There are 64 regs per node.  With Nahelem cpus (2 cores per node,
+ * 8 cpus per core, 2 threads per cpu) there are 32 cpu threads on
+ * a node.
+ *
+ * The window is located at top of ACPI MMR space
+ */
+#define SCIR_WINDOW_COUNT      64
+#define SCIR_LOCAL_MMR_BASE    (LOCAL_BUS_BASE + \
+                                LOCAL_BUS_SIZE - \
+                                SCIR_WINDOW_COUNT)
+
+#define SCIR_CPU_HEARTBEAT     0x01    /* timer interrupt */
+#define SCIR_CPU_ACTIVITY      0x02    /* not idle */
+#define SCIR_CPU_HB_INTERVAL   (HZ)    /* once per second */
+
 /*
  * Macros for converting between kernel virtual addresses, socket local physical
  * addresses, and UV global physical addresses.
@@ -277,6 +313,16 @@ static inline void uv_write_local_mmr(unsigned long offset, unsigned long val)
        *uv_local_mmr_address(offset) = val;
 }
 
+static inline unsigned char uv_read_local_mmr8(unsigned long offset)
+{
+       return *((unsigned char *)uv_local_mmr_address(offset));
+}
+
+static inline void uv_write_local_mmr8(unsigned long offset, unsigned char val)
+{
+       *((unsigned char *)uv_local_mmr_address(offset)) = val;
+}
+
 /*
  * Structures and definitions for converting between cpu, node, pnode, and blade
  * numbers.
@@ -351,5 +397,20 @@ static inline int uv_num_possible_blades(void)
        return uv_possible_blades;
 }
 
-#endif /* _ASM_X86_UV_UV_HUB_H */
+/* Update SCIR state */
+static inline void uv_set_scir_bits(unsigned char value)
+{
+       if (uv_hub_info->scir.state != value) {
+               uv_hub_info->scir.state = value;
+               uv_write_local_mmr8(uv_hub_info->scir.offset, value);
+       }
+}
+static inline void uv_set_cpu_scir_bits(int cpu, unsigned char value)
+{
+       if (uv_cpu_hub_info(cpu)->scir.state != value) {
+               uv_cpu_hub_info(cpu)->scir.state = value;
+               uv_write_local_mmr8(uv_cpu_hub_info(cpu)->scir.offset, value);
+       }
+}
 
+#endif /* _ASM_X86_UV_UV_HUB_H */
index f0dfe6f17e7eabbe41f90a8afd75c54766081fa2..d22d0f1bbea04c652c6cd3dbad12c1bcc0b864fa 100644 (file)
@@ -69,10 +69,10 @@ s64 uv_bios_call_reentrant(enum uv_bios_cmd which, u64 a1, u64 a2, u64 a3,
 
 long sn_partition_id;
 EXPORT_SYMBOL_GPL(sn_partition_id);
-long uv_coherency_id;
-EXPORT_SYMBOL_GPL(uv_coherency_id);
-long uv_region_size;
-EXPORT_SYMBOL_GPL(uv_region_size);
+long sn_coherency_id;
+EXPORT_SYMBOL_GPL(sn_coherency_id);
+long sn_region_size;
+EXPORT_SYMBOL_GPL(sn_region_size);
 int uv_type;
 
 
@@ -100,6 +100,58 @@ s64 uv_bios_get_sn_info(int fc, int *uvtype, long *partid, long *coher,
        return ret;
 }
 
+int
+uv_bios_mq_watchlist_alloc(int blade, void *mq, unsigned int mq_size,
+                          unsigned long *intr_mmr_offset)
+{
+       union uv_watchlist_u size_blade;
+       unsigned long addr;
+       u64 watchlist;
+       s64 ret;
+
+       addr = (unsigned long)mq;
+       size_blade.size = mq_size;
+       size_blade.blade = blade;
+
+       /*
+        * bios returns watchlist number or negative error number.
+        */
+       ret = (int)uv_bios_call_irqsave(UV_BIOS_WATCHLIST_ALLOC, addr,
+                       size_blade.val, (u64)intr_mmr_offset,
+                       (u64)&watchlist, 0);
+       if (ret < BIOS_STATUS_SUCCESS)
+               return ret;
+
+       return watchlist;
+}
+EXPORT_SYMBOL_GPL(uv_bios_mq_watchlist_alloc);
+
+int
+uv_bios_mq_watchlist_free(int blade, int watchlist_num)
+{
+       return (int)uv_bios_call_irqsave(UV_BIOS_WATCHLIST_FREE,
+                               blade, watchlist_num, 0, 0, 0);
+}
+EXPORT_SYMBOL_GPL(uv_bios_mq_watchlist_free);
+
+s64
+uv_bios_change_memprotect(u64 paddr, u64 len, enum uv_memprotect perms)
+{
+       return uv_bios_call_irqsave(UV_BIOS_MEMPROTECT, paddr, len,
+                                       perms, 0, 0);
+}
+EXPORT_SYMBOL_GPL(uv_bios_change_memprotect);
+
+s64
+uv_bios_reserved_page_pa(u64 buf, u64 *cookie, u64 *addr, u64 *len)
+{
+       s64 ret;
+
+       ret = uv_bios_call_irqsave(UV_BIOS_GET_PARTITION_ADDR, (u64)cookie,
+                                       (u64)addr, buf, (u64)len, 0);
+       return ret;
+}
+EXPORT_SYMBOL_GPL(uv_bios_reserved_page_pa);
 
 s64 uv_bios_freq_base(u64 clock_type, u64 *ticks_per_second)
 {
index 2c7dbdb98278316a6fd7228e9a7b2eaf76f58562..221299f4509f5ec6e2e545eaf7841c1cdb093149 100644 (file)
@@ -10,6 +10,7 @@
 
 #include <linux/kernel.h>
 #include <linux/threads.h>
+#include <linux/cpu.h>
 #include <linux/cpumask.h>
 #include <linux/string.h>
 #include <linux/ctype.h>
@@ -17,6 +18,9 @@
 #include <linux/sched.h>
 #include <linux/module.h>
 #include <linux/hardirq.h>
+#include <linux/timer.h>
+#include <linux/proc_fs.h>
+#include <asm/current.h>
 #include <asm/smp.h>
 #include <asm/ipi.h>
 #include <asm/genapic.h>
@@ -355,6 +359,103 @@ static __init void uv_rtc_init(void)
                sn_rtc_cycles_per_second = ticks_per_sec;
 }
 
+/*
+ * percpu heartbeat timer
+ */
+static void uv_heartbeat(unsigned long ignored)
+{
+       struct timer_list *timer = &uv_hub_info->scir.timer;
+       unsigned char bits = uv_hub_info->scir.state;
+
+       /* flip heartbeat bit */
+       bits ^= SCIR_CPU_HEARTBEAT;
+
+       /* is this cpu idle? */
+       if (idle_cpu(raw_smp_processor_id()))
+               bits &= ~SCIR_CPU_ACTIVITY;
+       else
+               bits |= SCIR_CPU_ACTIVITY;
+
+       /* update system controller interface reg */
+       uv_set_scir_bits(bits);
+
+       /* enable next timer period */
+       mod_timer(timer, jiffies + SCIR_CPU_HB_INTERVAL);
+}
+
+static void __cpuinit uv_heartbeat_enable(int cpu)
+{
+       if (!uv_cpu_hub_info(cpu)->scir.enabled) {
+               struct timer_list *timer = &uv_cpu_hub_info(cpu)->scir.timer;
+
+               uv_set_cpu_scir_bits(cpu, SCIR_CPU_HEARTBEAT|SCIR_CPU_ACTIVITY);
+               setup_timer(timer, uv_heartbeat, cpu);
+               timer->expires = jiffies + SCIR_CPU_HB_INTERVAL;
+               add_timer_on(timer, cpu);
+               uv_cpu_hub_info(cpu)->scir.enabled = 1;
+       }
+
+       /* check boot cpu */
+       if (!uv_cpu_hub_info(0)->scir.enabled)
+               uv_heartbeat_enable(0);
+}
+
+#ifdef CONFIG_HOTPLUG_CPU
+static void __cpuinit uv_heartbeat_disable(int cpu)
+{
+       if (uv_cpu_hub_info(cpu)->scir.enabled) {
+               uv_cpu_hub_info(cpu)->scir.enabled = 0;
+               del_timer(&uv_cpu_hub_info(cpu)->scir.timer);
+       }
+       uv_set_cpu_scir_bits(cpu, 0xff);
+}
+
+/*
+ * cpu hotplug notifier
+ */
+static __cpuinit int uv_scir_cpu_notify(struct notifier_block *self,
+                                      unsigned long action, void *hcpu)
+{
+       long cpu = (long)hcpu;
+
+       switch (action) {
+       case CPU_ONLINE:
+               uv_heartbeat_enable(cpu);
+               break;
+       case CPU_DOWN_PREPARE:
+               uv_heartbeat_disable(cpu);
+               break;
+       default:
+               break;
+       }
+       return NOTIFY_OK;
+}
+
+static __init void uv_scir_register_cpu_notifier(void)
+{
+       hotcpu_notifier(uv_scir_cpu_notify, 0);
+}
+
+#else /* !CONFIG_HOTPLUG_CPU */
+
+static __init void uv_scir_register_cpu_notifier(void)
+{
+}
+
+static __init int uv_init_heartbeat(void)
+{
+       int cpu;
+
+       if (is_uv_system())
+               for_each_online_cpu(cpu)
+                       uv_heartbeat_enable(cpu);
+       return 0;
+}
+
+late_initcall(uv_init_heartbeat);
+
+#endif /* !CONFIG_HOTPLUG_CPU */
+
 /*
  * Called on each cpu to initialize the per_cpu UV data area.
  *     ZZZ hotplug not supported yet
@@ -428,7 +529,7 @@ void __init uv_system_init(void)
 
        uv_bios_init();
        uv_bios_get_sn_info(0, &uv_type, &sn_partition_id,
-                           &uv_coherency_id, &uv_region_size);
+                           &sn_coherency_id, &sn_region_size);
        uv_rtc_init();
 
        for_each_present_cpu(cpu) {
@@ -450,7 +551,8 @@ void __init uv_system_init(void)
                uv_cpu_hub_info(cpu)->gpa_mask = (1 << (m_val + n_val)) - 1;
                uv_cpu_hub_info(cpu)->gnode_upper = gnode_upper;
                uv_cpu_hub_info(cpu)->global_mmr_base = mmr_base;
-               uv_cpu_hub_info(cpu)->coherency_domain_number = uv_coherency_id;
+               uv_cpu_hub_info(cpu)->coherency_domain_number = sn_coherency_id;
+               uv_cpu_hub_info(cpu)->scir.offset = SCIR_LOCAL_MMR_BASE + lcpu;
                uv_node_to_blade[nid] = blade;
                uv_cpu_to_blade[cpu] = blade;
                max_pnode = max(pnode, max_pnode);
@@ -467,4 +569,6 @@ void __init uv_system_init(void)
        map_mmioh_high(max_pnode);
 
        uv_cpu_init();
+       uv_scir_register_cpu_notifier();
+       proc_mkdir("sgi_uv", NULL);
 }
index 04431f34fd16f24946bb66b4eedf69ed7e51cf6f..6a00e5faaa74375d5b709488a0b9c371c3fe5c79 100644 (file)
@@ -566,14 +566,10 @@ static int __init uv_ptc_init(void)
        if (!is_uv_system())
                return 0;
 
-       if (!proc_mkdir("sgi_uv", NULL))
-               return -EINVAL;
-
        proc_uv_ptc = create_proc_entry(UV_PTC_BASENAME, 0444, NULL);
        if (!proc_uv_ptc) {
                printk(KERN_ERR "unable to create %s proc entry\n",
                       UV_PTC_BASENAME);
-               remove_proc_entry("sgi_uv", NULL);
                return -EINVAL;
        }
        proc_uv_ptc->proc_fops = &proc_uv_ptc_operations;
index 533923f83f1aa061635571b86934ad2715014be7..73b0ca061bb567401cc4d30a145c9daf05e72160 100644 (file)
@@ -317,7 +317,6 @@ int gru_proc_init(void)
 {
        struct proc_entry *p;
 
-       proc_mkdir("sgi_uv", NULL);
        proc_gru = proc_mkdir("sgi_uv/gru", NULL);
 
        for (p = proc_files; p->name; p++)
index ed1722e50049ff2aea256edbb574bb9969cb3bda..7b4cbd5e03e99a3e4297dc6d4b6de219d30f7389 100644 (file)
@@ -194,9 +194,10 @@ enum xp_retval {
        xpGruSendMqError,       /* 59: gru send message queue related error */
 
        xpBadChannelNumber,     /* 60: invalid channel number */
-       xpBadMsgType,           /* 60: invalid message type */
+       xpBadMsgType,           /* 61: invalid message type */
+       xpBiosError,            /* 62: BIOS error */
 
-       xpUnknownReason         /* 61: unknown reason - must be last in enum */
+       xpUnknownReason         /* 63: unknown reason - must be last in enum */
 };
 
 /*
@@ -345,6 +346,8 @@ extern unsigned long (*xp_pa) (void *);
 extern enum xp_retval (*xp_remote_memcpy) (unsigned long, const unsigned long,
                       size_t);
 extern int (*xp_cpu_to_nasid) (int);
+extern enum xp_retval (*xp_expand_memprotect) (unsigned long, unsigned long);
+extern enum xp_retval (*xp_restrict_memprotect) (unsigned long, unsigned long);
 
 extern u64 xp_nofault_PIOR_target;
 extern int xp_nofault_PIOR(void *);
index 66a1d19e08ad73ccd651f22efaddc4605fc0857a..9a2e77172d942f548402e9ce8ccce25d9e3bf935 100644 (file)
@@ -51,6 +51,13 @@ EXPORT_SYMBOL_GPL(xp_remote_memcpy);
 int (*xp_cpu_to_nasid) (int cpuid);
 EXPORT_SYMBOL_GPL(xp_cpu_to_nasid);
 
+enum xp_retval (*xp_expand_memprotect) (unsigned long phys_addr,
+                                       unsigned long size);
+EXPORT_SYMBOL_GPL(xp_expand_memprotect);
+enum xp_retval (*xp_restrict_memprotect) (unsigned long phys_addr,
+                                         unsigned long size);
+EXPORT_SYMBOL_GPL(xp_restrict_memprotect);
+
 /*
  * xpc_registrations[] keeps track of xpc_connect()'s done by the kernel-level
  * users of XPC.
index 1440134caf3166d25339771b73df741b814e35e1..fb3ec9d735a901fb7825c1d254f188d0334a2509 100644 (file)
@@ -120,6 +120,38 @@ xp_cpu_to_nasid_sn2(int cpuid)
        return cpuid_to_nasid(cpuid);
 }
 
+static enum xp_retval
+xp_expand_memprotect_sn2(unsigned long phys_addr, unsigned long size)
+{
+       u64 nasid_array = 0;
+       int ret;
+
+       ret = sn_change_memprotect(phys_addr, size, SN_MEMPROT_ACCESS_CLASS_1,
+                                  &nasid_array);
+       if (ret != 0) {
+               dev_err(xp, "sn_change_memprotect(,, "
+                       "SN_MEMPROT_ACCESS_CLASS_1,) failed ret=%d\n", ret);
+               return xpSalError;
+       }
+       return xpSuccess;
+}
+
+static enum xp_retval
+xp_restrict_memprotect_sn2(unsigned long phys_addr, unsigned long size)
+{
+       u64 nasid_array = 0;
+       int ret;
+
+       ret = sn_change_memprotect(phys_addr, size, SN_MEMPROT_ACCESS_CLASS_0,
+                                  &nasid_array);
+       if (ret != 0) {
+               dev_err(xp, "sn_change_memprotect(,, "
+                       "SN_MEMPROT_ACCESS_CLASS_0,) failed ret=%d\n", ret);
+               return xpSalError;
+       }
+       return xpSuccess;
+}
+
 enum xp_retval
 xp_init_sn2(void)
 {
@@ -132,6 +164,8 @@ xp_init_sn2(void)
        xp_pa = xp_pa_sn2;
        xp_remote_memcpy = xp_remote_memcpy_sn2;
        xp_cpu_to_nasid = xp_cpu_to_nasid_sn2;
+       xp_expand_memprotect = xp_expand_memprotect_sn2;
+       xp_restrict_memprotect = xp_restrict_memprotect_sn2;
 
        return xp_register_nofault_code_sn2();
 }
index d9f7ce2510bcc26fe6f7c31fd287ebd4ad91f96b..d238576b26fa9b6c8f5d96b8f41760db444d43f1 100644 (file)
 
 #include <linux/device.h>
 #include <asm/uv/uv_hub.h>
+#if defined CONFIG_X86_64
+#include <asm/uv/bios.h>
+#elif defined CONFIG_IA64_GENERIC || defined CONFIG_IA64_SGI_UV
+#include <asm/sn/sn_sal.h>
+#endif
 #include "../sgi-gru/grukservices.h"
 #include "xp.h"
 
@@ -49,18 +54,79 @@ xp_cpu_to_nasid_uv(int cpuid)
        return UV_PNODE_TO_NASID(uv_cpu_to_pnode(cpuid));
 }
 
+static enum xp_retval
+xp_expand_memprotect_uv(unsigned long phys_addr, unsigned long size)
+{
+       int ret;
+
+#if defined CONFIG_X86_64
+       ret = uv_bios_change_memprotect(phys_addr, size, UV_MEMPROT_ALLOW_RW);
+       if (ret != BIOS_STATUS_SUCCESS) {
+               dev_err(xp, "uv_bios_change_memprotect(,, "
+                       "UV_MEMPROT_ALLOW_RW) failed, ret=%d\n", ret);
+               return xpBiosError;
+       }
+
+#elif defined CONFIG_IA64_GENERIC || defined CONFIG_IA64_SGI_UV
+       u64 nasid_array;
+
+       ret = sn_change_memprotect(phys_addr, size, SN_MEMPROT_ACCESS_CLASS_1,
+                                  &nasid_array);
+       if (ret != 0) {
+               dev_err(xp, "sn_change_memprotect(,, "
+                       "SN_MEMPROT_ACCESS_CLASS_1,) failed ret=%d\n", ret);
+               return xpSalError;
+       }
+#else
+       #error not a supported configuration
+#endif
+       return xpSuccess;
+}
+
+static enum xp_retval
+xp_restrict_memprotect_uv(unsigned long phys_addr, unsigned long size)
+{
+       int ret;
+
+#if defined CONFIG_X86_64
+       ret = uv_bios_change_memprotect(phys_addr, size,
+                                       UV_MEMPROT_RESTRICT_ACCESS);
+       if (ret != BIOS_STATUS_SUCCESS) {
+               dev_err(xp, "uv_bios_change_memprotect(,, "
+                       "UV_MEMPROT_RESTRICT_ACCESS) failed, ret=%d\n", ret);
+               return xpBiosError;
+       }
+
+#elif defined CONFIG_IA64_GENERIC || defined CONFIG_IA64_SGI_UV
+       u64 nasid_array;
+
+       ret = sn_change_memprotect(phys_addr, size, SN_MEMPROT_ACCESS_CLASS_0,
+                                  &nasid_array);
+       if (ret != 0) {
+               dev_err(xp, "sn_change_memprotect(,, "
+                       "SN_MEMPROT_ACCESS_CLASS_0,) failed ret=%d\n", ret);
+               return xpSalError;
+       }
+#else
+       #error not a supported configuration
+#endif
+       return xpSuccess;
+}
+
 enum xp_retval
 xp_init_uv(void)
 {
        BUG_ON(!is_uv());
 
        xp_max_npartitions = XP_MAX_NPARTITIONS_UV;
-       xp_partition_id = 0;    /* !!! not correct value */
-       xp_region_size = 0;     /* !!! not correct value */
+       xp_partition_id = sn_partition_id;
+       xp_region_size = sn_region_size;
 
        xp_pa = xp_pa_uv;
        xp_remote_memcpy = xp_remote_memcpy_uv;
        xp_cpu_to_nasid = xp_cpu_to_nasid_uv;
+       xp_expand_memprotect = xp_expand_memprotect_uv;
+       xp_restrict_memprotect = xp_restrict_memprotect_uv;
 
        return xpSuccess;
 }
index 619208d618627692c8cbaec3e0448b1d15479577..a5bd658c2e83b13a5d9dee5805a71eb191750e21 100644 (file)
@@ -180,6 +180,18 @@ struct xpc_vars_part_sn2 {
                                 (XPC_RP_MACH_NASIDS(_rp) + \
                                  xpc_nasid_mask_nlongs))
 
+/*
+ * Info pertinent to a GRU message queue using a watch list for irq generation.
+ */
+struct xpc_gru_mq_uv {
+       void *address;          /* address of GRU message queue */
+       unsigned int order;     /* size of GRU message queue as a power of 2 */
+       int irq;                /* irq raised when message is received in mq */
+       int mmr_blade;          /* blade where watchlist was allocated from */
+       unsigned long mmr_offset; /* offset of irq mmr located on mmr_blade */
+       int watchlist_num;      /* number of watchlist allocatd by BIOS */
+};
+
 /*
  * The activate_mq is used to send/receive GRU messages that affect XPC's
  * heartbeat, partition active state, and channel state. This is UV only.
index b4882ccf6344a69607ee8c62f1ff3c84c32ce894..73b7fb8de47a32e0e38b8abe048ede7333fbe832 100644 (file)
@@ -553,22 +553,17 @@ static u64 xpc_prot_vec_sn2[MAX_NUMNODES];
 static enum xp_retval
 xpc_allow_amo_ops_sn2(struct amo *amos_page)
 {
-       u64 nasid_array = 0;
-       int ret;
+       enum xp_retval ret = xpSuccess;
 
        /*
         * On SHUB 1.1, we cannot call sn_change_memprotect() since the BIST
         * collides with memory operations. On those systems we call
         * xpc_allow_amo_ops_shub_wars_1_1_sn2() instead.
         */
-       if (!enable_shub_wars_1_1()) {
-               ret = sn_change_memprotect(ia64_tpa((u64)amos_page), PAGE_SIZE,
-                                          SN_MEMPROT_ACCESS_CLASS_1,
-                                          &nasid_array);
-               if (ret != 0)
-                       return xpSalError;
-       }
-       return xpSuccess;
+       if (!enable_shub_wars_1_1())
+               ret = xp_expand_memprotect(ia64_tpa((u64)amos_page), PAGE_SIZE);
+
+       return ret;
 }
 
 /*
index 1ac694c01623b99419e3b7d2afb9a2c16d8ceda7..684b2dd17583f76433f51f2b5befba3d4473394a 100644 (file)
 #include <linux/interrupt.h>
 #include <linux/delay.h>
 #include <linux/device.h>
+#include <linux/err.h>
 #include <asm/uv/uv_hub.h>
+#if defined CONFIG_X86_64
+#include <asm/uv/bios.h>
+#include <asm/uv/uv_irq.h>
+#elif defined CONFIG_IA64_GENERIC || defined CONFIG_IA64_SGI_UV
+#include <asm/sn/intr.h>
+#include <asm/sn/sn_sal.h>
+#endif
 #include "../sgi-gru/gru.h"
 #include "../sgi-gru/grukservices.h"
 #include "xpc.h"
@@ -27,15 +35,17 @@ static atomic64_t xpc_heartbeat_uv;
 static DECLARE_BITMAP(xpc_heartbeating_to_mask_uv, XP_MAX_NPARTITIONS_UV);
 
 #define XPC_ACTIVATE_MSG_SIZE_UV       (1 * GRU_CACHE_LINE_BYTES)
-#define XPC_NOTIFY_MSG_SIZE_UV         (2 * GRU_CACHE_LINE_BYTES)
+#define XPC_ACTIVATE_MQ_SIZE_UV                (4 * XP_MAX_NPARTITIONS_UV * \
+                                        XPC_ACTIVATE_MSG_SIZE_UV)
+#define XPC_ACTIVATE_IRQ_NAME          "xpc_activate"
 
-#define XPC_ACTIVATE_MQ_SIZE_UV        (4 * XP_MAX_NPARTITIONS_UV * \
-                                XPC_ACTIVATE_MSG_SIZE_UV)
-#define XPC_NOTIFY_MQ_SIZE_UV  (4 * XP_MAX_NPARTITIONS_UV * \
-                                XPC_NOTIFY_MSG_SIZE_UV)
+#define XPC_NOTIFY_MSG_SIZE_UV         (2 * GRU_CACHE_LINE_BYTES)
+#define XPC_NOTIFY_MQ_SIZE_UV          (4 * XP_MAX_NPARTITIONS_UV * \
+                                        XPC_NOTIFY_MSG_SIZE_UV)
+#define XPC_NOTIFY_IRQ_NAME            "xpc_notify"
 
-static void *xpc_activate_mq_uv;
-static void *xpc_notify_mq_uv;
+static struct xpc_gru_mq_uv *xpc_activate_mq_uv;
+static struct xpc_gru_mq_uv *xpc_notify_mq_uv;
 
 static int
 xpc_setup_partitions_sn_uv(void)
@@ -52,62 +62,209 @@ xpc_setup_partitions_sn_uv(void)
        return 0;
 }
 
-static void *
-xpc_create_gru_mq_uv(unsigned int mq_size, int cpuid, unsigned int irq,
+static int
+xpc_get_gru_mq_irq_uv(struct xpc_gru_mq_uv *mq, int cpu, char *irq_name)
+{
+#if defined CONFIG_X86_64
+       mq->irq = uv_setup_irq(irq_name, cpu, mq->mmr_blade, mq->mmr_offset);
+       if (mq->irq < 0) {
+               dev_err(xpc_part, "uv_setup_irq() returned error=%d\n",
+                       mq->irq);
+       }
+
+#elif defined CONFIG_IA64_GENERIC || defined CONFIG_IA64_SGI_UV
+       int mmr_pnode;
+       unsigned long mmr_value;
+
+       if (strcmp(irq_name, XPC_ACTIVATE_IRQ_NAME) == 0)
+               mq->irq = SGI_XPC_ACTIVATE;
+       else if (strcmp(irq_name, XPC_NOTIFY_IRQ_NAME) == 0)
+               mq->irq = SGI_XPC_NOTIFY;
+       else
+               return -EINVAL;
+
+       mmr_pnode = uv_blade_to_pnode(mq->mmr_blade);
+       mmr_value = (unsigned long)cpu_physical_id(cpu) << 32 | mq->irq;
+
+       uv_write_global_mmr64(mmr_pnode, mq->mmr_offset, mmr_value);
+#else
+       #error not a supported configuration
+#endif
+
+       return 0;
+}
+
+static void
+xpc_release_gru_mq_irq_uv(struct xpc_gru_mq_uv *mq)
+{
+#if defined CONFIG_X86_64
+       uv_teardown_irq(mq->irq, mq->mmr_blade, mq->mmr_offset);
+
+#elif defined CONFIG_IA64_GENERIC || defined CONFIG_IA64_SGI_UV
+       int mmr_pnode;
+       unsigned long mmr_value;
+
+       mmr_pnode = uv_blade_to_pnode(mq->mmr_blade);
+       mmr_value = 1UL << 16;
+
+       uv_write_global_mmr64(mmr_pnode, mq->mmr_offset, mmr_value);
+#else
+       #error not a supported configuration
+#endif
+}
+
+static int
+xpc_gru_mq_watchlist_alloc_uv(struct xpc_gru_mq_uv *mq)
+{
+       int ret;
+
+#if defined CONFIG_X86_64
+       ret = uv_bios_mq_watchlist_alloc(mq->mmr_blade, mq->address, mq->order,
+                                        &mq->mmr_offset);
+       if (ret < 0) {
+               dev_err(xpc_part, "uv_bios_mq_watchlist_alloc() failed, "
+                       "ret=%d\n", ret);
+               return ret;
+       }
+#elif defined CONFIG_IA64_GENERIC || defined CONFIG_IA64_SGI_UV
+       ret = sn_mq_watchlist_alloc(mq->mmr_blade, mq->address, mq->order,
+                                   &mq->mmr_offset);
+       if (ret < 0) {
+               dev_err(xpc_part, "sn_mq_watchlist_alloc() failed, ret=%d\n",
+                       ret);
+               return -EBUSY;
+       }
+#else
+       #error not a supported configuration
+#endif
+
+       mq->watchlist_num = ret;
+       return 0;
+}
+
+static void
+xpc_gru_mq_watchlist_free_uv(struct xpc_gru_mq_uv *mq)
+{
+       int ret;
+
+#if defined CONFIG_X86_64
+       ret = uv_bios_mq_watchlist_free(mq->mmr_blade, mq->watchlist_num);
+       BUG_ON(ret != BIOS_STATUS_SUCCESS);
+#elif defined CONFIG_IA64_GENERIC || defined CONFIG_IA64_SGI_UV
+       ret = sn_mq_watchlist_free(mq->mmr_blade, mq->watchlist_num);
+       BUG_ON(ret != SALRET_OK);
+#else
+       #error not a supported configuration
+#endif
+}
+
+static struct xpc_gru_mq_uv *
+xpc_create_gru_mq_uv(unsigned int mq_size, int cpu, char *irq_name,
                     irq_handler_t irq_handler)
 {
+       enum xp_retval xp_ret;
        int ret;
        int nid;
-       int mq_order;
+       int pg_order;
        struct page *page;
-       void *mq;
+       struct xpc_gru_mq_uv *mq;
+
+       mq = kmalloc(sizeof(struct xpc_gru_mq_uv), GFP_KERNEL);
+       if (mq == NULL) {
+               dev_err(xpc_part, "xpc_create_gru_mq_uv() failed to kmalloc() "
+                       "a xpc_gru_mq_uv structure\n");
+               ret = -ENOMEM;
+               goto out_1;
+       }
+
+       pg_order = get_order(mq_size);
+       mq->order = pg_order + PAGE_SHIFT;
+       mq_size = 1UL << mq->order;
+
+       mq->mmr_blade = uv_cpu_to_blade_id(cpu);
 
-       nid = cpu_to_node(cpuid);
-       mq_order = get_order(mq_size);
+       nid = cpu_to_node(cpu);
        page = alloc_pages_node(nid, GFP_KERNEL | __GFP_ZERO | GFP_THISNODE,
-                               mq_order);
+                               pg_order);
        if (page == NULL) {
                dev_err(xpc_part, "xpc_create_gru_mq_uv() failed to alloc %d "
                        "bytes of memory on nid=%d for GRU mq\n", mq_size, nid);
-               return NULL;
+               ret = -ENOMEM;
+               goto out_2;
        }
+       mq->address = page_address(page);
 
-       mq = page_address(page);
-       ret = gru_create_message_queue(mq, mq_size);
+       ret = gru_create_message_queue(mq->address, mq_size);
        if (ret != 0) {
                dev_err(xpc_part, "gru_create_message_queue() returned "
                        "error=%d\n", ret);
-               free_pages((unsigned long)mq, mq_order);
-               return NULL;
+               ret = -EINVAL;
+               goto out_3;
        }
 
-       /* !!! Need to do some other things to set up IRQ */
+       /* enable generation of irq when GRU mq operation occurs to this mq */
+       ret = xpc_gru_mq_watchlist_alloc_uv(mq);
+       if (ret != 0)
+               goto out_3;
 
-       ret = request_irq(irq, irq_handler, 0, "xpc", NULL);
+       ret = xpc_get_gru_mq_irq_uv(mq, cpu, irq_name);
+       if (ret != 0)
+               goto out_4;
+
+       ret = request_irq(mq->irq, irq_handler, 0, irq_name, NULL);
        if (ret != 0) {
                dev_err(xpc_part, "request_irq(irq=%d) returned error=%d\n",
-                       irq, ret);
-               free_pages((unsigned long)mq, mq_order);
-               return NULL;
+                       mq->irq, ret);
+               goto out_5;
        }
 
-       /* !!! enable generation of irq when GRU mq op occurs to this mq */
-
-       /* ??? allow other partitions to access GRU mq? */
+       /* allow other partitions to access this GRU mq */
+       xp_ret = xp_expand_memprotect(xp_pa(mq->address), mq_size);
+       if (xp_ret != xpSuccess) {
+               ret = -EACCES;
+               goto out_6;
+       }
 
        return mq;
+
+       /* something went wrong */
+out_6:
+       free_irq(mq->irq, NULL);
+out_5:
+       xpc_release_gru_mq_irq_uv(mq);
+out_4:
+       xpc_gru_mq_watchlist_free_uv(mq);
+out_3:
+       free_pages((unsigned long)mq->address, pg_order);
+out_2:
+       kfree(mq);
+out_1:
+       return ERR_PTR(ret);
 }
 
 static void
-xpc_destroy_gru_mq_uv(void *mq, unsigned int mq_size, unsigned int irq)
+xpc_destroy_gru_mq_uv(struct xpc_gru_mq_uv *mq)
 {
-       /* ??? disallow other partitions to access GRU mq? */
+       unsigned int mq_size;
+       int pg_order;
+       int ret;
+
+       /* disallow other partitions to access GRU mq */
+       mq_size = 1UL << mq->order;
+       ret = xp_restrict_memprotect(xp_pa(mq->address), mq_size);
+       BUG_ON(ret != xpSuccess);
 
-       /* !!! disable generation of irq when GRU mq op occurs to this mq */
+       /* unregister irq handler and release mq irq/vector mapping */
+       free_irq(mq->irq, NULL);
+       xpc_release_gru_mq_irq_uv(mq);
 
-       free_irq(irq, NULL);
+       /* disable generation of irq when GRU mq op occurs to this mq */
+       xpc_gru_mq_watchlist_free_uv(mq);
 
-       free_pages((unsigned long)mq, get_order(mq_size));
+       pg_order = mq->order - PAGE_SHIFT;
+       free_pages((unsigned long)mq->address, pg_order);
+
+       kfree(mq);
 }
 
 static enum xp_retval
@@ -402,7 +559,10 @@ xpc_handle_activate_IRQ_uv(int irq, void *dev_id)
        struct xpc_partition *part;
        int wakeup_hb_checker = 0;
 
-       while ((msg_hdr = gru_get_next_message(xpc_activate_mq_uv)) != NULL) {
+       while (1) {
+               msg_hdr = gru_get_next_message(xpc_activate_mq_uv->address);
+               if (msg_hdr == NULL)
+                       break;
 
                partid = msg_hdr->partid;
                if (partid < 0 || partid >= XP_MAX_NPARTITIONS_UV) {
@@ -418,7 +578,7 @@ xpc_handle_activate_IRQ_uv(int irq, void *dev_id)
                        }
                }
 
-               gru_free_message(xpc_activate_mq_uv, msg_hdr);
+               gru_free_message(xpc_activate_mq_uv->address, msg_hdr);
        }
 
        if (wakeup_hb_checker)
@@ -482,7 +642,7 @@ xpc_send_local_activate_IRQ_uv(struct xpc_partition *part, int act_state_req)
        struct xpc_partition_uv *part_uv = &part->sn.uv;
 
        /*
-        * !!! Make our side think that the remote parition sent an activate
+        * !!! Make our side think that the remote partition sent an activate
         * !!! message our way by doing what the activate IRQ handler would
         * !!! do had one really been sent.
         */
@@ -500,14 +660,39 @@ static enum xp_retval
 xpc_get_partition_rsvd_page_pa_uv(void *buf, u64 *cookie, unsigned long *rp_pa,
                                  size_t *len)
 {
-       /* !!! call the UV version of sn_partition_reserved_page_pa() */
-       return xpUnsupported;
+       s64 status;
+       enum xp_retval ret;
+
+#if defined CONFIG_X86_64
+       status = uv_bios_reserved_page_pa((u64)buf, cookie, (u64 *)rp_pa,
+                                         (u64 *)len);
+       if (status == BIOS_STATUS_SUCCESS)
+               ret = xpSuccess;
+       else if (status == BIOS_STATUS_MORE_PASSES)
+               ret = xpNeedMoreInfo;
+       else
+               ret = xpBiosError;
+
+#elif defined CONFIG_IA64_GENERIC || defined CONFIG_IA64_SGI_UV
+       status = sn_partition_reserved_page_pa((u64)buf, cookie, rp_pa, len);
+       if (status == SALRET_OK)
+               ret = xpSuccess;
+       else if (status == SALRET_MORE_PASSES)
+               ret = xpNeedMoreInfo;
+       else
+               ret = xpSalError;
+
+#else
+       #error not a supported configuration
+#endif
+
+       return ret;
 }
 
 static int
 xpc_setup_rsvd_page_sn_uv(struct xpc_rsvd_page *rp)
 {
-       rp->sn.activate_mq_gpa = uv_gpa(xpc_activate_mq_uv);
+       rp->sn.activate_mq_gpa = uv_gpa(xpc_activate_mq_uv->address);
        return 0;
 }
 
@@ -1411,22 +1596,18 @@ xpc_init_uv(void)
                return -E2BIG;
        }
 
-       /* ??? The cpuid argument's value is 0, is that what we want? */
-       /* !!! The irq argument's value isn't correct. */
-       xpc_activate_mq_uv = xpc_create_gru_mq_uv(XPC_ACTIVATE_MQ_SIZE_UV, 0, 0,
+       xpc_activate_mq_uv = xpc_create_gru_mq_uv(XPC_ACTIVATE_MQ_SIZE_UV, 0,
+                                                 XPC_ACTIVATE_IRQ_NAME,
                                                  xpc_handle_activate_IRQ_uv);
-       if (xpc_activate_mq_uv == NULL)
-               return -ENOMEM;
+       if (IS_ERR(xpc_activate_mq_uv))
+               return PTR_ERR(xpc_activate_mq_uv);
 
-       /* ??? The cpuid argument's value is 0, is that what we want? */
-       /* !!! The irq argument's value isn't correct. */
-       xpc_notify_mq_uv = xpc_create_gru_mq_uv(XPC_NOTIFY_MQ_SIZE_UV, 0, 0,
+       xpc_notify_mq_uv = xpc_create_gru_mq_uv(XPC_NOTIFY_MQ_SIZE_UV, 0,
+                                               XPC_NOTIFY_IRQ_NAME,
                                                xpc_handle_notify_IRQ_uv);
-       if (xpc_notify_mq_uv == NULL) {
-               /* !!! The irq argument's value isn't correct. */
-               xpc_destroy_gru_mq_uv(xpc_activate_mq_uv,
-                                     XPC_ACTIVATE_MQ_SIZE_UV, 0);
-               return -ENOMEM;
+       if (IS_ERR(xpc_notify_mq_uv)) {
+               xpc_destroy_gru_mq_uv(xpc_activate_mq_uv);
+               return PTR_ERR(xpc_notify_mq_uv);
        }
 
        return 0;
@@ -1435,9 +1616,6 @@ xpc_init_uv(void)
 void
 xpc_exit_uv(void)
 {
-       /* !!! The irq argument's value isn't correct. */
-       xpc_destroy_gru_mq_uv(xpc_notify_mq_uv, XPC_NOTIFY_MQ_SIZE_UV, 0);
-
-       /* !!! The irq argument's value isn't correct. */
-       xpc_destroy_gru_mq_uv(xpc_activate_mq_uv, XPC_ACTIVATE_MQ_SIZE_UV, 0);
+       xpc_destroy_gru_mq_uv(xpc_notify_mq_uv);
+       xpc_destroy_gru_mq_uv(xpc_activate_mq_uv);
 }