Merge branch 'release' of git://git.kernel.org/pub/scm/linux/kernel/git/aegl/linux-2.6

* 'release' of git://git.kernel.org/pub/scm/linux/kernel/git/aegl/linux-2.6:
  [IA64] run drivers/misc/sgi-xp through scripts/checkpatch.pl
  [IA64] run rest drivers/misc/sgi-xp through scripts/Lindent
  [IA64] run some drivers/misc/sgi-xp through scripts/Lindent
  [IA64] move XP and XPC to drivers/misc/sgi-xp
  [IA64] minor irq handler cleanups
  [IA64] simplify notify hooks in mca.c
  [IA64] do notify DIE_MCA_MONARCH_PROCESS for each monarchs
  [IA64] disable interrupts on exit of ia64_trace_syscall
This commit is contained in:
Linus Torvalds
2008-04-22 16:47:54 -07:00
20 changed files with 963 additions and 1418 deletions

View File

@@ -266,17 +266,6 @@ config IOSAPIC
depends on !IA64_HP_SIM depends on !IA64_HP_SIM
default y default y
config IA64_SGI_SN_XP
tristate "Support communication between SGI SSIs"
depends on IA64_GENERIC || IA64_SGI_SN2
select IA64_UNCACHED_ALLOCATOR
help
An SGI machine can be divided into multiple Single System
Images which act independently of each other and have
hardware based memory protection from the others. Enabling
this feature will allow for direct communication between SSIs
based on a network adapter and DMA messaging.
config FORCE_MAX_ZONEORDER config FORCE_MAX_ZONEORDER
int "MAX_ORDER (11 - 17)" if !HUGETLB_PAGE int "MAX_ORDER (11 - 17)" if !HUGETLB_PAGE
range 11 17 if !HUGETLB_PAGE range 11 17 if !HUGETLB_PAGE

View File

@@ -194,8 +194,8 @@ kdump_init_notifier(struct notifier_block *self, unsigned long val, void *data)
unw_init_running(kdump_cpu_freeze, NULL); unw_init_running(kdump_cpu_freeze, NULL);
break; break;
case DIE_MCA_MONARCH_LEAVE: case DIE_MCA_MONARCH_LEAVE:
/* die_register->signr indicate if MCA is recoverable */ /* *(nd->data) indicate if MCA is recoverable */
if (kdump_on_fatal_mca && !args->signr) { if (kdump_on_fatal_mca && !(*(nd->data))) {
atomic_set(&kdump_in_progress, 1); atomic_set(&kdump_in_progress, 1);
*(nd->monarch_cpu) = -1; *(nd->monarch_cpu) = -1;
machine_kdump_on_init(); machine_kdump_on_init();

View File

@@ -570,6 +570,7 @@ GLOBAL_ENTRY(ia64_trace_syscall)
br.call.sptk.many rp=syscall_trace_leave // give parent a chance to catch return value br.call.sptk.many rp=syscall_trace_leave // give parent a chance to catch return value
.ret3: .ret3:
(pUStk) cmp.eq.unc p6,p0=r0,r0 // p6 <- pUStk (pUStk) cmp.eq.unc p6,p0=r0,r0 // p6 <- pUStk
(pUStk) rsm psr.i // disable interrupts
br.cond.sptk .work_pending_syscall_end br.cond.sptk .work_pending_syscall_end
strace_error: strace_error:

View File

@@ -109,6 +109,20 @@
# define IA64_MCA_DEBUG(fmt...) # define IA64_MCA_DEBUG(fmt...)
#endif #endif
#define NOTIFY_INIT(event, regs, arg, spin) \
do { \
if ((notify_die((event), "INIT", (regs), (arg), 0, 0) \
== NOTIFY_STOP) && ((spin) == 1)) \
ia64_mca_spin(__func__); \
} while (0)
#define NOTIFY_MCA(event, regs, arg, spin) \
do { \
if ((notify_die((event), "MCA", (regs), (arg), 0, 0) \
== NOTIFY_STOP) && ((spin) == 1)) \
ia64_mca_spin(__func__); \
} while (0)
/* Used by mca_asm.S */ /* Used by mca_asm.S */
DEFINE_PER_CPU(u64, ia64_mca_data); /* == __per_cpu_mca[smp_processor_id()] */ DEFINE_PER_CPU(u64, ia64_mca_data); /* == __per_cpu_mca[smp_processor_id()] */
DEFINE_PER_CPU(u64, ia64_mca_per_cpu_pte); /* PTE to map per-CPU area */ DEFINE_PER_CPU(u64, ia64_mca_per_cpu_pte); /* PTE to map per-CPU area */
@@ -766,9 +780,8 @@ ia64_mca_rendez_int_handler(int rendez_irq, void *arg)
/* Mask all interrupts */ /* Mask all interrupts */
local_irq_save(flags); local_irq_save(flags);
if (notify_die(DIE_MCA_RENDZVOUS_ENTER, "MCA", get_irq_regs(),
(long)&nd, 0, 0) == NOTIFY_STOP) NOTIFY_MCA(DIE_MCA_RENDZVOUS_ENTER, get_irq_regs(), (long)&nd, 1);
ia64_mca_spin(__func__);
ia64_mc_info.imi_rendez_checkin[cpu] = IA64_MCA_RENDEZ_CHECKIN_DONE; ia64_mc_info.imi_rendez_checkin[cpu] = IA64_MCA_RENDEZ_CHECKIN_DONE;
/* Register with the SAL monarch that the slave has /* Register with the SAL monarch that the slave has
@@ -776,17 +789,13 @@ ia64_mca_rendez_int_handler(int rendez_irq, void *arg)
*/ */
ia64_sal_mc_rendez(); ia64_sal_mc_rendez();
if (notify_die(DIE_MCA_RENDZVOUS_PROCESS, "MCA", get_irq_regs(), NOTIFY_MCA(DIE_MCA_RENDZVOUS_PROCESS, get_irq_regs(), (long)&nd, 1);
(long)&nd, 0, 0) == NOTIFY_STOP)
ia64_mca_spin(__func__);
/* Wait for the monarch cpu to exit. */ /* Wait for the monarch cpu to exit. */
while (monarch_cpu != -1) while (monarch_cpu != -1)
cpu_relax(); /* spin until monarch leaves */ cpu_relax(); /* spin until monarch leaves */
if (notify_die(DIE_MCA_RENDZVOUS_LEAVE, "MCA", get_irq_regs(), NOTIFY_MCA(DIE_MCA_RENDZVOUS_LEAVE, get_irq_regs(), (long)&nd, 1);
(long)&nd, 0, 0) == NOTIFY_STOP)
ia64_mca_spin(__func__);
ia64_mc_info.imi_rendez_checkin[cpu] = IA64_MCA_RENDEZ_CHECKIN_NOTDONE; ia64_mc_info.imi_rendez_checkin[cpu] = IA64_MCA_RENDEZ_CHECKIN_NOTDONE;
/* Enable all interrupts */ /* Enable all interrupts */
@@ -1256,7 +1265,7 @@ ia64_mca_handler(struct pt_regs *regs, struct switch_stack *sw,
int recover, cpu = smp_processor_id(); int recover, cpu = smp_processor_id();
struct task_struct *previous_current; struct task_struct *previous_current;
struct ia64_mca_notify_die nd = struct ia64_mca_notify_die nd =
{ .sos = sos, .monarch_cpu = &monarch_cpu }; { .sos = sos, .monarch_cpu = &monarch_cpu, .data = &recover };
static atomic_t mca_count; static atomic_t mca_count;
static cpumask_t mca_cpu; static cpumask_t mca_cpu;
@@ -1272,9 +1281,7 @@ ia64_mca_handler(struct pt_regs *regs, struct switch_stack *sw,
previous_current = ia64_mca_modify_original_stack(regs, sw, sos, "MCA"); previous_current = ia64_mca_modify_original_stack(regs, sw, sos, "MCA");
if (notify_die(DIE_MCA_MONARCH_ENTER, "MCA", regs, (long)&nd, 0, 0) NOTIFY_MCA(DIE_MCA_MONARCH_ENTER, regs, (long)&nd, 1);
== NOTIFY_STOP)
ia64_mca_spin(__func__);
ia64_mc_info.imi_rendez_checkin[cpu] = IA64_MCA_RENDEZ_CHECKIN_CONCURRENT_MCA; ia64_mc_info.imi_rendez_checkin[cpu] = IA64_MCA_RENDEZ_CHECKIN_CONCURRENT_MCA;
if (sos->monarch) { if (sos->monarch) {
@@ -1288,14 +1295,13 @@ ia64_mca_handler(struct pt_regs *regs, struct switch_stack *sw,
* does not work. * does not work.
*/ */
ia64_mca_wakeup_all(); ia64_mca_wakeup_all();
if (notify_die(DIE_MCA_MONARCH_PROCESS, "MCA", regs, (long)&nd, 0, 0)
== NOTIFY_STOP)
ia64_mca_spin(__func__);
} else { } else {
while (cpu_isset(cpu, mca_cpu)) while (cpu_isset(cpu, mca_cpu))
cpu_relax(); /* spin until monarch wakes us */ cpu_relax(); /* spin until monarch wakes us */
} }
NOTIFY_MCA(DIE_MCA_MONARCH_PROCESS, regs, (long)&nd, 1);
/* Get the MCA error record and log it */ /* Get the MCA error record and log it */
ia64_mca_log_sal_error_record(SAL_INFO_TYPE_MCA); ia64_mca_log_sal_error_record(SAL_INFO_TYPE_MCA);
@@ -1320,9 +1326,7 @@ ia64_mca_handler(struct pt_regs *regs, struct switch_stack *sw,
mca_insert_tr(0x2); /*Reload dynamic itrs*/ mca_insert_tr(0x2); /*Reload dynamic itrs*/
} }
if (notify_die(DIE_MCA_MONARCH_LEAVE, "MCA", regs, (long)&nd, 0, recover) NOTIFY_MCA(DIE_MCA_MONARCH_LEAVE, regs, (long)&nd, 1);
== NOTIFY_STOP)
ia64_mca_spin(__func__);
if (atomic_dec_return(&mca_count) > 0) { if (atomic_dec_return(&mca_count) > 0) {
int i; int i;
@@ -1643,7 +1647,7 @@ ia64_init_handler(struct pt_regs *regs, struct switch_stack *sw,
struct ia64_mca_notify_die nd = struct ia64_mca_notify_die nd =
{ .sos = sos, .monarch_cpu = &monarch_cpu }; { .sos = sos, .monarch_cpu = &monarch_cpu };
(void) notify_die(DIE_INIT_ENTER, "INIT", regs, (long)&nd, 0, 0); NOTIFY_INIT(DIE_INIT_ENTER, regs, (long)&nd, 0);
mprintk(KERN_INFO "Entered OS INIT handler. PSP=%lx cpu=%d monarch=%ld\n", mprintk(KERN_INFO "Entered OS INIT handler. PSP=%lx cpu=%d monarch=%ld\n",
sos->proc_state_param, cpu, sos->monarch); sos->proc_state_param, cpu, sos->monarch);
@@ -1680,17 +1684,15 @@ ia64_init_handler(struct pt_regs *regs, struct switch_stack *sw,
ia64_mc_info.imi_rendez_checkin[cpu] = IA64_MCA_RENDEZ_CHECKIN_INIT; ia64_mc_info.imi_rendez_checkin[cpu] = IA64_MCA_RENDEZ_CHECKIN_INIT;
while (monarch_cpu == -1) while (monarch_cpu == -1)
cpu_relax(); /* spin until monarch enters */ cpu_relax(); /* spin until monarch enters */
if (notify_die(DIE_INIT_SLAVE_ENTER, "INIT", regs, (long)&nd, 0, 0)
== NOTIFY_STOP) NOTIFY_INIT(DIE_INIT_SLAVE_ENTER, regs, (long)&nd, 1);
ia64_mca_spin(__func__); NOTIFY_INIT(DIE_INIT_SLAVE_PROCESS, regs, (long)&nd, 1);
if (notify_die(DIE_INIT_SLAVE_PROCESS, "INIT", regs, (long)&nd, 0, 0)
== NOTIFY_STOP)
ia64_mca_spin(__func__);
while (monarch_cpu != -1) while (monarch_cpu != -1)
cpu_relax(); /* spin until monarch leaves */ cpu_relax(); /* spin until monarch leaves */
if (notify_die(DIE_INIT_SLAVE_LEAVE, "INIT", regs, (long)&nd, 0, 0)
== NOTIFY_STOP) NOTIFY_INIT(DIE_INIT_SLAVE_LEAVE, regs, (long)&nd, 1);
ia64_mca_spin(__func__);
mprintk("Slave on cpu %d returning to normal service.\n", cpu); mprintk("Slave on cpu %d returning to normal service.\n", cpu);
set_curr_task(cpu, previous_current); set_curr_task(cpu, previous_current);
ia64_mc_info.imi_rendez_checkin[cpu] = IA64_MCA_RENDEZ_CHECKIN_NOTDONE; ia64_mc_info.imi_rendez_checkin[cpu] = IA64_MCA_RENDEZ_CHECKIN_NOTDONE;
@@ -1699,9 +1701,7 @@ ia64_init_handler(struct pt_regs *regs, struct switch_stack *sw,
} }
monarch_cpu = cpu; monarch_cpu = cpu;
if (notify_die(DIE_INIT_MONARCH_ENTER, "INIT", regs, (long)&nd, 0, 0) NOTIFY_INIT(DIE_INIT_MONARCH_ENTER, regs, (long)&nd, 1);
== NOTIFY_STOP)
ia64_mca_spin(__func__);
/* /*
* Wait for a bit. On some machines (e.g., HP's zx2000 and zx6000, INIT can be * Wait for a bit. On some machines (e.g., HP's zx2000 and zx6000, INIT can be
@@ -1716,12 +1716,9 @@ ia64_init_handler(struct pt_regs *regs, struct switch_stack *sw,
* to default_monarch_init_process() above and just print all the * to default_monarch_init_process() above and just print all the
* tasks. * tasks.
*/ */
if (notify_die(DIE_INIT_MONARCH_PROCESS, "INIT", regs, (long)&nd, 0, 0) NOTIFY_INIT(DIE_INIT_MONARCH_PROCESS, regs, (long)&nd, 1);
== NOTIFY_STOP) NOTIFY_INIT(DIE_INIT_MONARCH_LEAVE, regs, (long)&nd, 1);
ia64_mca_spin(__func__);
if (notify_die(DIE_INIT_MONARCH_LEAVE, "INIT", regs, (long)&nd, 0, 0)
== NOTIFY_STOP)
ia64_mca_spin(__func__);
mprintk("\nINIT dump complete. Monarch on cpu %d returning to normal service.\n", cpu); mprintk("\nINIT dump complete. Monarch on cpu %d returning to normal service.\n", cpu);
atomic_dec(&monarchs); atomic_dec(&monarchs);
set_curr_task(cpu, previous_current); set_curr_task(cpu, previous_current);
@@ -1953,7 +1950,7 @@ ia64_mca_init(void)
printk(KERN_INFO "Increasing MCA rendezvous timeout from " printk(KERN_INFO "Increasing MCA rendezvous timeout from "
"%ld to %ld milliseconds\n", timeout, isrv.v0); "%ld to %ld milliseconds\n", timeout, isrv.v0);
timeout = isrv.v0; timeout = isrv.v0;
(void) notify_die(DIE_MCA_NEW_TIMEOUT, "MCA", NULL, timeout, 0, 0); NOTIFY_MCA(DIE_MCA_NEW_TIMEOUT, NULL, timeout, 0);
continue; continue;
} }
printk(KERN_ERR "Failed to register rendezvous interrupt " printk(KERN_ERR "Failed to register rendezvous interrupt "

View File

@@ -5511,7 +5511,7 @@ stop_monitoring:
} }
static int static int
pfm_do_interrupt_handler(int irq, void *arg, struct pt_regs *regs) pfm_do_interrupt_handler(void *arg, struct pt_regs *regs)
{ {
struct task_struct *task; struct task_struct *task;
pfm_context_t *ctx; pfm_context_t *ctx;
@@ -5591,7 +5591,7 @@ pfm_interrupt_handler(int irq, void *arg)
start_cycles = ia64_get_itc(); start_cycles = ia64_get_itc();
ret = pfm_do_interrupt_handler(irq, arg, regs); ret = pfm_do_interrupt_handler(arg, regs);
total_cycles = ia64_get_itc(); total_cycles = ia64_get_itc();

View File

@@ -4,7 +4,7 @@
# License. See the file "COPYING" in the main directory of this archive # License. See the file "COPYING" in the main directory of this archive
# for more details. # for more details.
# #
# Copyright (C) 1999,2001-2006 Silicon Graphics, Inc. All Rights Reserved. # Copyright (C) 1999,2001-2006,2008 Silicon Graphics, Inc. All Rights Reserved.
# #
EXTRA_CFLAGS += -Iarch/ia64/sn/include EXTRA_CFLAGS += -Iarch/ia64/sn/include
@@ -15,9 +15,4 @@ obj-y += setup.o bte.o bte_error.o irq.o mca.o idle.o \
sn2/ sn2/
obj-$(CONFIG_IA64_GENERIC) += machvec.o obj-$(CONFIG_IA64_GENERIC) += machvec.o
obj-$(CONFIG_SGI_TIOCX) += tiocx.o obj-$(CONFIG_SGI_TIOCX) += tiocx.o
obj-$(CONFIG_IA64_SGI_SN_XP) += xp.o
xp-y := xp_main.o xp_nofault.o
obj-$(CONFIG_IA64_SGI_SN_XP) += xpc.o
xpc-y := xpc_main.o xpc_channel.o xpc_partition.o
obj-$(CONFIG_IA64_SGI_SN_XP) += xpnet.o
obj-$(CONFIG_PCI_MSI) += msi_sn.o obj-$(CONFIG_PCI_MSI) += msi_sn.o

View File

@@ -187,8 +187,8 @@ void hub_error_init(struct hubdev_info *hubdev_info)
{ {
if (request_irq(SGI_II_ERROR, hub_eint_handler, IRQF_SHARED, if (request_irq(SGI_II_ERROR, hub_eint_handler, IRQF_SHARED,
"SN_hub_error", (void *)hubdev_info)) { "SN_hub_error", hubdev_info)) {
printk("hub_error_init: Failed to request_irq for 0x%p\n", printk(KERN_ERR "hub_error_init: Failed to request_irq for 0x%p\n",
hubdev_info); hubdev_info);
return; return;
} }

View File

@@ -655,7 +655,8 @@ tioce_dma(struct pci_dev *pdev, u64 paddr, size_t byte_count, int dma_flags)
* *
* Simply call tioce_do_dma_map() to create a map with the barrier bit set * Simply call tioce_do_dma_map() to create a map with the barrier bit set
* in the address. * in the address.
*/ static u64 */
static u64
tioce_dma_consistent(struct pci_dev *pdev, u64 paddr, size_t byte_count, int dma_flags) tioce_dma_consistent(struct pci_dev *pdev, u64 paddr, size_t byte_count, int dma_flags)
{ {
return tioce_do_dma_map(pdev, paddr, byte_count, 1, dma_flags); return tioce_do_dma_map(pdev, paddr, byte_count, 1, dma_flags);
@@ -668,7 +669,8 @@ tioce_dma_consistent(struct pci_dev *pdev, u64 paddr, size_t byte_count, int dma
* *
* Handle a CE error interrupt. Simply a wrapper around a SAL call which * Handle a CE error interrupt. Simply a wrapper around a SAL call which
* defers processing to the SGI prom. * defers processing to the SGI prom.
*/ static irqreturn_t */
static irqreturn_t
tioce_error_intr_handler(int irq, void *arg) tioce_error_intr_handler(int irq, void *arg)
{ {
struct tioce_common *soft = arg; struct tioce_common *soft = arg;

View File

@@ -360,4 +360,16 @@ config ENCLOSURE_SERVICES
driver (SCSI/ATA) which supports enclosures driver (SCSI/ATA) which supports enclosures
or a SCSI enclosure device (SES) to use these services. or a SCSI enclosure device (SES) to use these services.
config SGI_XP
tristate "Support communication between SGI SSIs"
depends on IA64_GENERIC || IA64_SGI_SN2
select IA64_UNCACHED_ALLOCATOR if IA64_GENERIC || IA64_SGI_SN2
select GENERIC_ALLOCATOR if IA64_GENERIC || IA64_SGI_SN2
---help---
An SGI machine can be divided into multiple Single System
Images which act independently of each other and have
hardware based memory protection from the others. Enabling
this feature will allow for direct communication between SSIs
based on a network adapter and DMA messaging.
endif # MISC_DEVICES endif # MISC_DEVICES

View File

@@ -24,3 +24,4 @@ obj-$(CONFIG_EEPROM_93CX6) += eeprom_93cx6.o
obj-$(CONFIG_INTEL_MENLOW) += intel_menlow.o obj-$(CONFIG_INTEL_MENLOW) += intel_menlow.o
obj-$(CONFIG_ENCLOSURE_SERVICES) += enclosure.o obj-$(CONFIG_ENCLOSURE_SERVICES) += enclosure.o
obj-$(CONFIG_KGDB_TESTS) += kgdbts.o obj-$(CONFIG_KGDB_TESTS) += kgdbts.o
obj-$(CONFIG_SGI_XP) += sgi-xp/

View File

@@ -0,0 +1,11 @@
#
# Makefile for SGI's XP devices.
#
obj-$(CONFIG_SGI_XP) += xp.o
xp-y := xp_main.o xp_nofault.o
obj-$(CONFIG_SGI_XP) += xpc.o
xpc-y := xpc_main.o xpc_channel.o xpc_partition.o
obj-$(CONFIG_SGI_XP) += xpnet.o

View File

@@ -3,18 +3,15 @@
* License. See the file "COPYING" in the main directory of this archive * License. See the file "COPYING" in the main directory of this archive
* for more details. * for more details.
* *
* Copyright (C) 2004-2005 Silicon Graphics, Inc. All rights reserved. * Copyright (C) 2004-2008 Silicon Graphics, Inc. All rights reserved.
*/ */
/* /*
* External Cross Partition (XP) structures and defines. * External Cross Partition (XP) structures and defines.
*/ */
#ifndef _DRIVERS_MISC_SGIXP_XP_H
#ifndef _ASM_IA64_SN_XP_H #define _DRIVERS_MISC_SGIXP_XP_H
#define _ASM_IA64_SN_XP_H
#include <linux/cache.h> #include <linux/cache.h>
#include <linux/hardirq.h> #include <linux/hardirq.h>
@@ -22,14 +19,12 @@
#include <asm/sn/types.h> #include <asm/sn/types.h>
#include <asm/sn/bte.h> #include <asm/sn/bte.h>
#ifdef USE_DBUG_ON #ifdef USE_DBUG_ON
#define DBUG_ON(condition) BUG_ON(condition) #define DBUG_ON(condition) BUG_ON(condition)
#else #else
#define DBUG_ON(condition) #define DBUG_ON(condition)
#endif #endif
/* /*
* Define the maximum number of logically defined partitions the system * Define the maximum number of logically defined partitions the system
* can support. It is constrained by the maximum number of hardware * can support. It is constrained by the maximum number of hardware
@@ -43,7 +38,6 @@
*/ */
#define XP_MAX_PARTITIONS 64 #define XP_MAX_PARTITIONS 64
/* /*
* Define the number of u64s required to represent all the C-brick nasids * Define the number of u64s required to represent all the C-brick nasids
* as a bitmap. The cross-partition kernel modules deal only with * as a bitmap. The cross-partition kernel modules deal only with
@@ -54,7 +48,6 @@
#define XP_NASID_MASK_BYTES ((XP_MAX_PHYSNODE_ID + 7) / 8) #define XP_NASID_MASK_BYTES ((XP_MAX_PHYSNODE_ID + 7) / 8)
#define XP_NASID_MASK_WORDS ((XP_MAX_PHYSNODE_ID + 63) / 64) #define XP_NASID_MASK_WORDS ((XP_MAX_PHYSNODE_ID + 63) / 64)
/* /*
* Wrapper for bte_copy() that should it return a failure status will retry * Wrapper for bte_copy() that should it return a failure status will retry
* the bte_copy() once in the hope that the failure was due to a temporary * the bte_copy() once in the hope that the failure was due to a temporary
@@ -74,7 +67,6 @@ xp_bte_copy(u64 src, u64 vdst, u64 len, u64 mode, void *notification)
bte_result_t ret; bte_result_t ret;
u64 pdst = ia64_tpa(vdst); u64 pdst = ia64_tpa(vdst);
/* /*
* Ensure that the physically mapped memory is contiguous. * Ensure that the physically mapped memory is contiguous.
* *
@@ -87,16 +79,15 @@ xp_bte_copy(u64 src, u64 vdst, u64 len, u64 mode, void *notification)
ret = bte_copy(src, pdst, len, mode, notification); ret = bte_copy(src, pdst, len, mode, notification);
if ((ret != BTE_SUCCESS) && BTE_ERROR_RETRY(ret)) { if ((ret != BTE_SUCCESS) && BTE_ERROR_RETRY(ret)) {
if (!in_interrupt()) { if (!in_interrupt())
cond_resched(); cond_resched();
}
ret = bte_copy(src, pdst, len, mode, notification); ret = bte_copy(src, pdst, len, mode, notification);
} }
return ret; return ret;
} }
/* /*
* XPC establishes channel connections between the local partition and any * XPC establishes channel connections between the local partition and any
* other partition that is currently up. Over these channels, kernel-level * other partition that is currently up. Over these channels, kernel-level
@@ -122,7 +113,6 @@ xp_bte_copy(u64 src, u64 vdst, u64 len, u64 mode, void *notification)
#error XPC_NCHANNELS exceeds MAXIMUM allowed. #error XPC_NCHANNELS exceeds MAXIMUM allowed.
#endif #endif
/* /*
* The format of an XPC message is as follows: * The format of an XPC message is as follows:
* *
@@ -160,12 +150,10 @@ struct xpc_msg {
u64 payload; /* user defined portion of message */ u64 payload; /* user defined portion of message */
}; };
#define XPC_MSG_PAYLOAD_OFFSET (u64) (&((struct xpc_msg *)0)->payload) #define XPC_MSG_PAYLOAD_OFFSET (u64) (&((struct xpc_msg *)0)->payload)
#define XPC_MSG_SIZE(_payload_size) \ #define XPC_MSG_SIZE(_payload_size) \
L1_CACHE_ALIGN(XPC_MSG_PAYLOAD_OFFSET + (_payload_size)) L1_CACHE_ALIGN(XPC_MSG_PAYLOAD_OFFSET + (_payload_size))
/* /*
* Define the return values and values passed to user's callout functions. * Define the return values and values passed to user's callout functions.
* (It is important to add new value codes at the end just preceding * (It is important to add new value codes at the end just preceding
@@ -267,10 +255,9 @@ enum xpc_retval {
/* 115: BTE end */ /* 115: BTE end */
xpcBteSh2End = xpcBteSh2Start + BTEFAIL_SH2_ALL, xpcBteSh2End = xpcBteSh2Start + BTEFAIL_SH2_ALL,
xpcUnknownReason /* 116: unknown reason -- must be last in list */ xpcUnknownReason /* 116: unknown reason - must be last in enum */
}; };
/* /*
* Define the callout function types used by XPC to update the user on * Define the callout function types used by XPC to update the user on
* connection activity and state changes (via the user function registered by * connection activity and state changes (via the user function registered by
@@ -381,7 +368,6 @@ typedef void (*xpc_channel_func)(enum xpc_retval reason, partid_t partid,
typedef void (*xpc_notify_func) (enum xpc_retval reason, partid_t partid, typedef void (*xpc_notify_func) (enum xpc_retval reason, partid_t partid,
int ch_number, void *key); int ch_number, void *key);
/* /*
* The following is a registration entry. There is a global array of these, * The following is a registration entry. There is a global array of these,
* one per channel. It is used to record the connection registration made * one per channel. It is used to record the connection registration made
@@ -406,15 +392,12 @@ struct xpc_registration {
u32 idle_limit; /* limit on #of idle kthreads */ u32 idle_limit; /* limit on #of idle kthreads */
} ____cacheline_aligned; } ____cacheline_aligned;
#define XPC_CHANNEL_REGISTERED(_c) (xpc_registrations[_c].func != NULL) #define XPC_CHANNEL_REGISTERED(_c) (xpc_registrations[_c].func != NULL)
/* the following are valid xpc_allocate() flags */ /* the following are valid xpc_allocate() flags */
#define XPC_WAIT 0 /* wait flag */ #define XPC_WAIT 0 /* wait flag */
#define XPC_NOWAIT 1 /* no wait flag */ #define XPC_NOWAIT 1 /* no wait flag */
struct xpc_interface { struct xpc_interface {
void (*connect) (int); void (*connect) (int);
void (*disconnect) (int); void (*disconnect) (int);
@@ -426,20 +409,18 @@ struct xpc_interface {
enum xpc_retval (*partid_to_nasids) (partid_t, void *); enum xpc_retval (*partid_to_nasids) (partid_t, void *);
}; };
extern struct xpc_interface xpc_interface; extern struct xpc_interface xpc_interface;
extern void xpc_set_interface(void (*)(int), extern void xpc_set_interface(void (*)(int),
void (*)(int), void (*)(int),
enum xpc_retval (*)(partid_t, int, u32, void **), enum xpc_retval (*)(partid_t, int, u32, void **),
enum xpc_retval (*)(partid_t, int, void *), enum xpc_retval (*)(partid_t, int, void *),
enum xpc_retval (*)(partid_t, int, void *, xpc_notify_func, enum xpc_retval (*)(partid_t, int, void *,
void *), xpc_notify_func, void *),
void (*)(partid_t, int, void *), void (*)(partid_t, int, void *),
enum xpc_retval (*)(partid_t, void *)); enum xpc_retval (*)(partid_t, void *));
extern void xpc_clear_interface(void); extern void xpc_clear_interface(void);
extern enum xpc_retval xpc_connect(int, xpc_channel_func, void *, u16, extern enum xpc_retval xpc_connect(int, xpc_channel_func, void *, u16,
u16, u32, u32); u16, u32, u32);
extern void xpc_disconnect(int); extern void xpc_disconnect(int);
@@ -475,11 +456,8 @@ xpc_partid_to_nasids(partid_t partid, void *nasids)
return xpc_interface.partid_to_nasids(partid, nasids); return xpc_interface.partid_to_nasids(partid, nasids);
} }
extern u64 xp_nofault_PIOR_target; extern u64 xp_nofault_PIOR_target;
extern int xp_nofault_PIOR(void *); extern int xp_nofault_PIOR(void *);
extern int xp_error_PIOR(void); extern int xp_error_PIOR(void);
#endif /* _DRIVERS_MISC_SGIXP_XP_H */
#endif /* _ASM_IA64_SN_XP_H */

View File

@@ -3,10 +3,9 @@
* License. See the file "COPYING" in the main directory of this archive * License. See the file "COPYING" in the main directory of this archive
* for more details. * for more details.
* *
* Copyright (c) 2004-2005 Silicon Graphics, Inc. All Rights Reserved. * Copyright (c) 2004-2008 Silicon Graphics, Inc. All Rights Reserved.
*/ */
/* /*
* Cross Partition (XP) base. * Cross Partition (XP) base.
* *
@@ -15,33 +14,39 @@
* *
*/ */
#include <linux/kernel.h> #include <linux/kernel.h>
#include <linux/interrupt.h> #include <linux/interrupt.h>
#include <linux/module.h> #include <linux/module.h>
#include <linux/mutex.h> #include <linux/mutex.h>
#include <asm/sn/intr.h> #include <asm/sn/intr.h>
#include <asm/sn/sn_sal.h> #include <asm/sn/sn_sal.h>
#include <asm/sn/xp.h> #include "xp.h"
/* /*
* Target of nofault PIO read. * The export of xp_nofault_PIOR needs to happen here since it is defined
* in drivers/misc/sgi-xp/xp_nofault.S. The target of the nofault read is
* defined here.
*/ */
u64 xp_nofault_PIOR_target; EXPORT_SYMBOL_GPL(xp_nofault_PIOR);
u64 xp_nofault_PIOR_target;
EXPORT_SYMBOL_GPL(xp_nofault_PIOR_target);
/* /*
* xpc_registrations[] keeps track of xpc_connect()'s done by the kernel-level * xpc_registrations[] keeps track of xpc_connect()'s done by the kernel-level
* users of XPC. * users of XPC.
*/ */
struct xpc_registration xpc_registrations[XPC_NCHANNELS]; struct xpc_registration xpc_registrations[XPC_NCHANNELS];
EXPORT_SYMBOL_GPL(xpc_registrations);
/* /*
* Initialize the XPC interface to indicate that XPC isn't loaded. * Initialize the XPC interface to indicate that XPC isn't loaded.
*/ */
static enum xpc_retval xpc_notloaded(void) { return xpcNotLoaded; } static enum xpc_retval
xpc_notloaded(void)
{
return xpcNotLoaded;
}
struct xpc_interface xpc_interface = { struct xpc_interface xpc_interface = {
(void (*)(int))xpc_notloaded, (void (*)(int))xpc_notloaded,
@@ -53,7 +58,7 @@ struct xpc_interface xpc_interface = {
(void (*)(partid_t, int, void *))xpc_notloaded, (void (*)(partid_t, int, void *))xpc_notloaded,
(enum xpc_retval(*)(partid_t, void *))xpc_notloaded (enum xpc_retval(*)(partid_t, void *))xpc_notloaded
}; };
EXPORT_SYMBOL_GPL(xpc_interface);
/* /*
* XPC calls this when it (the XPC module) has been loaded. * XPC calls this when it (the XPC module) has been loaded.
@@ -76,7 +81,7 @@ xpc_set_interface(void (*connect)(int),
xpc_interface.received = received; xpc_interface.received = received;
xpc_interface.partid_to_nasids = partid_to_nasids; xpc_interface.partid_to_nasids = partid_to_nasids;
} }
EXPORT_SYMBOL_GPL(xpc_set_interface);
/* /*
* XPC calls this when it (the XPC module) is being unloaded. * XPC calls this when it (the XPC module) is being unloaded.
@@ -91,13 +96,14 @@ xpc_clear_interface(void)
xpc_interface.send = (enum xpc_retval(*)(partid_t, int, void *)) xpc_interface.send = (enum xpc_retval(*)(partid_t, int, void *))
xpc_notloaded; xpc_notloaded;
xpc_interface.send_notify = (enum xpc_retval(*)(partid_t, int, void *, xpc_interface.send_notify = (enum xpc_retval(*)(partid_t, int, void *,
xpc_notify_func, void *)) xpc_notloaded; xpc_notify_func,
void *))xpc_notloaded;
xpc_interface.received = (void (*)(partid_t, int, void *)) xpc_interface.received = (void (*)(partid_t, int, void *))
xpc_notloaded; xpc_notloaded;
xpc_interface.partid_to_nasids = (enum xpc_retval(*)(partid_t, void *)) xpc_interface.partid_to_nasids = (enum xpc_retval(*)(partid_t, void *))
xpc_notloaded; xpc_notloaded;
} }
EXPORT_SYMBOL_GPL(xpc_clear_interface);
/* /*
* Register for automatic establishment of a channel connection whenever * Register for automatic establishment of a channel connection whenever
@@ -129,7 +135,6 @@ xpc_connect(int ch_number, xpc_channel_func func, void *key, u16 payload_size,
{ {
struct xpc_registration *registration; struct xpc_registration *registration;
DBUG_ON(ch_number < 0 || ch_number >= XPC_NCHANNELS); DBUG_ON(ch_number < 0 || ch_number >= XPC_NCHANNELS);
DBUG_ON(payload_size == 0 || nentries == 0); DBUG_ON(payload_size == 0 || nentries == 0);
DBUG_ON(func == NULL); DBUG_ON(func == NULL);
@@ -137,9 +142,8 @@ xpc_connect(int ch_number, xpc_channel_func func, void *key, u16 payload_size,
registration = &xpc_registrations[ch_number]; registration = &xpc_registrations[ch_number];
if (mutex_lock_interruptible(&registration->mutex) != 0) { if (mutex_lock_interruptible(&registration->mutex) != 0)
return xpcInterrupted; return xpcInterrupted;
}
/* if XPC_CHANNEL_REGISTERED(ch_number) */ /* if XPC_CHANNEL_REGISTERED(ch_number) */
if (registration->func != NULL) { if (registration->func != NULL) {
@@ -161,7 +165,7 @@ xpc_connect(int ch_number, xpc_channel_func func, void *key, u16 payload_size,
return xpcSuccess; return xpcSuccess;
} }
EXPORT_SYMBOL_GPL(xpc_connect);
/* /*
* Remove the registration for automatic connection of the specified channel * Remove the registration for automatic connection of the specified channel
@@ -181,7 +185,6 @@ xpc_disconnect(int ch_number)
{ {
struct xpc_registration *registration; struct xpc_registration *registration;
DBUG_ON(ch_number < 0 || ch_number >= XPC_NCHANNELS); DBUG_ON(ch_number < 0 || ch_number >= XPC_NCHANNELS);
registration = &xpc_registrations[ch_number]; registration = &xpc_registrations[ch_number];
@@ -213,7 +216,7 @@ xpc_disconnect(int ch_number)
return; return;
} }
EXPORT_SYMBOL_GPL(xpc_disconnect);
int __init int __init
xp_init(void) xp_init(void)
@@ -222,10 +225,8 @@ xp_init(void)
u64 func_addr = *(u64 *)xp_nofault_PIOR; u64 func_addr = *(u64 *)xp_nofault_PIOR;
u64 err_func_addr = *(u64 *)xp_error_PIOR; u64 err_func_addr = *(u64 *)xp_error_PIOR;
if (!ia64_platform_is("sn2"))
if (!ia64_platform_is("sn2")) {
return -ENODEV; return -ENODEV;
}
/* /*
* Register a nofault code region which performs a cross-partition * Register a nofault code region which performs a cross-partition
@@ -236,8 +237,9 @@ xp_init(void)
* least some CPUs on Shubs <= v1.2, which unfortunately we have to * least some CPUs on Shubs <= v1.2, which unfortunately we have to
* work around). * work around).
*/ */
if ((ret = sn_register_nofault_code(func_addr, err_func_addr, ret = sn_register_nofault_code(func_addr, err_func_addr, err_func_addr,
err_func_addr, 1, 1)) != 0) { 1, 1);
if (ret != 0) {
printk(KERN_ERR "XP: can't register nofault code, error=%d\n", printk(KERN_ERR "XP: can't register nofault code, error=%d\n",
ret); ret);
} }
@@ -245,21 +247,19 @@ xp_init(void)
* Setup the nofault PIO read target. (There is no special reason why * Setup the nofault PIO read target. (There is no special reason why
* SH_IPI_ACCESS was selected.) * SH_IPI_ACCESS was selected.)
*/ */
if (is_shub2()) { if (is_shub2())
xp_nofault_PIOR_target = SH2_IPI_ACCESS0; xp_nofault_PIOR_target = SH2_IPI_ACCESS0;
} else { else
xp_nofault_PIOR_target = SH1_IPI_ACCESS; xp_nofault_PIOR_target = SH1_IPI_ACCESS;
}
/* initialize the connection registration mutex */ /* initialize the connection registration mutex */
for (ch_number = 0; ch_number < XPC_NCHANNELS; ch_number++) { for (ch_number = 0; ch_number < XPC_NCHANNELS; ch_number++)
mutex_init(&xpc_registrations[ch_number].mutex); mutex_init(&xpc_registrations[ch_number].mutex);
}
return 0; return 0;
} }
module_init(xp_init);
module_init(xp_init);
void __exit void __exit
xp_exit(void) xp_exit(void)
@@ -267,24 +267,13 @@ xp_exit(void)
u64 func_addr = *(u64 *)xp_nofault_PIOR; u64 func_addr = *(u64 *)xp_nofault_PIOR;
u64 err_func_addr = *(u64 *)xp_error_PIOR; u64 err_func_addr = *(u64 *)xp_error_PIOR;
/* unregister the PIO read nofault code region */ /* unregister the PIO read nofault code region */
(void)sn_register_nofault_code(func_addr, err_func_addr, (void)sn_register_nofault_code(func_addr, err_func_addr,
err_func_addr, 1, 0); err_func_addr, 1, 0);
} }
module_exit(xp_exit);
module_exit(xp_exit);
MODULE_AUTHOR("Silicon Graphics, Inc."); MODULE_AUTHOR("Silicon Graphics, Inc.");
MODULE_DESCRIPTION("Cross Partition (XP) base"); MODULE_DESCRIPTION("Cross Partition (XP) base");
MODULE_LICENSE("GPL"); MODULE_LICENSE("GPL");
EXPORT_SYMBOL(xp_nofault_PIOR);
EXPORT_SYMBOL(xp_nofault_PIOR_target);
EXPORT_SYMBOL(xpc_registrations);
EXPORT_SYMBOL(xpc_interface);
EXPORT_SYMBOL(xpc_clear_interface);
EXPORT_SYMBOL(xpc_set_interface);
EXPORT_SYMBOL(xpc_connect);
EXPORT_SYMBOL(xpc_disconnect);

View File

@@ -3,10 +3,9 @@
* License. See the file "COPYING" in the main directory of this archive * License. See the file "COPYING" in the main directory of this archive
* for more details. * for more details.
* *
* Copyright (c) 2004-2007 Silicon Graphics, Inc. All Rights Reserved. * Copyright (c) 2004-2008 Silicon Graphics, Inc. All Rights Reserved.
*/ */
/* /*
* The xp_nofault_PIOR function takes a pointer to a remote PIO register * The xp_nofault_PIOR function takes a pointer to a remote PIO register
* and attempts to load and consume a value from it. This function * and attempts to load and consume a value from it. This function

View File

@@ -3,17 +3,15 @@
* License. See the file "COPYING" in the main directory of this archive * License. See the file "COPYING" in the main directory of this archive
* for more details. * for more details.
* *
* Copyright (c) 2004-2007 Silicon Graphics, Inc. All Rights Reserved. * Copyright (c) 2004-2008 Silicon Graphics, Inc. All Rights Reserved.
*/ */
/* /*
* Cross Partition Communication (XPC) structures and macros. * Cross Partition Communication (XPC) structures and macros.
*/ */
#ifndef _ASM_IA64_SN_XPC_H #ifndef _DRIVERS_MISC_SGIXP_XPC_H
#define _ASM_IA64_SN_XPC_H #define _DRIVERS_MISC_SGIXP_XPC_H
#include <linux/interrupt.h> #include <linux/interrupt.h>
#include <linux/sysctl.h> #include <linux/sysctl.h>
@@ -27,8 +25,7 @@
#include <asm/sn/addrs.h> #include <asm/sn/addrs.h>
#include <asm/sn/mspec.h> #include <asm/sn/mspec.h>
#include <asm/sn/shub_mmr.h> #include <asm/sn/shub_mmr.h>
#include <asm/sn/xp.h> #include "xp.h"
/* /*
* XPC Version numbers consist of a major and minor number. XPC can always * XPC Version numbers consist of a major and minor number. XPC can always
@@ -39,7 +36,6 @@
#define XPC_VERSION_MAJOR(_v) ((_v) >> 4) #define XPC_VERSION_MAJOR(_v) ((_v) >> 4)
#define XPC_VERSION_MINOR(_v) ((_v) & 0xf) #define XPC_VERSION_MINOR(_v) ((_v) & 0xf)
/* /*
* The next macros define word or bit representations for given * The next macros define word or bit representations for given
* C-brick nasid in either the SAL provided bit array representing * C-brick nasid in either the SAL provided bit array representing
@@ -67,7 +63,6 @@
/* define the process name of the discovery thread */ /* define the process name of the discovery thread */
#define XPC_DISCOVERY_THREAD_NAME "xpc_discovery" #define XPC_DISCOVERY_THREAD_NAME "xpc_discovery"
/* /*
* the reserved page * the reserved page
* *
@@ -115,7 +110,7 @@ struct xpc_rsvd_page {
u8 partid; /* SAL: partition ID */ u8 partid; /* SAL: partition ID */
u8 version; u8 version;
u8 pad1[6]; /* align to next u64 in cacheline */ u8 pad1[6]; /* align to next u64 in cacheline */
volatile u64 vars_pa; u64 vars_pa; /* physical address of struct xpc_vars */
struct timespec stamp; /* time when reserved page was setup by XPC */ struct timespec stamp; /* time when reserved page was setup by XPC */
u64 pad2[9]; /* align to last u64 in cacheline */ u64 pad2[9]; /* align to last u64 in cacheline */
u64 nasids_size; /* SAL: size of each nasid mask in bytes */ u64 nasids_size; /* SAL: size of each nasid mask in bytes */
@@ -138,14 +133,13 @@ xpc_compare_stamps(struct timespec *stamp1, struct timespec *stamp2)
{ {
int ret; int ret;
ret = stamp1->tv_sec - stamp2->tv_sec;
if ((ret = stamp1->tv_sec - stamp2->tv_sec) == 0) { if (ret == 0)
ret = stamp1->tv_nsec - stamp2->tv_nsec; ret = stamp1->tv_nsec - stamp2->tv_nsec;
}
return ret; return ret;
} }
/* /*
* Define the structures by which XPC variables can be exported to other * Define the structures by which XPC variables can be exported to other
* partitions. (There are two: struct xpc_vars and struct xpc_vars_part) * partitions. (There are two: struct xpc_vars and struct xpc_vars_part)
@@ -177,7 +171,6 @@ struct xpc_vars {
#define XPC_SUPPORTS_DISENGAGE_REQUEST(_version) \ #define XPC_SUPPORTS_DISENGAGE_REQUEST(_version) \
(_version >= _XPC_VERSION(3, 1)) (_version >= _XPC_VERSION(3, 1))
static inline int static inline int
xpc_hb_allowed(partid_t partid, struct xpc_vars *vars) xpc_hb_allowed(partid_t partid, struct xpc_vars *vars)
{ {
@@ -208,7 +201,6 @@ xpc_disallow_hb(partid_t partid, struct xpc_vars *vars)
old_mask); old_mask);
} }
/* /*
* The AMOs page consists of a number of AMO variables which are divided into * The AMOs page consists of a number of AMO variables which are divided into
* four groups, The first two groups are used to identify an IRQ's sender. * four groups, The first two groups are used to identify an IRQ's sender.
@@ -222,7 +214,6 @@ xpc_disallow_hb(partid_t partid, struct xpc_vars *vars)
#define XPC_ENGAGED_PARTITIONS_AMO (XPC_ACTIVATE_IRQ_AMOS + XP_NASID_MASK_WORDS) #define XPC_ENGAGED_PARTITIONS_AMO (XPC_ACTIVATE_IRQ_AMOS + XP_NASID_MASK_WORDS)
#define XPC_DISENGAGE_REQUEST_AMO (XPC_ENGAGED_PARTITIONS_AMO + 1) #define XPC_DISENGAGE_REQUEST_AMO (XPC_ENGAGED_PARTITIONS_AMO + 1)
/* /*
* The following structure describes the per partition specific variables. * The following structure describes the per partition specific variables.
* *
@@ -234,7 +225,7 @@ xpc_disallow_hb(partid_t partid, struct xpc_vars *vars)
* occupies half a cacheline. * occupies half a cacheline.
*/ */
struct xpc_vars_part { struct xpc_vars_part {
volatile u64 magic; u64 magic;
u64 openclose_args_pa; /* physical address of open and close args */ u64 openclose_args_pa; /* physical address of open and close args */
u64 GPs_pa; /* physical address of Get/Put values */ u64 GPs_pa; /* physical address of Get/Put values */
@@ -260,17 +251,17 @@ struct xpc_vars_part {
#define XPC_VP_MAGIC1 0x0053524156435058L /* 'XPCVARS\0'L (little endian) */ #define XPC_VP_MAGIC1 0x0053524156435058L /* 'XPCVARS\0'L (little endian) */
#define XPC_VP_MAGIC2 0x0073726176435058L /* 'XPCvars\0'L (little endian) */ #define XPC_VP_MAGIC2 0x0073726176435058L /* 'XPCvars\0'L (little endian) */
/* the reserved page sizes and offsets */ /* the reserved page sizes and offsets */
#define XPC_RP_HEADER_SIZE L1_CACHE_ALIGN(sizeof(struct xpc_rsvd_page)) #define XPC_RP_HEADER_SIZE L1_CACHE_ALIGN(sizeof(struct xpc_rsvd_page))
#define XPC_RP_VARS_SIZE L1_CACHE_ALIGN(sizeof(struct xpc_vars)) #define XPC_RP_VARS_SIZE L1_CACHE_ALIGN(sizeof(struct xpc_vars))
#define XPC_RP_PART_NASIDS(_rp) (u64 *) ((u8 *) _rp + XPC_RP_HEADER_SIZE) #define XPC_RP_PART_NASIDS(_rp) ((u64 *)((u8 *)(_rp) + XPC_RP_HEADER_SIZE))
#define XPC_RP_MACH_NASIDS(_rp) (XPC_RP_PART_NASIDS(_rp) + xp_nasid_mask_words) #define XPC_RP_MACH_NASIDS(_rp) (XPC_RP_PART_NASIDS(_rp) + xp_nasid_mask_words)
#define XPC_RP_VARS(_rp) ((struct xpc_vars *) XPC_RP_MACH_NASIDS(_rp) + xp_nasid_mask_words) #define XPC_RP_VARS(_rp) ((struct xpc_vars *)(XPC_RP_MACH_NASIDS(_rp) + \
#define XPC_RP_VARS_PART(_rp) (struct xpc_vars_part *) ((u8 *) XPC_RP_VARS(rp) + XPC_RP_VARS_SIZE) xp_nasid_mask_words))
#define XPC_RP_VARS_PART(_rp) ((struct xpc_vars_part *) \
((u8 *)XPC_RP_VARS(_rp) + XPC_RP_VARS_SIZE))
/* /*
* Functions registered by add_timer() or called by kernel_thread() only * Functions registered by add_timer() or called by kernel_thread() only
@@ -285,21 +276,17 @@ struct xpc_vars_part {
#define XPC_UNPACK_ARG1(_args) (((u64) _args) & 0xffffffff) #define XPC_UNPACK_ARG1(_args) (((u64) _args) & 0xffffffff)
#define XPC_UNPACK_ARG2(_args) ((((u64) _args) >> 32) & 0xffffffff) #define XPC_UNPACK_ARG2(_args) ((((u64) _args) >> 32) & 0xffffffff)
/* /*
* Define a Get/Put value pair (pointers) used with a message queue. * Define a Get/Put value pair (pointers) used with a message queue.
*/ */
struct xpc_gp { struct xpc_gp {
volatile s64 get; /* Get value */ s64 get; /* Get value */
volatile s64 put; /* Put value */ s64 put; /* Put value */
}; };
#define XPC_GP_SIZE \ #define XPC_GP_SIZE \
L1_CACHE_ALIGN(sizeof(struct xpc_gp) * XPC_NCHANNELS) L1_CACHE_ALIGN(sizeof(struct xpc_gp) * XPC_NCHANNELS)
/* /*
* Define a structure that contains arguments associated with opening and * Define a structure that contains arguments associated with opening and
* closing a channel. * closing a channel.
@@ -315,20 +302,15 @@ struct xpc_openclose_args {
#define XPC_OPENCLOSE_ARGS_SIZE \ #define XPC_OPENCLOSE_ARGS_SIZE \
L1_CACHE_ALIGN(sizeof(struct xpc_openclose_args) * XPC_NCHANNELS) L1_CACHE_ALIGN(sizeof(struct xpc_openclose_args) * XPC_NCHANNELS)
/* struct xpc_msg flags */ /* struct xpc_msg flags */
#define XPC_M_DONE 0x01 /* msg has been received/consumed */ #define XPC_M_DONE 0x01 /* msg has been received/consumed */
#define XPC_M_READY 0x02 /* msg is ready to be sent */ #define XPC_M_READY 0x02 /* msg is ready to be sent */
#define XPC_M_INTERRUPT 0x04 /* send interrupt when msg consumed */ #define XPC_M_INTERRUPT 0x04 /* send interrupt when msg consumed */
#define XPC_MSG_ADDRESS(_payload) \ #define XPC_MSG_ADDRESS(_payload) \
((struct xpc_msg *)((u8 *)(_payload) - XPC_MSG_PAYLOAD_OFFSET)) ((struct xpc_msg *)((u8 *)(_payload) - XPC_MSG_PAYLOAD_OFFSET))
/* /*
* Defines notify entry. * Defines notify entry.
* *
@@ -336,7 +318,7 @@ struct xpc_openclose_args {
* and consumed by the intended recipient. * and consumed by the intended recipient.
*/ */
struct xpc_notify { struct xpc_notify {
volatile u8 type; /* type of notification */ u8 type; /* type of notification */
/* the following two fields are only used if type == XPC_N_CALL */ /* the following two fields are only used if type == XPC_N_CALL */
xpc_notify_func func; /* user's notify function */ xpc_notify_func func; /* user's notify function */
@@ -347,8 +329,6 @@ struct xpc_notify {
#define XPC_N_CALL 0x01 /* notify function provided by user */ #define XPC_N_CALL 0x01 /* notify function provided by user */
/* /*
* Define the structure that manages all the stuff required by a channel. In * Define the structure that manages all the stuff required by a channel. In
* particular, they are used to manage the messages sent across the channel. * particular, they are used to manage the messages sent across the channel.
@@ -481,22 +461,16 @@ struct xpc_channel {
/* kthread management related fields */ /* kthread management related fields */
// >>> rethink having kthreads_assigned_limit and kthreads_idle_limit; perhaps
// >>> allow the assigned limit be unbounded and let the idle limit be dynamic
// >>> dependent on activity over the last interval of time
atomic_t kthreads_assigned; /* #of kthreads assigned to channel */ atomic_t kthreads_assigned; /* #of kthreads assigned to channel */
u32 kthreads_assigned_limit; /* limit on #of kthreads assigned */ u32 kthreads_assigned_limit; /* limit on #of kthreads assigned */
atomic_t kthreads_idle; /* #of kthreads idle waiting for work */ atomic_t kthreads_idle; /* #of kthreads idle waiting for work */
u32 kthreads_idle_limit; /* limit on #of kthreads idle */ u32 kthreads_idle_limit; /* limit on #of kthreads idle */
atomic_t kthreads_active; /* #of kthreads actively working */ atomic_t kthreads_active; /* #of kthreads actively working */
// >>> following field is temporary
u32 kthreads_created; /* total #of kthreads created */
wait_queue_head_t idle_wq; /* idle kthread wait queue */ wait_queue_head_t idle_wq; /* idle kthread wait queue */
} ____cacheline_aligned; } ____cacheline_aligned;
/* struct xpc_channel flags */ /* struct xpc_channel flags */
#define XPC_C_WASCONNECTED 0x00000001 /* channel was connected */ #define XPC_C_WASCONNECTED 0x00000001 /* channel was connected */
@@ -526,8 +500,6 @@ struct xpc_channel {
0x00020000 /* disconnecting callout completed */ 0x00020000 /* disconnecting callout completed */
#define XPC_C_WDISCONNECT 0x00040000 /* waiting for channel disconnect */ #define XPC_C_WDISCONNECT 0x00040000 /* waiting for channel disconnect */
/* /*
* Manages channels on a partition basis. There is one of these structures * Manages channels on a partition basis. There is one of these structures
* for each partition (a partition will never utilize the structure that * for each partition (a partition will never utilize the structure that
@@ -557,14 +529,12 @@ struct xpc_partition {
unsigned long disengage_request_timeout; /* timeout in jiffies */ unsigned long disengage_request_timeout; /* timeout in jiffies */
struct timer_list disengage_request_timer; struct timer_list disengage_request_timer;
/* XPC infrastructure referencing and teardown control */ /* XPC infrastructure referencing and teardown control */
volatile u8 setup_state; /* infrastructure setup state */ u8 setup_state; /* infrastructure setup state */
wait_queue_head_t teardown_wq; /* kthread waiting to teardown infra */ wait_queue_head_t teardown_wq; /* kthread waiting to teardown infra */
atomic_t references; /* #of references to infrastructure */ atomic_t references; /* #of references to infrastructure */
/* /*
* NONE OF THE PRECEDING FIELDS OF THIS STRUCTURE WILL BE CLEARED WHEN * NONE OF THE PRECEDING FIELDS OF THIS STRUCTURE WILL BE CLEARED WHEN
* XPC SETS UP THE NECESSARY INFRASTRUCTURE TO SUPPORT CROSS PARTITION * XPC SETS UP THE NECESSARY INFRASTRUCTURE TO SUPPORT CROSS PARTITION
@@ -572,7 +542,6 @@ struct xpc_partition {
* 'nchannels' FIELD MUST BE THE FIRST OF THE FIELDS TO BE CLEARED.) * 'nchannels' FIELD MUST BE THE FIRST OF THE FIELDS TO BE CLEARED.)
*/ */
u8 nchannels; /* #of defined channels supported */ u8 nchannels; /* #of defined channels supported */
atomic_t nchannels_active; /* #of channels that are not DISCONNECTED */ atomic_t nchannels_active; /* #of channels that are not DISCONNECTED */
atomic_t nchannels_engaged; /* #of channels engaged with remote part */ atomic_t nchannels_engaged; /* #of channels engaged with remote part */
@@ -581,12 +550,11 @@ struct xpc_partition {
void *local_GPs_base; /* base address of kmalloc'd space */ void *local_GPs_base; /* base address of kmalloc'd space */
struct xpc_gp *local_GPs; /* local Get/Put values */ struct xpc_gp *local_GPs; /* local Get/Put values */
void *remote_GPs_base; /* base address of kmalloc'd space */ void *remote_GPs_base; /* base address of kmalloc'd space */
struct xpc_gp *remote_GPs;/* copy of remote partition's local Get/Put */ struct xpc_gp *remote_GPs; /* copy of remote partition's local */
/* values */ /* Get/Put values */
u64 remote_GPs_pa; /* phys address of remote partition's local */ u64 remote_GPs_pa; /* phys address of remote partition's local */
/* Get/Put values */ /* Get/Put values */
/* fields used to pass args when opening or closing a channel */ /* fields used to pass args when opening or closing a channel */
void *local_openclose_args_base; /* base address of kmalloc'd space */ void *local_openclose_args_base; /* base address of kmalloc'd space */
@@ -596,7 +564,6 @@ struct xpc_partition {
/* args */ /* args */
u64 remote_openclose_args_pa; /* phys addr of remote's args */ u64 remote_openclose_args_pa; /* phys addr of remote's args */
/* IPI sending, receiving and handling related fields */ /* IPI sending, receiving and handling related fields */
int remote_IPI_nasid; /* nasid of where to send IPIs */ int remote_IPI_nasid; /* nasid of where to send IPIs */
@@ -610,7 +577,6 @@ struct xpc_partition {
spinlock_t IPI_lock; /* IPI handler lock */ spinlock_t IPI_lock; /* IPI handler lock */
/* channel manager related fields */ /* channel manager related fields */
atomic_t channel_mgr_requests; /* #of requests to activate chan mgr */ atomic_t channel_mgr_requests; /* #of requests to activate chan mgr */
@@ -618,7 +584,6 @@ struct xpc_partition {
} ____cacheline_aligned; } ____cacheline_aligned;
/* struct xpc_partition act_state values (for XPC HB) */ /* struct xpc_partition act_state values (for XPC HB) */
#define XPC_P_INACTIVE 0x00 /* partition is not active */ #define XPC_P_INACTIVE 0x00 /* partition is not active */
@@ -627,11 +592,9 @@ struct xpc_partition {
#define XPC_P_ACTIVE 0x03 /* xpc_partition_up() was called */ #define XPC_P_ACTIVE 0x03 /* xpc_partition_up() was called */
#define XPC_P_DEACTIVATING 0x04 /* partition deactivation initiated */ #define XPC_P_DEACTIVATING 0x04 /* partition deactivation initiated */
#define XPC_DEACTIVATE_PARTITION(_p, _reason) \ #define XPC_DEACTIVATE_PARTITION(_p, _reason) \
xpc_deactivate_partition(__LINE__, (_p), (_reason)) xpc_deactivate_partition(__LINE__, (_p), (_reason))
/* struct xpc_partition setup_state values */ /* struct xpc_partition setup_state values */
#define XPC_P_UNSET 0x00 /* infrastructure was never setup */ #define XPC_P_UNSET 0x00 /* infrastructure was never setup */
@@ -639,8 +602,6 @@ struct xpc_partition {
#define XPC_P_WTEARDOWN 0x02 /* waiting to teardown infrastructure */ #define XPC_P_WTEARDOWN 0x02 /* waiting to teardown infrastructure */
#define XPC_P_TORNDOWN 0x03 /* infrastructure is torndown */ #define XPC_P_TORNDOWN 0x03 /* infrastructure is torndown */
/* /*
* struct xpc_partition IPI_timer #of seconds to wait before checking for * struct xpc_partition IPI_timer #of seconds to wait before checking for
* dropped IPIs. These occur whenever an IPI amo write doesn't complete until * dropped IPIs. These occur whenever an IPI amo write doesn't complete until
@@ -648,22 +609,17 @@ struct xpc_partition {
*/ */
#define XPC_P_DROPPED_IPI_WAIT (0.25 * HZ) #define XPC_P_DROPPED_IPI_WAIT (0.25 * HZ)
/* number of seconds to wait for other partitions to disengage */ /* number of seconds to wait for other partitions to disengage */
#define XPC_DISENGAGE_REQUEST_DEFAULT_TIMELIMIT 90 #define XPC_DISENGAGE_REQUEST_DEFAULT_TIMELIMIT 90
/* interval in seconds to print 'waiting disengagement' messages */ /* interval in seconds to print 'waiting disengagement' messages */
#define XPC_DISENGAGE_PRINTMSG_INTERVAL 10 #define XPC_DISENGAGE_PRINTMSG_INTERVAL 10
#define XPC_PARTID(_p) ((partid_t) ((_p) - &xpc_partitions[0])) #define XPC_PARTID(_p) ((partid_t) ((_p) - &xpc_partitions[0]))
/* found in xp_main.c */ /* found in xp_main.c */
extern struct xpc_registration xpc_registrations[]; extern struct xpc_registration xpc_registrations[];
/* found in xpc_main.c */ /* found in xpc_main.c */
extern struct device *xpc_part; extern struct device *xpc_part;
extern struct device *xpc_chan; extern struct device *xpc_chan;
@@ -676,7 +632,6 @@ extern void xpc_activate_kthreads(struct xpc_channel *, int);
extern void xpc_create_kthreads(struct xpc_channel *, int, int); extern void xpc_create_kthreads(struct xpc_channel *, int, int);
extern void xpc_disconnect_wait(int); extern void xpc_disconnect_wait(int);
/* found in xpc_partition.c */ /* found in xpc_partition.c */
extern int xpc_exiting; extern int xpc_exiting;
extern struct xpc_vars *xpc_vars; extern struct xpc_vars *xpc_vars;
@@ -699,7 +654,6 @@ extern void xpc_deactivate_partition(const int, struct xpc_partition *,
enum xpc_retval); enum xpc_retval);
extern enum xpc_retval xpc_initiate_partid_to_nasids(partid_t, void *); extern enum xpc_retval xpc_initiate_partid_to_nasids(partid_t, void *);
/* found in xpc_channel.c */ /* found in xpc_channel.c */
extern void xpc_initiate_connect(int); extern void xpc_initiate_connect(int);
extern void xpc_initiate_disconnect(int); extern void xpc_initiate_disconnect(int);
@@ -719,17 +673,12 @@ extern void xpc_disconnect_callout(struct xpc_channel *, enum xpc_retval);
extern void xpc_partition_going_down(struct xpc_partition *, enum xpc_retval); extern void xpc_partition_going_down(struct xpc_partition *, enum xpc_retval);
extern void xpc_teardown_infrastructure(struct xpc_partition *); extern void xpc_teardown_infrastructure(struct xpc_partition *);
static inline void static inline void
xpc_wakeup_channel_mgr(struct xpc_partition *part) xpc_wakeup_channel_mgr(struct xpc_partition *part)
{ {
if (atomic_inc_return(&part->channel_mgr_requests) == 1) { if (atomic_inc_return(&part->channel_mgr_requests) == 1)
wake_up(&part->channel_mgr_wq); wake_up(&part->channel_mgr_wq);
} }
}
/* /*
* These next two inlines are used to keep us from tearing down a channel's * These next two inlines are used to keep us from tearing down a channel's
@@ -747,17 +696,13 @@ xpc_msgqueue_deref(struct xpc_channel *ch)
s32 refs = atomic_dec_return(&ch->references); s32 refs = atomic_dec_return(&ch->references);
DBUG_ON(refs < 0); DBUG_ON(refs < 0);
if (refs == 0) { if (refs == 0)
xpc_wakeup_channel_mgr(&xpc_partitions[ch->partid]); xpc_wakeup_channel_mgr(&xpc_partitions[ch->partid]);
} }
}
#define XPC_DISCONNECT_CHANNEL(_ch, _reason, _irqflgs) \ #define XPC_DISCONNECT_CHANNEL(_ch, _reason, _irqflgs) \
xpc_disconnect_channel(__LINE__, _ch, _reason, _irqflgs) xpc_disconnect_channel(__LINE__, _ch, _reason, _irqflgs)
/* /*
* These two inlines are used to keep us from tearing down a partition's * These two inlines are used to keep us from tearing down a partition's
* setup infrastructure while a thread may be referencing it. * setup infrastructure while a thread may be referencing it.
@@ -767,29 +712,24 @@ xpc_part_deref(struct xpc_partition *part)
{ {
s32 refs = atomic_dec_return(&part->references); s32 refs = atomic_dec_return(&part->references);
DBUG_ON(refs < 0); DBUG_ON(refs < 0);
if (refs == 0 && part->setup_state == XPC_P_WTEARDOWN) { if (refs == 0 && part->setup_state == XPC_P_WTEARDOWN)
wake_up(&part->teardown_wq); wake_up(&part->teardown_wq);
} }
}
static inline int static inline int
xpc_part_ref(struct xpc_partition *part) xpc_part_ref(struct xpc_partition *part)
{ {
int setup; int setup;
atomic_inc(&part->references); atomic_inc(&part->references);
setup = (part->setup_state == XPC_P_SETUP); setup = (part->setup_state == XPC_P_SETUP);
if (!setup) { if (!setup)
xpc_part_deref(part); xpc_part_deref(part);
}
return setup; return setup;
} }
/* /*
* The following macro is to be used for the setting of the reason and * The following macro is to be used for the setting of the reason and
* reason_line fields in both the struct xpc_channel and struct xpc_partition * reason_line fields in both the struct xpc_channel and struct xpc_partition
@@ -801,8 +741,6 @@ xpc_part_ref(struct xpc_partition *part)
(_p)->reason_line = _line; \ (_p)->reason_line = _line; \
} }
/* /*
* This next set of inlines are used to keep track of when a partition is * This next set of inlines are used to keep track of when a partition is
* potentially engaged in accessing memory belonging to another partition. * potentially engaged in accessing memory belonging to another partition.
@@ -813,8 +751,8 @@ xpc_mark_partition_engaged(struct xpc_partition *part)
{ {
unsigned long irq_flags; unsigned long irq_flags;
AMO_t *amo = (AMO_t *)__va(part->remote_amos_page_pa + AMO_t *amo = (AMO_t *)__va(part->remote_amos_page_pa +
(XPC_ENGAGED_PARTITIONS_AMO * sizeof(AMO_t))); (XPC_ENGAGED_PARTITIONS_AMO *
sizeof(AMO_t)));
local_irq_save(irq_flags); local_irq_save(irq_flags);
@@ -828,7 +766,8 @@ xpc_mark_partition_engaged(struct xpc_partition *part)
* keep sending IPIs and AMOs to it until the heartbeat times out. * keep sending IPIs and AMOs to it until the heartbeat times out.
*/ */
(void)xp_nofault_PIOR((u64 *)GLOBAL_MMR_ADDR(NASID_GET(&amo-> (void)xp_nofault_PIOR((u64 *)GLOBAL_MMR_ADDR(NASID_GET(&amo->
variable), xp_nofault_PIOR_target)); variable),
xp_nofault_PIOR_target));
local_irq_restore(irq_flags); local_irq_restore(irq_flags);
} }
@@ -838,8 +777,8 @@ xpc_mark_partition_disengaged(struct xpc_partition *part)
{ {
unsigned long irq_flags; unsigned long irq_flags;
AMO_t *amo = (AMO_t *)__va(part->remote_amos_page_pa + AMO_t *amo = (AMO_t *)__va(part->remote_amos_page_pa +
(XPC_ENGAGED_PARTITIONS_AMO * sizeof(AMO_t))); (XPC_ENGAGED_PARTITIONS_AMO *
sizeof(AMO_t)));
local_irq_save(irq_flags); local_irq_save(irq_flags);
@@ -853,7 +792,8 @@ xpc_mark_partition_disengaged(struct xpc_partition *part)
* keep sending IPIs and AMOs to it until the heartbeat times out. * keep sending IPIs and AMOs to it until the heartbeat times out.
*/ */
(void)xp_nofault_PIOR((u64 *)GLOBAL_MMR_ADDR(NASID_GET(&amo-> (void)xp_nofault_PIOR((u64 *)GLOBAL_MMR_ADDR(NASID_GET(&amo->
variable), xp_nofault_PIOR_target)); variable),
xp_nofault_PIOR_target));
local_irq_restore(irq_flags); local_irq_restore(irq_flags);
} }
@@ -865,7 +805,6 @@ xpc_request_partition_disengage(struct xpc_partition *part)
AMO_t *amo = (AMO_t *)__va(part->remote_amos_page_pa + AMO_t *amo = (AMO_t *)__va(part->remote_amos_page_pa +
(XPC_DISENGAGE_REQUEST_AMO * sizeof(AMO_t))); (XPC_DISENGAGE_REQUEST_AMO * sizeof(AMO_t)));
local_irq_save(irq_flags); local_irq_save(irq_flags);
/* set bit corresponding to our partid in remote partition's AMO */ /* set bit corresponding to our partid in remote partition's AMO */
@@ -878,7 +817,8 @@ xpc_request_partition_disengage(struct xpc_partition *part)
* keep sending IPIs and AMOs to it until the heartbeat times out. * keep sending IPIs and AMOs to it until the heartbeat times out.
*/ */
(void)xp_nofault_PIOR((u64 *)GLOBAL_MMR_ADDR(NASID_GET(&amo-> (void)xp_nofault_PIOR((u64 *)GLOBAL_MMR_ADDR(NASID_GET(&amo->
variable), xp_nofault_PIOR_target)); variable),
xp_nofault_PIOR_target));
local_irq_restore(irq_flags); local_irq_restore(irq_flags);
} }
@@ -890,7 +830,6 @@ xpc_cancel_partition_disengage_request(struct xpc_partition *part)
AMO_t *amo = (AMO_t *)__va(part->remote_amos_page_pa + AMO_t *amo = (AMO_t *)__va(part->remote_amos_page_pa +
(XPC_DISENGAGE_REQUEST_AMO * sizeof(AMO_t))); (XPC_DISENGAGE_REQUEST_AMO * sizeof(AMO_t)));
local_irq_save(irq_flags); local_irq_save(irq_flags);
/* clear bit corresponding to our partid in remote partition's AMO */ /* clear bit corresponding to our partid in remote partition's AMO */
@@ -903,7 +842,8 @@ xpc_cancel_partition_disengage_request(struct xpc_partition *part)
* keep sending IPIs and AMOs to it until the heartbeat times out. * keep sending IPIs and AMOs to it until the heartbeat times out.
*/ */
(void)xp_nofault_PIOR((u64 *)GLOBAL_MMR_ADDR(NASID_GET(&amo-> (void)xp_nofault_PIOR((u64 *)GLOBAL_MMR_ADDR(NASID_GET(&amo->
variable), xp_nofault_PIOR_target)); variable),
xp_nofault_PIOR_target));
local_irq_restore(irq_flags); local_irq_restore(irq_flags);
} }
@@ -913,7 +853,6 @@ xpc_partition_engaged(u64 partid_mask)
{ {
AMO_t *amo = xpc_vars->amos_page + XPC_ENGAGED_PARTITIONS_AMO; AMO_t *amo = xpc_vars->amos_page + XPC_ENGAGED_PARTITIONS_AMO;
/* return our partition's AMO variable ANDed with partid_mask */ /* return our partition's AMO variable ANDed with partid_mask */
return (FETCHOP_LOAD_OP(TO_AMO((u64)&amo->variable), FETCHOP_LOAD) & return (FETCHOP_LOAD_OP(TO_AMO((u64)&amo->variable), FETCHOP_LOAD) &
partid_mask); partid_mask);
@@ -924,7 +863,6 @@ xpc_partition_disengage_requested(u64 partid_mask)
{ {
AMO_t *amo = xpc_vars->amos_page + XPC_DISENGAGE_REQUEST_AMO; AMO_t *amo = xpc_vars->amos_page + XPC_DISENGAGE_REQUEST_AMO;
/* return our partition's AMO variable ANDed with partid_mask */ /* return our partition's AMO variable ANDed with partid_mask */
return (FETCHOP_LOAD_OP(TO_AMO((u64)&amo->variable), FETCHOP_LOAD) & return (FETCHOP_LOAD_OP(TO_AMO((u64)&amo->variable), FETCHOP_LOAD) &
partid_mask); partid_mask);
@@ -935,7 +873,6 @@ xpc_clear_partition_engaged(u64 partid_mask)
{ {
AMO_t *amo = xpc_vars->amos_page + XPC_ENGAGED_PARTITIONS_AMO; AMO_t *amo = xpc_vars->amos_page + XPC_ENGAGED_PARTITIONS_AMO;
/* clear bit(s) based on partid_mask in our partition's AMO */ /* clear bit(s) based on partid_mask in our partition's AMO */
FETCHOP_STORE_OP(TO_AMO((u64)&amo->variable), FETCHOP_AND, FETCHOP_STORE_OP(TO_AMO((u64)&amo->variable), FETCHOP_AND,
~partid_mask); ~partid_mask);
@@ -946,14 +883,11 @@ xpc_clear_partition_disengage_request(u64 partid_mask)
{ {
AMO_t *amo = xpc_vars->amos_page + XPC_DISENGAGE_REQUEST_AMO; AMO_t *amo = xpc_vars->amos_page + XPC_DISENGAGE_REQUEST_AMO;
/* clear bit(s) based on partid_mask in our partition's AMO */ /* clear bit(s) based on partid_mask in our partition's AMO */
FETCHOP_STORE_OP(TO_AMO((u64)&amo->variable), FETCHOP_AND, FETCHOP_STORE_OP(TO_AMO((u64)&amo->variable), FETCHOP_AND,
~partid_mask); ~partid_mask);
} }
/* /*
* The following set of macros and inlines are used for the sending and * The following set of macros and inlines are used for the sending and
* receiving of IPIs (also known as IRQs). There are two flavors of IPIs, * receiving of IPIs (also known as IRQs). There are two flavors of IPIs,
@@ -967,14 +901,12 @@ xpc_IPI_receive(AMO_t *amo)
return FETCHOP_LOAD_OP(TO_AMO((u64)&amo->variable), FETCHOP_CLEAR); return FETCHOP_LOAD_OP(TO_AMO((u64)&amo->variable), FETCHOP_CLEAR);
} }
static inline enum xpc_retval static inline enum xpc_retval
xpc_IPI_send(AMO_t *amo, u64 flag, int nasid, int phys_cpuid, int vector) xpc_IPI_send(AMO_t *amo, u64 flag, int nasid, int phys_cpuid, int vector)
{ {
int ret = 0; int ret = 0;
unsigned long irq_flags; unsigned long irq_flags;
local_irq_save(irq_flags); local_irq_save(irq_flags);
FETCHOP_STORE_OP(TO_AMO((u64)&amo->variable), FETCHOP_OR, flag); FETCHOP_STORE_OP(TO_AMO((u64)&amo->variable), FETCHOP_OR, flag);
@@ -994,7 +926,6 @@ xpc_IPI_send(AMO_t *amo, u64 flag, int nasid, int phys_cpuid, int vector)
return ((ret == 0) ? xpcSuccess : xpcPioReadError); return ((ret == 0) ? xpcSuccess : xpcPioReadError);
} }
/* /*
* IPIs associated with SGI_XPC_ACTIVATE IRQ. * IPIs associated with SGI_XPC_ACTIVATE IRQ.
*/ */
@@ -1011,7 +942,6 @@ xpc_activate_IRQ_send(u64 amos_page_pa, int from_nasid, int to_nasid,
AMO_t *amos = (AMO_t *)__va(amos_page_pa + AMO_t *amos = (AMO_t *)__va(amos_page_pa +
(XPC_ACTIVATE_IRQ_AMOS * sizeof(AMO_t))); (XPC_ACTIVATE_IRQ_AMOS * sizeof(AMO_t)));
(void)xpc_IPI_send(&amos[w_index], (1UL << b_index), to_nasid, (void)xpc_IPI_send(&amos[w_index], (1UL << b_index), to_nasid,
to_phys_cpuid, SGI_XPC_ACTIVATE); to_phys_cpuid, SGI_XPC_ACTIVATE);
} }
@@ -1027,7 +957,8 @@ static inline void
xpc_IPI_send_activated(struct xpc_partition *part) xpc_IPI_send_activated(struct xpc_partition *part)
{ {
xpc_activate_IRQ_send(part->remote_amos_page_pa, cnodeid_to_nasid(0), xpc_activate_IRQ_send(part->remote_amos_page_pa, cnodeid_to_nasid(0),
part->remote_act_nasid, part->remote_act_phys_cpuid); part->remote_act_nasid,
part->remote_act_phys_cpuid);
} }
static inline void static inline void
@@ -1041,10 +972,10 @@ static inline void
xpc_IPI_send_disengage(struct xpc_partition *part) xpc_IPI_send_disengage(struct xpc_partition *part)
{ {
xpc_activate_IRQ_send(part->remote_amos_page_pa, cnodeid_to_nasid(0), xpc_activate_IRQ_send(part->remote_amos_page_pa, cnodeid_to_nasid(0),
part->remote_act_nasid, part->remote_act_phys_cpuid); part->remote_act_nasid,
part->remote_act_phys_cpuid);
} }
/* /*
* IPIs associated with SGI_XPC_NOTIFY IRQ. * IPIs associated with SGI_XPC_NOTIFY IRQ.
*/ */
@@ -1063,27 +994,22 @@ xpc_notify_IRQ_send(struct xpc_channel *ch, u8 ipi_flag, char *ipi_flag_string,
struct xpc_partition *part = &xpc_partitions[ch->partid]; struct xpc_partition *part = &xpc_partitions[ch->partid];
enum xpc_retval ret; enum xpc_retval ret;
if (likely(part->act_state != XPC_P_DEACTIVATING)) { if (likely(part->act_state != XPC_P_DEACTIVATING)) {
ret = xpc_IPI_send(part->remote_IPI_amo_va, ret = xpc_IPI_send(part->remote_IPI_amo_va,
(u64)ipi_flag << (ch->number * 8), (u64)ipi_flag << (ch->number * 8),
part->remote_IPI_nasid, part->remote_IPI_nasid,
part->remote_IPI_phys_cpuid, part->remote_IPI_phys_cpuid, SGI_XPC_NOTIFY);
SGI_XPC_NOTIFY);
dev_dbg(xpc_chan, "%s sent to partid=%d, channel=%d, ret=%d\n", dev_dbg(xpc_chan, "%s sent to partid=%d, channel=%d, ret=%d\n",
ipi_flag_string, ch->partid, ch->number, ret); ipi_flag_string, ch->partid, ch->number, ret);
if (unlikely(ret != xpcSuccess)) { if (unlikely(ret != xpcSuccess)) {
if (irq_flags != NULL) { if (irq_flags != NULL)
spin_unlock_irqrestore(&ch->lock, *irq_flags); spin_unlock_irqrestore(&ch->lock, *irq_flags);
}
XPC_DEACTIVATE_PARTITION(part, ret); XPC_DEACTIVATE_PARTITION(part, ret);
if (irq_flags != NULL) { if (irq_flags != NULL)
spin_lock_irqsave(&ch->lock, *irq_flags); spin_lock_irqsave(&ch->lock, *irq_flags);
} }
} }
} }
}
/* /*
* Make it look like the remote partition, which is associated with the * Make it look like the remote partition, which is associated with the
@@ -1099,14 +1025,12 @@ xpc_notify_IRQ_send_local(struct xpc_channel *ch, u8 ipi_flag,
{ {
struct xpc_partition *part = &xpc_partitions[ch->partid]; struct xpc_partition *part = &xpc_partitions[ch->partid];
FETCHOP_STORE_OP(TO_AMO((u64)&part->local_IPI_amo_va->variable), FETCHOP_STORE_OP(TO_AMO((u64)&part->local_IPI_amo_va->variable),
FETCHOP_OR, ((u64)ipi_flag << (ch->number * 8))); FETCHOP_OR, ((u64)ipi_flag << (ch->number * 8)));
dev_dbg(xpc_chan, "%s sent local from partid=%d, channel=%d\n", dev_dbg(xpc_chan, "%s sent local from partid=%d, channel=%d\n",
ipi_flag_string, ch->partid, ch->number); ipi_flag_string, ch->partid, ch->number);
} }
/* /*
* The sending and receiving of IPIs includes the setting of an AMO variable * The sending and receiving of IPIs includes the setting of an AMO variable
* to indicate the reason the IPI was sent. The 64-bit variable is divided * to indicate the reason the IPI was sent. The 64-bit variable is divided
@@ -1121,21 +1045,18 @@ xpc_notify_IRQ_send_local(struct xpc_channel *ch, u8 ipi_flag,
#define XPC_IPI_OPENREPLY 0x08 #define XPC_IPI_OPENREPLY 0x08
#define XPC_IPI_MSGREQUEST 0x10 #define XPC_IPI_MSGREQUEST 0x10
/* given an AMO variable and a channel#, get its associated IPI flags */ /* given an AMO variable and a channel#, get its associated IPI flags */
#define XPC_GET_IPI_FLAGS(_amo, _c) ((u8) (((_amo) >> ((_c) * 8)) & 0xff)) #define XPC_GET_IPI_FLAGS(_amo, _c) ((u8) (((_amo) >> ((_c) * 8)) & 0xff))
#define XPC_SET_IPI_FLAGS(_amo, _c, _f) (_amo) |= ((u64) (_f) << ((_c) * 8)) #define XPC_SET_IPI_FLAGS(_amo, _c, _f) (_amo) |= ((u64) (_f) << ((_c) * 8))
#define XPC_ANY_OPENCLOSE_IPI_FLAGS_SET(_amo) ((_amo) & __IA64_UL_CONST(0x0f0f0f0f0f0f0f0f)) #define XPC_ANY_OPENCLOSE_IPI_FLAGS_SET(_amo) ((_amo) & 0x0f0f0f0f0f0f0f0fUL)
#define XPC_ANY_MSG_IPI_FLAGS_SET(_amo) ((_amo) & __IA64_UL_CONST(0x1010101010101010)) #define XPC_ANY_MSG_IPI_FLAGS_SET(_amo) ((_amo) & 0x1010101010101010UL)
static inline void static inline void
xpc_IPI_send_closerequest(struct xpc_channel *ch, unsigned long *irq_flags) xpc_IPI_send_closerequest(struct xpc_channel *ch, unsigned long *irq_flags)
{ {
struct xpc_openclose_args *args = ch->local_openclose_args; struct xpc_openclose_args *args = ch->local_openclose_args;
args->reason = ch->reason; args->reason = ch->reason;
XPC_NOTIFY_IRQ_SEND(ch, XPC_IPI_CLOSEREQUEST, irq_flags); XPC_NOTIFY_IRQ_SEND(ch, XPC_IPI_CLOSEREQUEST, irq_flags);
@@ -1152,7 +1073,6 @@ xpc_IPI_send_openrequest(struct xpc_channel *ch, unsigned long *irq_flags)
{ {
struct xpc_openclose_args *args = ch->local_openclose_args; struct xpc_openclose_args *args = ch->local_openclose_args;
args->msg_size = ch->msg_size; args->msg_size = ch->msg_size;
args->local_nentries = ch->local_nentries; args->local_nentries = ch->local_nentries;
@@ -1164,7 +1084,6 @@ xpc_IPI_send_openreply(struct xpc_channel *ch, unsigned long *irq_flags)
{ {
struct xpc_openclose_args *args = ch->local_openclose_args; struct xpc_openclose_args *args = ch->local_openclose_args;
args->remote_nentries = ch->remote_nentries; args->remote_nentries = ch->remote_nentries;
args->local_nentries = ch->local_nentries; args->local_nentries = ch->local_nentries;
args->local_msgqueue_pa = __pa(ch->local_msgqueue); args->local_msgqueue_pa = __pa(ch->local_msgqueue);
@@ -1184,7 +1103,6 @@ xpc_IPI_send_local_msgrequest(struct xpc_channel *ch)
XPC_NOTIFY_IRQ_SEND_LOCAL(ch, XPC_IPI_MSGREQUEST); XPC_NOTIFY_IRQ_SEND_LOCAL(ch, XPC_IPI_MSGREQUEST);
} }
/* /*
* Memory for XPC's AMO variables is allocated by the MSPEC driver. These * Memory for XPC's AMO variables is allocated by the MSPEC driver. These
* pages are located in the lowest granule. The lowest granule uses 4k pages * pages are located in the lowest granule. The lowest granule uses 4k pages
@@ -1201,13 +1119,10 @@ xpc_IPI_init(int index)
{ {
AMO_t *amo = xpc_vars->amos_page + index; AMO_t *amo = xpc_vars->amos_page + index;
(void)xpc_IPI_receive(amo); /* clear AMO variable */ (void)xpc_IPI_receive(amo); /* clear AMO variable */
return amo; return amo;
} }
static inline enum xpc_retval static inline enum xpc_retval
xpc_map_bte_errors(bte_result_t error) xpc_map_bte_errors(bte_result_t error)
{ {
@@ -1220,22 +1135,31 @@ xpc_map_bte_errors(bte_result_t error)
return xpcBteUnmappedError; return xpcBteUnmappedError;
} }
switch (error) { switch (error) {
case BTE_SUCCESS: return xpcSuccess; case BTE_SUCCESS:
case BTEFAIL_DIR: return xpcBteDirectoryError; return xpcSuccess;
case BTEFAIL_POISON: return xpcBtePoisonError; case BTEFAIL_DIR:
case BTEFAIL_WERR: return xpcBteWriteError; return xpcBteDirectoryError;
case BTEFAIL_ACCESS: return xpcBteAccessError; case BTEFAIL_POISON:
case BTEFAIL_PWERR: return xpcBtePWriteError; return xpcBtePoisonError;
case BTEFAIL_PRERR: return xpcBtePReadError; case BTEFAIL_WERR:
case BTEFAIL_TOUT: return xpcBteTimeOutError; return xpcBteWriteError;
case BTEFAIL_XTERR: return xpcBteXtalkError; case BTEFAIL_ACCESS:
case BTEFAIL_NOTAVAIL: return xpcBteNotAvailable; return xpcBteAccessError;
default: return xpcBteUnmappedError; case BTEFAIL_PWERR:
return xpcBtePWriteError;
case BTEFAIL_PRERR:
return xpcBtePReadError;
case BTEFAIL_TOUT:
return xpcBteTimeOutError;
case BTEFAIL_XTERR:
return xpcBteXtalkError;
case BTEFAIL_NOTAVAIL:
return xpcBteNotAvailable;
default:
return xpcBteUnmappedError;
} }
} }
/* /*
* Check to see if there is any channel activity to/from the specified * Check to see if there is any channel activity to/from the specified
* partition. * partition.
@@ -1246,11 +1170,9 @@ xpc_check_for_channel_activity(struct xpc_partition *part)
u64 IPI_amo; u64 IPI_amo;
unsigned long irq_flags; unsigned long irq_flags;
IPI_amo = xpc_IPI_receive(part->local_IPI_amo_va); IPI_amo = xpc_IPI_receive(part->local_IPI_amo_va);
if (IPI_amo == 0) { if (IPI_amo == 0)
return; return;
}
spin_lock_irqsave(&part->IPI_lock, irq_flags); spin_lock_irqsave(&part->IPI_lock, irq_flags);
part->local_IPI_amo |= IPI_amo; part->local_IPI_amo |= IPI_amo;
@@ -1262,6 +1184,4 @@ xpc_check_for_channel_activity(struct xpc_partition *part)
xpc_wakeup_channel_mgr(part); xpc_wakeup_channel_mgr(part);
} }
#endif /* _DRIVERS_MISC_SGIXP_XPC_H */
#endif /* _ASM_IA64_SN_XPC_H */

View File

@@ -3,10 +3,9 @@
* License. See the file "COPYING" in the main directory of this archive * License. See the file "COPYING" in the main directory of this archive
* for more details. * for more details.
* *
* Copyright (c) 2004-2007 Silicon Graphics, Inc. All Rights Reserved. * Copyright (c) 2004-2008 Silicon Graphics, Inc. All Rights Reserved.
*/ */
/* /*
* Cross Partition Communication (XPC) support - standard version. * Cross Partition Communication (XPC) support - standard version.
* *
@@ -44,23 +43,20 @@
* *
*/ */
#include <linux/kernel.h> #include <linux/kernel.h>
#include <linux/module.h> #include <linux/module.h>
#include <linux/init.h> #include <linux/init.h>
#include <linux/sched.h>
#include <linux/syscalls.h>
#include <linux/cache.h> #include <linux/cache.h>
#include <linux/interrupt.h> #include <linux/interrupt.h>
#include <linux/delay.h> #include <linux/delay.h>
#include <linux/reboot.h> #include <linux/reboot.h>
#include <linux/completion.h> #include <linux/completion.h>
#include <linux/kdebug.h> #include <linux/kdebug.h>
#include <linux/kthread.h>
#include <linux/uaccess.h>
#include <asm/sn/intr.h> #include <asm/sn/intr.h>
#include <asm/sn/sn_sal.h> #include <asm/sn/sn_sal.h>
#include <asm/uaccess.h> #include "xpc.h"
#include <asm/sn/xpc.h>
/* define two XPC debug device structures to be used with dev_dbg() et al */ /* define two XPC debug device structures to be used with dev_dbg() et al */
@@ -81,10 +77,8 @@ struct device xpc_chan_dbg_subname = {
struct device *xpc_part = &xpc_part_dbg_subname; struct device *xpc_part = &xpc_part_dbg_subname;
struct device *xpc_chan = &xpc_chan_dbg_subname; struct device *xpc_chan = &xpc_chan_dbg_subname;
static int xpc_kdebug_ignore; static int xpc_kdebug_ignore;
/* systune related variables for /proc/sys directories */ /* systune related variables for /proc/sys directories */
static int xpc_hb_interval = XPC_HB_DEFAULT_INTERVAL; static int xpc_hb_interval = XPC_HB_DEFAULT_INTERVAL;
@@ -96,7 +90,7 @@ static int xpc_hb_check_min_interval = 10;
static int xpc_hb_check_max_interval = 120; static int xpc_hb_check_max_interval = 120;
int xpc_disengage_request_timelimit = XPC_DISENGAGE_REQUEST_DEFAULT_TIMELIMIT; int xpc_disengage_request_timelimit = XPC_DISENGAGE_REQUEST_DEFAULT_TIMELIMIT;
static int xpc_disengage_request_min_timelimit = 0; static int xpc_disengage_request_min_timelimit; /* = 0 */
static int xpc_disengage_request_max_timelimit = 120; static int xpc_disengage_request_max_timelimit = 120;
static ctl_table xpc_sys_xpc_hb_dir[] = { static ctl_table xpc_sys_xpc_hb_dir[] = {
@@ -109,8 +103,7 @@ static ctl_table xpc_sys_xpc_hb_dir[] = {
.proc_handler = &proc_dointvec_minmax, .proc_handler = &proc_dointvec_minmax,
.strategy = &sysctl_intvec, .strategy = &sysctl_intvec,
.extra1 = &xpc_hb_min_interval, .extra1 = &xpc_hb_min_interval,
.extra2 = &xpc_hb_max_interval .extra2 = &xpc_hb_max_interval},
},
{ {
.ctl_name = CTL_UNNUMBERED, .ctl_name = CTL_UNNUMBERED,
.procname = "hb_check_interval", .procname = "hb_check_interval",
@@ -120,8 +113,7 @@ static ctl_table xpc_sys_xpc_hb_dir[] = {
.proc_handler = &proc_dointvec_minmax, .proc_handler = &proc_dointvec_minmax,
.strategy = &sysctl_intvec, .strategy = &sysctl_intvec,
.extra1 = &xpc_hb_check_min_interval, .extra1 = &xpc_hb_check_min_interval,
.extra2 = &xpc_hb_check_max_interval .extra2 = &xpc_hb_check_max_interval},
},
{} {}
}; };
static ctl_table xpc_sys_xpc_dir[] = { static ctl_table xpc_sys_xpc_dir[] = {
@@ -129,8 +121,7 @@ static ctl_table xpc_sys_xpc_dir[] = {
.ctl_name = CTL_UNNUMBERED, .ctl_name = CTL_UNNUMBERED,
.procname = "hb", .procname = "hb",
.mode = 0555, .mode = 0555,
.child = xpc_sys_xpc_hb_dir .child = xpc_sys_xpc_hb_dir},
},
{ {
.ctl_name = CTL_UNNUMBERED, .ctl_name = CTL_UNNUMBERED,
.procname = "disengage_request_timelimit", .procname = "disengage_request_timelimit",
@@ -140,8 +131,7 @@ static ctl_table xpc_sys_xpc_dir[] = {
.proc_handler = &proc_dointvec_minmax, .proc_handler = &proc_dointvec_minmax,
.strategy = &sysctl_intvec, .strategy = &sysctl_intvec,
.extra1 = &xpc_disengage_request_min_timelimit, .extra1 = &xpc_disengage_request_min_timelimit,
.extra2 = &xpc_disengage_request_max_timelimit .extra2 = &xpc_disengage_request_max_timelimit},
},
{} {}
}; };
static ctl_table xpc_sys_dir[] = { static ctl_table xpc_sys_dir[] = {
@@ -149,8 +139,7 @@ static ctl_table xpc_sys_dir[] = {
.ctl_name = CTL_UNNUMBERED, .ctl_name = CTL_UNNUMBERED,
.procname = "xpc", .procname = "xpc",
.mode = 0555, .mode = 0555,
.child = xpc_sys_xpc_dir .child = xpc_sys_xpc_dir},
},
{} {}
}; };
static struct ctl_table_header *xpc_sysctl; static struct ctl_table_header *xpc_sysctl;
@@ -172,13 +161,10 @@ static DECLARE_COMPLETION(xpc_hb_checker_exited);
/* notification that the xpc_discovery thread has exited */ /* notification that the xpc_discovery thread has exited */
static DECLARE_COMPLETION(xpc_discovery_exited); static DECLARE_COMPLETION(xpc_discovery_exited);
static struct timer_list xpc_hb_timer; static struct timer_list xpc_hb_timer;
static void xpc_kthread_waitmsgs(struct xpc_partition *, struct xpc_channel *); static void xpc_kthread_waitmsgs(struct xpc_partition *, struct xpc_channel *);
static int xpc_system_reboot(struct notifier_block *, unsigned long, void *); static int xpc_system_reboot(struct notifier_block *, unsigned long, void *);
static struct notifier_block xpc_reboot_notifier = { static struct notifier_block xpc_reboot_notifier = {
.notifier_call = xpc_system_reboot, .notifier_call = xpc_system_reboot,
@@ -189,7 +175,6 @@ static struct notifier_block xpc_die_notifier = {
.notifier_call = xpc_system_die, .notifier_call = xpc_system_die,
}; };
/* /*
* Timer function to enforce the timelimit on the partition disengage request. * Timer function to enforce the timelimit on the partition disengage request.
*/ */
@@ -198,7 +183,6 @@ xpc_timeout_partition_disengage_request(unsigned long data)
{ {
struct xpc_partition *part = (struct xpc_partition *)data; struct xpc_partition *part = (struct xpc_partition *)data;
DBUG_ON(time_before(jiffies, part->disengage_request_timeout)); DBUG_ON(time_before(jiffies, part->disengage_request_timeout));
(void)xpc_partition_disengaged(part); (void)xpc_partition_disengaged(part);
@@ -207,7 +191,6 @@ xpc_timeout_partition_disengage_request(unsigned long data)
DBUG_ON(xpc_partition_engaged(1UL << XPC_PARTID(part)) != 0); DBUG_ON(xpc_partition_engaged(1UL << XPC_PARTID(part)) != 0);
} }
/* /*
* Notify the heartbeat check thread that an IRQ has been received. * Notify the heartbeat check thread that an IRQ has been received.
*/ */
@@ -219,7 +202,6 @@ xpc_act_IRQ_handler(int irq, void *dev_id)
return IRQ_HANDLED; return IRQ_HANDLED;
} }
/* /*
* Timer to produce the heartbeat. The timer structures function is * Timer to produce the heartbeat. The timer structures function is
* already set when this is initially called. A tunable is used to * already set when this is initially called. A tunable is used to
@@ -230,15 +212,13 @@ xpc_hb_beater(unsigned long dummy)
{ {
xpc_vars->heartbeat++; xpc_vars->heartbeat++;
if (time_after_eq(jiffies, xpc_hb_check_timeout)) { if (time_after_eq(jiffies, xpc_hb_check_timeout))
wake_up_interruptible(&xpc_act_IRQ_wq); wake_up_interruptible(&xpc_act_IRQ_wq);
}
xpc_hb_timer.expires = jiffies + (xpc_hb_interval * HZ); xpc_hb_timer.expires = jiffies + (xpc_hb_interval * HZ);
add_timer(&xpc_hb_timer); add_timer(&xpc_hb_timer);
} }
/* /*
* This thread is responsible for nearly all of the partition * This thread is responsible for nearly all of the partition
* activation/deactivation. * activation/deactivation.
@@ -250,25 +230,21 @@ xpc_hb_checker(void *ignore)
int new_IRQ_count; int new_IRQ_count;
int force_IRQ = 0; int force_IRQ = 0;
/* this thread was marked active by xpc_hb_init() */ /* this thread was marked active by xpc_hb_init() */
daemonize(XPC_HB_CHECK_THREAD_NAME);
set_cpus_allowed(current, cpumask_of_cpu(XPC_HB_CHECK_CPU)); set_cpus_allowed(current, cpumask_of_cpu(XPC_HB_CHECK_CPU));
/* set our heartbeating to other partitions into motion */ /* set our heartbeating to other partitions into motion */
xpc_hb_check_timeout = jiffies + (xpc_hb_check_interval * HZ); xpc_hb_check_timeout = jiffies + (xpc_hb_check_interval * HZ);
xpc_hb_beater(0); xpc_hb_beater(0);
while (!(volatile int) xpc_exiting) { while (!xpc_exiting) {
dev_dbg(xpc_part, "woke up with %d ticks rem; %d IRQs have " dev_dbg(xpc_part, "woke up with %d ticks rem; %d IRQs have "
"been received\n", "been received\n",
(int)(xpc_hb_check_timeout - jiffies), (int)(xpc_hb_check_timeout - jiffies),
atomic_read(&xpc_act_IRQ_rcvd) - last_IRQ_count); atomic_read(&xpc_act_IRQ_rcvd) - last_IRQ_count);
/* checking of remote heartbeats is skewed by IRQ handling */ /* checking of remote heartbeats is skewed by IRQ handling */
if (time_after_eq(jiffies, xpc_hb_check_timeout)) { if (time_after_eq(jiffies, xpc_hb_check_timeout)) {
dev_dbg(xpc_part, "checking remote heartbeats\n"); dev_dbg(xpc_part, "checking remote heartbeats\n");
@@ -282,7 +258,6 @@ xpc_hb_checker(void *ignore)
force_IRQ = 1; force_IRQ = 1;
} }
/* check for outstanding IRQs */ /* check for outstanding IRQs */
new_IRQ_count = atomic_read(&xpc_act_IRQ_rcvd); new_IRQ_count = atomic_read(&xpc_act_IRQ_rcvd);
if (last_IRQ_count < new_IRQ_count || force_IRQ != 0) { if (last_IRQ_count < new_IRQ_count || force_IRQ != 0) {
@@ -304,20 +279,20 @@ xpc_hb_checker(void *ignore)
/* wait for IRQ or timeout */ /* wait for IRQ or timeout */
(void)wait_event_interruptible(xpc_act_IRQ_wq, (void)wait_event_interruptible(xpc_act_IRQ_wq,
(last_IRQ_count < atomic_read(&xpc_act_IRQ_rcvd) || (last_IRQ_count <
time_after_eq(jiffies, xpc_hb_check_timeout) || atomic_read(&xpc_act_IRQ_rcvd)
(volatile int) xpc_exiting)); || time_after_eq(jiffies,
xpc_hb_check_timeout) ||
xpc_exiting));
} }
dev_dbg(xpc_part, "heartbeat checker is exiting\n"); dev_dbg(xpc_part, "heartbeat checker is exiting\n");
/* mark this thread as having exited */ /* mark this thread as having exited */
complete(&xpc_hb_checker_exited); complete(&xpc_hb_checker_exited);
return 0; return 0;
} }
/* /*
* This thread will attempt to discover other partitions to activate * This thread will attempt to discover other partitions to activate
* based on info provided by SAL. This new thread is short lived and * based on info provided by SAL. This new thread is short lived and
@@ -326,8 +301,6 @@ xpc_hb_checker(void *ignore)
static int static int
xpc_initiate_discovery(void *ignore) xpc_initiate_discovery(void *ignore)
{ {
daemonize(XPC_DISCOVERY_THREAD_NAME);
xpc_discovery(); xpc_discovery();
dev_dbg(xpc_part, "discovery thread is exiting\n"); dev_dbg(xpc_part, "discovery thread is exiting\n");
@@ -337,7 +310,6 @@ xpc_initiate_discovery(void *ignore)
return 0; return 0;
} }
/* /*
* Establish first contact with the remote partititon. This involves pulling * Establish first contact with the remote partititon. This involves pulling
* the XPC per partition variables from the remote partition and waiting for * the XPC per partition variables from the remote partition and waiting for
@@ -348,7 +320,6 @@ xpc_make_first_contact(struct xpc_partition *part)
{ {
enum xpc_retval ret; enum xpc_retval ret;
while ((ret = xpc_pull_remote_vars_part(part)) != xpcSuccess) { while ((ret = xpc_pull_remote_vars_part(part)) != xpcSuccess) {
if (ret != xpcRetry) { if (ret != xpcRetry) {
XPC_DEACTIVATE_PARTITION(part, ret); XPC_DEACTIVATE_PARTITION(part, ret);
@@ -361,15 +332,13 @@ xpc_make_first_contact(struct xpc_partition *part)
/* wait a 1/4 of a second or so */ /* wait a 1/4 of a second or so */
(void)msleep_interruptible(250); (void)msleep_interruptible(250);
if (part->act_state == XPC_P_DEACTIVATING) { if (part->act_state == XPC_P_DEACTIVATING)
return part->reason; return part->reason;
} }
}
return xpc_mark_partition_active(part); return xpc_mark_partition_active(part);
} }
/* /*
* The first kthread assigned to a newly activated partition is the one * The first kthread assigned to a newly activated partition is the one
* created by XPC HB with which it calls xpc_partition_up(). XPC hangs on to * created by XPC HB with which it calls xpc_partition_up(). XPC hangs on to
@@ -391,7 +360,6 @@ xpc_channel_mgr(struct xpc_partition *part)
xpc_process_channel_activity(part); xpc_process_channel_activity(part);
/* /*
* Wait until we've been requested to activate kthreads or * Wait until we've been requested to activate kthreads or
* all of the channel's message queues have been torn down or * all of the channel's message queues have been torn down or
@@ -408,19 +376,14 @@ xpc_channel_mgr(struct xpc_partition *part)
atomic_dec(&part->channel_mgr_requests); atomic_dec(&part->channel_mgr_requests);
(void)wait_event_interruptible(part->channel_mgr_wq, (void)wait_event_interruptible(part->channel_mgr_wq,
(atomic_read(&part->channel_mgr_requests) > 0 || (atomic_read(&part->channel_mgr_requests) > 0 ||
(volatile u64) part->local_IPI_amo != 0 || part->local_IPI_amo != 0 ||
((volatile u8) part->act_state == (part->act_state == XPC_P_DEACTIVATING &&
XPC_P_DEACTIVATING &&
atomic_read(&part->nchannels_active) == 0 && atomic_read(&part->nchannels_active) == 0 &&
xpc_partition_disengaged(part)))); xpc_partition_disengaged(part))));
atomic_set(&part->channel_mgr_requests, 1); atomic_set(&part->channel_mgr_requests, 1);
// >>> Does it need to wakeup periodically as well? In case we
// >>> miscalculated the #of kthreads to wakeup or create?
} }
} }
/* /*
* When XPC HB determines that a partition has come up, it will create a new * When XPC HB determines that a partition has come up, it will create a new
* kthread and that kthread will call this function to attempt to set up the * kthread and that kthread will call this function to attempt to set up the
@@ -443,9 +406,8 @@ xpc_partition_up(struct xpc_partition *part)
dev_dbg(xpc_chan, "activating partition %d\n", XPC_PARTID(part)); dev_dbg(xpc_chan, "activating partition %d\n", XPC_PARTID(part));
if (xpc_setup_infrastructure(part) != xpcSuccess) { if (xpc_setup_infrastructure(part) != xpcSuccess)
return; return;
}
/* /*
* The kthread that XPC HB called us with will become the * The kthread that XPC HB called us with will become the
@@ -456,25 +418,20 @@ xpc_partition_up(struct xpc_partition *part)
(void)xpc_part_ref(part); /* this will always succeed */ (void)xpc_part_ref(part); /* this will always succeed */
if (xpc_make_first_contact(part) == xpcSuccess) { if (xpc_make_first_contact(part) == xpcSuccess)
xpc_channel_mgr(part); xpc_channel_mgr(part);
}
xpc_part_deref(part); xpc_part_deref(part);
xpc_teardown_infrastructure(part); xpc_teardown_infrastructure(part);
} }
static int static int
xpc_activating(void *__partid) xpc_activating(void *__partid)
{ {
partid_t partid = (u64)__partid; partid_t partid = (u64)__partid;
struct xpc_partition *part = &xpc_partitions[partid]; struct xpc_partition *part = &xpc_partitions[partid];
unsigned long irq_flags; unsigned long irq_flags;
struct sched_param param = { .sched_priority = MAX_RT_PRIO - 1 };
int ret;
DBUG_ON(partid <= 0 || partid >= XP_MAX_PARTITIONS); DBUG_ON(partid <= 0 || partid >= XP_MAX_PARTITIONS);
@@ -496,21 +453,6 @@ xpc_activating(void *__partid)
dev_dbg(xpc_part, "bringing partition %d up\n", partid); dev_dbg(xpc_part, "bringing partition %d up\n", partid);
daemonize("xpc%02d", partid);
/*
* This thread needs to run at a realtime priority to prevent a
* significant performance degradation.
*/
ret = sched_setscheduler(current, SCHED_FIFO, &param);
if (ret != 0) {
dev_warn(xpc_part, "unable to set pid %d to a realtime "
"priority, ret=%d\n", current->pid, ret);
}
/* allow this thread and its children to run on any CPU */
set_cpus_allowed(current, CPU_MASK_ALL);
/* /*
* Register the remote partition's AMOs with SAL so it can handle * Register the remote partition's AMOs with SAL so it can handle
* and cleanup errors within that address range should the remote * and cleanup errors within that address range should the remote
@@ -537,7 +479,6 @@ xpc_activating(void *__partid)
xpc_allow_hb(partid, xpc_vars); xpc_allow_hb(partid, xpc_vars);
xpc_IPI_send_activated(part); xpc_IPI_send_activated(part);
/* /*
* xpc_partition_up() holds this thread and marks this partition as * xpc_partition_up() holds this thread and marks this partition as
* XPC_P_ACTIVE by calling xpc_hb_mark_active(). * XPC_P_ACTIVE by calling xpc_hb_mark_active().
@@ -555,14 +496,12 @@ xpc_activating(void *__partid)
return 0; return 0;
} }
void void
xpc_activate_partition(struct xpc_partition *part) xpc_activate_partition(struct xpc_partition *part)
{ {
partid_t partid = XPC_PARTID(part); partid_t partid = XPC_PARTID(part);
unsigned long irq_flags; unsigned long irq_flags;
pid_t pid; struct task_struct *kthread;
spin_lock_irqsave(&part->act_lock, irq_flags); spin_lock_irqsave(&part->act_lock, irq_flags);
@@ -573,9 +512,9 @@ xpc_activate_partition(struct xpc_partition *part)
spin_unlock_irqrestore(&part->act_lock, irq_flags); spin_unlock_irqrestore(&part->act_lock, irq_flags);
pid = kernel_thread(xpc_activating, (void *) ((u64) partid), 0); kthread = kthread_run(xpc_activating, (void *)((u64)partid), "xpc%02d",
partid);
if (unlikely(pid <= 0)) { if (IS_ERR(kthread)) {
spin_lock_irqsave(&part->act_lock, irq_flags); spin_lock_irqsave(&part->act_lock, irq_flags);
part->act_state = XPC_P_INACTIVE; part->act_state = XPC_P_INACTIVE;
XPC_SET_REASON(part, xpcCloneKThreadFailed, __LINE__); XPC_SET_REASON(part, xpcCloneKThreadFailed, __LINE__);
@@ -583,12 +522,11 @@ xpc_activate_partition(struct xpc_partition *part)
} }
} }
/* /*
* Handle the receipt of a SGI_XPC_NOTIFY IRQ by seeing whether the specified * Handle the receipt of a SGI_XPC_NOTIFY IRQ by seeing whether the specified
* partition actually sent it. Since SGI_XPC_NOTIFY IRQs may be shared by more * partition actually sent it. Since SGI_XPC_NOTIFY IRQs may be shared by more
* than one partition, we use an AMO_t structure per partition to indicate * than one partition, we use an AMO_t structure per partition to indicate
* whether a partition has sent an IPI or not. >>> If it has, then wake up the * whether a partition has sent an IPI or not. If it has, then wake up the
* associated kthread to handle it. * associated kthread to handle it.
* *
* All SGI_XPC_NOTIFY IRQs received by XPC are the result of IPIs sent by XPC * All SGI_XPC_NOTIFY IRQs received by XPC are the result of IPIs sent by XPC
@@ -606,7 +544,6 @@ xpc_notify_IRQ_handler(int irq, void *dev_id)
partid_t partid = (partid_t) (u64)dev_id; partid_t partid = (partid_t) (u64)dev_id;
struct xpc_partition *part = &xpc_partitions[partid]; struct xpc_partition *part = &xpc_partitions[partid];
DBUG_ON(partid <= 0 || partid >= XP_MAX_PARTITIONS); DBUG_ON(partid <= 0 || partid >= XP_MAX_PARTITIONS);
if (xpc_part_ref(part)) { if (xpc_part_ref(part)) {
@@ -617,7 +554,6 @@ xpc_notify_IRQ_handler(int irq, void *dev_id)
return IRQ_HANDLED; return IRQ_HANDLED;
} }
/* /*
* Check to see if xpc_notify_IRQ_handler() dropped any IPIs on the floor * Check to see if xpc_notify_IRQ_handler() dropped any IPIs on the floor
* because the write to their associated IPI amo completed after the IRQ/IPI * because the write to their associated IPI amo completed after the IRQ/IPI
@@ -636,7 +572,6 @@ xpc_dropped_IPI_check(struct xpc_partition *part)
} }
} }
void void
xpc_activate_kthreads(struct xpc_channel *ch, int needed) xpc_activate_kthreads(struct xpc_channel *ch, int needed)
{ {
@@ -644,7 +579,6 @@ xpc_activate_kthreads(struct xpc_channel *ch, int needed)
int assigned = atomic_read(&ch->kthreads_assigned); int assigned = atomic_read(&ch->kthreads_assigned);
int wakeup; int wakeup;
DBUG_ON(needed <= 0); DBUG_ON(needed <= 0);
if (idle > 0) { if (idle > 0) {
@@ -658,17 +592,14 @@ xpc_activate_kthreads(struct xpc_channel *ch, int needed)
wake_up_nr(&ch->idle_wq, wakeup); wake_up_nr(&ch->idle_wq, wakeup);
} }
if (needed <= 0) { if (needed <= 0)
return; return;
}
if (needed + assigned > ch->kthreads_assigned_limit) { if (needed + assigned > ch->kthreads_assigned_limit) {
needed = ch->kthreads_assigned_limit - assigned; needed = ch->kthreads_assigned_limit - assigned;
// >>>should never be less than 0 if (needed <= 0)
if (needed <= 0) {
return; return;
} }
}
dev_dbg(xpc_chan, "create %d new kthreads, partid=%d, channel=%d\n", dev_dbg(xpc_chan, "create %d new kthreads, partid=%d, channel=%d\n",
needed, ch->partid, ch->number); needed, ch->partid, ch->number);
@@ -676,7 +607,6 @@ xpc_activate_kthreads(struct xpc_channel *ch, int needed)
xpc_create_kthreads(ch, needed, 0); xpc_create_kthreads(ch, needed, 0);
} }
/* /*
* This function is where XPC's kthreads wait for messages to deliver. * This function is where XPC's kthreads wait for messages to deliver.
*/ */
@@ -686,10 +616,8 @@ xpc_kthread_waitmsgs(struct xpc_partition *part, struct xpc_channel *ch)
do { do {
/* deliver messages to their intended recipients */ /* deliver messages to their intended recipients */
while ((volatile s64) ch->w_local_GP.get < while (ch->w_local_GP.get < ch->w_remote_GP.put &&
(volatile s64) ch->w_remote_GP.put && !(ch->flags & XPC_C_DISCONNECTING)) {
!((volatile u32) ch->flags &
XPC_C_DISCONNECTING)) {
xpc_deliver_msg(ch); xpc_deliver_msg(ch);
} }
@@ -704,19 +632,16 @@ xpc_kthread_waitmsgs(struct xpc_partition *part, struct xpc_channel *ch)
"wait_event_interruptible_exclusive()\n"); "wait_event_interruptible_exclusive()\n");
(void)wait_event_interruptible_exclusive(ch->idle_wq, (void)wait_event_interruptible_exclusive(ch->idle_wq,
((volatile s64) ch->w_local_GP.get < (ch->w_local_GP.get < ch->w_remote_GP.put ||
(volatile s64) ch->w_remote_GP.put || (ch->flags & XPC_C_DISCONNECTING)));
((volatile u32) ch->flags &
XPC_C_DISCONNECTING)));
atomic_dec(&ch->kthreads_idle); atomic_dec(&ch->kthreads_idle);
} while (!((volatile u32) ch->flags & XPC_C_DISCONNECTING)); } while (!(ch->flags & XPC_C_DISCONNECTING));
} }
static int static int
xpc_daemonize_kthread(void *args) xpc_kthread_start(void *args)
{ {
partid_t partid = XPC_UNPACK_ARG1(args); partid_t partid = XPC_UNPACK_ARG1(args);
u16 ch_number = XPC_UNPACK_ARG2(args); u16 ch_number = XPC_UNPACK_ARG2(args);
@@ -725,9 +650,6 @@ xpc_daemonize_kthread(void *args)
int n_needed; int n_needed;
unsigned long irq_flags; unsigned long irq_flags;
daemonize("xpc%02dc%d", partid, ch_number);
dev_dbg(xpc_chan, "kthread starting, partid=%d, channel=%d\n", dev_dbg(xpc_chan, "kthread starting, partid=%d, channel=%d\n",
partid, ch_number); partid, ch_number);
@@ -756,10 +678,9 @@ xpc_daemonize_kthread(void *args)
* need one less than total #of messages to deliver. * need one less than total #of messages to deliver.
*/ */
n_needed = ch->w_remote_GP.put - ch->w_local_GP.get - 1; n_needed = ch->w_remote_GP.put - ch->w_local_GP.get - 1;
if (n_needed > 0 && if (n_needed > 0 && !(ch->flags & XPC_C_DISCONNECTING))
!(ch->flags & XPC_C_DISCONNECTING)) {
xpc_activate_kthreads(ch, n_needed); xpc_activate_kthreads(ch, n_needed);
}
} else { } else {
spin_unlock_irqrestore(&ch->lock, irq_flags); spin_unlock_irqrestore(&ch->lock, irq_flags);
} }
@@ -798,7 +719,6 @@ xpc_daemonize_kthread(void *args)
return 0; return 0;
} }
/* /*
* For each partition that XPC has established communications with, there is * For each partition that XPC has established communications with, there is
* a minimum of one kernel thread assigned to perform any operation that * a minimum of one kernel thread assigned to perform any operation that
@@ -816,10 +736,9 @@ xpc_create_kthreads(struct xpc_channel *ch, int needed,
int ignore_disconnecting) int ignore_disconnecting)
{ {
unsigned long irq_flags; unsigned long irq_flags;
pid_t pid;
u64 args = XPC_PACK_ARGS(ch->partid, ch->number); u64 args = XPC_PACK_ARGS(ch->partid, ch->number);
struct xpc_partition *part = &xpc_partitions[ch->partid]; struct xpc_partition *part = &xpc_partitions[ch->partid];
struct task_struct *kthread;
while (needed-- > 0) { while (needed-- > 0) {
@@ -846,8 +765,9 @@ xpc_create_kthreads(struct xpc_channel *ch, int needed,
(void)xpc_part_ref(part); (void)xpc_part_ref(part);
xpc_msgqueue_ref(ch); xpc_msgqueue_ref(ch);
pid = kernel_thread(xpc_daemonize_kthread, (void *) args, 0); kthread = kthread_run(xpc_kthread_start, (void *)args,
if (pid < 0) { "xpc%02dc%d", ch->partid, ch->number);
if (IS_ERR(kthread)) {
/* the fork failed */ /* the fork failed */
/* /*
@@ -857,7 +777,7 @@ xpc_create_kthreads(struct xpc_channel *ch, int needed,
* to this channel are blocked in the channel's * to this channel are blocked in the channel's
* registerer, because the only thing that will unblock * registerer, because the only thing that will unblock
* them is the xpcDisconnecting callout that this * them is the xpcDisconnecting callout that this
* failed kernel_thread would have made. * failed kthread_run() would have made.
*/ */
if (atomic_dec_return(&ch->kthreads_assigned) == 0 && if (atomic_dec_return(&ch->kthreads_assigned) == 0 &&
@@ -882,12 +802,9 @@ xpc_create_kthreads(struct xpc_channel *ch, int needed,
} }
break; break;
} }
ch->kthreads_created++; // >>> temporary debug only!!!
} }
} }
void void
xpc_disconnect_wait(int ch_number) xpc_disconnect_wait(int ch_number)
{ {
@@ -897,14 +814,12 @@ xpc_disconnect_wait(int ch_number)
struct xpc_channel *ch; struct xpc_channel *ch;
int wakeup_channel_mgr; int wakeup_channel_mgr;
/* now wait for all callouts to the caller's function to cease */ /* now wait for all callouts to the caller's function to cease */
for (partid = 1; partid < XP_MAX_PARTITIONS; partid++) { for (partid = 1; partid < XP_MAX_PARTITIONS; partid++) {
part = &xpc_partitions[partid]; part = &xpc_partitions[partid];
if (!xpc_part_ref(part)) { if (!xpc_part_ref(part))
continue; continue;
}
ch = &part->channels[ch_number]; ch = &part->channels[ch_number];
@@ -923,7 +838,8 @@ xpc_disconnect_wait(int ch_number)
if (part->act_state != XPC_P_DEACTIVATING) { if (part->act_state != XPC_P_DEACTIVATING) {
spin_lock(&part->IPI_lock); spin_lock(&part->IPI_lock);
XPC_SET_IPI_FLAGS(part->local_IPI_amo, XPC_SET_IPI_FLAGS(part->local_IPI_amo,
ch->number, ch->delayed_IPI_flags); ch->number,
ch->delayed_IPI_flags);
spin_unlock(&part->IPI_lock); spin_unlock(&part->IPI_lock);
wakeup_channel_mgr = 1; wakeup_channel_mgr = 1;
} }
@@ -933,15 +849,13 @@ xpc_disconnect_wait(int ch_number)
ch->flags &= ~XPC_C_WDISCONNECT; ch->flags &= ~XPC_C_WDISCONNECT;
spin_unlock_irqrestore(&ch->lock, irq_flags); spin_unlock_irqrestore(&ch->lock, irq_flags);
if (wakeup_channel_mgr) { if (wakeup_channel_mgr)
xpc_wakeup_channel_mgr(part); xpc_wakeup_channel_mgr(part);
}
xpc_part_deref(part); xpc_part_deref(part);
} }
} }
static void static void
xpc_do_exit(enum xpc_retval reason) xpc_do_exit(enum xpc_retval reason)
{ {
@@ -950,7 +864,6 @@ xpc_do_exit(enum xpc_retval reason)
struct xpc_partition *part; struct xpc_partition *part;
unsigned long printmsg_time, disengage_request_timeout = 0; unsigned long printmsg_time, disengage_request_timeout = 0;
/* a 'rmmod XPC' and a 'reboot' cannot both end up here together */ /* a 'rmmod XPC' and a 'reboot' cannot both end up here together */
DBUG_ON(xpc_exiting == 1); DBUG_ON(xpc_exiting == 1);
@@ -971,11 +884,9 @@ xpc_do_exit(enum xpc_retval reason)
/* wait for the heartbeat checker thread to exit */ /* wait for the heartbeat checker thread to exit */
wait_for_completion(&xpc_hb_checker_exited); wait_for_completion(&xpc_hb_checker_exited);
/* sleep for a 1/3 of a second or so */ /* sleep for a 1/3 of a second or so */
(void)msleep_interruptible(300); (void)msleep_interruptible(300);
/* wait for all partitions to become inactive */ /* wait for all partitions to become inactive */
printmsg_time = jiffies + (XPC_DISENGAGE_PRINTMSG_INTERVAL * HZ); printmsg_time = jiffies + (XPC_DISENGAGE_PRINTMSG_INTERVAL * HZ);
@@ -1037,7 +948,6 @@ xpc_do_exit(enum xpc_retval reason)
DBUG_ON(xpc_partition_engaged(-1UL)); DBUG_ON(xpc_partition_engaged(-1UL));
/* indicate to others that our reserved page is uninitialized */ /* indicate to others that our reserved page is uninitialized */
xpc_rsvd_page->vars_pa = 0; xpc_rsvd_page->vars_pa = 0;
@@ -1056,18 +966,15 @@ xpc_do_exit(enum xpc_retval reason)
/* close down protections for IPI operations */ /* close down protections for IPI operations */
xpc_restrict_IPI_ops(); xpc_restrict_IPI_ops();
/* clear the interface to XPC's functions */ /* clear the interface to XPC's functions */
xpc_clear_interface(); xpc_clear_interface();
if (xpc_sysctl) { if (xpc_sysctl)
unregister_sysctl_table(xpc_sysctl); unregister_sysctl_table(xpc_sysctl);
}
kfree(xpc_remote_copy_buffer_base); kfree(xpc_remote_copy_buffer_base);
} }
/* /*
* This function is called when the system is being rebooted. * This function is called when the system is being rebooted.
*/ */
@@ -1076,7 +983,6 @@ xpc_system_reboot(struct notifier_block *nb, unsigned long event, void *unused)
{ {
enum xpc_retval reason; enum xpc_retval reason;
switch (event) { switch (event) {
case SYS_RESTART: case SYS_RESTART:
reason = xpcSystemReboot; reason = xpcSystemReboot;
@@ -1095,7 +1001,6 @@ xpc_system_reboot(struct notifier_block *nb, unsigned long event, void *unused)
return NOTIFY_DONE; return NOTIFY_DONE;
} }
/* /*
* Notify other partitions to disengage from all references to our memory. * Notify other partitions to disengage from all references to our memory.
*/ */
@@ -1107,7 +1012,6 @@ xpc_die_disengage(void)
unsigned long engaged; unsigned long engaged;
long time, printmsg_time, disengage_request_timeout; long time, printmsg_time, disengage_request_timeout;
/* keep xpc_hb_checker thread from doing anything (just in case) */ /* keep xpc_hb_checker thread from doing anything (just in case) */
xpc_exiting = 1; xpc_exiting = 1;
@@ -1171,7 +1075,6 @@ xpc_die_disengage(void)
} }
} }
/* /*
* This function is called when the system is being restarted or halted due * This function is called when the system is being restarted or halted due
* to some sort of system failure. If this is the case we need to notify the * to some sort of system failure. If this is the case we need to notify the
@@ -1191,9 +1094,9 @@ xpc_system_die(struct notifier_block *nb, unsigned long event, void *unused)
case DIE_KDEBUG_ENTER: case DIE_KDEBUG_ENTER:
/* Should lack of heartbeat be ignored by other partitions? */ /* Should lack of heartbeat be ignored by other partitions? */
if (!xpc_kdebug_ignore) { if (!xpc_kdebug_ignore)
break; break;
}
/* fall through */ /* fall through */
case DIE_MCA_MONARCH_ENTER: case DIE_MCA_MONARCH_ENTER:
case DIE_INIT_MONARCH_ENTER: case DIE_INIT_MONARCH_ENTER:
@@ -1203,9 +1106,9 @@ xpc_system_die(struct notifier_block *nb, unsigned long event, void *unused)
case DIE_KDEBUG_LEAVE: case DIE_KDEBUG_LEAVE:
/* Is lack of heartbeat being ignored by other partitions? */ /* Is lack of heartbeat being ignored by other partitions? */
if (!xpc_kdebug_ignore) { if (!xpc_kdebug_ignore)
break; break;
}
/* fall through */ /* fall through */
case DIE_MCA_MONARCH_LEAVE: case DIE_MCA_MONARCH_LEAVE:
case DIE_INIT_MONARCH_LEAVE: case DIE_INIT_MONARCH_LEAVE:
@@ -1217,26 +1120,23 @@ xpc_system_die(struct notifier_block *nb, unsigned long event, void *unused)
return NOTIFY_DONE; return NOTIFY_DONE;
} }
int __init int __init
xpc_init(void) xpc_init(void)
{ {
int ret; int ret;
partid_t partid; partid_t partid;
struct xpc_partition *part; struct xpc_partition *part;
pid_t pid; struct task_struct *kthread;
size_t buf_size; size_t buf_size;
if (!ia64_platform_is("sn2"))
if (!ia64_platform_is("sn2")) {
return -ENODEV; return -ENODEV;
}
buf_size = max(XPC_RP_VARS_SIZE, buf_size = max(XPC_RP_VARS_SIZE,
XPC_RP_HEADER_SIZE + XP_NASID_MASK_BYTES); XPC_RP_HEADER_SIZE + XP_NASID_MASK_BYTES);
xpc_remote_copy_buffer = xpc_kmalloc_cacheline_aligned(buf_size, xpc_remote_copy_buffer = xpc_kmalloc_cacheline_aligned(buf_size,
GFP_KERNEL, &xpc_remote_copy_buffer_base); GFP_KERNEL,
&xpc_remote_copy_buffer_base);
if (xpc_remote_copy_buffer == NULL) if (xpc_remote_copy_buffer == NULL)
return -ENOMEM; return -ENOMEM;
@@ -1299,9 +1199,8 @@ xpc_init(void)
xpc_restrict_IPI_ops(); xpc_restrict_IPI_ops();
if (xpc_sysctl) { if (xpc_sysctl)
unregister_sysctl_table(xpc_sysctl); unregister_sysctl_table(xpc_sysctl);
}
kfree(xpc_remote_copy_buffer_base); kfree(xpc_remote_copy_buffer_base);
return -EBUSY; return -EBUSY;
@@ -1319,26 +1218,22 @@ xpc_init(void)
free_irq(SGI_XPC_ACTIVATE, NULL); free_irq(SGI_XPC_ACTIVATE, NULL);
xpc_restrict_IPI_ops(); xpc_restrict_IPI_ops();
if (xpc_sysctl) { if (xpc_sysctl)
unregister_sysctl_table(xpc_sysctl); unregister_sysctl_table(xpc_sysctl);
}
kfree(xpc_remote_copy_buffer_base); kfree(xpc_remote_copy_buffer_base);
return -EBUSY; return -EBUSY;
} }
/* add ourselves to the reboot_notifier_list */ /* add ourselves to the reboot_notifier_list */
ret = register_reboot_notifier(&xpc_reboot_notifier); ret = register_reboot_notifier(&xpc_reboot_notifier);
if (ret != 0) { if (ret != 0)
dev_warn(xpc_part, "can't register reboot notifier\n"); dev_warn(xpc_part, "can't register reboot notifier\n");
}
/* add ourselves to the die_notifier list */ /* add ourselves to the die_notifier list */
ret = register_die_notifier(&xpc_die_notifier); ret = register_die_notifier(&xpc_die_notifier);
if (ret != 0) { if (ret != 0)
dev_warn(xpc_part, "can't register die notifier\n"); dev_warn(xpc_part, "can't register die notifier\n");
}
init_timer(&xpc_hb_timer); init_timer(&xpc_hb_timer);
xpc_hb_timer.function = xpc_hb_beater; xpc_hb_timer.function = xpc_hb_beater;
@@ -1347,8 +1242,8 @@ xpc_init(void)
* The real work-horse behind xpc. This processes incoming * The real work-horse behind xpc. This processes incoming
* interrupts and monitors remote heartbeats. * interrupts and monitors remote heartbeats.
*/ */
pid = kernel_thread(xpc_hb_checker, NULL, 0); kthread = kthread_run(xpc_hb_checker, NULL, XPC_HB_CHECK_THREAD_NAME);
if (pid < 0) { if (IS_ERR(kthread)) {
dev_err(xpc_part, "failed while forking hb check thread\n"); dev_err(xpc_part, "failed while forking hb check thread\n");
/* indicate to others that our reserved page is uninitialized */ /* indicate to others that our reserved page is uninitialized */
@@ -1364,22 +1259,21 @@ xpc_init(void)
free_irq(SGI_XPC_ACTIVATE, NULL); free_irq(SGI_XPC_ACTIVATE, NULL);
xpc_restrict_IPI_ops(); xpc_restrict_IPI_ops();
if (xpc_sysctl) { if (xpc_sysctl)
unregister_sysctl_table(xpc_sysctl); unregister_sysctl_table(xpc_sysctl);
}
kfree(xpc_remote_copy_buffer_base); kfree(xpc_remote_copy_buffer_base);
return -EBUSY; return -EBUSY;
} }
/* /*
* Startup a thread that will attempt to discover other partitions to * Startup a thread that will attempt to discover other partitions to
* activate based on info provided by SAL. This new thread is short * activate based on info provided by SAL. This new thread is short
* lived and will exit once discovery is complete. * lived and will exit once discovery is complete.
*/ */
pid = kernel_thread(xpc_initiate_discovery, NULL, 0); kthread = kthread_run(xpc_initiate_discovery, NULL,
if (pid < 0) { XPC_DISCOVERY_THREAD_NAME);
if (IS_ERR(kthread)) {
dev_err(xpc_part, "failed while forking discovery thread\n"); dev_err(xpc_part, "failed while forking discovery thread\n");
/* mark this new thread as a non-starter */ /* mark this new thread as a non-starter */
@@ -1389,7 +1283,6 @@ xpc_init(void)
return -EBUSY; return -EBUSY;
} }
/* set the interface to point at XPC's functions */ /* set the interface to point at XPC's functions */
xpc_set_interface(xpc_initiate_connect, xpc_initiate_disconnect, xpc_set_interface(xpc_initiate_connect, xpc_initiate_disconnect,
xpc_initiate_allocate, xpc_initiate_send, xpc_initiate_allocate, xpc_initiate_send,
@@ -1398,16 +1291,16 @@ xpc_init(void)
return 0; return 0;
} }
module_init(xpc_init);
module_init(xpc_init);
void __exit void __exit
xpc_exit(void) xpc_exit(void)
{ {
xpc_do_exit(xpcUnloading); xpc_do_exit(xpcUnloading);
} }
module_exit(xpc_exit);
module_exit(xpc_exit);
MODULE_AUTHOR("Silicon Graphics, Inc."); MODULE_AUTHOR("Silicon Graphics, Inc.");
MODULE_DESCRIPTION("Cross Partition Communication (XPC) support"); MODULE_DESCRIPTION("Cross Partition Communication (XPC) support");
@@ -1428,4 +1321,3 @@ MODULE_PARM_DESC(xpc_disengage_request_timelimit, "Number of seconds to wait "
module_param(xpc_kdebug_ignore, int, 0); module_param(xpc_kdebug_ignore, int, 0);
MODULE_PARM_DESC(xpc_kdebug_ignore, "Should lack of heartbeat be ignored by " MODULE_PARM_DESC(xpc_kdebug_ignore, "Should lack of heartbeat be ignored by "
"other partitions when dropping into kdebug."); "other partitions when dropping into kdebug.");

View File

@@ -3,10 +3,9 @@
* License. See the file "COPYING" in the main directory of this archive * License. See the file "COPYING" in the main directory of this archive
* for more details. * for more details.
* *
* Copyright (c) 2004-2006 Silicon Graphics, Inc. All Rights Reserved. * Copyright (c) 2004-2008 Silicon Graphics, Inc. All Rights Reserved.
*/ */
/* /*
* Cross Partition Communication (XPC) partition support. * Cross Partition Communication (XPC) partition support.
* *
@@ -16,7 +15,6 @@
* *
*/ */
#include <linux/kernel.h> #include <linux/kernel.h>
#include <linux/sysctl.h> #include <linux/sysctl.h>
#include <linux/cache.h> #include <linux/cache.h>
@@ -28,13 +26,11 @@
#include <asm/sn/sn_sal.h> #include <asm/sn/sn_sal.h>
#include <asm/sn/nodepda.h> #include <asm/sn/nodepda.h>
#include <asm/sn/addrs.h> #include <asm/sn/addrs.h>
#include <asm/sn/xpc.h> #include "xpc.h"
/* XPC is exiting flag */ /* XPC is exiting flag */
int xpc_exiting; int xpc_exiting;
/* SH_IPI_ACCESS shub register value on startup */ /* SH_IPI_ACCESS shub register value on startup */
static u64 xpc_sh1_IPI_access; static u64 xpc_sh1_IPI_access;
static u64 xpc_sh2_IPI_access0; static u64 xpc_sh2_IPI_access0;
@@ -42,11 +38,9 @@ static u64 xpc_sh2_IPI_access1;
static u64 xpc_sh2_IPI_access2; static u64 xpc_sh2_IPI_access2;
static u64 xpc_sh2_IPI_access3; static u64 xpc_sh2_IPI_access3;
/* original protection values for each node */ /* original protection values for each node */
u64 xpc_prot_vec[MAX_NUMNODES]; u64 xpc_prot_vec[MAX_NUMNODES];
/* this partition's reserved page pointers */ /* this partition's reserved page pointers */
struct xpc_rsvd_page *xpc_rsvd_page; struct xpc_rsvd_page *xpc_rsvd_page;
static u64 *xpc_part_nasids; static u64 *xpc_part_nasids;
@@ -57,7 +51,6 @@ struct xpc_vars_part *xpc_vars_part;
static int xp_nasid_mask_bytes; /* actual size in bytes of nasid mask */ static int xp_nasid_mask_bytes; /* actual size in bytes of nasid mask */
static int xp_nasid_mask_words; /* actual size in words of nasid mask */ static int xp_nasid_mask_words; /* actual size in words of nasid mask */
/* /*
* For performance reasons, each entry of xpc_partitions[] is cacheline * For performance reasons, each entry of xpc_partitions[] is cacheline
* aligned. And xpc_partitions[] is padded with an additional entry at the * aligned. And xpc_partitions[] is padded with an additional entry at the
@@ -66,7 +59,6 @@ static int xp_nasid_mask_words; /* actual size in words of nasid mask */
*/ */
struct xpc_partition xpc_partitions[XP_MAX_PARTITIONS + 1]; struct xpc_partition xpc_partitions[XP_MAX_PARTITIONS + 1];
/* /*
* Generic buffer used to store a local copy of portions of a remote * Generic buffer used to store a local copy of portions of a remote
* partition's reserved page (either its header and part_nasids mask, * partition's reserved page (either its header and part_nasids mask,
@@ -75,7 +67,6 @@ struct xpc_partition xpc_partitions[XP_MAX_PARTITIONS + 1];
char *xpc_remote_copy_buffer; char *xpc_remote_copy_buffer;
void *xpc_remote_copy_buffer_base; void *xpc_remote_copy_buffer_base;
/* /*
* Guarantee that the kmalloc'd memory is cacheline aligned. * Guarantee that the kmalloc'd memory is cacheline aligned.
*/ */
@@ -84,23 +75,22 @@ xpc_kmalloc_cacheline_aligned(size_t size, gfp_t flags, void **base)
{ {
/* see if kmalloc will give us cachline aligned memory by default */ /* see if kmalloc will give us cachline aligned memory by default */
*base = kmalloc(size, flags); *base = kmalloc(size, flags);
if (*base == NULL) { if (*base == NULL)
return NULL; return NULL;
}
if ((u64) *base == L1_CACHE_ALIGN((u64) *base)) { if ((u64)*base == L1_CACHE_ALIGN((u64)*base))
return *base; return *base;
}
kfree(*base); kfree(*base);
/* nope, we'll have to do it ourselves */ /* nope, we'll have to do it ourselves */
*base = kmalloc(size + L1_CACHE_BYTES, flags); *base = kmalloc(size + L1_CACHE_BYTES, flags);
if (*base == NULL) { if (*base == NULL)
return NULL; return NULL;
}
return (void *)L1_CACHE_ALIGN((u64)*base); return (void *)L1_CACHE_ALIGN((u64)*base);
} }
/* /*
* Given a nasid, get the physical address of the partition's reserved page * Given a nasid, get the physical address of the partition's reserved page
* for that nasid. This function returns 0 on any error. * for that nasid. This function returns 0 on any error.
@@ -117,7 +107,6 @@ xpc_get_rsvd_page_pa(int nasid)
u64 buf_len = 0; u64 buf_len = 0;
void *buf_base = NULL; void *buf_base = NULL;
while (1) { while (1) {
status = sn_partition_reserved_page_pa(buf, &cookie, &rp_pa, status = sn_partition_reserved_page_pa(buf, &cookie, &rp_pa,
@@ -127,15 +116,15 @@ xpc_get_rsvd_page_pa(int nasid)
"0x%016lx, address=0x%016lx, len=0x%016lx\n", "0x%016lx, address=0x%016lx, len=0x%016lx\n",
status, cookie, rp_pa, len); status, cookie, rp_pa, len);
if (status != SALRET_MORE_PASSES) { if (status != SALRET_MORE_PASSES)
break; break;
}
if (L1_CACHE_ALIGN(len) > buf_len) { if (L1_CACHE_ALIGN(len) > buf_len) {
kfree(buf_base); kfree(buf_base);
buf_len = L1_CACHE_ALIGN(len); buf_len = L1_CACHE_ALIGN(len);
buf = (u64)xpc_kmalloc_cacheline_aligned(buf_len, buf = (u64)xpc_kmalloc_cacheline_aligned(buf_len,
GFP_KERNEL, &buf_base); GFP_KERNEL,
&buf_base);
if (buf_base == NULL) { if (buf_base == NULL) {
dev_err(xpc_part, "unable to kmalloc " dev_err(xpc_part, "unable to kmalloc "
"len=0x%016lx\n", buf_len); "len=0x%016lx\n", buf_len);
@@ -155,14 +144,13 @@ xpc_get_rsvd_page_pa(int nasid)
kfree(buf_base); kfree(buf_base);
if (status != SALRET_OK) { if (status != SALRET_OK)
rp_pa = 0; rp_pa = 0;
}
dev_dbg(xpc_part, "reserved page at phys address 0x%016lx\n", rp_pa); dev_dbg(xpc_part, "reserved page at phys address 0x%016lx\n", rp_pa);
return rp_pa; return rp_pa;
} }
/* /*
* Fill the partition reserved page with the information needed by * Fill the partition reserved page with the information needed by
* other partitions to discover we are alive and establish initial * other partitions to discover we are alive and establish initial
@@ -176,7 +164,6 @@ xpc_rsvd_page_init(void)
u64 rp_pa, nasid_array = 0; u64 rp_pa, nasid_array = 0;
int i, ret; int i, ret;
/* get the local reserved page's address */ /* get the local reserved page's address */
preempt_disable(); preempt_disable();
@@ -222,7 +209,8 @@ xpc_rsvd_page_init(void)
* on subsequent loads of XPC. This AMO page is never freed, and its * on subsequent loads of XPC. This AMO page is never freed, and its
* memory protections are never restricted. * memory protections are never restricted.
*/ */
if ((amos_page = xpc_vars->amos_page) == NULL) { amos_page = xpc_vars->amos_page;
if (amos_page == NULL) {
amos_page = (AMO_t *)TO_AMO(uncached_alloc_page(0)); amos_page = (AMO_t *)TO_AMO(uncached_alloc_page(0));
if (amos_page == NULL) { if (amos_page == NULL) {
dev_err(xpc_part, "can't allocate page of AMOs\n"); dev_err(xpc_part, "can't allocate page of AMOs\n");
@@ -235,7 +223,8 @@ xpc_rsvd_page_init(void)
*/ */
if (!enable_shub_wars_1_1()) { if (!enable_shub_wars_1_1()) {
ret = sn_change_memprotect(ia64_tpa((u64)amos_page), ret = sn_change_memprotect(ia64_tpa((u64)amos_page),
PAGE_SIZE, SN_MEMPROT_ACCESS_CLASS_1, PAGE_SIZE,
SN_MEMPROT_ACCESS_CLASS_1,
&nasid_array); &nasid_array);
if (ret != 0) { if (ret != 0) {
dev_err(xpc_part, "can't change memory " dev_err(xpc_part, "can't change memory "
@@ -270,15 +259,13 @@ xpc_rsvd_page_init(void)
xpc_vars->amos_page_pa = ia64_tpa((u64)amos_page); xpc_vars->amos_page_pa = ia64_tpa((u64)amos_page);
xpc_vars->amos_page = amos_page; /* save for next load of XPC */ xpc_vars->amos_page = amos_page; /* save for next load of XPC */
/* clear xpc_vars_part */ /* clear xpc_vars_part */
memset((u64 *)xpc_vars_part, 0, sizeof(struct xpc_vars_part) * memset((u64 *)xpc_vars_part, 0, sizeof(struct xpc_vars_part) *
XP_MAX_PARTITIONS); XP_MAX_PARTITIONS);
/* initialize the activate IRQ related AMO variables */ /* initialize the activate IRQ related AMO variables */
for (i = 0; i < xp_nasid_mask_words; i++) { for (i = 0; i < xp_nasid_mask_words; i++)
(void)xpc_IPI_init(XPC_ACTIVATE_IRQ_AMOS + i); (void)xpc_IPI_init(XPC_ACTIVATE_IRQ_AMOS + i);
}
/* initialize the engaged remote partitions related AMO variables */ /* initialize the engaged remote partitions related AMO variables */
(void)xpc_IPI_init(XPC_ENGAGED_PARTITIONS_AMO); (void)xpc_IPI_init(XPC_ENGAGED_PARTITIONS_AMO);
@@ -296,7 +283,6 @@ xpc_rsvd_page_init(void)
return rp; return rp;
} }
/* /*
* Change protections to allow IPI operations (and AMO operations on * Change protections to allow IPI operations (and AMO operations on
* Shub 1.1 systems). * Shub 1.1 systems).
@@ -307,8 +293,7 @@ xpc_allow_IPI_ops(void)
int node; int node;
int nasid; int nasid;
/* >>> Change SH_IPI_ACCESS code to use SAL call once it is available */
// >>> Change SH_IPI_ACCESS code to use SAL call once it is available.
if (is_shub2()) { if (is_shub2()) {
xpc_sh2_IPI_access0 = xpc_sh2_IPI_access0 =
@@ -348,12 +333,15 @@ xpc_allow_IPI_ops(void)
if (enable_shub_wars_1_1()) { if (enable_shub_wars_1_1()) {
/* open up everything */ /* open up everything */
xpc_prot_vec[node] = (u64)HUB_L((u64 *) xpc_prot_vec[node] = (u64)HUB_L((u64 *)
GLOBAL_MMR_ADDR(nasid, GLOBAL_MMR_ADDR
(nasid,
SH1_MD_DQLP_MMR_DIR_PRIVEC0)); SH1_MD_DQLP_MMR_DIR_PRIVEC0));
HUB_S((u64 *) GLOBAL_MMR_ADDR(nasid, HUB_S((u64 *)
GLOBAL_MMR_ADDR(nasid,
SH1_MD_DQLP_MMR_DIR_PRIVEC0), SH1_MD_DQLP_MMR_DIR_PRIVEC0),
-1UL); -1UL);
HUB_S((u64 *) GLOBAL_MMR_ADDR(nasid, HUB_S((u64 *)
GLOBAL_MMR_ADDR(nasid,
SH1_MD_DQRP_MMR_DIR_PRIVEC0), SH1_MD_DQRP_MMR_DIR_PRIVEC0),
-1UL); -1UL);
} }
@@ -361,7 +349,6 @@ xpc_allow_IPI_ops(void)
} }
} }
/* /*
* Restrict protections to disallow IPI operations (and AMO operations on * Restrict protections to disallow IPI operations (and AMO operations on
* Shub 1.1 systems). * Shub 1.1 systems).
@@ -372,8 +359,7 @@ xpc_restrict_IPI_ops(void)
int node; int node;
int nasid; int nasid;
/* >>> Change SH_IPI_ACCESS code to use SAL call once it is available */
// >>> Change SH_IPI_ACCESS code to use SAL call once it is available.
if (is_shub2()) { if (is_shub2()) {
@@ -408,7 +394,6 @@ xpc_restrict_IPI_ops(void)
} }
} }
/* /*
* At periodic intervals, scan through all active partitions and ensure * At periodic intervals, scan through all active partitions and ensure
* their heartbeat is still active. If not, the partition is deactivated. * their heartbeat is still active. If not, the partition is deactivated.
@@ -421,18 +406,15 @@ xpc_check_remote_hb(void)
partid_t partid; partid_t partid;
bte_result_t bres; bte_result_t bres;
remote_vars = (struct xpc_vars *)xpc_remote_copy_buffer; remote_vars = (struct xpc_vars *)xpc_remote_copy_buffer;
for (partid = 1; partid < XP_MAX_PARTITIONS; partid++) { for (partid = 1; partid < XP_MAX_PARTITIONS; partid++) {
if (xpc_exiting) { if (xpc_exiting)
break; break;
}
if (partid == sn_partition_id) { if (partid == sn_partition_id)
continue; continue;
}
part = &xpc_partitions[partid]; part = &xpc_partitions[partid];
@@ -470,7 +452,6 @@ xpc_check_remote_hb(void)
} }
} }
/* /*
* Get a copy of a portion of the remote partition's rsvd page. * Get a copy of a portion of the remote partition's rsvd page.
* *
@@ -484,33 +465,25 @@ xpc_get_remote_rp(int nasid, u64 *discovered_nasids,
{ {
int bres, i; int bres, i;
/* get the reserved page's physical address */ /* get the reserved page's physical address */
*remote_rp_pa = xpc_get_rsvd_page_pa(nasid); *remote_rp_pa = xpc_get_rsvd_page_pa(nasid);
if (*remote_rp_pa == 0) { if (*remote_rp_pa == 0)
return xpcNoRsvdPageAddr; return xpcNoRsvdPageAddr;
}
/* pull over the reserved page header and part_nasids mask */ /* pull over the reserved page header and part_nasids mask */
bres = xp_bte_copy(*remote_rp_pa, (u64)remote_rp, bres = xp_bte_copy(*remote_rp_pa, (u64)remote_rp,
XPC_RP_HEADER_SIZE + xp_nasid_mask_bytes, XPC_RP_HEADER_SIZE + xp_nasid_mask_bytes,
(BTE_NOTIFY | BTE_WACQUIRE), NULL); (BTE_NOTIFY | BTE_WACQUIRE), NULL);
if (bres != BTE_SUCCESS) { if (bres != BTE_SUCCESS)
return xpc_map_bte_errors(bres); return xpc_map_bte_errors(bres);
}
if (discovered_nasids != NULL) { if (discovered_nasids != NULL) {
u64 *remote_part_nasids = XPC_RP_PART_NASIDS(remote_rp); u64 *remote_part_nasids = XPC_RP_PART_NASIDS(remote_rp);
for (i = 0; i < xp_nasid_mask_words; i++)
for (i = 0; i < xp_nasid_mask_words; i++) {
discovered_nasids[i] |= remote_part_nasids[i]; discovered_nasids[i] |= remote_part_nasids[i];
} }
}
/* check that the partid is for another partition */ /* check that the partid is for another partition */
@@ -519,10 +492,8 @@ xpc_get_remote_rp(int nasid, u64 *discovered_nasids,
return xpcInvalidPartid; return xpcInvalidPartid;
} }
if (remote_rp->partid == sn_partition_id) { if (remote_rp->partid == sn_partition_id)
return xpcLocalPartid; return xpcLocalPartid;
}
if (XPC_VERSION_MAJOR(remote_rp->version) != if (XPC_VERSION_MAJOR(remote_rp->version) !=
XPC_VERSION_MAJOR(XPC_RP_VERSION)) { XPC_VERSION_MAJOR(XPC_RP_VERSION)) {
@@ -532,7 +503,6 @@ xpc_get_remote_rp(int nasid, u64 *discovered_nasids,
return xpcSuccess; return xpcSuccess;
} }
/* /*
* Get a copy of the remote partition's XPC variables from the reserved page. * Get a copy of the remote partition's XPC variables from the reserved page.
* *
@@ -544,17 +514,14 @@ xpc_get_remote_vars(u64 remote_vars_pa, struct xpc_vars *remote_vars)
{ {
int bres; int bres;
if (remote_vars_pa == 0)
if (remote_vars_pa == 0) {
return xpcVarsNotSet; return xpcVarsNotSet;
}
/* pull over the cross partition variables */ /* pull over the cross partition variables */
bres = xp_bte_copy(remote_vars_pa, (u64)remote_vars, XPC_RP_VARS_SIZE, bres = xp_bte_copy(remote_vars_pa, (u64)remote_vars, XPC_RP_VARS_SIZE,
(BTE_NOTIFY | BTE_WACQUIRE), NULL); (BTE_NOTIFY | BTE_WACQUIRE), NULL);
if (bres != BTE_SUCCESS) { if (bres != BTE_SUCCESS)
return xpc_map_bte_errors(bres); return xpc_map_bte_errors(bres);
}
if (XPC_VERSION_MAJOR(remote_vars->version) != if (XPC_VERSION_MAJOR(remote_vars->version) !=
XPC_VERSION_MAJOR(XPC_V_VERSION)) { XPC_VERSION_MAJOR(XPC_V_VERSION)) {
@@ -564,7 +531,6 @@ xpc_get_remote_vars(u64 remote_vars_pa, struct xpc_vars *remote_vars)
return xpcSuccess; return xpcSuccess;
} }
/* /*
* Update the remote partition's info. * Update the remote partition's info.
*/ */
@@ -613,7 +579,6 @@ xpc_update_partition_info(struct xpc_partition *part, u8 remote_rp_version,
part->remote_vars_version); part->remote_vars_version);
} }
/* /*
* Prior code has determined the nasid which generated an IPI. Inspect * Prior code has determined the nasid which generated an IPI. Inspect
* that nasid to determine if its partition needs to be activated or * that nasid to determine if its partition needs to be activated or
@@ -643,7 +608,6 @@ xpc_identify_act_IRQ_req(int nasid)
struct xpc_partition *part; struct xpc_partition *part;
enum xpc_retval ret; enum xpc_retval ret;
/* pull over the reserved page structure */ /* pull over the reserved page structure */
remote_rp = (struct xpc_rsvd_page *)xpc_remote_copy_buffer; remote_rp = (struct xpc_rsvd_page *)xpc_remote_copy_buffer;
@@ -657,13 +621,12 @@ xpc_identify_act_IRQ_req(int nasid)
remote_vars_pa = remote_rp->vars_pa; remote_vars_pa = remote_rp->vars_pa;
remote_rp_version = remote_rp->version; remote_rp_version = remote_rp->version;
if (XPC_SUPPORTS_RP_STAMP(remote_rp_version)) { if (XPC_SUPPORTS_RP_STAMP(remote_rp_version))
remote_rp_stamp = remote_rp->stamp; remote_rp_stamp = remote_rp->stamp;
}
partid = remote_rp->partid; partid = remote_rp->partid;
part = &xpc_partitions[partid]; part = &xpc_partitions[partid];
/* pull over the cross partition variables */ /* pull over the cross partition variables */
remote_vars = (struct xpc_vars *)xpc_remote_copy_buffer; remote_vars = (struct xpc_vars *)xpc_remote_copy_buffer;
@@ -678,7 +641,6 @@ xpc_identify_act_IRQ_req(int nasid)
return; return;
} }
part->act_IRQ_rcvd++; part->act_IRQ_rcvd++;
dev_dbg(xpc_part, "partid for nasid %d is %d; IRQs = %d; HB = " dev_dbg(xpc_part, "partid for nasid %d is %d; IRQs = %d; HB = "
@@ -722,8 +684,7 @@ xpc_identify_act_IRQ_req(int nasid)
/* see if the other side rebooted */ /* see if the other side rebooted */
if (part->remote_amos_page_pa == if (part->remote_amos_page_pa ==
remote_vars->amos_page_pa && remote_vars->amos_page_pa &&
xpc_hb_allowed(sn_partition_id, xpc_hb_allowed(sn_partition_id, remote_vars)) {
remote_vars)) {
/* doesn't look that way, so ignore the IPI */ /* doesn't look that way, so ignore the IPI */
return; return;
} }
@@ -778,8 +739,9 @@ xpc_identify_act_IRQ_req(int nasid)
partid)); partid));
xpc_update_partition_info(part, remote_rp_version, xpc_update_partition_info(part, remote_rp_version,
&remote_rp_stamp, remote_rp_pa, &remote_rp_stamp,
remote_vars_pa, remote_vars); remote_rp_pa, remote_vars_pa,
remote_vars);
reactivate = 1; reactivate = 1;
} }
} }
@@ -800,7 +762,6 @@ xpc_identify_act_IRQ_req(int nasid)
} }
} }
/* /*
* Loop through the activation AMO variables and process any bits * Loop through the activation AMO variables and process any bits
* which are set. Each bit indicates a nasid sending a partition * which are set. Each bit indicates a nasid sending a partition
@@ -817,16 +778,13 @@ xpc_identify_act_IRQ_sender(void)
int n_IRQs_detected = 0; int n_IRQs_detected = 0;
AMO_t *act_amos; AMO_t *act_amos;
act_amos = xpc_vars->amos_page + XPC_ACTIVATE_IRQ_AMOS; act_amos = xpc_vars->amos_page + XPC_ACTIVATE_IRQ_AMOS;
/* scan through act AMO variable looking for non-zero entries */ /* scan through act AMO variable looking for non-zero entries */
for (word = 0; word < xp_nasid_mask_words; word++) { for (word = 0; word < xp_nasid_mask_words; word++) {
if (xpc_exiting) { if (xpc_exiting)
break; break;
}
nasid_mask = xpc_IPI_receive(&act_amos[word]); nasid_mask = xpc_IPI_receive(&act_amos[word]);
if (nasid_mask == 0) { if (nasid_mask == 0) {
@@ -837,7 +795,6 @@ xpc_identify_act_IRQ_sender(void)
dev_dbg(xpc_part, "AMO[%d] gave back 0x%lx\n", word, dev_dbg(xpc_part, "AMO[%d] gave back 0x%lx\n", word,
nasid_mask); nasid_mask);
/* /*
* If this nasid has been added to the machine since * If this nasid has been added to the machine since
* our partition was reset, this will retain the * our partition was reset, this will retain the
@@ -846,7 +803,6 @@ xpc_identify_act_IRQ_sender(void)
*/ */
xpc_mach_nasids[word] |= nasid_mask; xpc_mach_nasids[word] |= nasid_mask;
/* locate the nasid(s) which sent interrupts */ /* locate the nasid(s) which sent interrupts */
for (bit = 0; bit < (8 * sizeof(u64)); bit++) { for (bit = 0; bit < (8 * sizeof(u64)); bit++) {
@@ -862,7 +818,6 @@ xpc_identify_act_IRQ_sender(void)
return n_IRQs_detected; return n_IRQs_detected;
} }
/* /*
* See if the other side has responded to a partition disengage request * See if the other side has responded to a partition disengage request
* from us. * from us.
@@ -873,11 +828,11 @@ xpc_partition_disengaged(struct xpc_partition *part)
partid_t partid = XPC_PARTID(part); partid_t partid = XPC_PARTID(part);
int disengaged; int disengaged;
disengaged = (xpc_partition_engaged(1UL << partid) == 0); disengaged = (xpc_partition_engaged(1UL << partid) == 0);
if (part->disengage_request_timeout) { if (part->disengage_request_timeout) {
if (!disengaged) { if (!disengaged) {
if (time_before(jiffies, part->disengage_request_timeout)) { if (time_before(jiffies,
part->disengage_request_timeout)) {
/* timelimit hasn't been reached yet */ /* timelimit hasn't been reached yet */
return 0; return 0;
} }
@@ -903,18 +858,15 @@ xpc_partition_disengaged(struct xpc_partition *part)
DBUG_ON(part->act_state != XPC_P_DEACTIVATING && DBUG_ON(part->act_state != XPC_P_DEACTIVATING &&
part->act_state != XPC_P_INACTIVE); part->act_state != XPC_P_INACTIVE);
if (part->act_state != XPC_P_INACTIVE) { if (part->act_state != XPC_P_INACTIVE)
xpc_wakeup_channel_mgr(part); xpc_wakeup_channel_mgr(part);
}
if (XPC_SUPPORTS_DISENGAGE_REQUEST(part->remote_vars_version)) { if (XPC_SUPPORTS_DISENGAGE_REQUEST(part->remote_vars_version))
xpc_cancel_partition_disengage_request(part); xpc_cancel_partition_disengage_request(part);
} }
}
return disengaged; return disengaged;
} }
/* /*
* Mark specified partition as active. * Mark specified partition as active.
*/ */
@@ -924,7 +876,6 @@ xpc_mark_partition_active(struct xpc_partition *part)
unsigned long irq_flags; unsigned long irq_flags;
enum xpc_retval ret; enum xpc_retval ret;
dev_dbg(xpc_part, "setting partition %d to ACTIVE\n", XPC_PARTID(part)); dev_dbg(xpc_part, "setting partition %d to ACTIVE\n", XPC_PARTID(part));
spin_lock_irqsave(&part->act_lock, irq_flags); spin_lock_irqsave(&part->act_lock, irq_flags);
@@ -940,7 +891,6 @@ xpc_mark_partition_active(struct xpc_partition *part)
return ret; return ret;
} }
/* /*
* Notify XPC that the partition is down. * Notify XPC that the partition is down.
*/ */
@@ -950,7 +900,6 @@ xpc_deactivate_partition(const int line, struct xpc_partition *part,
{ {
unsigned long irq_flags; unsigned long irq_flags;
spin_lock_irqsave(&part->act_lock, irq_flags); spin_lock_irqsave(&part->act_lock, irq_flags);
if (part->act_state == XPC_P_INACTIVE) { if (part->act_state == XPC_P_INACTIVE) {
@@ -994,7 +943,6 @@ xpc_deactivate_partition(const int line, struct xpc_partition *part,
xpc_partition_going_down(part, reason); xpc_partition_going_down(part, reason);
} }
/* /*
* Mark specified partition as inactive. * Mark specified partition as inactive.
*/ */
@@ -1003,7 +951,6 @@ xpc_mark_partition_inactive(struct xpc_partition *part)
{ {
unsigned long irq_flags; unsigned long irq_flags;
dev_dbg(xpc_part, "setting partition %d to INACTIVE\n", dev_dbg(xpc_part, "setting partition %d to INACTIVE\n",
XPC_PARTID(part)); XPC_PARTID(part));
@@ -1013,7 +960,6 @@ xpc_mark_partition_inactive(struct xpc_partition *part)
part->remote_rp_pa = 0; part->remote_rp_pa = 0;
} }
/* /*
* SAL has provided a partition and machine mask. The partition mask * SAL has provided a partition and machine mask. The partition mask
* contains a bit for each even nasid in our partition. The machine * contains a bit for each even nasid in our partition. The machine
@@ -1041,15 +987,13 @@ xpc_discovery(void)
u64 *discovered_nasids; u64 *discovered_nasids;
enum xpc_retval ret; enum xpc_retval ret;
remote_rp = xpc_kmalloc_cacheline_aligned(XPC_RP_HEADER_SIZE + remote_rp = xpc_kmalloc_cacheline_aligned(XPC_RP_HEADER_SIZE +
xp_nasid_mask_bytes, xp_nasid_mask_bytes,
GFP_KERNEL, &remote_rp_base); GFP_KERNEL, &remote_rp_base);
if (remote_rp == NULL) { if (remote_rp == NULL)
return; return;
}
remote_vars = (struct xpc_vars *) remote_rp;
remote_vars = (struct xpc_vars *)remote_rp;
discovered_nasids = kzalloc(sizeof(u64) * xp_nasid_mask_words, discovered_nasids = kzalloc(sizeof(u64) * xp_nasid_mask_words,
GFP_KERNEL); GFP_KERNEL);
@@ -1081,23 +1025,19 @@ xpc_discovery(void)
for (region = 0; region < max_regions; region++) { for (region = 0; region < max_regions; region++) {
if ((volatile int) xpc_exiting) { if (xpc_exiting)
break; break;
}
dev_dbg(xpc_part, "searching region %d\n", region); dev_dbg(xpc_part, "searching region %d\n", region);
for (nasid = (region * region_size * 2); for (nasid = (region * region_size * 2);
nasid < ((region + 1) * region_size * 2); nasid < ((region + 1) * region_size * 2); nasid += 2) {
nasid += 2) {
if ((volatile int) xpc_exiting) { if (xpc_exiting)
break; break;
}
dev_dbg(xpc_part, "checking nasid %d\n", nasid); dev_dbg(xpc_part, "checking nasid %d\n", nasid);
if (XPC_NASID_IN_ARRAY(nasid, xpc_part_nasids)) { if (XPC_NASID_IN_ARRAY(nasid, xpc_part_nasids)) {
dev_dbg(xpc_part, "PROM indicates Nasid %d is " dev_dbg(xpc_part, "PROM indicates Nasid %d is "
"part of the local partition; skipping " "part of the local partition; skipping "
@@ -1119,7 +1059,6 @@ xpc_discovery(void)
continue; continue;
} }
/* pull over the reserved page structure */ /* pull over the reserved page structure */
ret = xpc_get_remote_rp(nasid, discovered_nasids, ret = xpc_get_remote_rp(nasid, discovered_nasids,
@@ -1129,9 +1068,9 @@ xpc_discovery(void)
"from nasid %d, reason=%d\n", nasid, "from nasid %d, reason=%d\n", nasid,
ret); ret);
if (ret == xpcLocalPartid) { if (ret == xpcLocalPartid)
break; break;
}
continue; continue;
} }
@@ -1140,7 +1079,6 @@ xpc_discovery(void)
partid = remote_rp->partid; partid = remote_rp->partid;
part = &xpc_partitions[partid]; part = &xpc_partitions[partid];
/* pull over the cross partition variables */ /* pull over the cross partition variables */
ret = xpc_get_remote_vars(remote_vars_pa, remote_vars); ret = xpc_get_remote_vars(remote_vars_pa, remote_vars);
@@ -1171,10 +1109,10 @@ xpc_discovery(void)
* get the same page for remote_act_amos_pa after * get the same page for remote_act_amos_pa after
* module reloads and system reboots. * module reloads and system reboots.
*/ */
if (sn_register_xp_addr_region( if (sn_register_xp_addr_region
remote_vars->amos_page_pa, (remote_vars->amos_page_pa, PAGE_SIZE, 1) < 0) {
PAGE_SIZE, 1) < 0) { dev_dbg(xpc_part,
dev_dbg(xpc_part, "partition %d failed to " "partition %d failed to "
"register xp_addr region 0x%016lx\n", "register xp_addr region 0x%016lx\n",
partid, remote_vars->amos_page_pa); partid, remote_vars->amos_page_pa);
@@ -1209,7 +1147,6 @@ xpc_discovery(void)
kfree(remote_rp_base); kfree(remote_rp_base);
} }
/* /*
* Given a partid, get the nasids owned by that partition from the * Given a partid, get the nasids owned by that partition from the
* remote partition's reserved page. * remote partition's reserved page.
@@ -1221,19 +1158,17 @@ xpc_initiate_partid_to_nasids(partid_t partid, void *nasid_mask)
u64 part_nasid_pa; u64 part_nasid_pa;
int bte_res; int bte_res;
part = &xpc_partitions[partid]; part = &xpc_partitions[partid];
if (part->remote_rp_pa == 0) { if (part->remote_rp_pa == 0)
return xpcPartitionDown; return xpcPartitionDown;
}
memset(nasid_mask, 0, XP_NASID_MASK_BYTES); memset(nasid_mask, 0, XP_NASID_MASK_BYTES);
part_nasid_pa = (u64)XPC_RP_PART_NASIDS(part->remote_rp_pa); part_nasid_pa = (u64)XPC_RP_PART_NASIDS(part->remote_rp_pa);
bte_res = xp_bte_copy(part_nasid_pa, (u64)nasid_mask, bte_res = xp_bte_copy(part_nasid_pa, (u64)nasid_mask,
xp_nasid_mask_bytes, (BTE_NOTIFY | BTE_WACQUIRE), NULL); xp_nasid_mask_bytes, (BTE_NOTIFY | BTE_WACQUIRE),
NULL);
return xpc_map_bte_errors(bte_res); return xpc_map_bte_errors(bte_res);
} }

View File

@@ -3,10 +3,9 @@
* License. See the file "COPYING" in the main directory of this archive * License. See the file "COPYING" in the main directory of this archive
* for more details. * for more details.
* *
* Copyright (C) 1999,2001-2005 Silicon Graphics, Inc. All rights reserved. * Copyright (C) 1999-2008 Silicon Graphics, Inc. All rights reserved.
*/ */
/* /*
* Cross Partition Network Interface (XPNET) support * Cross Partition Network Interface (XPNET) support
* *
@@ -21,8 +20,8 @@
* *
*/ */
#include <linux/module.h> #include <linux/module.h>
#include <linux/types.h>
#include <linux/kernel.h> #include <linux/kernel.h>
#include <linux/init.h> #include <linux/init.h>
#include <linux/ioport.h> #include <linux/ioport.h>
@@ -36,10 +35,8 @@
#include <asm/sn/bte.h> #include <asm/sn/bte.h>
#include <asm/sn/io.h> #include <asm/sn/io.h>
#include <asm/sn/sn_sal.h> #include <asm/sn/sn_sal.h>
#include <asm/types.h>
#include <asm/atomic.h> #include <asm/atomic.h>
#include <asm/sn/xp.h> #include "xp.h"
/* /*
* The message payload transferred by XPC. * The message payload transferred by XPC.
@@ -79,7 +76,6 @@ struct xpnet_message {
#define XPNET_MSG_ALIGNED_SIZE (L1_CACHE_ALIGN(XPNET_MSG_SIZE)) #define XPNET_MSG_ALIGNED_SIZE (L1_CACHE_ALIGN(XPNET_MSG_SIZE))
#define XPNET_MSG_NENTRIES (PAGE_SIZE / XPNET_MSG_ALIGNED_SIZE) #define XPNET_MSG_NENTRIES (PAGE_SIZE / XPNET_MSG_ALIGNED_SIZE)
#define XPNET_MAX_KTHREADS (XPNET_MSG_NENTRIES + 1) #define XPNET_MAX_KTHREADS (XPNET_MSG_NENTRIES + 1)
#define XPNET_MAX_IDLE_KTHREADS (XPNET_MSG_NENTRIES + 1) #define XPNET_MAX_IDLE_KTHREADS (XPNET_MSG_NENTRIES + 1)
@@ -101,7 +97,6 @@ struct xpnet_message {
#define XPNET_DEVICE_NAME "xp0" #define XPNET_DEVICE_NAME "xp0"
/* /*
* When messages are queued with xpc_send_notify, a kmalloc'd buffer * When messages are queued with xpc_send_notify, a kmalloc'd buffer
* of the following type is passed as a notification cookie. When the * of the following type is passed as a notification cookie. When the
@@ -145,7 +140,6 @@ static DEFINE_SPINLOCK(xpnet_broadcast_lock);
/* 32KB has been determined to be the ideal */ /* 32KB has been determined to be the ideal */
#define XPNET_DEF_MTU (0x8000UL) #define XPNET_DEF_MTU (0x8000UL)
/* /*
* The partition id is encapsulated in the MAC address. The following * The partition id is encapsulated in the MAC address. The following
* define locates the octet the partid is in. * define locates the octet the partid is in.
@@ -153,7 +147,6 @@ static DEFINE_SPINLOCK(xpnet_broadcast_lock);
#define XPNET_PARTID_OCTET 1 #define XPNET_PARTID_OCTET 1
#define XPNET_LICENSE_OCTET 2 #define XPNET_LICENSE_OCTET 2
/* /*
* Define the XPNET debug device structure that is to be used with dev_dbg(), * Define the XPNET debug device structure that is to be used with dev_dbg(),
* dev_err(), dev_warn(), and dev_info(). * dev_err(), dev_warn(), and dev_info().
@@ -180,7 +173,6 @@ xpnet_receive(partid_t partid, int channel, struct xpnet_message *msg)
struct xpnet_dev_private *priv = struct xpnet_dev_private *priv =
(struct xpnet_dev_private *)xpnet_device->priv; (struct xpnet_dev_private *)xpnet_device->priv;
if (!XPNET_VALID_MSG(msg)) { if (!XPNET_VALID_MSG(msg)) {
/* /*
* Packet with a different XPC version. Ignore. * Packet with a different XPC version. Ignore.
@@ -194,7 +186,6 @@ xpnet_receive(partid_t partid, int channel, struct xpnet_message *msg)
dev_dbg(xpnet, "received 0x%lx, %d, %d, %d\n", msg->buf_pa, msg->size, dev_dbg(xpnet, "received 0x%lx, %d, %d, %d\n", msg->buf_pa, msg->size,
msg->leadin_ignore, msg->tailout_ignore); msg->leadin_ignore, msg->tailout_ignore);
/* reserve an extra cache line */ /* reserve an extra cache line */
skb = dev_alloc_skb(msg->size + L1_CACHE_BYTES); skb = dev_alloc_skb(msg->size + L1_CACHE_BYTES);
if (!skb) { if (!skb) {
@@ -232,7 +223,8 @@ xpnet_receive(partid_t partid, int channel, struct xpnet_message *msg)
"%lu)\n", skb->data, &msg->data, "%lu)\n", skb->data, &msg->data,
(size_t)msg->embedded_bytes); (size_t)msg->embedded_bytes);
skb_copy_to_linear_data(skb, &msg->data, (size_t)msg->embedded_bytes); skb_copy_to_linear_data(skb, &msg->data,
(size_t)msg->embedded_bytes);
} else { } else {
dev_dbg(xpnet, "transferring buffer to the skb->data area;\n\t" dev_dbg(xpnet, "transferring buffer to the skb->data area;\n\t"
"bte_copy(0x%p, 0x%p, %hu)\n", (void *)msg->buf_pa, "bte_copy(0x%p, 0x%p, %hu)\n", (void *)msg->buf_pa,
@@ -244,9 +236,11 @@ xpnet_receive(partid_t partid, int channel, struct xpnet_message *msg)
msg->size, (BTE_NOTIFY | BTE_WACQUIRE), NULL); msg->size, (BTE_NOTIFY | BTE_WACQUIRE), NULL);
if (bret != BTE_SUCCESS) { if (bret != BTE_SUCCESS) {
// >>> Need better way of cleaning skb. Currently skb /*
// >>> appears in_use and we can't just call * >>> Need better way of cleaning skb. Currently skb
// >>> dev_kfree_skb. * >>> appears in_use and we can't just call
* >>> dev_kfree_skb.
*/
dev_err(xpnet, "bte_copy(0x%p, 0x%p, 0x%hx) returned " dev_err(xpnet, "bte_copy(0x%p, 0x%p, 0x%hx) returned "
"error=0x%x\n", (void *)msg->buf_pa, "error=0x%x\n", (void *)msg->buf_pa,
(void *)__pa((u64)skb->data & (void *)__pa((u64)skb->data &
@@ -275,7 +269,6 @@ xpnet_receive(partid_t partid, int channel, struct xpnet_message *msg)
(void *)skb->head, (void *)skb->data, skb_tail_pointer(skb), (void *)skb->head, (void *)skb->data, skb_tail_pointer(skb),
skb_end_pointer(skb), skb->len); skb_end_pointer(skb), skb->len);
xpnet_device->last_rx = jiffies; xpnet_device->last_rx = jiffies;
priv->stats.rx_packets++; priv->stats.rx_packets++;
priv->stats.rx_bytes += skb->len + ETH_HLEN; priv->stats.rx_bytes += skb->len + ETH_HLEN;
@@ -284,7 +277,6 @@ xpnet_receive(partid_t partid, int channel, struct xpnet_message *msg)
xpc_received(partid, channel, (void *)msg); xpc_received(partid, channel, (void *)msg);
} }
/* /*
* This is the handler which XPC calls during any sort of change in * This is the handler which XPC calls during any sort of change in
* state or message reception on a connection. * state or message reception on a connection.
@@ -295,7 +287,6 @@ xpnet_connection_activity(enum xpc_retval reason, partid_t partid, int channel,
{ {
long bp; long bp;
DBUG_ON(partid <= 0 || partid >= XP_MAX_PARTITIONS); DBUG_ON(partid <= 0 || partid >= XP_MAX_PARTITIONS);
DBUG_ON(channel != XPC_NET_CHANNEL); DBUG_ON(channel != XPC_NET_CHANNEL);
@@ -325,9 +316,8 @@ xpnet_connection_activity(enum xpc_retval reason, partid_t partid, int channel,
bp = xpnet_broadcast_partitions; bp = xpnet_broadcast_partitions;
spin_unlock_bh(&xpnet_broadcast_lock); spin_unlock_bh(&xpnet_broadcast_lock);
if (bp == 0) { if (bp == 0)
netif_carrier_off(xpnet_device); netif_carrier_off(xpnet_device);
}
dev_dbg(xpnet, "%s disconnected from partition %d; " dev_dbg(xpnet, "%s disconnected from partition %d; "
"xpnet_broadcast_partitions=0x%lx\n", "xpnet_broadcast_partitions=0x%lx\n",
@@ -337,13 +327,11 @@ xpnet_connection_activity(enum xpc_retval reason, partid_t partid, int channel,
} }
} }
static int static int
xpnet_dev_open(struct net_device *dev) xpnet_dev_open(struct net_device *dev)
{ {
enum xpc_retval ret; enum xpc_retval ret;
dev_dbg(xpnet, "calling xpc_connect(%d, 0x%p, NULL, %ld, %ld, %ld, " dev_dbg(xpnet, "calling xpc_connect(%d, 0x%p, NULL, %ld, %ld, %ld, "
"%ld)\n", XPC_NET_CHANNEL, xpnet_connection_activity, "%ld)\n", XPC_NET_CHANNEL, xpnet_connection_activity,
XPNET_MSG_SIZE, XPNET_MSG_NENTRIES, XPNET_MAX_KTHREADS, XPNET_MSG_SIZE, XPNET_MSG_NENTRIES, XPNET_MAX_KTHREADS,
@@ -364,7 +352,6 @@ xpnet_dev_open(struct net_device *dev)
return 0; return 0;
} }
static int static int
xpnet_dev_stop(struct net_device *dev) xpnet_dev_stop(struct net_device *dev)
{ {
@@ -375,7 +362,6 @@ xpnet_dev_stop(struct net_device *dev)
return 0; return 0;
} }
static int static int
xpnet_dev_change_mtu(struct net_device *dev, int new_mtu) xpnet_dev_change_mtu(struct net_device *dev, int new_mtu)
{ {
@@ -392,7 +378,6 @@ xpnet_dev_change_mtu(struct net_device *dev, int new_mtu)
return 0; return 0;
} }
/* /*
* Required for the net_device structure. * Required for the net_device structure.
*/ */
@@ -402,7 +387,6 @@ xpnet_dev_set_config(struct net_device *dev, struct ifmap *new_map)
return 0; return 0;
} }
/* /*
* Return statistics to the caller. * Return statistics to the caller.
*/ */
@@ -411,13 +395,11 @@ xpnet_dev_get_stats(struct net_device *dev)
{ {
struct xpnet_dev_private *priv; struct xpnet_dev_private *priv;
priv = (struct xpnet_dev_private *)dev->priv; priv = (struct xpnet_dev_private *)dev->priv;
return &priv->stats; return &priv->stats;
} }
/* /*
* Notification that the other end has received the message and * Notification that the other end has received the message and
* DMA'd the skb information. At this point, they are done with * DMA'd the skb information. At this point, they are done with
@@ -428,9 +410,7 @@ static void
xpnet_send_completed(enum xpc_retval reason, partid_t partid, int channel, xpnet_send_completed(enum xpc_retval reason, partid_t partid, int channel,
void *__qm) void *__qm)
{ {
struct xpnet_pending_msg *queued_msg = struct xpnet_pending_msg *queued_msg = (struct xpnet_pending_msg *)__qm;
(struct xpnet_pending_msg *) __qm;
DBUG_ON(queued_msg == NULL); DBUG_ON(queued_msg == NULL);
@@ -446,7 +426,6 @@ xpnet_send_completed(enum xpc_retval reason, partid_t partid, int channel,
} }
} }
/* /*
* Network layer has formatted a packet (skb) and is ready to place it * Network layer has formatted a packet (skb) and is ready to place it
* "on the wire". Prepare and send an xpnet_message to all partitions * "on the wire". Prepare and send an xpnet_message to all partitions
@@ -469,16 +448,13 @@ xpnet_dev_hard_start_xmit(struct sk_buff *skb, struct net_device *dev)
struct xpnet_dev_private *priv; struct xpnet_dev_private *priv;
u16 embedded_bytes; u16 embedded_bytes;
priv = (struct xpnet_dev_private *)dev->priv; priv = (struct xpnet_dev_private *)dev->priv;
dev_dbg(xpnet, ">skb->head=0x%p skb->data=0x%p skb->tail=0x%p " dev_dbg(xpnet, ">skb->head=0x%p skb->data=0x%p skb->tail=0x%p "
"skb->end=0x%p skb->len=%d\n", (void *)skb->head, "skb->end=0x%p skb->len=%d\n", (void *)skb->head,
(void *)skb->data, skb_tail_pointer(skb), skb_end_pointer(skb), (void *)skb->data, skb_tail_pointer(skb), skb_end_pointer(skb),
skb->len); skb->len);
/* /*
* The xpnet_pending_msg tracks how many outstanding * The xpnet_pending_msg tracks how many outstanding
* xpc_send_notifies are relying on this skb. When none * xpc_send_notifies are relying on this skb. When none
@@ -494,7 +470,6 @@ xpnet_dev_hard_start_xmit(struct sk_buff *skb, struct net_device *dev)
return -ENOMEM; return -ENOMEM;
} }
/* get the beginning of the first cacheline and end of last */ /* get the beginning of the first cacheline and end of last */
start_addr = ((u64)skb->data & ~(L1_CACHE_BYTES - 1)); start_addr = ((u64)skb->data & ~(L1_CACHE_BYTES - 1));
end_addr = L1_CACHE_ALIGN((u64)skb_tail_pointer(skb)); end_addr = L1_CACHE_ALIGN((u64)skb_tail_pointer(skb));
@@ -506,7 +481,6 @@ xpnet_dev_hard_start_xmit(struct sk_buff *skb, struct net_device *dev)
embedded_bytes = skb->len; embedded_bytes = skb->len;
} }
/* /*
* Since the send occurs asynchronously, we set the count to one * Since the send occurs asynchronously, we set the count to one
* and begin sending. Any sends that happen to complete before * and begin sending. Any sends that happen to complete before
@@ -517,7 +491,6 @@ xpnet_dev_hard_start_xmit(struct sk_buff *skb, struct net_device *dev)
atomic_set(&queued_msg->use_count, 1); atomic_set(&queued_msg->use_count, 1);
queued_msg->skb = skb; queued_msg->skb = skb;
second_mac_octet = skb->data[XPNET_PARTID_OCTET]; second_mac_octet = skb->data[XPNET_PARTID_OCTET];
if (second_mac_octet == 0xff) { if (second_mac_octet == 0xff) {
/* we are being asked to broadcast to all partitions */ /* we are being asked to broadcast to all partitions */
@@ -543,7 +516,6 @@ xpnet_dev_hard_start_xmit(struct sk_buff *skb, struct net_device *dev)
for (dest_partid = 1; dp && dest_partid < XP_MAX_PARTITIONS; for (dest_partid = 1; dp && dest_partid < XP_MAX_PARTITIONS;
dest_partid++) { dest_partid++) {
if (!(dp & (1UL << (dest_partid - 1)))) { if (!(dp & (1UL << (dest_partid - 1)))) {
/* not destined for this partition */ /* not destined for this partition */
continue; continue;
@@ -552,14 +524,12 @@ xpnet_dev_hard_start_xmit(struct sk_buff *skb, struct net_device *dev)
/* remove this partition from the destinations mask */ /* remove this partition from the destinations mask */
dp &= ~(1UL << (dest_partid - 1)); dp &= ~(1UL << (dest_partid - 1));
/* found a partition to send to */ /* found a partition to send to */
ret = xpc_allocate(dest_partid, XPC_NET_CHANNEL, ret = xpc_allocate(dest_partid, XPC_NET_CHANNEL,
XPC_NOWAIT, (void **)&msg); XPC_NOWAIT, (void **)&msg);
if (unlikely(ret != xpcSuccess)) { if (unlikely(ret != xpcSuccess))
continue; continue;
}
msg->embedded_bytes = embedded_bytes; msg->embedded_bytes = embedded_bytes;
if (unlikely(embedded_bytes != 0)) { if (unlikely(embedded_bytes != 0)) {
@@ -583,7 +553,6 @@ xpnet_dev_hard_start_xmit(struct sk_buff *skb, struct net_device *dev)
dest_partid, XPC_NET_CHANNEL, msg->buf_pa, msg->size, dest_partid, XPC_NET_CHANNEL, msg->buf_pa, msg->size,
msg->leadin_ignore, msg->tailout_ignore); msg->leadin_ignore, msg->tailout_ignore);
atomic_inc(&queued_msg->use_count); atomic_inc(&queued_msg->use_count);
ret = xpc_send_notify(dest_partid, XPC_NET_CHANNEL, msg, ret = xpc_send_notify(dest_partid, XPC_NET_CHANNEL, msg,
@@ -592,14 +561,12 @@ xpnet_dev_hard_start_xmit(struct sk_buff *skb, struct net_device *dev)
atomic_dec(&queued_msg->use_count); atomic_dec(&queued_msg->use_count);
continue; continue;
} }
} }
if (atomic_dec_return(&queued_msg->use_count) == 0) { if (atomic_dec_return(&queued_msg->use_count) == 0) {
dev_dbg(xpnet, "no partitions to receive packet destined for " dev_dbg(xpnet, "no partitions to receive packet destined for "
"%d\n", dest_partid); "%d\n", dest_partid);
dev_kfree_skb(skb); dev_kfree_skb(skb);
kfree(queued_msg); kfree(queued_msg);
} }
@@ -610,7 +577,6 @@ xpnet_dev_hard_start_xmit(struct sk_buff *skb, struct net_device *dev)
return 0; return 0;
} }
/* /*
* Deal with transmit timeouts coming from the network layer. * Deal with transmit timeouts coming from the network layer.
*/ */
@@ -619,14 +585,12 @@ xpnet_dev_tx_timeout (struct net_device *dev)
{ {
struct xpnet_dev_private *priv; struct xpnet_dev_private *priv;
priv = (struct xpnet_dev_private *)dev->priv; priv = (struct xpnet_dev_private *)dev->priv;
priv->stats.tx_errors++; priv->stats.tx_errors++;
return; return;
} }
static int __init static int __init
xpnet_init(void) xpnet_init(void)
{ {
@@ -634,10 +598,8 @@ xpnet_init(void)
u32 license_num; u32 license_num;
int result = -ENOMEM; int result = -ENOMEM;
if (!ia64_platform_is("sn2"))
if (!ia64_platform_is("sn2")) {
return -ENODEV; return -ENODEV;
}
dev_info(xpnet, "registering network device %s\n", XPNET_DEVICE_NAME); dev_info(xpnet, "registering network device %s\n", XPNET_DEVICE_NAME);
@@ -647,9 +609,8 @@ xpnet_init(void)
*/ */
xpnet_device = alloc_netdev(sizeof(struct xpnet_dev_private), xpnet_device = alloc_netdev(sizeof(struct xpnet_dev_private),
XPNET_DEVICE_NAME, ether_setup); XPNET_DEVICE_NAME, ether_setup);
if (xpnet_device == NULL) { if (xpnet_device == NULL)
return -ENOMEM; return -ENOMEM;
}
netif_carrier_off(xpnet_device); netif_carrier_off(xpnet_device);
@@ -690,14 +651,13 @@ xpnet_init(void)
xpnet_device->features = NETIF_F_NO_CSUM; xpnet_device->features = NETIF_F_NO_CSUM;
result = register_netdev(xpnet_device); result = register_netdev(xpnet_device);
if (result != 0) { if (result != 0)
free_netdev(xpnet_device); free_netdev(xpnet_device);
}
return result; return result;
} }
module_init(xpnet_init);
module_init(xpnet_init);
static void __exit static void __exit
xpnet_exit(void) xpnet_exit(void)
@@ -709,10 +669,9 @@ xpnet_exit(void)
free_netdev(xpnet_device); free_netdev(xpnet_device);
} }
module_exit(xpnet_exit);
module_exit(xpnet_exit);
MODULE_AUTHOR("Silicon Graphics, Inc."); MODULE_AUTHOR("Silicon Graphics, Inc.");
MODULE_DESCRIPTION("Cross Partition Network adapter (XPNET)"); MODULE_DESCRIPTION("Cross Partition Network adapter (XPNET)");
MODULE_LICENSE("GPL"); MODULE_LICENSE("GPL");

View File

@@ -157,6 +157,7 @@ extern void ia64_mca_printk(const char * fmt, ...)
struct ia64_mca_notify_die { struct ia64_mca_notify_die {
struct ia64_sal_os_state *sos; struct ia64_sal_os_state *sos;
int *monarch_cpu; int *monarch_cpu;
int *data;
}; };
DECLARE_PER_CPU(u64, ia64_mca_pal_base); DECLARE_PER_CPU(u64, ia64_mca_pal_base);