Commit 8e85c23e authored by Dean Nelson's avatar Dean Nelson Committed by Linus Torvalds

sgi-xp: add _sn2 suffix to a few variables

Add an '_sn2' suffix to some variables found in xpc_sn2.c.
Signed-off-by: default avatarDean Nelson <dcn@sgi.com>
Signed-off-by: default avatarAndrew Morton <akpm@linux-foundation.org>
Signed-off-by: default avatarLinus Torvalds <torvalds@linux-foundation.org>
parent ee6665e3
...@@ -56,15 +56,15 @@ ...@@ -56,15 +56,15 @@
static char *xpc_remote_copy_buffer_sn2; static char *xpc_remote_copy_buffer_sn2;
static void *xpc_remote_copy_buffer_base_sn2; static void *xpc_remote_copy_buffer_base_sn2;
static struct xpc_vars_sn2 *xpc_vars; /* >>> Add _sn2 suffix? */ static struct xpc_vars_sn2 *xpc_vars_sn2;
static struct xpc_vars_part_sn2 *xpc_vars_part; /* >>> Add _sn2 suffix? */ static struct xpc_vars_part_sn2 *xpc_vars_part_sn2;
/* SH_IPI_ACCESS shub register value on startup */ /* SH_IPI_ACCESS shub register value on startup */
static u64 xpc_sh1_IPI_access; static u64 xpc_sh1_IPI_access_sn2;
static u64 xpc_sh2_IPI_access0; static u64 xpc_sh2_IPI_access0_sn2;
static u64 xpc_sh2_IPI_access1; static u64 xpc_sh2_IPI_access1_sn2;
static u64 xpc_sh2_IPI_access2; static u64 xpc_sh2_IPI_access2_sn2;
static u64 xpc_sh2_IPI_access3; static u64 xpc_sh2_IPI_access3_sn2;
/* /*
* Change protections to allow IPI operations. * Change protections to allow IPI operations.
...@@ -77,13 +77,13 @@ xpc_allow_IPI_ops_sn2(void) ...@@ -77,13 +77,13 @@ xpc_allow_IPI_ops_sn2(void)
/* >>> The following should get moved into SAL. */ /* >>> The following should get moved into SAL. */
if (is_shub2()) { if (is_shub2()) {
xpc_sh2_IPI_access0 = xpc_sh2_IPI_access0_sn2 =
(u64)HUB_L((u64 *)LOCAL_MMR_ADDR(SH2_IPI_ACCESS0)); (u64)HUB_L((u64 *)LOCAL_MMR_ADDR(SH2_IPI_ACCESS0));
xpc_sh2_IPI_access1 = xpc_sh2_IPI_access1_sn2 =
(u64)HUB_L((u64 *)LOCAL_MMR_ADDR(SH2_IPI_ACCESS1)); (u64)HUB_L((u64 *)LOCAL_MMR_ADDR(SH2_IPI_ACCESS1));
xpc_sh2_IPI_access2 = xpc_sh2_IPI_access2_sn2 =
(u64)HUB_L((u64 *)LOCAL_MMR_ADDR(SH2_IPI_ACCESS2)); (u64)HUB_L((u64 *)LOCAL_MMR_ADDR(SH2_IPI_ACCESS2));
xpc_sh2_IPI_access3 = xpc_sh2_IPI_access3_sn2 =
(u64)HUB_L((u64 *)LOCAL_MMR_ADDR(SH2_IPI_ACCESS3)); (u64)HUB_L((u64 *)LOCAL_MMR_ADDR(SH2_IPI_ACCESS3));
for_each_online_node(node) { for_each_online_node(node) {
...@@ -98,7 +98,7 @@ xpc_allow_IPI_ops_sn2(void) ...@@ -98,7 +98,7 @@ xpc_allow_IPI_ops_sn2(void)
-1UL); -1UL);
} }
} else { } else {
xpc_sh1_IPI_access = xpc_sh1_IPI_access_sn2 =
(u64)HUB_L((u64 *)LOCAL_MMR_ADDR(SH1_IPI_ACCESS)); (u64)HUB_L((u64 *)LOCAL_MMR_ADDR(SH1_IPI_ACCESS));
for_each_online_node(node) { for_each_online_node(node) {
...@@ -123,19 +123,19 @@ xpc_disallow_IPI_ops_sn2(void) ...@@ -123,19 +123,19 @@ xpc_disallow_IPI_ops_sn2(void)
for_each_online_node(node) { for_each_online_node(node) {
nasid = cnodeid_to_nasid(node); nasid = cnodeid_to_nasid(node);
HUB_S((u64 *)GLOBAL_MMR_ADDR(nasid, SH2_IPI_ACCESS0), HUB_S((u64 *)GLOBAL_MMR_ADDR(nasid, SH2_IPI_ACCESS0),
xpc_sh2_IPI_access0); xpc_sh2_IPI_access0_sn2);
HUB_S((u64 *)GLOBAL_MMR_ADDR(nasid, SH2_IPI_ACCESS1), HUB_S((u64 *)GLOBAL_MMR_ADDR(nasid, SH2_IPI_ACCESS1),
xpc_sh2_IPI_access1); xpc_sh2_IPI_access1_sn2);
HUB_S((u64 *)GLOBAL_MMR_ADDR(nasid, SH2_IPI_ACCESS2), HUB_S((u64 *)GLOBAL_MMR_ADDR(nasid, SH2_IPI_ACCESS2),
xpc_sh2_IPI_access2); xpc_sh2_IPI_access2_sn2);
HUB_S((u64 *)GLOBAL_MMR_ADDR(nasid, SH2_IPI_ACCESS3), HUB_S((u64 *)GLOBAL_MMR_ADDR(nasid, SH2_IPI_ACCESS3),
xpc_sh2_IPI_access3); xpc_sh2_IPI_access3_sn2);
} }
} else { } else {
for_each_online_node(node) { for_each_online_node(node) {
nasid = cnodeid_to_nasid(node); nasid = cnodeid_to_nasid(node);
HUB_S((u64 *)GLOBAL_MMR_ADDR(nasid, SH1_IPI_ACCESS), HUB_S((u64 *)GLOBAL_MMR_ADDR(nasid, SH1_IPI_ACCESS),
xpc_sh1_IPI_access); xpc_sh1_IPI_access_sn2);
} }
} }
} }
...@@ -182,7 +182,7 @@ xpc_send_IRQ_sn2(struct amo *amo, u64 flag, int nasid, int phys_cpuid, ...@@ -182,7 +182,7 @@ xpc_send_IRQ_sn2(struct amo *amo, u64 flag, int nasid, int phys_cpuid,
static struct amo * static struct amo *
xpc_init_IRQ_amo_sn2(int index) xpc_init_IRQ_amo_sn2(int index)
{ {
struct amo *amo = xpc_vars->amos_page + index; struct amo *amo = xpc_vars_sn2->amos_page + index;
(void)xpc_receive_IRQ_amo_sn2(amo); /* clear amo variable */ (void)xpc_receive_IRQ_amo_sn2(amo); /* clear amo variable */
return amo; return amo;
...@@ -225,7 +225,7 @@ xpc_send_local_activate_IRQ_sn2(int from_nasid) ...@@ -225,7 +225,7 @@ xpc_send_local_activate_IRQ_sn2(int from_nasid)
{ {
int w_index = XPC_NASID_W_INDEX(from_nasid); int w_index = XPC_NASID_W_INDEX(from_nasid);
int b_index = XPC_NASID_B_INDEX(from_nasid); int b_index = XPC_NASID_B_INDEX(from_nasid);
struct amo *amos = (struct amo *)__va(xpc_vars->amos_page_pa + struct amo *amos = (struct amo *)__va(xpc_vars_sn2->amos_page_pa +
(XPC_ACTIVATE_IRQ_AMOS_SN2 * (XPC_ACTIVATE_IRQ_AMOS_SN2 *
sizeof(struct amo))); sizeof(struct amo)));
...@@ -492,7 +492,8 @@ xpc_indicate_partition_disengaged_sn2(struct xpc_partition *part) ...@@ -492,7 +492,8 @@ xpc_indicate_partition_disengaged_sn2(struct xpc_partition *part)
static int static int
xpc_partition_engaged_sn2(short partid) xpc_partition_engaged_sn2(short partid)
{ {
struct amo *amo = xpc_vars->amos_page + XPC_ENGAGED_PARTITIONS_AMO_SN2; struct amo *amo = xpc_vars_sn2->amos_page +
XPC_ENGAGED_PARTITIONS_AMO_SN2;
/* our partition's amo variable ANDed with partid mask */ /* our partition's amo variable ANDed with partid mask */
return (FETCHOP_LOAD_OP(TO_AMO((u64)&amo->variable), FETCHOP_LOAD) & return (FETCHOP_LOAD_OP(TO_AMO((u64)&amo->variable), FETCHOP_LOAD) &
...@@ -502,7 +503,8 @@ xpc_partition_engaged_sn2(short partid) ...@@ -502,7 +503,8 @@ xpc_partition_engaged_sn2(short partid)
static int static int
xpc_any_partition_engaged_sn2(void) xpc_any_partition_engaged_sn2(void)
{ {
struct amo *amo = xpc_vars->amos_page + XPC_ENGAGED_PARTITIONS_AMO_SN2; struct amo *amo = xpc_vars_sn2->amos_page +
XPC_ENGAGED_PARTITIONS_AMO_SN2;
/* our partition's amo variable */ /* our partition's amo variable */
return FETCHOP_LOAD_OP(TO_AMO((u64)&amo->variable), FETCHOP_LOAD) != 0; return FETCHOP_LOAD_OP(TO_AMO((u64)&amo->variable), FETCHOP_LOAD) != 0;
...@@ -511,7 +513,8 @@ xpc_any_partition_engaged_sn2(void) ...@@ -511,7 +513,8 @@ xpc_any_partition_engaged_sn2(void)
static void static void
xpc_assume_partition_disengaged_sn2(short partid) xpc_assume_partition_disengaged_sn2(short partid)
{ {
struct amo *amo = xpc_vars->amos_page + XPC_ENGAGED_PARTITIONS_AMO_SN2; struct amo *amo = xpc_vars_sn2->amos_page +
XPC_ENGAGED_PARTITIONS_AMO_SN2;
/* clear bit(s) based on partid mask in our partition's amo */ /* clear bit(s) based on partid mask in our partition's amo */
FETCHOP_STORE_OP(TO_AMO((u64)&amo->variable), FETCHOP_AND, FETCHOP_STORE_OP(TO_AMO((u64)&amo->variable), FETCHOP_AND,
...@@ -580,27 +583,27 @@ xpc_rsvd_page_init_sn2(struct xpc_rsvd_page *rp) ...@@ -580,27 +583,27 @@ xpc_rsvd_page_init_sn2(struct xpc_rsvd_page *rp)
int i; int i;
int ret; int ret;
xpc_vars = XPC_RP_VARS(rp); xpc_vars_sn2 = XPC_RP_VARS(rp);
rp->sn.vars_pa = __pa(xpc_vars); rp->sn.vars_pa = __pa(xpc_vars_sn2);
/* vars_part array follows immediately after vars */ /* vars_part array follows immediately after vars */
xpc_vars_part = (struct xpc_vars_part_sn2 *)((u8 *)XPC_RP_VARS(rp) + xpc_vars_part_sn2 = (struct xpc_vars_part_sn2 *)((u8 *)XPC_RP_VARS(rp) +
XPC_RP_VARS_SIZE); XPC_RP_VARS_SIZE);
/* /*
* Before clearing xpc_vars, see if a page of amos had been previously * Before clearing xpc_vars_sn2, see if a page of amos had been
* allocated. If not we'll need to allocate one and set permissions * previously allocated. If not we'll need to allocate one and set
* so that cross-partition amos are allowed. * permissions so that cross-partition amos are allowed.
* *
* The allocated amo page needs MCA reporting to remain disabled after * The allocated amo page needs MCA reporting to remain disabled after
* XPC has unloaded. To make this work, we keep a copy of the pointer * XPC has unloaded. To make this work, we keep a copy of the pointer
* to this page (i.e., amos_page) in the struct xpc_vars structure, * to this page (i.e., amos_page) in the struct xpc_vars_sn2 structure,
* which is pointed to by the reserved page, and re-use that saved copy * which is pointed to by the reserved page, and re-use that saved copy
* on subsequent loads of XPC. This amo page is never freed, and its * on subsequent loads of XPC. This amo page is never freed, and its
* memory protections are never restricted. * memory protections are never restricted.
*/ */
amos_page = xpc_vars->amos_page; amos_page = xpc_vars_sn2->amos_page;
if (amos_page == NULL) { if (amos_page == NULL) {
amos_page = (struct amo *)TO_AMO(uncached_alloc_page(0, 1)); amos_page = (struct amo *)TO_AMO(uncached_alloc_page(0, 1));
if (amos_page == NULL) { if (amos_page == NULL) {
...@@ -621,18 +624,18 @@ xpc_rsvd_page_init_sn2(struct xpc_rsvd_page *rp) ...@@ -621,18 +624,18 @@ xpc_rsvd_page_init_sn2(struct xpc_rsvd_page *rp)
} }
} }
/* clear xpc_vars */ /* clear xpc_vars_sn2 */
memset(xpc_vars, 0, sizeof(struct xpc_vars_sn2)); memset(xpc_vars_sn2, 0, sizeof(struct xpc_vars_sn2));
xpc_vars->version = XPC_V_VERSION; xpc_vars_sn2->version = XPC_V_VERSION;
xpc_vars->activate_IRQ_nasid = cpuid_to_nasid(0); xpc_vars_sn2->activate_IRQ_nasid = cpuid_to_nasid(0);
xpc_vars->activate_IRQ_phys_cpuid = cpu_physical_id(0); xpc_vars_sn2->activate_IRQ_phys_cpuid = cpu_physical_id(0);
xpc_vars->vars_part_pa = __pa(xpc_vars_part); xpc_vars_sn2->vars_part_pa = __pa(xpc_vars_part_sn2);
xpc_vars->amos_page_pa = ia64_tpa((u64)amos_page); xpc_vars_sn2->amos_page_pa = ia64_tpa((u64)amos_page);
xpc_vars->amos_page = amos_page; /* save for next load of XPC */ xpc_vars_sn2->amos_page = amos_page; /* save for next load of XPC */
/* clear xpc_vars_part */ /* clear xpc_vars_part_sn2 */
memset((u64 *)xpc_vars_part, 0, sizeof(struct xpc_vars_part_sn2) * memset((u64 *)xpc_vars_part_sn2, 0, sizeof(struct xpc_vars_part_sn2) *
xp_max_npartitions); xp_max_npartitions);
/* initialize the activate IRQ related amo variables */ /* initialize the activate IRQ related amo variables */
...@@ -649,30 +652,30 @@ xpc_rsvd_page_init_sn2(struct xpc_rsvd_page *rp) ...@@ -649,30 +652,30 @@ xpc_rsvd_page_init_sn2(struct xpc_rsvd_page *rp)
static void static void
xpc_increment_heartbeat_sn2(void) xpc_increment_heartbeat_sn2(void)
{ {
xpc_vars->heartbeat++; xpc_vars_sn2->heartbeat++;
} }
static void static void
xpc_offline_heartbeat_sn2(void) xpc_offline_heartbeat_sn2(void)
{ {
xpc_increment_heartbeat_sn2(); xpc_increment_heartbeat_sn2();
xpc_vars->heartbeat_offline = 1; xpc_vars_sn2->heartbeat_offline = 1;
} }
static void static void
xpc_online_heartbeat_sn2(void) xpc_online_heartbeat_sn2(void)
{ {
xpc_increment_heartbeat_sn2(); xpc_increment_heartbeat_sn2();
xpc_vars->heartbeat_offline = 0; xpc_vars_sn2->heartbeat_offline = 0;
} }
static void static void
xpc_heartbeat_init_sn2(void) xpc_heartbeat_init_sn2(void)
{ {
DBUG_ON(xpc_vars == NULL); DBUG_ON(xpc_vars_sn2 == NULL);
bitmap_zero(xpc_vars->heartbeating_to_mask, XP_MAX_NPARTITIONS_SN2); bitmap_zero(xpc_vars_sn2->heartbeating_to_mask, XP_MAX_NPARTITIONS_SN2);
xpc_heartbeating_to_mask = &xpc_vars->heartbeating_to_mask[0]; xpc_heartbeating_to_mask = &xpc_vars_sn2->heartbeating_to_mask[0];
xpc_online_heartbeat_sn2(); xpc_online_heartbeat_sn2();
} }
...@@ -845,7 +848,8 @@ xpc_cancel_partition_deactivation_request_sn2(struct xpc_partition *part) ...@@ -845,7 +848,8 @@ xpc_cancel_partition_deactivation_request_sn2(struct xpc_partition *part)
static int static int
xpc_partition_deactivation_requested_sn2(short partid) xpc_partition_deactivation_requested_sn2(short partid)
{ {
struct amo *amo = xpc_vars->amos_page + XPC_DEACTIVATE_REQUEST_AMO_SN2; struct amo *amo = xpc_vars_sn2->amos_page +
XPC_DEACTIVATE_REQUEST_AMO_SN2;
/* our partition's amo variable ANDed with partid mask */ /* our partition's amo variable ANDed with partid mask */
return (FETCHOP_LOAD_OP(TO_AMO((u64)&amo->variable), FETCHOP_LOAD) & return (FETCHOP_LOAD_OP(TO_AMO((u64)&amo->variable), FETCHOP_LOAD) &
...@@ -1033,7 +1037,7 @@ xpc_identify_activate_IRQ_sender_sn2(void) ...@@ -1033,7 +1037,7 @@ xpc_identify_activate_IRQ_sender_sn2(void)
int n_IRQs_detected = 0; int n_IRQs_detected = 0;
struct amo *act_amos; struct amo *act_amos;
act_amos = xpc_vars->amos_page + XPC_ACTIVATE_IRQ_AMOS_SN2; act_amos = xpc_vars_sn2->amos_page + XPC_ACTIVATE_IRQ_AMOS_SN2;
/* scan through act amo variable looking for non-zero entries */ /* scan through act amo variable looking for non-zero entries */
for (word = 0; word < xpc_nasid_mask_words; word++) { for (word = 0; word < xpc_nasid_mask_words; word++) {
...@@ -1261,15 +1265,17 @@ xpc_setup_infrastructure_sn2(struct xpc_partition *part) ...@@ -1261,15 +1265,17 @@ xpc_setup_infrastructure_sn2(struct xpc_partition *part)
* The setting of the magic # indicates that these per partition * The setting of the magic # indicates that these per partition
* specific variables are ready to be used. * specific variables are ready to be used.
*/ */
xpc_vars_part[partid].GPs_pa = __pa(part_sn2->local_GPs); xpc_vars_part_sn2[partid].GPs_pa = __pa(part_sn2->local_GPs);
xpc_vars_part[partid].openclose_args_pa = xpc_vars_part_sn2[partid].openclose_args_pa =
__pa(part->local_openclose_args); __pa(part->local_openclose_args);
xpc_vars_part[partid].chctl_amo_pa = __pa(part_sn2->local_chctl_amo_va); xpc_vars_part_sn2[partid].chctl_amo_pa =
__pa(part_sn2->local_chctl_amo_va);
cpuid = raw_smp_processor_id(); /* any CPU in this partition will do */ cpuid = raw_smp_processor_id(); /* any CPU in this partition will do */
xpc_vars_part[partid].notify_IRQ_nasid = cpuid_to_nasid(cpuid); xpc_vars_part_sn2[partid].notify_IRQ_nasid = cpuid_to_nasid(cpuid);
xpc_vars_part[partid].notify_IRQ_phys_cpuid = cpu_physical_id(cpuid); xpc_vars_part_sn2[partid].notify_IRQ_phys_cpuid =
xpc_vars_part[partid].nchannels = part->nchannels; cpu_physical_id(cpuid);
xpc_vars_part[partid].magic = XPC_VP_MAGIC1; xpc_vars_part_sn2[partid].nchannels = part->nchannels;
xpc_vars_part_sn2[partid].magic = XPC_VP_MAGIC1;
return xpSuccess; return xpSuccess;
...@@ -1316,7 +1322,7 @@ xpc_teardown_infrastructure_sn2(struct xpc_partition *part) ...@@ -1316,7 +1322,7 @@ xpc_teardown_infrastructure_sn2(struct xpc_partition *part)
DBUG_ON(part->setup_state != XPC_P_SETUP); DBUG_ON(part->setup_state != XPC_P_SETUP);
part->setup_state = XPC_P_WTEARDOWN; part->setup_state = XPC_P_WTEARDOWN;
xpc_vars_part[partid].magic = 0; xpc_vars_part_sn2[partid].magic = 0;
free_irq(SGI_XPC_NOTIFY, (void *)(u64)partid); free_irq(SGI_XPC_NOTIFY, (void *)(u64)partid);
...@@ -1432,7 +1438,7 @@ xpc_pull_remote_vars_part_sn2(struct xpc_partition *part) ...@@ -1432,7 +1438,7 @@ xpc_pull_remote_vars_part_sn2(struct xpc_partition *part)
return xpRetry; return xpRetry;
} }
if (xpc_vars_part[partid].magic == XPC_VP_MAGIC1) { if (xpc_vars_part_sn2[partid].magic == XPC_VP_MAGIC1) {
/* validate the variables */ /* validate the variables */
...@@ -1462,7 +1468,7 @@ xpc_pull_remote_vars_part_sn2(struct xpc_partition *part) ...@@ -1462,7 +1468,7 @@ xpc_pull_remote_vars_part_sn2(struct xpc_partition *part)
/* let the other side know that we've pulled their variables */ /* let the other side know that we've pulled their variables */
xpc_vars_part[partid].magic = XPC_VP_MAGIC2; xpc_vars_part_sn2[partid].magic = XPC_VP_MAGIC2;
} }
if (pulled_entry->magic == XPC_VP_MAGIC1) if (pulled_entry->magic == XPC_VP_MAGIC1)
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment