return NULL;
 }
 
+static DEFINE_PER_CPU(cpumask_var_t, uv_flush_tlb_mask);
+
 /**
  * uv_flush_tlb_others - globally purge translation cache of a virtual
  * address or all TLB's
                                          struct mm_struct *mm,
                                          unsigned long va, unsigned int cpu)
 {
-       static DEFINE_PER_CPU(cpumask_t, flush_tlb_mask);
-       struct cpumask *flush_mask = &__get_cpu_var(flush_tlb_mask);
+       struct cpumask *flush_mask = __get_cpu_var(uv_flush_tlb_mask);
        int i;
        int bit;
        int blade;
        if (!is_uv_system())
                return 0;
 
+       for_each_possible_cpu(cur_cpu)
+               alloc_cpumask_var_node(&per_cpu(uv_flush_tlb_mask, cur_cpu),
+                                      GFP_KERNEL, cpu_to_node(cur_cpu));
+
        uv_bau_retry_limit = 1;
        uv_nshift = uv_hub_info->n_val;
        uv_mmask = (1UL << uv_hub_info->n_val) - 1;