forked from luck/tmp_suning_uos_patched
Miscellaneous ia64 changes for 3.11 merge window
-----BEGIN PGP SIGNATURE----- Version: GnuPG v1.4.12 (GNU/Linux) iQIcBAABAgAGBQJRzg4kAAoJEKurIx+X31iBGWkP/RflCjUdJ11l8wj/KTpbiP1R jW0Jx3T8SIqMEtDRVgXRT9Q51lpz1rmHmFBd2+Gw8H0FkoOT/8glyU3rBNW+gSK0 +14M/ebfiBNj4X6pIzKEGbVr8TgpajjRXvjjQCgyQIxBgvvv+hZ2Ui8XIcRiAk6d Ikme6YICti8ikgY18onsY6t7kI9zWclj0opt4uPKn4fkulyFWQJszLgTOx1EbFRp WltW/coCzigCbZRYOhdJpszwII73ItfyFCvk8CCuQDzqQ0PwdSE6lbIJRyT/UqHi f7HUxb6j6Nai5gAhXJ3OeJ8PlxBRNwiYFvgRCKLzsiX+74vdeIkCV0scwwbAYORO 8qE+btgw7npQ8zlRuE6mOIJ9/zIkBm7S0aWDZmPv5fMGMYAL0D/ZYMMvXl/IYdPu NovS86unw7c/lz1+VEfzFeQb3K8aIYpApup7pVxAZ70+dvCyM7IfHjngQNsCbkU9 u8n/yuhuZXzTDo/gq/ctV/A10DFGXoJVHxMyjsaN8bhvrLSb/OBaJL/hljB1VEsF +IJsnpi/V7VS917mM2F3tfar5khlINLcoYFrOkkYGJY+Zr28llqozsJ3rQQXKF2g LdFQM0Cl7UQaQtQtV47n+9F0AsExjJWn0G5asn5Mh29y7RmGU+iEQ8QDbhRNYCO4 51+FOWHIArCrA2lchCz9 =hiGp -----END PGP SIGNATURE----- Merge tag 'please-pull-misc-3.11' of git://git.kernel.org/pub/scm/linux/kernel/git/aegl/linux Pull misc ia64 updates from Tony Luck: "Miscellaneous ia64 changes for 3.11 merge window" * tag 'please-pull-misc-3.11' of git://git.kernel.org/pub/scm/linux/kernel/git/aegl/linux: [IA64] Delete __cpuinit usage from all ia64 users [IA64] hpsim: Fix check for overlong simscsi prefix. [IA64] pci: Remove unused fallback_dev [IA64] perfmon: Use %*phD specifier to dump small buffers [IA64] Fix trap #45 handling
This commit is contained in:
commit
e39dfe52f8
|
@ -88,8 +88,8 @@ simscsi_setup (char *s)
|
|||
if (strlen(s) > MAX_ROOT_LEN) {
|
||||
printk(KERN_ERR "simscsi_setup: prefix too long---using default %s\n",
|
||||
simscsi_root);
|
||||
}
|
||||
simscsi_root = s;
|
||||
} else
|
||||
simscsi_root = s;
|
||||
return 1;
|
||||
}
|
||||
|
||||
|
|
|
@ -807,7 +807,7 @@ int acpi_isa_irq_to_gsi(unsigned isa_irq, u32 *gsi)
|
|||
* ACPI based hotplug CPU support
|
||||
*/
|
||||
#ifdef CONFIG_ACPI_HOTPLUG_CPU
|
||||
static __cpuinit
|
||||
static
|
||||
int acpi_map_cpu2node(acpi_handle handle, int cpu, int physid)
|
||||
{
|
||||
#ifdef CONFIG_ACPI_NUMA
|
||||
|
@ -882,7 +882,7 @@ __init void prefill_possible_map(void)
|
|||
set_cpu_possible(i, true);
|
||||
}
|
||||
|
||||
static int __cpuinit _acpi_map_lsapic(acpi_handle handle, int *pcpu)
|
||||
static int _acpi_map_lsapic(acpi_handle handle, int *pcpu)
|
||||
{
|
||||
struct acpi_buffer buffer = { ACPI_ALLOCATE_BUFFER, NULL };
|
||||
union acpi_object *obj;
|
||||
|
|
|
@ -225,17 +225,17 @@ static struct attribute_group err_inject_attr_group = {
|
|||
.name = "err_inject"
|
||||
};
|
||||
/* Add/Remove err_inject interface for CPU device */
|
||||
static int __cpuinit err_inject_add_dev(struct device * sys_dev)
|
||||
static int err_inject_add_dev(struct device *sys_dev)
|
||||
{
|
||||
return sysfs_create_group(&sys_dev->kobj, &err_inject_attr_group);
|
||||
}
|
||||
|
||||
static int __cpuinit err_inject_remove_dev(struct device * sys_dev)
|
||||
static int err_inject_remove_dev(struct device *sys_dev)
|
||||
{
|
||||
sysfs_remove_group(&sys_dev->kobj, &err_inject_attr_group);
|
||||
return 0;
|
||||
}
|
||||
static int __cpuinit err_inject_cpu_callback(struct notifier_block *nfb,
|
||||
static int err_inject_cpu_callback(struct notifier_block *nfb,
|
||||
unsigned long action, void *hcpu)
|
||||
{
|
||||
unsigned int cpu = (unsigned long)hcpu;
|
||||
|
@ -256,7 +256,7 @@ static int __cpuinit err_inject_cpu_callback(struct notifier_block *nfb,
|
|||
return NOTIFY_OK;
|
||||
}
|
||||
|
||||
static struct notifier_block __cpuinitdata err_inject_cpu_notifier =
|
||||
static struct notifier_block err_inject_cpu_notifier =
|
||||
{
|
||||
.notifier_call = err_inject_cpu_callback,
|
||||
};
|
||||
|
|
|
@ -631,7 +631,7 @@ ia64_mca_register_cpev (int cpev)
|
|||
* Outputs
|
||||
* None
|
||||
*/
|
||||
void __cpuinit
|
||||
void
|
||||
ia64_mca_cmc_vector_setup (void)
|
||||
{
|
||||
cmcv_reg_t cmcv;
|
||||
|
@ -1814,7 +1814,7 @@ static struct irqaction mca_cpep_irqaction = {
|
|||
* format most of the fields.
|
||||
*/
|
||||
|
||||
static void __cpuinit
|
||||
static void
|
||||
format_mca_init_stack(void *mca_data, unsigned long offset,
|
||||
const char *type, int cpu)
|
||||
{
|
||||
|
@ -1844,7 +1844,7 @@ static void * __init_refok mca_bootmem(void)
|
|||
}
|
||||
|
||||
/* Do per-CPU MCA-related initialization. */
|
||||
void __cpuinit
|
||||
void
|
||||
ia64_mca_cpu_init(void *cpu_data)
|
||||
{
|
||||
void *pal_vaddr;
|
||||
|
@ -1896,7 +1896,7 @@ ia64_mca_cpu_init(void *cpu_data)
|
|||
PAGE_KERNEL));
|
||||
}
|
||||
|
||||
static void __cpuinit ia64_mca_cmc_vector_adjust(void *dummy)
|
||||
static void ia64_mca_cmc_vector_adjust(void *dummy)
|
||||
{
|
||||
unsigned long flags;
|
||||
|
||||
|
@ -1906,7 +1906,7 @@ static void __cpuinit ia64_mca_cmc_vector_adjust(void *dummy)
|
|||
local_irq_restore(flags);
|
||||
}
|
||||
|
||||
static int __cpuinit mca_cpu_callback(struct notifier_block *nfb,
|
||||
static int mca_cpu_callback(struct notifier_block *nfb,
|
||||
unsigned long action,
|
||||
void *hcpu)
|
||||
{
|
||||
|
@ -1922,7 +1922,7 @@ static int __cpuinit mca_cpu_callback(struct notifier_block *nfb,
|
|||
return NOTIFY_OK;
|
||||
}
|
||||
|
||||
static struct notifier_block mca_cpu_notifier __cpuinitdata = {
|
||||
static struct notifier_block mca_cpu_notifier = {
|
||||
.notifier_call = mca_cpu_callback
|
||||
};
|
||||
|
||||
|
|
|
@ -30,7 +30,7 @@ EXPORT_SYMBOL(cpu_to_node_map);
|
|||
cpumask_t node_to_cpu_mask[MAX_NUMNODES] __cacheline_aligned;
|
||||
EXPORT_SYMBOL(node_to_cpu_mask);
|
||||
|
||||
void __cpuinit map_cpu_to_node(int cpu, int nid)
|
||||
void map_cpu_to_node(int cpu, int nid)
|
||||
{
|
||||
int oldnid;
|
||||
if (nid < 0) { /* just initialize by zero */
|
||||
|
@ -51,7 +51,7 @@ void __cpuinit map_cpu_to_node(int cpu, int nid)
|
|||
return;
|
||||
}
|
||||
|
||||
void __cpuinit unmap_cpu_from_node(int cpu, int nid)
|
||||
void unmap_cpu_from_node(int cpu, int nid)
|
||||
{
|
||||
WARN_ON(!cpu_isset(cpu, node_to_cpu_mask[nid]));
|
||||
WARN_ON(cpu_to_node_map[cpu] != nid);
|
||||
|
|
|
@ -932,7 +932,7 @@ static const struct file_operations proc_palinfo_fops = {
|
|||
.release = single_release,
|
||||
};
|
||||
|
||||
static void __cpuinit
|
||||
static void
|
||||
create_palinfo_proc_entries(unsigned int cpu)
|
||||
{
|
||||
pal_func_cpu_u_t f;
|
||||
|
@ -962,7 +962,7 @@ remove_palinfo_proc_entries(unsigned int hcpu)
|
|||
remove_proc_subtree(cpustr, palinfo_dir);
|
||||
}
|
||||
|
||||
static int __cpuinit palinfo_cpu_callback(struct notifier_block *nfb,
|
||||
static int palinfo_cpu_callback(struct notifier_block *nfb,
|
||||
unsigned long action, void *hcpu)
|
||||
{
|
||||
unsigned int hotcpu = (unsigned long)hcpu;
|
||||
|
|
|
@ -33,15 +33,6 @@ int force_iommu __read_mostly;
|
|||
|
||||
int iommu_pass_through;
|
||||
|
||||
/* Dummy device used for NULL arguments (normally ISA). Better would
|
||||
be probably a smaller DMA mask, but this is bug-to-bug compatible
|
||||
to i386. */
|
||||
struct device fallback_dev = {
|
||||
.init_name = "fallback device",
|
||||
.coherent_dma_mask = DMA_BIT_MASK(32),
|
||||
.dma_mask = &fallback_dev.coherent_dma_mask,
|
||||
};
|
||||
|
||||
extern struct dma_map_ops intel_dma_ops;
|
||||
|
||||
static int __init pci_iommu_init(void)
|
||||
|
|
|
@ -5647,24 +5647,8 @@ pfm_proc_show_header(struct seq_file *m)
|
|||
|
||||
list_for_each(pos, &pfm_buffer_fmt_list) {
|
||||
entry = list_entry(pos, pfm_buffer_fmt_t, fmt_list);
|
||||
seq_printf(m, "format : %02x-%02x-%02x-%02x-%02x-%02x-%02x-%02x-%02x-%02x-%02x-%02x-%02x-%02x-%02x-%02x %s\n",
|
||||
entry->fmt_uuid[0],
|
||||
entry->fmt_uuid[1],
|
||||
entry->fmt_uuid[2],
|
||||
entry->fmt_uuid[3],
|
||||
entry->fmt_uuid[4],
|
||||
entry->fmt_uuid[5],
|
||||
entry->fmt_uuid[6],
|
||||
entry->fmt_uuid[7],
|
||||
entry->fmt_uuid[8],
|
||||
entry->fmt_uuid[9],
|
||||
entry->fmt_uuid[10],
|
||||
entry->fmt_uuid[11],
|
||||
entry->fmt_uuid[12],
|
||||
entry->fmt_uuid[13],
|
||||
entry->fmt_uuid[14],
|
||||
entry->fmt_uuid[15],
|
||||
entry->fmt_name);
|
||||
seq_printf(m, "format : %16phD %s\n",
|
||||
entry->fmt_uuid, entry->fmt_name);
|
||||
}
|
||||
spin_unlock(&pfm_buffer_fmt_lock);
|
||||
|
||||
|
|
|
@ -568,7 +568,7 @@ static const struct file_operations salinfo_data_fops = {
|
|||
.llseek = default_llseek,
|
||||
};
|
||||
|
||||
static int __cpuinit
|
||||
static int
|
||||
salinfo_cpu_callback(struct notifier_block *nb, unsigned long action, void *hcpu)
|
||||
{
|
||||
unsigned int i, cpu = (unsigned long)hcpu;
|
||||
|
@ -609,7 +609,7 @@ salinfo_cpu_callback(struct notifier_block *nb, unsigned long action, void *hcpu
|
|||
return NOTIFY_OK;
|
||||
}
|
||||
|
||||
static struct notifier_block salinfo_cpu_notifier __cpuinitdata =
|
||||
static struct notifier_block salinfo_cpu_notifier =
|
||||
{
|
||||
.notifier_call = salinfo_cpu_callback,
|
||||
.priority = 0,
|
||||
|
|
|
@ -748,7 +748,7 @@ const struct seq_operations cpuinfo_op = {
|
|||
#define MAX_BRANDS 8
|
||||
static char brandname[MAX_BRANDS][128];
|
||||
|
||||
static char * __cpuinit
|
||||
static char *
|
||||
get_model_name(__u8 family, __u8 model)
|
||||
{
|
||||
static int overflow;
|
||||
|
@ -778,7 +778,7 @@ get_model_name(__u8 family, __u8 model)
|
|||
return "Unknown";
|
||||
}
|
||||
|
||||
static void __cpuinit
|
||||
static void
|
||||
identify_cpu (struct cpuinfo_ia64 *c)
|
||||
{
|
||||
union {
|
||||
|
@ -850,7 +850,7 @@ identify_cpu (struct cpuinfo_ia64 *c)
|
|||
* 2. the minimum of the i-cache stride sizes for "flush_icache_range()".
|
||||
* 3. the minimum of the cache stride sizes for "clflush_cache_range()".
|
||||
*/
|
||||
static void __cpuinit
|
||||
static void
|
||||
get_cache_info(void)
|
||||
{
|
||||
unsigned long line_size, max = 1;
|
||||
|
@ -915,10 +915,10 @@ get_cache_info(void)
|
|||
* cpu_init() initializes state that is per-CPU. This function acts
|
||||
* as a 'CPU state barrier', nothing should get across.
|
||||
*/
|
||||
void __cpuinit
|
||||
void
|
||||
cpu_init (void)
|
||||
{
|
||||
extern void __cpuinit ia64_mmu_init (void *);
|
||||
extern void ia64_mmu_init(void *);
|
||||
static unsigned long max_num_phys_stacked = IA64_NUM_PHYS_STACK_REG;
|
||||
unsigned long num_phys_stacked;
|
||||
pal_vm_info_2_u_t vmi;
|
||||
|
|
|
@ -351,7 +351,7 @@ static inline void smp_setup_percpu_timer(void)
|
|||
{
|
||||
}
|
||||
|
||||
static void __cpuinit
|
||||
static void
|
||||
smp_callin (void)
|
||||
{
|
||||
int cpuid, phys_id, itc_master;
|
||||
|
@ -442,7 +442,7 @@ smp_callin (void)
|
|||
/*
|
||||
* Activate a secondary processor. head.S calls this.
|
||||
*/
|
||||
int __cpuinit
|
||||
int
|
||||
start_secondary (void *unused)
|
||||
{
|
||||
/* Early console may use I/O ports */
|
||||
|
@ -459,7 +459,7 @@ start_secondary (void *unused)
|
|||
return 0;
|
||||
}
|
||||
|
||||
static int __cpuinit
|
||||
static int
|
||||
do_boot_cpu (int sapicid, int cpu, struct task_struct *idle)
|
||||
{
|
||||
int timeout;
|
||||
|
@ -728,7 +728,7 @@ static inline void set_cpu_sibling_map(int cpu)
|
|||
}
|
||||
}
|
||||
|
||||
int __cpuinit
|
||||
int
|
||||
__cpu_up(unsigned int cpu, struct task_struct *tidle)
|
||||
{
|
||||
int ret;
|
||||
|
|
|
@ -135,11 +135,11 @@ struct cpu_cache_info {
|
|||
struct kobject kobj;
|
||||
};
|
||||
|
||||
static struct cpu_cache_info all_cpu_cache_info[NR_CPUS] __cpuinitdata;
|
||||
static struct cpu_cache_info all_cpu_cache_info[NR_CPUS];
|
||||
#define LEAF_KOBJECT_PTR(x,y) (&all_cpu_cache_info[x].cache_leaves[y])
|
||||
|
||||
#ifdef CONFIG_SMP
|
||||
static void __cpuinit cache_shared_cpu_map_setup( unsigned int cpu,
|
||||
static void cache_shared_cpu_map_setup(unsigned int cpu,
|
||||
struct cache_info * this_leaf)
|
||||
{
|
||||
pal_cache_shared_info_t csi;
|
||||
|
@ -174,7 +174,7 @@ static void __cpuinit cache_shared_cpu_map_setup( unsigned int cpu,
|
|||
&csi) == PAL_STATUS_SUCCESS);
|
||||
}
|
||||
#else
|
||||
static void __cpuinit cache_shared_cpu_map_setup(unsigned int cpu,
|
||||
static void cache_shared_cpu_map_setup(unsigned int cpu,
|
||||
struct cache_info * this_leaf)
|
||||
{
|
||||
cpu_set(cpu, this_leaf->shared_cpu_map);
|
||||
|
@ -298,7 +298,7 @@ static struct kobj_type cache_ktype_percpu_entry = {
|
|||
.sysfs_ops = &cache_sysfs_ops,
|
||||
};
|
||||
|
||||
static void __cpuinit cpu_cache_sysfs_exit(unsigned int cpu)
|
||||
static void cpu_cache_sysfs_exit(unsigned int cpu)
|
||||
{
|
||||
kfree(all_cpu_cache_info[cpu].cache_leaves);
|
||||
all_cpu_cache_info[cpu].cache_leaves = NULL;
|
||||
|
@ -307,7 +307,7 @@ static void __cpuinit cpu_cache_sysfs_exit(unsigned int cpu)
|
|||
return;
|
||||
}
|
||||
|
||||
static int __cpuinit cpu_cache_sysfs_init(unsigned int cpu)
|
||||
static int cpu_cache_sysfs_init(unsigned int cpu)
|
||||
{
|
||||
unsigned long i, levels, unique_caches;
|
||||
pal_cache_config_info_t cci;
|
||||
|
@ -351,7 +351,7 @@ static int __cpuinit cpu_cache_sysfs_init(unsigned int cpu)
|
|||
}
|
||||
|
||||
/* Add cache interface for CPU device */
|
||||
static int __cpuinit cache_add_dev(struct device * sys_dev)
|
||||
static int cache_add_dev(struct device *sys_dev)
|
||||
{
|
||||
unsigned int cpu = sys_dev->id;
|
||||
unsigned long i, j;
|
||||
|
@ -401,7 +401,7 @@ static int __cpuinit cache_add_dev(struct device * sys_dev)
|
|||
}
|
||||
|
||||
/* Remove cache interface for CPU device */
|
||||
static int __cpuinit cache_remove_dev(struct device * sys_dev)
|
||||
static int cache_remove_dev(struct device *sys_dev)
|
||||
{
|
||||
unsigned int cpu = sys_dev->id;
|
||||
unsigned long i;
|
||||
|
@ -425,7 +425,7 @@ static int __cpuinit cache_remove_dev(struct device * sys_dev)
|
|||
* When a cpu is hot-plugged, do a check and initiate
|
||||
* cache kobject if necessary
|
||||
*/
|
||||
static int __cpuinit cache_cpu_callback(struct notifier_block *nfb,
|
||||
static int cache_cpu_callback(struct notifier_block *nfb,
|
||||
unsigned long action, void *hcpu)
|
||||
{
|
||||
unsigned int cpu = (unsigned long)hcpu;
|
||||
|
@ -445,7 +445,7 @@ static int __cpuinit cache_cpu_callback(struct notifier_block *nfb,
|
|||
return NOTIFY_OK;
|
||||
}
|
||||
|
||||
static struct notifier_block __cpuinitdata cache_cpu_notifier =
|
||||
static struct notifier_block cache_cpu_notifier =
|
||||
{
|
||||
.notifier_call = cache_cpu_callback
|
||||
};
|
||||
|
|
|
@ -630,7 +630,7 @@ ia64_fault (unsigned long vector, unsigned long isr, unsigned long ifa,
|
|||
printk(KERN_ERR " iip - 0x%lx, ifa - 0x%lx, isr - 0x%lx\n",
|
||||
iip, ifa, isr);
|
||||
force_sig(SIGSEGV, current);
|
||||
break;
|
||||
return;
|
||||
|
||||
case 46:
|
||||
printk(KERN_ERR "Unexpected IA-32 intercept trap (Trap 46)\n");
|
||||
|
|
|
@ -156,8 +156,7 @@ static void *cpu_data;
|
|||
*
|
||||
* Allocate and setup per-cpu data areas.
|
||||
*/
|
||||
void * __cpuinit
|
||||
per_cpu_init (void)
|
||||
void *per_cpu_init(void)
|
||||
{
|
||||
static bool first_time = true;
|
||||
void *cpu0_data = __cpu0_per_cpu;
|
||||
|
|
|
@ -593,7 +593,7 @@ void __init find_memory(void)
|
|||
* find_pernode_space() does most of this already, we just need to set
|
||||
* local_per_cpu_offset
|
||||
*/
|
||||
void __cpuinit *per_cpu_init(void)
|
||||
void *per_cpu_init(void)
|
||||
{
|
||||
int cpu;
|
||||
static int first_time = 1;
|
||||
|
|
|
@ -86,7 +86,7 @@ int __meminit __early_pfn_to_nid(unsigned long pfn)
|
|||
return -1;
|
||||
}
|
||||
|
||||
void __cpuinit numa_clear_node(int cpu)
|
||||
void numa_clear_node(int cpu)
|
||||
{
|
||||
unmap_cpu_from_node(cpu, NUMA_NO_NODE);
|
||||
}
|
||||
|
|
|
@ -192,7 +192,7 @@ void __init early_sn_setup(void)
|
|||
}
|
||||
|
||||
extern int platform_intr_list[];
|
||||
static int __cpuinitdata shub_1_1_found;
|
||||
static int shub_1_1_found;
|
||||
|
||||
/*
|
||||
* sn_check_for_wars
|
||||
|
@ -200,7 +200,7 @@ static int __cpuinitdata shub_1_1_found;
|
|||
* Set flag for enabling shub specific wars
|
||||
*/
|
||||
|
||||
static inline int __cpuinit is_shub_1_1(int nasid)
|
||||
static inline int is_shub_1_1(int nasid)
|
||||
{
|
||||
unsigned long id;
|
||||
int rev;
|
||||
|
@ -212,7 +212,7 @@ static inline int __cpuinit is_shub_1_1(int nasid)
|
|||
return rev <= 2;
|
||||
}
|
||||
|
||||
static void __cpuinit sn_check_for_wars(void)
|
||||
static void sn_check_for_wars(void)
|
||||
{
|
||||
int cnode;
|
||||
|
||||
|
@ -558,7 +558,7 @@ static void __init sn_init_pdas(char **cmdline_p)
|
|||
* Also sets up a few fields in the nodepda. Also known as
|
||||
* platform_cpu_init() by the ia64 machvec code.
|
||||
*/
|
||||
void __cpuinit sn_cpu_init(void)
|
||||
void sn_cpu_init(void)
|
||||
{
|
||||
int cpuid;
|
||||
int cpuphyid;
|
||||
|
|
|
@ -74,7 +74,7 @@ void __init xen_setup_vcpu_info_placement(void)
|
|||
xen_vcpu_setup(cpu);
|
||||
}
|
||||
|
||||
void __cpuinit
|
||||
void
|
||||
xen_cpu_init(void)
|
||||
{
|
||||
xen_smp_intr_init();
|
||||
|
|
Loading…
Reference in New Issue
Block a user