Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux

cpumask: use new cpumask_ functions in core code.

Impact: cleanup

Time to clean up remaining laggards using the old cpu_ functions.

Signed-off-by: Rusty Russell <rusty@rustcorp.com.au>
Cc: Greg Kroah-Hartman <gregkh@suse.de>
Cc: Ingo Molnar <mingo@elte.hu>
Cc: Trond.Myklebust@netapp.com

+9 -9
+1 -1
drivers/base/cpu.c
··· 119 119 #define print_cpus_func(type) \ 120 120 static ssize_t print_cpus_##type(struct sysdev_class *class, char *buf) \ 121 121 { \ 122 - return print_cpus_map(buf, &cpu_##type##_map); \ 122 + return print_cpus_map(buf, cpu_##type##_mask); \ 123 123 } \ 124 124 static struct sysdev_class_attribute attr_##type##_map = \ 125 125 _SYSDEV_CLASS_ATTR(type, 0444, print_cpus_##type, NULL)
+2 -2
include/linux/cpuset.h
··· 90 90 static inline void cpuset_cpus_allowed(struct task_struct *p, 91 91 struct cpumask *mask) 92 92 { 93 - *mask = cpu_possible_map; 93 + cpumask_copy(mask, cpu_possible_mask); 94 94 } 95 95 static inline void cpuset_cpus_allowed_locked(struct task_struct *p, 96 96 struct cpumask *mask) 97 97 { 98 - *mask = cpu_possible_map; 98 + cpumask_copy(mask, cpu_possible_mask); 99 99 } 100 100 101 101 static inline nodemask_t cpuset_mems_allowed(struct task_struct *p)
+3 -3
kernel/workqueue.c
··· 416 416 might_sleep(); 417 417 lock_map_acquire(&wq->lockdep_map); 418 418 lock_map_release(&wq->lockdep_map); 419 - for_each_cpu_mask_nr(cpu, *cpu_map) 419 + for_each_cpu(cpu, cpu_map) 420 420 flush_cpu_workqueue(per_cpu_ptr(wq->cpu_wq, cpu)); 421 421 } 422 422 EXPORT_SYMBOL_GPL(flush_workqueue); ··· 547 547 wq = cwq->wq; 548 548 cpu_map = wq_cpu_map(wq); 549 549 550 - for_each_cpu_mask_nr(cpu, *cpu_map) 550 + for_each_cpu(cpu, cpu_map) 551 551 wait_on_cpu_work(per_cpu_ptr(wq->cpu_wq, cpu), work); 552 552 } 553 553 ··· 911 911 list_del(&wq->list); 912 912 spin_unlock(&workqueue_lock); 913 913 914 - for_each_cpu_mask_nr(cpu, *cpu_map) 914 + for_each_cpu(cpu, cpu_map) 915 915 cleanup_workqueue_thread(per_cpu_ptr(wq->cpu_wq, cpu)); 916 916 cpu_maps_update_done(); 917 917
+1 -1
mm/allocpercpu.c
··· 143 143 { 144 144 if (unlikely(!__pdata)) 145 145 return; 146 - __percpu_depopulate_mask(__pdata, &cpu_possible_map); 146 + __percpu_depopulate_mask(__pdata, cpu_possible_mask); 147 147 kfree(__percpu_disguise(__pdata)); 148 148 } 149 149 EXPORT_SYMBOL_GPL(free_percpu);
+1 -1
mm/vmstat.c
··· 27 27 28 28 memset(ret, 0, NR_VM_EVENT_ITEMS * sizeof(unsigned long)); 29 29 30 - for_each_cpu_mask_nr(cpu, *cpumask) { 30 + for_each_cpu(cpu, cpumask) { 31 31 struct vm_event_state *this = &per_cpu(vm_event_states, cpu); 32 32 33 33 for (i = 0; i < NR_VM_EVENT_ITEMS; i++)
+1 -1
net/sunrpc/svc.c
··· 312 312 switch (m->mode) { 313 313 case SVC_POOL_PERCPU: 314 314 { 315 - set_cpus_allowed_ptr(task, &cpumask_of_cpu(node)); 315 + set_cpus_allowed_ptr(task, cpumask_of(node)); 316 316 break; 317 317 } 318 318 case SVC_POOL_PERNODE: