fixes trailing whitespace and spaces before tab indents in 2.6.20-rc5-rt7 as
reported with: git-apply --whitespace=error-all
---
arch/arm/mach-omap1/time.c | 2 +-
arch/i386/kernel/entry.S | 2 +-
arch/i386/kernel/reboot.c | 2 +-
arch/ia64/Kconfig | 4 ++--
arch/ia64/kernel/fsys.S | 2 +-
arch/ia64/kernel/time.c | 4 ++--
arch/x86_64/kernel/entry.S | 2 +-
arch/x86_64/kernel/hpet.c | 2 +-
arch/x86_64/kernel/reboot.c | 2 +-
arch/x86_64/kernel/smpboot.c | 8 ++++----
arch/x86_64/kernel/tsc.c | 8 ++++----
arch/x86_64/kernel/vsyscall.c | 2 +-
drivers/char/lpptest.c | 2 +-
include/asm-arm/atomic.h | 2 +-
include/asm-generic/vmlinux.lds.h | 4 ++--
include/linux/irq.h | 6 +++---
include/linux/mutex.h | 2 +-
kernel/fork.c | 6 +++---
kernel/futex.c | 4 ++--
kernel/hrtimer.c | 2 +-
kernel/latency_trace.c | 6 +++---
kernel/spinlock.c | 2 +-
kernel/workqueue.c | 2 +-
mm/migrate.c | 2 +-
mm/slab.c | 14 +++++++-------
25 files changed, 47 insertions(+), 47 deletions(-)
diff --git a/arch/arm/mach-omap1/time.c b/arch/arm/mach-omap1/time.c
index 7baf0df..85f9f5e 100644
--- a/arch/arm/mach-omap1/time.c
+++ b/arch/arm/mach-omap1/time.c
@@ -105,7 +105,7 @@ static inline unsigned long omap_mpu_tim
static inline void omap_mpu_set_autoreset(int nr)
{
- volatile omap_mpu_timer_regs_t* timer = omap_mpu_timer_base(nr);
+ volatile omap_mpu_timer_regs_t* timer = omap_mpu_timer_base(nr);
timer->cntl = timer->cntl | MPU_TIMER_AR;
}
diff --git a/arch/i386/kernel/entry.S b/arch/i386/kernel/entry.S
index ce8a092..33d6b80 100644
--- a/arch/i386/kernel/entry.S
+++ b/arch/i386/kernel/entry.S
@@ -272,7 +272,7 @@ need_resched:
jz restore_nocheck
testl $IF_MASK,PT_EFLAGS(%esp) # interrupts off (exception path) ?
jz restore_nocheck
- DISABLE_INTERRUPTS(CLBR_ANY)
+ DISABLE_INTERRUPTS(CLBR_ANY)
call preempt_schedule_irq
jmp need_resched
diff --git a/arch/i386/kernel/reboot.c b/arch/i386/kernel/reboot.c
index 5f81e80..a3e7410 100644
--- a/arch/i386/kernel/reboot.c
+++ b/arch/i386/kernel/reboot.c
@@ -327,7 +327,7 @@ void machine_emergency_restart(void)
asm volatile (
"1: .byte 0x0f, 0x01, 0xc4 \n" /* vmxoff */
"2: \n"
- ".section __ex_table,\"a\" \n"
+ ".section __ex_table,\"a\" \n"
" .align 4 \n"
" .long 1b,2b \n"
".previous \n"
diff --git a/arch/ia64/Kconfig b/arch/ia64/Kconfig
index 333049c..a1a6bc3 100644
--- a/arch/ia64/Kconfig
+++ b/arch/ia64/Kconfig
@@ -300,7 +300,7 @@ config HIGH_RES_RESOLUTION
choice
prompt "Clock source"
depends on HIGH_RES_TIMERS
- default HIGH_RES_TIMER_ITC
+ default HIGH_RES_TIMER_ITC
help
This option allows you to choose the hardware source in charge
of generating high precision interruptions on your system.
@@ -308,7 +308,7 @@ choice
<timer> <resolution>
ITC Interval Time Counter 1/CPU clock
- HPET High Precision Event Timer ~ (XXX:have to check the spec)
+ HPET High Precision Event Timer ~ (XXX:have to check the spec)
The ITC timer is available on all the ia64 computers because
it is integrated directly into the processor. However it may not
diff --git a/arch/ia64/kernel/fsys.S b/arch/ia64/kernel/fsys.S
index 7b94c34..2abf920 100644
--- a/arch/ia64/kernel/fsys.S
+++ b/arch/ia64/kernel/fsys.S
@@ -219,7 +219,7 @@ ENTRY(fsys_gettimeofday)
(p6) br.cond.spnt.few .fail_einval // deferred branch
;;
ld8 r30 = [r10] // clocksource->mmio_ptr
- movl r19 = itc_lastcycle
+ movl r19 = itc_lastcycle
add r23 = IA64_CLOCKSOURCE_SHIFT_OFFSET,r20
cmp.ne p6, p0 = 0, r2 // Fallback if work is scheduled
(p6) br.cond.spnt.many fsys_fallback_syscall
diff --git a/arch/ia64/kernel/time.c b/arch/ia64/kernel/time.c
index 983ef26..4fc3670 100644
--- a/arch/ia64/kernel/time.c
+++ b/arch/ia64/kernel/time.c
@@ -262,10 +262,10 @@ ia64_init_itm (void)
ia64_cpu_local_tick();
if (!clocksource_itc_p) {
- /* Sort out mult/shift values: */
+ /* Sort out mult/shift values: */
clocksource_itc.mult = clocksource_hz2mult(local_cpu_data->itc_freq,
clocksource_itc.shift);
- clocksource_register(&clocksource_itc);
+ clocksource_register(&clocksource_itc);
clocksource_itc_p = &clocksource_itc;
}
}
diff --git a/arch/x86_64/kernel/entry.S b/arch/x86_64/kernel/entry.S
index fb0dace..75f2626 100644
--- a/arch/x86_64/kernel/entry.S
+++ b/arch/x86_64/kernel/entry.S
@@ -361,7 +361,7 @@ tracesys:
movq %r10,%rcx /* fixup for C */
TRACE_SYS_CALL
call *sys_call_table(,%rax,8)
- TRACE_SYS_RET
+ TRACE_SYS_RET
1: movq %rax,RAX-ARGOFFSET(%rsp)
/* Use IRET because user could have changed frame */
diff --git a/arch/x86_64/kernel/hpet.c b/arch/x86_64/kernel/hpet.c
index 9edc2e4..8ae86ff 100644
--- a/arch/x86_64/kernel/hpet.c
+++ b/arch/x86_64/kernel/hpet.c
@@ -63,7 +63,7 @@ static __init int late_hpet_init(void)
unsigned int ntimer;
if (!hpet_address)
- return 0;
+ return 0;
memset(&hd, 0, sizeof (hd));
diff --git a/arch/x86_64/kernel/reboot.c b/arch/x86_64/kernel/reboot.c
index dda0b88..36bad04 100644
--- a/arch/x86_64/kernel/reboot.c
+++ b/arch/x86_64/kernel/reboot.c
@@ -124,7 +124,7 @@ void machine_emergency_restart(void)
asm volatile (
"1: .byte 0x0f, 0x01, 0xc4 \n" /* vmxoff */
"2: \n"
- ".section __ex_table,\"a\" \n"
+ ".section __ex_table,\"a\" \n"
" .align 8 \n"
" .quad 1b,2b \n"
".previous \n"
diff --git a/arch/x86_64/kernel/smpboot.c b/arch/x86_64/kernel/smpboot.c
index f2c0475..8551845 100644
--- a/arch/x86_64/kernel/smpboot.c
+++ b/arch/x86_64/kernel/smpboot.c
@@ -335,8 +335,8 @@ void __cpuinit notrace start_secondary(v
barrier();
/*
- * Check TSC sync first:
- */
+ * Check TSC sync first:
+ */
check_tsc_sync_target();
Dprintk("cpu %d: setting up apic clock\n", smp_processor_id());
@@ -954,8 +954,8 @@ int __cpuinit __cpu_up(unsigned int cpu)
Dprintk("waiting for cpu %d\n", cpu);
/*
- * Make sure and check TSC sync:
- */
+ * Make sure and check TSC sync:
+ */
check_tsc_sync_source(cpu);
while (!cpu_isset(cpu, cpu_online_map))
diff --git a/arch/x86_64/kernel/tsc.c b/arch/x86_64/kernel/tsc.c
index ef2ec93..655d170 100644
--- a/arch/x86_64/kernel/tsc.c
+++ b/arch/x86_64/kernel/tsc.c
@@ -157,17 +157,17 @@ __cpuinit int unsynchronized_tsc(void)
#endif
/* Most intel systems have synchronized TSCs except for
multi node systems */
- if (boot_cpu_data.x86_vendor == X86_VENDOR_INTEL) {
+ if (boot_cpu_data.x86_vendor == X86_VENDOR_INTEL) {
#ifdef CONFIG_ACPI
/* But TSC doesn't tick in C3 so don't use it there */
if (acpi_fadt.length > 0 && acpi_fadt.plvl3_lat < 100)
return 1;
#endif
- return 0;
+ return 0;
}
- /* Assume multi socket systems are not synchronized */
- return num_present_cpus() > 1;
+ /* Assume multi socket systems are not synchronized */
+ return num_present_cpus() > 1;
}
int __init notsc_setup(char *s)
diff --git a/arch/x86_64/kernel/vsyscall.c b/arch/x86_64/kernel/vsyscall.c
index b681eee..ed7c882 100644
--- a/arch/x86_64/kernel/vsyscall.c
+++ b/arch/x86_64/kernel/vsyscall.c
@@ -73,7 +73,7 @@ void update_vsyscall(struct timespec* wa
vsyscall_gtod_data.wall_time_tv.tv_usec = wall_time->tv_nsec/1000;
vsyscall_gtod_data.sys_tz = sys_tz;
- if (cpu_has(&boot_cpu_data, X86_FEATURE_RDTSCP))
+ if (cpu_has(&boot_cpu_data, X86_FEATURE_RDTSCP))
vsyscall_gtod_data.vgetcpu_mode = VGETCPU_RDTSCP;
else
vsyscall_gtod_data.vgetcpu_mode = VGETCPU_LSL;
diff --git a/drivers/char/lpptest.c b/drivers/char/lpptest.c
index 02765fe..0c9116b 100644
--- a/drivers/char/lpptest.c
+++ b/drivers/char/lpptest.c
@@ -78,7 +78,7 @@ static cycles_t test_response(void)
outb(0x08, 0x378);
now = get_cycles();
while(1) {
- if (inb(0x379) != in)
+ if (inb(0x379) != in)
break;
if (timeout++ > 1000000) {
outb(0x00, 0x378);
diff --git a/include/asm-arm/atomic.h b/include/asm-arm/atomic.h
index e21494d..9846347 100644
--- a/include/asm-arm/atomic.h
+++ b/include/asm-arm/atomic.h
@@ -242,7 +242,7 @@ static inline unsigned long __cmpxchg(vo
__typeof__(*(ptr)) _o_ = (o); \
__typeof__(*(ptr)) _n_ = (n); \
(__typeof__(*(ptr))) __cmpxchg((ptr), (unsigned long)_o_, \
- (unsigned long)_n_, sizeof(*(ptr))); \
+ (unsigned long)_n_, sizeof(*(ptr))); \
})
#endif
diff --git a/include/asm-generic/vmlinux.lds.h b/include/asm-generic/vmlinux.lds.h
index 4c595cb..8ebba62 100644
--- a/include/asm-generic/vmlinux.lds.h
+++ b/include/asm-generic/vmlinux.lds.h
@@ -167,7 +167,7 @@
#define EH_FRAME \
/* Unwind data binary search table */ \
. = ALIGN(8); \
- .eh_frame_hdr : AT(ADDR(.eh_frame_hdr) - LOAD_OFFSET) { \
+ .eh_frame_hdr : AT(ADDR(.eh_frame_hdr) - LOAD_OFFSET) { \
VMLINUX_SYMBOL(__start_unwind_hdr) = .; \
*(.eh_frame_hdr) \
VMLINUX_SYMBOL(__end_unwind_hdr) = .; \
@@ -176,7 +176,7 @@
. = ALIGN(8); \
.eh_frame : AT(ADDR(.eh_frame) - LOAD_OFFSET) { \
VMLINUX_SYMBOL(__start_unwind) = .; \
- *(.eh_frame) \
+ *(.eh_frame) \
VMLINUX_SYMBOL(__end_unwind) = .; \
}
#else
diff --git a/include/linux/irq.h b/include/linux/irq.h
index 28f049a..0f3bc01 100644
--- a/include/linux/irq.h
+++ b/include/linux/irq.h
@@ -161,9 +161,9 @@ struct irq_desc {
unsigned int wake_depth; /* nested wake enables */
unsigned int irq_count; /* For detecting broken IRQs */
unsigned int irqs_unhandled;
- struct task_struct *thread;
- wait_queue_head_t wait_for_handler;
- cycles_t timestamp;
+ struct task_struct *thread;
+ wait_queue_head_t wait_for_handler;
+ cycles_t timestamp;
raw_spinlock_t lock;
#ifdef CONFIG_SMP
cpumask_t affinity;
diff --git a/include/linux/mutex.h b/include/linux/mutex.h
index 7a5d1fe..25d9df8 100644
--- a/include/linux/mutex.h
+++ b/include/linux/mutex.h
@@ -30,7 +30,7 @@ struct mutex {
};
#define DEFINE_MUTEX(mutexname) \
- struct mutex mutexname = { \
+ struct mutex mutexname = { \
.lock = __RT_MUTEX_INITIALIZER(mutexname.lock) \
}
diff --git a/kernel/fork.c b/kernel/fork.c
index 075067e..1630673 100644
--- a/kernel/fork.c
+++ b/kernel/fork.c
@@ -1876,9 +1876,9 @@ static int __devinit cpu_callback(struct
printk("desched_thread for %i failed\n", hotcpu);
return NOTIFY_BAD;
}
- per_cpu(desched_task, hotcpu) = p;
+ per_cpu(desched_task, hotcpu) = p;
kthread_bind(p, hotcpu);
- break;
+ break;
case CPU_ONLINE:
wake_up_process(per_cpu(desched_task, hotcpu));
@@ -1896,7 +1896,7 @@ static int __devinit cpu_callback(struct
takeover_tasklets(hotcpu);
break;
#endif /* CONFIG_HOTPLUG_CPU */
- }
+ }
return NOTIFY_OK;
}
diff --git a/kernel/futex.c b/kernel/futex.c
index b7ca553..9507575 100644
--- a/kernel/futex.c
+++ b/kernel/futex.c
@@ -1358,7 +1358,7 @@ static int fixup_pi_state_owner(u32 __us
curval = futex_atomic_cmpxchg_inatomic(uaddr,
uval, newval);
if (curval == -EFAULT)
- ret = -EFAULT;
+ ret = -EFAULT;
if (curval == uval)
break;
uval = curval;
@@ -2380,7 +2380,7 @@ static int fixup_pi_state_owner64(u64 __
curval = futex_atomic_cmpxchg_inatomic64(uaddr,
uval, newval);
if (curval == -EFAULT)
- ret = -EFAULT;
+ ret = -EFAULT;
if (curval == uval)
break;
uval = curval;
diff --git a/kernel/hrtimer.c b/kernel/hrtimer.c
index fb7731f..dfa9d0f 100644
--- a/kernel/hrtimer.c
+++ b/kernel/hrtimer.c
@@ -767,7 +767,7 @@ static enum hrtimer_restart hrtimer_sche
{
struct hrtimer_cpu_base *cpu_base =
container_of(timer, struct hrtimer_cpu_base, sched_timer);
- struct pt_regs *regs = get_irq_regs();
+ struct pt_regs *regs = get_irq_regs();
ktime_t now = ktime_get();
/* Check, if the jiffies need an update */
diff --git a/kernel/latency_trace.c b/kernel/latency_trace.c
index 936238a..7f6c304 100644
--- a/kernel/latency_trace.c
+++ b/kernel/latency_trace.c
@@ -546,7 +546,7 @@ static void notrace early_print_entry(st
(entry->flags & TRACE_FLAG_IRQS_OFF) ? 'd' :
(entry->flags & TRACE_FLAG_IRQS_HARD_OFF) ? 'D' : '.',
(entry->flags & TRACE_FLAG_NEED_RESCHED_DELAYED) ? 'n' :
- ((entry->flags & TRACE_FLAG_NEED_RESCHED) ? 'N' : '.'));
+ ((entry->flags & TRACE_FLAG_NEED_RESCHED) ? 'N' : '.'));
hardirq = entry->flags & TRACE_FLAG_HARDIRQ;
softirq = entry->flags & TRACE_FLAG_SOFTIRQ;
@@ -1321,7 +1321,7 @@ print_generic(struct seq_file *m, struct
(entry->flags & TRACE_FLAG_IRQS_OFF) ? 'd' :
(entry->flags & TRACE_FLAG_IRQS_HARD_OFF) ? 'D' : '.',
(entry->flags & TRACE_FLAG_NEED_RESCHED_DELAYED) ? 'n' :
- ((entry->flags & TRACE_FLAG_NEED_RESCHED) ? 'N' : '.'));
+ ((entry->flags & TRACE_FLAG_NEED_RESCHED) ? 'N' : '.'));
hardirq = entry->flags & TRACE_FLAG_HARDIRQ;
softirq = entry->flags & TRACE_FLAG_SOFTIRQ;
@@ -2483,7 +2483,7 @@ static void print_entry(struct trace_ent
(entry->flags & TRACE_FLAG_IRQS_OFF) ? 'd' :
(entry->flags & TRACE_FLAG_IRQS_HARD_OFF) ? 'D' : '.',
(entry->flags & TRACE_FLAG_NEED_RESCHED_DELAYED) ? 'n' :
- ((entry->flags & TRACE_FLAG_NEED_RESCHED) ? 'N' : '.'));
+ ((entry->flags & TRACE_FLAG_NEED_RESCHED) ? 'N' : '.'));
hardirq = entry->flags & TRACE_FLAG_HARDIRQ;
softirq = entry->flags & TRACE_FLAG_SOFTIRQ;
diff --git a/kernel/spinlock.c b/kernel/spinlock.c
index 95a4b60..2b0b792 100644
--- a/kernel/spinlock.c
+++ b/kernel/spinlock.c
@@ -252,7 +252,7 @@ void __lockfunc __##op##_lock(locktype##
if (!(lock)->break_lock) \
(lock)->break_lock = 1; \
while (!__raw_##op##_can_lock(&(lock)->raw_lock) && \
- (lock)->break_lock) \
+ (lock)->break_lock) \
__raw_##op##_relax(&lock->raw_lock); \
} \
(lock)->break_lock = 0; \
diff --git a/kernel/workqueue.c b/kernel/workqueue.c
index b532f38..e685d6d 100644
--- a/kernel/workqueue.c
+++ b/kernel/workqueue.c
@@ -309,7 +309,7 @@ static void leak_check(void *func)
printk(KERN_ERR "BUG: workqueue leaked lock or atomic: "
"%s/0x%08x/%d\n",
current->comm, preempt_count(),
- current->pid);
+ current->pid);
printk(KERN_ERR " last function: ");
print_symbol("%s\n", (unsigned long)func);
debug_show_held_locks(current);
diff --git a/mm/migrate.c b/mm/migrate.c
index e35e95e..7bf090c 100644
--- a/mm/migrate.c
+++ b/mm/migrate.c
@@ -329,7 +329,7 @@ static int migrate_page_move_mapping(str
radix_tree_replace_slot(pslot, newpage);
page->mapping = NULL;
- write_unlock_irq(&mapping->tree_lock);
+ write_unlock_irq(&mapping->tree_lock);
end_page_no_new_refs(page);
/*
diff --git a/mm/slab.c b/mm/slab.c
index 038812c..bec4763 100644
--- a/mm/slab.c
+++ b/mm/slab.c
@@ -143,13 +143,13 @@
# define slab_irq_disable_rt(flags) do { (void)(flags); } while (0)
# define slab_irq_enable_rt(flags) do { (void)(flags); } while (0)
# define slab_spin_lock_irq(lock, cpu) \
- do { spin_lock_irq(lock); (cpu) = smp_processor_id(); } while (0)
+ do { spin_lock_irq(lock); (cpu) = smp_processor_id(); } while (0)
# define slab_spin_unlock_irq(lock, cpu) \
- spin_unlock_irq(lock)
+ spin_unlock_irq(lock)
# define slab_spin_lock_irqsave(lock, flags, cpu) \
- do { spin_lock_irqsave(lock, flags); (cpu) = smp_processor_id(); } while (0)
+ do { spin_lock_irqsave(lock, flags); (cpu) = smp_processor_id(); } while (0)
# define slab_spin_unlock_irqrestore(lock, flags, cpu) \
- do { spin_unlock_irqrestore(lock, flags); } while (0)
+ do { spin_unlock_irqrestore(lock, flags); } while (0)
#else
DEFINE_PER_CPU_LOCKED(int, slab_irq_locks) = { 0, };
# define slab_irq_disable(cpu) (void)get_cpu_var_locked(slab_irq_locks, &(cpu))
@@ -167,9 +167,9 @@ DEFINE_PER_CPU_LOCKED(int, slab_irq_lock
# define slab_spin_unlock_irq(lock, cpu) \
do { spin_unlock(lock); slab_irq_enable(cpu); } while (0)
# define slab_spin_lock_irqsave(lock, flags, cpu) \
- do { slab_irq_disable(cpu); spin_lock_irqsave(lock, flags); } while (0)
+ do { slab_irq_disable(cpu); spin_lock_irqsave(lock, flags); } while (0)
# define slab_spin_unlock_irqrestore(lock, flags, cpu) \
- do { spin_unlock_irqrestore(lock, flags); slab_irq_enable(cpu); } while (0)
+ do { spin_unlock_irqrestore(lock, flags); slab_irq_enable(cpu); } while (0)
#endif
/*
@@ -3335,7 +3335,7 @@ static __always_inline void *__cache_all
* ____cache_alloc_node() knows how to locate memory on other nodes
*/
if (NUMA_BUILD && !objp)
- objp = ____cache_alloc_node(cachep, flags, cpu_to_node(this_cpu), &this_cpu);
+ objp = ____cache_alloc_node(cachep, flags, cpu_to_node(this_cpu), &this_cpu);
slab_irq_restore(irqflags, this_cpu);
objp = cache_alloc_debugcheck_after(cachep, flags, objp,
caller);
--
1.4.4.1
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [email protected]
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
[Index of Archives]
[Kernel Newbies]
[Netfilter]
[Bugtraq]
[Photo]
[Stuff]
[Gimp]
[Yosemite News]
[MIPS Linux]
[ARM Linux]
[Linux Security]
[Linux RAID]
[Video 4 Linux]
[Linux for the blind]
[Linux Resources]