x86/asm/entry, perf: Explicitly optimize vm86 handling in code_segment_base()
authorAndy Lutomirski <luto@kernel.org>
Thu, 19 Mar 2015 01:33:30 +0000 (18:33 -0700)
committerIngo Molnar <mingo@kernel.org>
Mon, 23 Mar 2015 10:13:41 +0000 (11:13 +0100)
commit383f3af3f88aadafe1fcf1948987ad538683fb8c
tree026e66b686487ce83da463d2f2e6183e6ec20520
parenta67e7277d01ccfd39b0db5a198c2643cc19dd79c
x86/asm/entry, perf: Explicitly optimize vm86 handling in code_segment_base()

There's no point in checking the VM bit on 64-bit, and, since
we're explicitly checking it, we can use user_mode_ignore_vm86()
after the check.

While we're at it, rearrange the #ifdef slightly to make the code
flow a bit clearer.

Signed-off-by: Andy Lutomirski <luto@kernel.org>
Cc: Borislav Petkov <bp@alien8.de>
Cc: Brad Spengler <spender@grsecurity.net>
Cc: Denys Vlasenko <dvlasenk@redhat.com>
Cc: H. Peter Anvin <hpa@zytor.com>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Peter Zijlstra <a.p.zijlstra@chello.nl>
Cc: Thomas Gleixner <tglx@linutronix.de>
Link: http://lkml.kernel.org/r/dc1457a734feccd03a19bb3538a7648582f57cdd.1426728647.git.luto@kernel.org
Signed-off-by: Ingo Molnar <mingo@kernel.org>
arch/x86/kernel/cpu/perf_event.c