1. 28 Feb, 2016 11 commits
  2. 21 Jul, 2015 1 commit
    • Mark Rutland's avatar
      arm64: entry: fix context tracking for el0_sp_pc · f6b01e50
      Mark Rutland authored
      commit 46b0567c upstream.
      Commit 6c81fe79 ("arm64: enable context tracking") did not
      update el0_sp_pc to use ct_user_exit, but this appears to have been
      unintentional. In commit 6ab6463a
       ("arm64: adjust el0_sync so
      that a function can be called") we made x0 available, and in the return
      to userspace we call ct_user_enter in the kernel_exit macro.
      Due to this, we currently don't correctly inform RCU of the user->kernel
      transition, and may erroneously account for time spent in the kernel as
      if we were in an extended quiescent state when CONFIG_CONTEXT_TRACKING
      is enabled.
      As we do record the kernel->user transition, a userspace application
      making accesses from an unaligned stack pointer can demonstrate the
      imbalance, provoking the following warning:
      ------------[ cut here ]------------
      WARNING: CPU: 2 PID: 3660 at kernel/context_tracking.c:75 context_tracking_enter+0xd8/0xe4()
      Modules linked in:
      CPU: 2 PID: 3660 Comm: a.out Not tainted 4.1.0-rc7+ #8
      Hardware name: ARM Juno development board (r0) (DT)
      Call trace:
      [<ffffffc000089914>] dump_backtrace+0x0/0x124
      [<ffffffc000089a48>] show_stack+0x10/0x1c
      [<ffffffc0005b3cbc>] dump_stack+0x84/0xc8
      [<ffffffc0000b3214>] warn_slowpath_common+0x98/0xd0
      [<ffffffc0000b330c>] warn_slowpath_null+0x14/0x20
      [<ffffffc00013ada4>] context_tracking_enter+0xd4/0xe4
      [<ffffffc0005b534c>] preempt_schedule_irq+0xd4/0x114
      [<ffffffc00008561c>] el1_preempt+0x4/0x28
      [<ffffffc0001b8040>] exit_files+0x38/0x4c
      [<ffffffc0000b5b94>] do_exit+0x430/0x978
      [<ffffffc0000b614c>] do_group_exit+0x40/0xd4
      [<ffffffc0000c0208>] get_signal+0x23c/0x4f4
      [<ffffffc0000890b4>] do_signal+0x1ac/0x518
      [<ffffffc000089650>] do_notify_resume+0x5c/0x68
      ---[ end trace 963c192600337066 ]---
      This patch adds the missing ct_user_exit to the el0_sp_pc entry path,
      correcting the context tracking for this case.
      Signed-off-by: default avatarMark Rutland <mark.rutland@arm.com>
      Acked-by: default avatarWill Deacon <will.deacon@arm.com>
      Fixes: 6c81fe79
       ("arm64: enable context tracking")
      Signed-off-by: default avatarCatalin Marinas <catalin.marinas@arm.com>
      Signed-off-by: default avatarGreg Kroah-Hartman <gregkh@linuxfoundation.org>
  3. 01 Apr, 2015 1 commit
  4. 27 Jan, 2015 1 commit
  5. 15 Jan, 2015 1 commit
  6. 28 Nov, 2014 1 commit
    • AKASHI Takahiro's avatar
      arm64: ptrace: allow tracer to skip a system call · 1014c81d
      AKASHI Takahiro authored
      If tracer modifies a syscall number to -1, this traced system call should
      be skipped with a return value specified in x0.
      This patch implements this semantics.
      Please note:
      * syscall entry tracing and syscall exit tracing (ftrace tracepoint and
        audit) are always executed, if enabled, even when skipping a system call
        (that is, -1).
        In this way, we can avoid a potential bug where audit_syscall_entry()
        might be called without audit_syscall_exit() at the previous system call
        being called, that would cause OOPs in audit_syscall_entry().
      Signed-off-by: default avatarAKASHI Takahiro <takahiro.akashi@linaro.org>
      [will: fixed up conflict with blr rework]
      Signed-off-by: default avatarWill Deacon <will.deacon@arm.com>
  7. 25 Nov, 2014 1 commit
  8. 14 Nov, 2014 2 commits
    • Will Deacon's avatar
      arm64: entry: use ldp/stp instead of push/pop when saving/restoring regs · 63648dd2
      Will Deacon authored
      The push/pop instructions can be suboptimal when saving/restoring large
      amounts of data to/from the stack, for example on entry/exit from the
      kernel. This is because:
        (1) They act on descending addresses (i.e. the newly decremented sp),
            which may defeat some hardware prefetchers
        (2) They introduce an implicit dependency between each instruction, as
            the sp has to be updated in order to resolve the address of the
            next access.
      This patch removes the push/pop instructions from our kernel entry/exit
      macros in favour of ldp/stp plus offset.
      Signed-off-by: default avatarWill Deacon <will.deacon@arm.com>
    • Will Deacon's avatar
      arm64: entry: avoid writing lr explicitly for constructing return paths · d54e81f9
      Will Deacon authored
      Using an explicit adr instruction to set the link register to point at
      ret_fast_syscall/ret_to_user can defeat branch and return stack predictors.
      Instead, use the standard calling instructions (bl, blr) and have an
      unconditional branch as the following instruction.
      Signed-off-by: default avatarWill Deacon <will.deacon@arm.com>
  9. 23 Sep, 2014 1 commit
    • Will Deacon's avatar
      arm64: debug: don't re-enable debug exceptions on return from el1_dbg · 1059c6bf
      Will Deacon authored
      When returning from a debug exception taken from EL1, we unmask debug
      exceptions after handling the exception. This is crucial for debug
      exceptions taken from EL0, so that any kernel work on the ret_to_user
      path can be debugged by kgdb.
      However, when returning back to EL1 the only thing left to do is to
      restore the original register state before the exception return. If
      single-step has been enabled by the debug exception handler, we will
      get stuck in an infinite debug exception loop, since we will take the
      step exception as soon as we unmask debug exceptions.
      This patch avoids unmasking debug exceptions on the debug exception
      return path when the exception was taken from EL1.
      Fixes: 2a283070
       (arm64: debug: avoid accessing mdscr_el1 on fault paths where possible)
      Cc: <stable@vger.kernel.org> #3.16+
      Reported-by: default avatarDavid Long <dave.long@linaro.org>
      Reported-by: default avatarAKASHI Takahiro <takahiro.akashi@linaro.org>
      Signed-off-by: default avatarWill Deacon <will.deacon@arm.com>
      Signed-off-by: default avatarCatalin Marinas <catalin.marinas@arm.com>
  10. 10 Jul, 2014 3 commits
  11. 18 Jun, 2014 1 commit
  12. 12 May, 2014 3 commits
  13. 08 May, 2014 1 commit
    • Ard Biesheuvel's avatar
      arm64: defer reloading a task's FPSIMD state to userland resume · 005f78cd
      Ard Biesheuvel authored
      If a task gets scheduled out and back in again and nothing has touched
      its FPSIMD state in the mean time, there is really no reason to reload
      it from memory. Similarly, repeated calls to kernel_neon_begin() and
      kernel_neon_end() will preserve and restore the FPSIMD state every time.
      This patch defers the FPSIMD state restore to the last possible moment,
      i.e., right before the task returns to userland. If a task does not return to
      userland at all (for any reason), the existing FPSIMD state is preserved
      and may be reused by the owning task if it gets scheduled in again on the
      same CPU.
      This patch adds two more functions to abstract away from straight FPSIMD
      register file saves and restores:
      - fpsimd_restore_current_state -> ensure current's FPSIMD state is loaded
      - fpsimd_flush_task_state -> invalidate live copies of a task's FPSIMD state
      Signed-off-by: default avatarArd Biesheuvel <ard.biesheuvel@linaro.org>
  14. 13 Jan, 2014 1 commit
  15. 19 Dec, 2013 1 commit
  16. 25 Nov, 2013 1 commit
    • Marc Zyngier's avatar
      arm64: let the core code deal with preempt_count · 64681787
      Marc Zyngier authored
      Commit f27dde8d
       (sched: Add NEED_RESCHED to the preempt_count)
      introduced the use of bit 31 in preempt_count for obscure scheduling
      This causes interrupts taken from EL0 to hit the (open coded) BUG when
      this flag is flipped while handling the interrupt (we compare the
      values before and after, and kill the kernel if they are different).
      The fix is to stop messing with the preempt count entirely, as this
      is already being dealt with in the generic code (irq_enter/irq_exit).
      Tested on a dual A53 FPGA running cyclictest.
      Signed-off-by: default avatarMarc Zyngier <marc.zyngier@arm.com>
      Signed-off-by: default avatarCatalin Marinas <catalin.marinas@arm.com>
  17. 05 Nov, 2013 1 commit
  18. 03 Sep, 2013 1 commit
  19. 22 Aug, 2013 1 commit
  20. 26 Jul, 2013 1 commit
  21. 31 May, 2013 1 commit
    • Mark Rutland's avatar
      arm64: treat unhandled compat el0 traps as undef · 381cc2b9
      Mark Rutland authored
      Currently, if a compat process reads or writes from/to a disabled
      cp15/cp14 register, the trap is not handled by the el0_sync_compat
      handler, and the kernel will head to bad_mode, where it will die(), and
      oops(). For 64 bit processes, disabled system register accesses are
      currently treated as unhandled instructions.
      This patch modifies entry.S to treat these unhandled traps as undefined
      instructions, sending a SIGILL to userspace. This gives processes a
      chance to handle this and stop using inaccessible registers, and
      prevents further issues in the kernel as a result of the die().
      Reported-by: default avatarJohannes Jensen <Johannes.Jensen@arm.com>
      Signed-off-by: default avatarMark Rutland <mark.rutland@arm.com>
      Signed-off-by: default avatarCatalin Marinas <catalin.marinas@arm.com>
  22. 17 Apr, 2013 1 commit
  23. 14 Feb, 2013 1 commit
  24. 05 Dec, 2012 1 commit
  25. 22 Oct, 2012 1 commit