Skip to content
  1. Mar 05, 2023
    • Linus Torvalds's avatar
      cpumask: re-introduce constant-sized cpumask optimizations · 596ff4a0
      Linus Torvalds authored
      
      
      Commit aa47a7c2 ("lib/cpumask: deprecate nr_cpumask_bits") resulted
      in the cpumask operations potentially becoming hugely less efficient,
      because suddenly the cpumask was always considered to be variable-sized.
      
      The optimization was then later added back in a limited form by commit
      6f9c07be ("lib/cpumask: add FORCE_NR_CPUS config option"), but that
      FORCE_NR_CPUS option is not useful in a generic kernel and more of a
      special case for embedded situations with fixed hardware.
      
      Instead, just re-introduce the optimization, with some changes.
      
      Instead of depending on CPUMASK_OFFSTACK being false, and then always
      using the full constant cpumask width, this introduces three different
      cpumask "sizes":
      
       - the exact size (nr_cpumask_bits) remains identical to nr_cpu_ids.
      
         This is used for situations where we should use the exact size.
      
       - the "small" size (small_cpumask_bits) is the NR_CPUS constant if it
         fits in a single word and the bitmap operations thus end up able
         to trigger the "small_const_nbits()" optimizations.
      
         This is used for the operations that have optimized single-word
         cases that get inlined, notably the bit find and scanning functions.
      
       - the "large" size (large_cpumask_bits) is the NR_CPUS constant if it
         is an sufficiently small constant that makes simple "copy" and
         "clear" operations more efficient.
      
         This is arbitrarily set at four words or less.
      
      As a an example of this situation, without this fixed size optimization,
      cpumask_clear() will generate code like
      
              movl    nr_cpu_ids(%rip), %edx
              addq    $63, %rdx
              shrq    $3, %rdx
              andl    $-8, %edx
              callq   memset@PLT
      
      on x86-64, because it would calculate the "exact" number of longwords
      that need to be cleared.
      
      In contrast, with this patch, using a MAX_CPU of 64 (which is quite a
      reasonable value to use), the above becomes a single
      
      	movq $0,cpumask
      
      instruction instead, because instead of caring to figure out exactly how
      many CPU's the system has, it just knows that the cpumask will be a
      single word and can just clear it all.
      
      Note that this does end up tightening the rules a bit from the original
      version in another way: operations that set bits in the cpumask are now
      limited to the actual nr_cpu_ids limit, whereas we used to do the
      nr_cpumask_bits thing almost everywhere in the cpumask code.
      
      But if you just clear bits, or scan for bits, we can use the simpler
      compile-time constants.
      
      In the process, remove 'cpumask_complement()' and 'for_each_cpu_not()'
      which were not useful, and which fundamentally have to be limited to
      'nr_cpu_ids'.  Better remove them now than have somebody introduce use
      of them later.
      
      Of course, on x86-64 with MAXSMP there is no sane small compile-time
      constant for the cpumask sizes, and we end up using the actual CPU bits,
      and will generate the above kind of horrors regardless.  Please don't
      use MAXSMP unless you really expect to have machines with thousands of
      cores.
      
      Signed-off-by: default avatarLinus Torvalds <torvalds@linux-foundation.org>
      596ff4a0
    • Masahiro Yamada's avatar
      Remove Intel compiler support · 95207db8
      Masahiro Yamada authored
      include/linux/compiler-intel.h had no update in the past 3 years.
      
      We often forget about the third C compiler to build the kernel.
      
      For example, commit a0a12c3e ("asm goto: eradicate CC_HAS_ASM_GOTO")
      only mentioned GCC and Clang.
      
      init/Kconfig defines CC_IS_GCC and CC_IS_CLANG but not CC_IS_ICC,
      and nobody has reported any issue.
      
      I guess the Intel Compiler support is broken, and nobody is caring
      about it.
      
      Harald Arnesen pointed out ICC (classic Intel C/C++ compiler) is
      deprecated:
      
          $ icc -v
          icc: remark #10441: The Intel(R) C++ Compiler Classic (ICC) is
          deprecated and will be removed from product release in the second half
          of 2023. The Intel(R) oneAPI DPC++/C++ Compiler (ICX) is the recommended
          compiler moving forward. Please transition to use this compiler. Use
          '-diag-disable=10441' to disable this message.
          icc version 2021.7.0 (gcc version 12.1.0 compatibility)
      
      Arnd Bergmann provided a link to the article, "Intel C/C++ compilers
      complete adoption of LLVM".
      
      lib/zstd/common/compiler.h and lib/zstd/compress/zstd_fast.c were kept
      untouched for better sync with https://github.com/facebook/zstd
      
      Link: https://www.intel.com/content/www/us/en/developer/articles/technical/adoption-of-llvm-complete-icx.html
      
      
      Signed-off-by: default avatarMasahiro Yamada <masahiroy@kernel.org>
      Acked-by: default avatarArnd Bergmann <arnd@arndb.de>
      Reviewed-by: default avatarNick Desaulniers <ndesaulniers@google.com>
      Reviewed-by: default avatarNathan Chancellor <nathan@kernel.org>
      Reviewed-by: default avatarMiguel Ojeda <ojeda@kernel.org>
      Signed-off-by: default avatarLinus Torvalds <torvalds@linux-foundation.org>
      95207db8
  2. Mar 03, 2023
    • Marco Elver's avatar
      kasan, x86: don't rename memintrinsics in uninstrumented files · 4ec4190b
      Marco Elver authored
      Now that memcpy/memset/memmove are no longer overridden by KASAN, we can
      just use the normal symbol names in uninstrumented files.
      
      Drop the preprocessor redefinitions.
      
      Link: https://lkml.kernel.org/r/20230224085942.1791837-4-elver@google.com
      
      
      Fixes: 69d4c0d3 ("entry, kasan, x86: Disallow overriding mem*() functions")
      Signed-off-by: default avatarMarco Elver <elver@google.com>
      Reviewed-by: default avatarAndrey Konovalov <andreyknvl@gmail.com>
      Cc: Alexander Potapenko <glider@google.com>
      Cc: Andrey Ryabinin <ryabinin.a.a@gmail.com>
      Cc: Borislav Petkov (AMD) <bp@alien8.de>
      Cc: Dave Hansen <dave.hansen@linux.intel.com>
      Cc: Dmitry Vyukov <dvyukov@google.com>
      Cc: Ingo Molnar <mingo@redhat.com>
      Cc: Jakub Jelinek <jakub@redhat.com>
      Cc: Kees Cook <keescook@chromium.org>
      Cc: Linux Kernel Functional Testing <lkft@linaro.org>
      Cc: Naresh Kamboju <naresh.kamboju@linaro.org>
      Cc: Nathan Chancellor <nathan@kernel.org>
      Cc: Nick Desaulniers <ndesaulniers@google.com>
      Cc: Nicolas Schier <nicolas@fjasle.eu>
      Cc: Peter Zijlstra (Intel) <peterz@infradead.org>
      Cc: Thomas Gleixner <tglx@linutronix.de>
      Cc: Vincenzo Frascino <vincenzo.frascino@arm.com>
      Signed-off-by: default avatarAndrew Morton <akpm@linux-foundation.org>
      4ec4190b
  3. Mar 02, 2023
    • Al Viro's avatar
      openrisc: fix livelock in uaccess · caa82ae7
      Al Viro authored
      
      
      openrisc equivalent of 26178ec1 "x86: mm: consolidate VM_FAULT_RETRY handling"
      If e.g. get_user() triggers a page fault and a fatal signal is caught, we might
      end up with handle_mm_fault() returning VM_FAULT_RETRY and not doing anything
      to page tables.  In such case we must *not* return to the faulting insn -
      that would repeat the entire thing without making any progress; what we need
      instead is to treat that as failed (user) memory access.
      
      Signed-off-by: default avatarAl Viro <viro@zeniv.linux.org.uk>
      caa82ae7
    • Al Viro's avatar
      nios2: fix livelock in uaccess · e902e508
      Al Viro authored
      
      
      nios2 equivalent of 26178ec1 "x86: mm: consolidate VM_FAULT_RETRY handling"
      If e.g. get_user() triggers a page fault and a fatal signal is caught, we might
      end up with handle_mm_fault() returning VM_FAULT_RETRY and not doing anything
      to page tables.  In such case we must *not* return to the faulting insn -
      that would repeat the entire thing without making any progress; what we need
      instead is to treat that as failed (user) memory access.
      
      Signed-off-by: default avatarAl Viro <viro@zeniv.linux.org.uk>
      e902e508
    • Al Viro's avatar
      microblaze: fix livelock in uaccess · a1179ac7
      Al Viro authored
      
      
      microblaze equivalent of 26178ec1 "x86: mm: consolidate VM_FAULT_RETRY handling"
      If e.g. get_user() triggers a page fault and a fatal signal is caught, we might
      end up with handle_mm_fault() returning VM_FAULT_RETRY and not doing anything
      to page tables.  In such case we must *not* return to the faulting insn -
      that would repeat the entire thing without making any progress; what we need
      instead is to treat that as failed (user) memory access.
      
      Signed-off-by: default avatarAl Viro <viro@zeniv.linux.org.uk>
      a1179ac7
    • Al Viro's avatar
      ia64: fix livelock in uaccess · d088af1e
      Al Viro authored
      
      
      ia64 equivalent of 26178ec1 "x86: mm: consolidate VM_FAULT_RETRY handling"
      If e.g. get_user() triggers a page fault and a fatal signal is caught, we might
      end up with handle_mm_fault() returning VM_FAULT_RETRY and not doing anything
      to page tables.  In such case we must *not* return to the faulting insn -
      that would repeat the entire thing without making any progress; what we need
      instead is to treat that as failed (user) memory access.
      
      Signed-off-by: default avatarAl Viro <viro@zeniv.linux.org.uk>
      d088af1e
    • Al Viro's avatar
      sparc: fix livelock in uaccess · 79c54c97
      Al Viro authored
      
      
      sparc equivalent of 26178ec1 "x86: mm: consolidate VM_FAULT_RETRY handling"
      If e.g. get_user() triggers a page fault and a fatal signal is caught, we might
      end up with handle_mm_fault() returning VM_FAULT_RETRY and not doing anything
      to page tables.  In such case we must *not* return to the faulting insn -
      that would repeat the entire thing without making any progress; what we need
      instead is to treat that as failed (user) memory access.
      
      Signed-off-by: default avatarAl Viro <viro@zeniv.linux.org.uk>
      79c54c97
    • Al Viro's avatar
      alpha: fix livelock in uaccess · dce45493
      Al Viro authored
      
      
      alpha equivalent of 26178ec1 "x86: mm: consolidate VM_FAULT_RETRY handling"
      If e.g. get_user() triggers a page fault and a fatal signal is caught, we might
      end up with handle_mm_fault() returning VM_FAULT_RETRY and not doing anything
      to page tables.  In such case we must *not* return to the faulting insn -
      that would repeat the entire thing without making any progress; what we need
      instead is to treat that as failed (user) memory access.
      
      Signed-off-by: default avatarAl Viro <viro@zeniv.linux.org.uk>
      dce45493
    • Al Viro's avatar
      parisc: fix livelock in uaccess · 15261678
      Al Viro authored
      
      
      parisc equivalent of 26178ec1 "x86: mm: consolidate VM_FAULT_RETRY handling"
      If e.g. get_user() triggers a page fault and a fatal signal is caught, we might
      end up with handle_mm_fault() returning VM_FAULT_RETRY and not doing anything
      to page tables.  In such case we must *not* return to the faulting insn -
      that would repeat the entire thing without making any progress; what we need
      instead is to treat that as failed (user) memory access.
      
      Tested-by: default avatarHelge Deller <deller@gmx.de>
      Signed-off-by: default avatarAl Viro <viro@zeniv.linux.org.uk>
      15261678
    • Al Viro's avatar
      hexagon: fix livelock in uaccess · 0b92ed09
      Al Viro authored
      
      
      hexagon equivalent of 26178ec1 "x86: mm: consolidate VM_FAULT_RETRY handling"
      If e.g. get_user() triggers a page fault and a fatal signal is caught, we might
      end up with handle_mm_fault() returning VM_FAULT_RETRY and not doing anything
      to page tables.  In such case we must *not* return to the faulting insn -
      that would repeat the entire thing without making any progress; what we need
      instead is to treat that as failed (user) memory access.
      
      Acked-by: default avatarBrian Cain <bcain@quicinc.com>
      Signed-off-by: default avatarAl Viro <viro@zeniv.linux.org.uk>
      0b92ed09
    • Al Viro's avatar
      riscv: fix livelock in uaccess · d835eb3a
      Al Viro authored
      
      
      riscv equivalent of 26178ec1 "x86: mm: consolidate VM_FAULT_RETRY handling"
      If e.g. get_user() triggers a page fault and a fatal signal is caught, we might
      end up with handle_mm_fault() returning VM_FAULT_RETRY and not doing anything
      to page tables.  In such case we must *not* return to the faulting insn -
      that would repeat the entire thing without making any progress; what we need
      instead is to treat that as failed (user) memory access.
      
      Tested-by: default avatarBjörn Töpel <bjorn@kernel.org>
      Tested-by: default avatarGeert Uytterhoeven <geert+renesas@glider.be>
      Signed-off-by: default avatarAl Viro <viro@zeniv.linux.org.uk>
      d835eb3a
    • Al Viro's avatar
      m68k: fix livelock in uaccess · bd75497a
      Al Viro authored
      
      
      m68k equivalent of 26178ec1 "x86: mm: consolidate VM_FAULT_RETRY handling"
      If e.g. get_user() triggers a page fault and a fatal signal is caught, we might
      end up with handle_mm_fault() returning VM_FAULT_RETRY and not doing anything
      to page tables.  In such case we must *not* return to the faulting insn -
      that would repeat the entire thing without making any progress; what we need
      instead is to treat that as failed (user) memory access.
      
      Tested-by: default avatarFinn Thain <fthain@linux-m68k.org>
      Tested-by: default avatarGeert Uytterhoeven <geert@linux-m68k.org>
      Acked-by: default avatarGeert Uytterhoeven <geert@linux-m68k.org>
      Signed-off-by: default avatarAl Viro <viro@zeniv.linux.org.uk>
      bd75497a
    • Vasily Gorbik's avatar
      s390/kprobes: fix current_kprobe never cleared after kprobes reenter · cd579539
      Vasily Gorbik authored
      
      
      Recent test_kprobe_missed kprobes kunit test uncovers the following
      problem. Once kprobe is triggered from another kprobe (kprobe reenter),
      all future kprobes on this cpu are considered as kprobe reenter, thus
      pre_handler and post_handler are not being called and kprobes are counted
      as "missed".
      
      Commit b9599798 ("[S390] kprobes: activation and deactivation")
      introduced a simpler scheme for kprobes (de)activation and status
      tracking by using push_kprobe/pop_kprobe, which supposed to work for
      both initial kprobe entry as well as kprobe reentry and helps to avoid
      handling those two cases differently. The problem is that a sequence of
      calls in case of kprobes reenter:
      push_kprobe() <- NULL (current_kprobe)
      push_kprobe() <- kprobe1 (current_kprobe)
      pop_kprobe() -> kprobe1 (current_kprobe)
      pop_kprobe() -> kprobe1 (current_kprobe)
      leaves "kprobe1" as "current_kprobe" on this cpu, instead of setting it
      to NULL. In fact push_kprobe/pop_kprobe can only store a single state
      (there is just one prev_kprobe in kprobe_ctlblk). Which is a hack but
      sufficient, there is no need to have another prev_kprobe just to store
      NULL. To make a simple and backportable fix simply reset "prev_kprobe"
      when kprobe is poped from this "stack". No need to worry about
      "kprobe_status" in this case, because its value is only checked when
      current_kprobe != NULL.
      
      Cc: stable@vger.kernel.org
      Fixes: b9599798 ("[S390] kprobes: activation and deactivation")
      Reviewed-by: default avatarHeiko Carstens <hca@linux.ibm.com>
      Signed-off-by: default avatarVasily Gorbik <gor@linux.ibm.com>
      Signed-off-by: default avatarHeiko Carstens <hca@linux.ibm.com>
      cd579539
    • Vasily Gorbik's avatar
      s390/kprobes: fix irq mask clobbering on kprobe reenter from post_handler · 42e19e6f
      Vasily Gorbik authored
      
      
      Recent test_kprobe_missed kprobes kunit test uncovers the following error
      (reported when CONFIG_DEBUG_ATOMIC_SLEEP is enabled):
      
      BUG: sleeping function called from invalid context at kernel/locking/mutex.c:580
      in_atomic(): 0, irqs_disabled(): 1, non_block: 0, pid: 662, name: kunit_try_catch
      preempt_count: 0, expected: 0
      RCU nest depth: 0, expected: 0
      no locks held by kunit_try_catch/662.
      irq event stamp: 280
      hardirqs last  enabled at (279): [<00000003e60a3d42>] __do_pgm_check+0x17a/0x1c0
      hardirqs last disabled at (280): [<00000003e3bd774a>] kprobe_exceptions_notify+0x27a/0x318
      softirqs last  enabled at (0): [<00000003e3c5c890>] copy_process+0x14a8/0x4c80
      softirqs last disabled at (0): [<0000000000000000>] 0x0
      CPU: 46 PID: 662 Comm: kunit_try_catch Tainted: G                 N 6.2.0-173644-g44c18d77f0c0 #2
      Hardware name: IBM 3931 A01 704 (LPAR)
      Call Trace:
       [<00000003e60a3a00>] dump_stack_lvl+0x120/0x198
       [<00000003e3d02e82>] __might_resched+0x60a/0x668
       [<00000003e60b9908>] __mutex_lock+0xc0/0x14e0
       [<00000003e60bad5a>] mutex_lock_nested+0x32/0x40
       [<00000003e3f7b460>] unregister_kprobe+0x30/0xd8
       [<00000003e51b2602>] test_kprobe_missed+0xf2/0x268
       [<00000003e51b5406>] kunit_try_run_case+0x10e/0x290
       [<00000003e51b7dfa>] kunit_generic_run_threadfn_adapter+0x62/0xb8
       [<00000003e3ce30f8>] kthread+0x2d0/0x398
       [<00000003e3b96afa>] __ret_from_fork+0x8a/0xe8
       [<00000003e60ccada>] ret_from_fork+0xa/0x40
      
      The reason for this error report is that kprobes handling code failed
      to restore irqs.
      
      The problem is that when kprobe is triggered from another kprobe
      post_handler current sequence of enable_singlestep / disable_singlestep
      is the following:
      enable_singlestep  <- original kprobe (saves kprobe_saved_imask)
      enable_singlestep  <- kprobe triggered from post_handler (clobbers kprobe_saved_imask)
      disable_singlestep <- kprobe triggered from post_handler (restores kprobe_saved_imask)
      disable_singlestep <- original kprobe (restores wrong clobbered kprobe_saved_imask)
      
      There is just one kprobe_ctlblk per cpu and both calls saves and
      loads irq mask to kprobe_saved_imask. To fix the problem simply move
      resume_execution (which calls disable_singlestep) before calling
      post_handler. This also fixes the problem that post_handler is called
      with pt_regs which were not yet adjusted after single-stepping.
      
      Cc: stable@vger.kernel.org
      Fixes: 4ba069b8 ("[S390] add kprobes support.")
      Reviewed-by: default avatarHeiko Carstens <hca@linux.ibm.com>
      Signed-off-by: default avatarVasily Gorbik <gor@linux.ibm.com>
      Signed-off-by: default avatarHeiko Carstens <hca@linux.ibm.com>
      42e19e6f
    • Alexandre Ghiti's avatar
      riscv: Bump COMMAND_LINE_SIZE value to 1024 · 61fc1ee8
      Alexandre Ghiti authored
      
      
      Increase COMMAND_LINE_SIZE as the current default value is too low
      for syzbot kernel command line.
      
      There has been considerable discussion on this patch that has led to a
      larger patch set removing COMMAND_LINE_SIZE from the uapi headers on all
      ports.  That's not quite done yet, but it's gotten far enough we're
      confident this is not a uABI change so this is safe.
      
      Reported-by: default avatarDmitry Vyukov <dvyukov@google.com>
      Signed-off-by: default avatarAlexandre Ghiti <alex@ghiti.fr>
      Link: https://lore.kernel.org/r/20210316193420.904-1-alex@ghiti.fr
      [Palmer: it's not uabi]
      Link: https://lore.kernel.org/linux-riscv/874b8076-b0d1-4aaa-bcd8-05d523060152@app.fastmail.com/#t
      
      
      Signed-off-by: default avatarPalmer Dabbelt <palmer@rivosinc.com>
      61fc1ee8
  4. Mar 01, 2023
  5. Feb 28, 2023
  6. Feb 27, 2023
Loading