* incoming
@ 2021-06-29 2:32 Andrew Morton
2021-06-29 2:33 ` [patch 001/192] mm/gup: fix try_grab_compound_head() race with split_huge_page() Andrew Morton
` (191 more replies)
0 siblings, 192 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:32 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
192 patches, based on 7cf3dead1ad70c72edb03e2d98e1f3dcd332cdb2.
Subsystems affected by this patch series:
mm/gup
mm/pagealloc
kthread
ia64
scripts
ntfs
squashfs
ocfs2
z
kernel/watchdog
mm/slab
mm/slub
mm/kmemleak
mm/dax
mm/debug
mm/pagecache
mm/gup
mm/swap
mm/memcg
mm/pagemap
mm/mprotect
mm/bootmem
mm/dma
mm/tracing
mm/vmalloc
mm/kasan
mm/initialization
mm/pagealloc
mm/memory-failure
Subsystem: mm/gup
Jann Horn <jannh@google.com>:
mm/gup: fix try_grab_compound_head() race with split_huge_page()
Subsystem: mm/pagealloc
Mike Rapoport <rppt@linux.ibm.com>:
mm/page_alloc: fix memory map initialization for descending nodes
Mel Gorman <mgorman@techsingularity.net>:
mm/page_alloc: correct return value of populated elements if bulk array is populated
Subsystem: kthread
Jonathan Neuschäfer <j.neuschaefer@gmx.net>:
kthread: switch to new kerneldoc syntax for named variable macro argument
Petr Mladek <pmladek@suse.com>:
kthread_worker: fix return value when kthread_mod_delayed_work() races with kthread_cancel_delayed_work_sync()
Subsystem: ia64
Randy Dunlap <rdunlap@infradead.org>:
ia64: headers: drop duplicated words
Arnd Bergmann <arnd@arndb.de>:
ia64: mca_drv: fix incorrect array size calculation
Subsystem: scripts
"Steven Rostedt (VMware)" <rostedt@goodmis.org>:
Patch series "streamline_config.pl: Fix Perl spacing":
streamline_config.pl: make spacing consistent
streamline_config.pl: add softtabstop=4 for vim users
Colin Ian King <colin.king@canonical.com>:
scripts/spelling.txt: add more spellings to spelling.txt
Subsystem: ntfs
Desmond Cheong Zhi Xi <desmondcheongzx@gmail.com>:
ntfs: fix validity check for file name attribute
Subsystem: squashfs
Vincent Whitchurch <vincent.whitchurch@axis.com>:
squashfs: add option to panic on errors
Subsystem: ocfs2
Yang Yingliang <yangyingliang@huawei.com>:
ocfs2: remove unnecessary INIT_LIST_HEAD()
Subsystem: z
Dan Carpenter <dan.carpenter@oracle.com>:
ocfs2: fix snprintf() checking
Colin Ian King <colin.king@canonical.com>:
ocfs2: remove redundant assignment to pointer queue
Wan Jiabing <wanjiabing@vivo.com>:
ocfs2: remove repeated uptodate check for buffer
Chen Huang <chenhuang5@huawei.com>:
ocfs2: replace simple_strtoull() with kstrtoull()
Colin Ian King <colin.king@canonical.com>:
ocfs2: remove redundant initialization of variable ret
Subsystem: kernel/watchdog
Wang Qing <wangqing@vivo.com>:
kernel: watchdog: modify the explanation related to watchdog thread
doc: watchdog: modify the explanation related to watchdog thread
doc: watchdog: modify the doc related to "watchdog/%u"
Subsystem: mm/slab
gumingtao <gumingtao1225@gmail.com>:
slab: use __func__ to trace function name
Subsystem: mm/slub
Vlastimil Babka <vbabka@suse.cz>:
kunit: make test->lock irq safe
Oliver Glitta <glittao@gmail.com>:
mm/slub, kunit: add a KUnit test for SLUB debugging functionality
slub: remove resiliency_test() function
Hyeonggon Yoo <42.hyeyoo@gmail.com>:
mm, slub: change run-time assertion in kmalloc_index() to compile-time
Stephen Boyd <swboyd@chromium.org>:
slub: restore slub_debug=- behavior
slub: actually use 'message' in restore_bytes()
Joe Perches <joe@perches.com>:
slub: indicate slab_fix() uses printf formats
Stephen Boyd <swboyd@chromium.org>:
slub: force on no_hash_pointers when slub_debug is enabled
Faiyaz Mohammed <faiyazm@codeaurora.org>:
mm: slub: move sysfs slab alloc/free interfaces to debugfs
Georgi Djakov <quic_c_gdjako@quicinc.com>:
mm/slub: add taint after the errors are printed
Subsystem: mm/kmemleak
Yanfei Xu <yanfei.xu@windriver.com>:
mm/kmemleak: fix possible wrong memory scanning period
Subsystem: mm/dax
Jan Kara <jack@suse.cz>:
dax: fix ENOMEM handling in grab_mapping_entry()
Subsystem: mm/debug
Tang Bin <tangbin@cmss.chinamobile.com>:
tools/vm/page_owner_sort.c: check malloc() return
Anshuman Khandual <anshuman.khandual@arm.com>:
mm/debug_vm_pgtable: ensure THP availability via has_transparent_hugepage()
Nicolas Saenz Julienne <nsaenzju@redhat.com>:
mm: mmap_lock: use local locks instead of disabling preemption
Gavin Shan <gshan@redhat.com>:
Patch series "mm/page_reporting: Make page reporting work on arm64 with 64KB page size", v4:
mm/page_reporting: fix code style in __page_reporting_request()
mm/page_reporting: export reporting order as module parameter
mm/page_reporting: allow driver to specify reporting order
virtio_balloon: specify page reporting order if needed
Subsystem: mm/pagecache
Kefeng Wang <wangkefeng.wang@huawei.com>:
mm: page-writeback: kill get_writeback_state() comments
Chi Wu <wuchi.zero@gmail.com>:
mm/page-writeback: Fix performance when BDI's share of ratio is 0.
mm/page-writeback: update the comment of Dirty position control
mm/page-writeback: use __this_cpu_inc() in account_page_dirtied()
Roman Gushchin <guro@fb.com>:
Patch series "cgroup, blkcg: prevent dirty inodes to pin dying memory cgroups", v9:
writeback, cgroup: do not switch inodes with I_WILL_FREE flag
writeback, cgroup: add smp_mb() to cgroup_writeback_umount()
writeback, cgroup: increment isw_nr_in_flight before grabbing an inode
writeback, cgroup: switch to rcu_work API in inode_switch_wbs()
writeback, cgroup: keep list of inodes attached to bdi_writeback
writeback, cgroup: split out the functional part of inode_switch_wbs_work_fn()
writeback, cgroup: support switching multiple inodes at once
writeback, cgroup: release dying cgwbs by switching attached inodes
Christoph Hellwig <hch@lst.de>:
Patch series "remove the implicit .set_page_dirty default":
fs: unexport __set_page_dirty
fs: move ramfs_aops to libfs
mm: require ->set_page_dirty to be explicitly wired up
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
Patch series "Further set_page_dirty cleanups":
mm/writeback: move __set_page_dirty() to core mm
mm/writeback: use __set_page_dirty in __set_page_dirty_nobuffers
iomap: use __set_page_dirty_nobuffers
fs: remove anon_set_page_dirty()
fs: remove noop_set_page_dirty()
mm: move page dirtying prototypes from mm.h
Subsystem: mm/gup
Peter Xu <peterx@redhat.com>:
Patch series "mm/gup: Fix pin page write cache bouncing on has_pinned", v2:
mm/gup_benchmark: support threading
Andrea Arcangeli <aarcange@redhat.com>:
mm: gup: allow FOLL_PIN to scale in SMP
mm: gup: pack has_pinned in MMF_HAS_PINNED
Christophe Leroy <christophe.leroy@csgroup.eu>:
mm: pagewalk: fix walk for hugepage tables
Subsystem: mm/swap
Miaohe Lin <linmiaohe@huawei.com>:
Patch series "close various race windows for swap", v6:
mm/swapfile: use percpu_ref to serialize against concurrent swapoff
swap: fix do_swap_page() race with swapoff
mm/swap: remove confusing checking for non_swap_entry() in swap_ra_info()
mm/shmem: fix shmem_swapin() race with swapoff
Patch series "Cleanups for swap", v2:
mm/swapfile: move get_swap_page_of_type() under CONFIG_HIBERNATION
mm/swap: remove unused local variable nr_shadows
mm/swap_slots.c: delete meaningless forward declarations
Huang Ying <ying.huang@intel.com>:
mm, swap: remove unnecessary smp_rmb() in swap_type_to_swap_info()
mm: free idle swap cache page after COW
swap: check mapping_empty() for swap cache before being freed
Subsystem: mm/memcg
Waiman Long <longman@redhat.com>:
Patch series "mm/memcg: Reduce kmemcache memory accounting overhead", v6:
mm/memcg: move mod_objcg_state() to memcontrol.c
mm/memcg: cache vmstat data in percpu memcg_stock_pcp
mm/memcg: improve refill_obj_stock() performance
mm/memcg: optimize user context object stock access
Patch series "mm: memcg/slab: Fix objcg pointer array handling problem", v4:
mm: memcg/slab: properly set up gfp flags for objcg pointer array
mm: memcg/slab: create a new set of kmalloc-cg-<n> caches
mm: memcg/slab: disable cache merging for KMALLOC_NORMAL caches
Muchun Song <songmuchun@bytedance.com>:
mm: memcontrol: fix root_mem_cgroup charging
Patch series "memcontrol code cleanup and simplification", v3:
mm: memcontrol: fix page charging in page replacement
mm: memcontrol: bail out early when !mm in get_mem_cgroup_from_mm
mm: memcontrol: remove the pgdata parameter of mem_cgroup_page_lruvec
mm: memcontrol: simplify lruvec_holds_page_lru_lock
mm: memcontrol: rename lruvec_holds_page_lru_lock to page_matches_lruvec
mm: memcontrol: simplify the logic of objcg pinning memcg
mm: memcontrol: move obj_cgroup_uncharge_pages() out of css_set_lock
mm: vmscan: remove noinline_for_stack
wenhuizhang <wenhui@gwmail.gwu.edu>:
memcontrol: use flexible-array member
Dan Schatzberg <schatzberg.dan@gmail.com>:
Patch series "Charge loop device i/o to issuing cgroup", v14:
loop: use worker per cgroup instead of kworker
mm: charge active memcg when no mm is set
loop: charge i/o to mem and blk cg
Huilong Deng <denghuilong@cdjrlc.com>:
mm: memcontrol: remove trailing semicolon in macros
Subsystem: mm/pagemap
David Hildenbrand <david@redhat.com>:
Patch series "perf/binfmt/mm: remove in-tree usage of MAP_EXECUTABLE":
perf: MAP_EXECUTABLE does not indicate VM_MAYEXEC
binfmt: remove in-tree usage of MAP_EXECUTABLE
mm: ignore MAP_EXECUTABLE in ksys_mmap_pgoff()
Gonzalo Matias Juarez Tello <gmjuareztello@gmail.com>:
mm/mmap.c: logic of find_vma_intersection repeated in __do_munmap
Liam Howlett <liam.howlett@oracle.com>:
mm/mmap: introduce unlock_range() for code cleanup
mm/mmap: use find_vma_intersection() in do_mmap() for overlap
Liu Xiang <liu.xiang@zlingsmart.com>:
mm/memory.c: fix comment of finish_mkwrite_fault()
Liam Howlett <liam.howlett@oracle.com>:
Patch series "mm: Add vma_lookup()", v2:
mm: add vma_lookup(), update find_vma_intersection() comments
drm/i915/selftests: use vma_lookup() in __igt_mmap()
arch/arc/kernel/troubleshoot: use vma_lookup() instead of find_vma()
arch/arm64/kvm: use vma_lookup() instead of find_vma_intersection()
arch/powerpc/kvm/book3s_hv_uvmem: use vma_lookup() instead of find_vma_intersection()
arch/powerpc/kvm/book3s: use vma_lookup() in kvmppc_hv_setup_htab_rma()
arch/mips/kernel/traps: use vma_lookup() instead of find_vma()
arch/m68k/kernel/sys_m68k: use vma_lookup() in sys_cacheflush()
x86/sgx: use vma_lookup() in sgx_encl_find()
virt/kvm: use vma_lookup() instead of find_vma_intersection()
vfio: use vma_lookup() instead of find_vma_intersection()
net/ipv5/tcp: use vma_lookup() in tcp_zerocopy_receive()
drm/amdgpu: use vma_lookup() in amdgpu_ttm_tt_get_user_pages()
media: videobuf2: use vma_lookup() in get_vaddr_frames()
misc/sgi-gru/grufault: use vma_lookup() in gru_find_vma()
kernel/events/uprobes: use vma_lookup() in find_active_uprobe()
lib/test_hmm: use vma_lookup() in dmirror_migrate()
mm/ksm: use vma_lookup() in find_mergeable_vma()
mm/migrate: use vma_lookup() in do_pages_stat_array()
mm/mremap: use vma_lookup() in vma_to_resize()
mm/memory.c: use vma_lookup() in __access_remote_vm()
mm/mempolicy: use vma_lookup() in __access_remote_vm()
Chen Li <chenli@uniontech.com>:
mm: update legacy flush_tlb_* to use vma
Subsystem: mm/mprotect
Peter Collingbourne <pcc@google.com>:
mm: improve mprotect(R|W) efficiency on pages referenced once
Subsystem: mm/bootmem
Souptick Joarder <jrdr.linux@gmail.com>:
h8300: remove unused variable
Subsystem: mm/dma
YueHaibing <yuehaibing@huawei.com>:
mm/dmapool: use DEVICE_ATTR_RO macro
Subsystem: mm/tracing
Vincent Whitchurch <vincent.whitchurch@axis.com>:
mm, tracing: unify PFN format strings
Subsystem: mm/vmalloc
"Uladzislau Rezki (Sony)" <urezki@gmail.com>:
Patch series "vmalloc() vs bulk allocator", v2:
mm/page_alloc: add an alloc_pages_bulk_array_node() helper
mm/vmalloc: switch to bulk allocator in __vmalloc_area_node()
mm/vmalloc: print a warning message first on failure
mm/vmalloc: remove quoted strings split across lines
Uladzislau Rezki <urezki@gmail.com>:
mm/vmalloc: fallback to a single page allocator
Rafael Aquini <aquini@redhat.com>:
mm: vmalloc: add cond_resched() in __vunmap()
Subsystem: mm/kasan
Alexander Potapenko <glider@google.com>:
printk: introduce dump_stack_lvl()
kasan: use dump_stack_lvl(KERN_ERR) to print stacks
David Gow <davidgow@google.com>:
kasan: test: improve failure message in KUNIT_EXPECT_KASAN_FAIL()
Daniel Axtens <dja@axtens.net>:
Patch series "KASAN core changes for ppc64 radix KASAN", v16:
kasan: allow an architecture to disable inline instrumentation
kasan: allow architectures to provide an outline readiness check
mm: define default MAX_PTRS_PER_* in include/pgtable.h
kasan: use MAX_PTRS_PER_* for early shadow tables
Kuan-Ying Lee <Kuan-Ying.Lee@mediatek.com>:
Patch series "kasan: add memory corruption identification support for hw tag-based kasan", v4:
kasan: rename CONFIG_KASAN_SW_TAGS_IDENTIFY to CONFIG_KASAN_TAGS_IDENTIFY
kasan: integrate the common part of two KASAN tag-based modes
kasan: add memory corruption identification support for hardware tag-based mode
Subsystem: mm/initialization
Jungseung Lee <js07.lee@samsung.com>:
mm: report which part of mem is being freed on initmem case
Subsystem: mm/pagealloc
Mike Rapoport <rppt@linux.ibm.com>:
mm/mmzone.h: simplify is_highmem_idx()
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
Patch series "Constify struct page arguments":
mm: make __dump_page static
Aaron Tomlin <atomlin@redhat.com>:
mm/page_alloc: bail out on fatal signal during reclaim/compaction retry attempt
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
mm/debug: factor PagePoisoned out of __dump_page
mm/page_owner: constify dump_page_owner
mm: make compound_head const-preserving
mm: constify get_pfnblock_flags_mask and get_pfnblock_migratetype
mm: constify page_count and page_ref_count
mm: optimise nth_page for contiguous memmap
Heiner Kallweit <hkallweit1@gmail.com>:
mm/page_alloc: switch to pr_debug
Andrii Nakryiko <andrii@kernel.org>:
kbuild: skip per-CPU BTF generation for pahole v1.18-v1.21
Mel Gorman <mgorman@techsingularity.net>:
mm/page_alloc: split per cpu page lists and zone stats
mm/page_alloc: convert per-cpu list protection to local_lock
mm/vmstat: convert NUMA statistics to basic NUMA counters
mm/vmstat: inline NUMA event counter updates
mm/page_alloc: batch the accounting updates in the bulk allocator
mm/page_alloc: reduce duration that IRQs are disabled for VM counters
mm/page_alloc: explicitly acquire the zone lock in __free_pages_ok
mm/page_alloc: avoid conflating IRQs disabled with zone->lock
mm/page_alloc: update PGFREE outside the zone lock in __free_pages_ok
Minchan Kim <minchan@kernel.org>:
mm: page_alloc: dump migrate-failed pages only at -EBUSY
Mel Gorman <mgorman@techsingularity.net>:
Patch series "Calculate pcp->high based on zone sizes and active CPUs", v2:
mm/page_alloc: delete vm.percpu_pagelist_fraction
mm/page_alloc: disassociate the pcp->high from pcp->batch
mm/page_alloc: adjust pcp->high after CPU hotplug events
mm/page_alloc: scale the number of pages that are batch freed
mm/page_alloc: limit the number of pages on PCP lists when reclaim is active
mm/page_alloc: introduce vm.percpu_pagelist_high_fraction
Dong Aisheng <aisheng.dong@nxp.com>:
mm: drop SECTION_SHIFT in code comments
mm/page_alloc: improve memmap_pages dbg msg
Liu Shixin <liushixin2@huawei.com>:
mm/page_alloc: fix counting of managed_pages
Mel Gorman <mgorman@techsingularity.net>:
Patch series "Allow high order pages to be stored on PCP", v2:
mm/page_alloc: move free_the_page
Mike Rapoport <rppt@linux.ibm.com>:
Patch series "Remove DISCONTIGMEM memory model", v3:
alpha: remove DISCONTIGMEM and NUMA
arc: update comment about HIGHMEM implementation
arc: remove support for DISCONTIGMEM
m68k: remove support for DISCONTIGMEM
mm: remove CONFIG_DISCONTIGMEM
arch, mm: remove stale mentions of DISCONIGMEM
docs: remove description of DISCONTIGMEM
mm: replace CONFIG_NEED_MULTIPLE_NODES with CONFIG_NUMA
mm: replace CONFIG_FLAT_NODE_MEM_MAP with CONFIG_FLATMEM
Mel Gorman <mgorman@techsingularity.net>:
mm/page_alloc: allow high-order pages to be stored on the per-cpu lists
mm/page_alloc: split pcp->high across all online CPUs for cpuless nodes
Subsystem: mm/memory-failure
Naoya Horiguchi <naoya.horiguchi@nec.com>:
mm,hwpoison: send SIGBUS with error virutal address
mm,hwpoison: make get_hwpoison_page() call get_any_page()
Documentation/admin-guide/kernel-parameters.txt | 6
Documentation/admin-guide/lockup-watchdogs.rst | 4
Documentation/admin-guide/sysctl/kernel.rst | 10
Documentation/admin-guide/sysctl/vm.rst | 52 -
Documentation/dev-tools/kasan.rst | 9
Documentation/vm/memory-model.rst | 45
arch/alpha/Kconfig | 22
arch/alpha/include/asm/machvec.h | 6
arch/alpha/include/asm/mmzone.h | 100 --
arch/alpha/include/asm/pgtable.h | 4
arch/alpha/include/asm/topology.h | 39
arch/alpha/kernel/core_marvel.c | 53 -
arch/alpha/kernel/core_wildfire.c | 29
arch/alpha/kernel/pci_iommu.c | 29
arch/alpha/kernel/proto.h | 8
arch/alpha/kernel/setup.c | 16
arch/alpha/kernel/sys_marvel.c | 5
arch/alpha/kernel/sys_wildfire.c | 5
arch/alpha/mm/Makefile | 2
arch/alpha/mm/init.c | 3
arch/alpha/mm/numa.c | 223 ----
arch/arc/Kconfig | 13
arch/arc/include/asm/mmzone.h | 40
arch/arc/kernel/troubleshoot.c | 8
arch/arc/mm/init.c | 21
arch/arm/include/asm/tlbflush.h | 13
arch/arm/mm/tlb-v6.S | 2
arch/arm/mm/tlb-v7.S | 2
arch/arm64/Kconfig | 2
arch/arm64/kvm/mmu.c | 2
arch/h8300/kernel/setup.c | 2
arch/ia64/Kconfig | 2
arch/ia64/include/asm/pal.h | 2
arch/ia64/include/asm/spinlock.h | 2
arch/ia64/include/asm/uv/uv_hub.h | 2
arch/ia64/kernel/efi_stub.S | 2
arch/ia64/kernel/mca_drv.c | 2
arch/ia64/kernel/topology.c | 5
arch/ia64/mm/numa.c | 5
arch/m68k/Kconfig.cpu | 10
arch/m68k/include/asm/mmzone.h | 10
arch/m68k/include/asm/page.h | 2
arch/m68k/include/asm/page_mm.h | 35
arch/m68k/include/asm/tlbflush.h | 2
arch/m68k/kernel/sys_m68k.c | 4
arch/m68k/mm/init.c | 20
arch/mips/Kconfig | 2
arch/mips/include/asm/mmzone.h | 8
arch/mips/include/asm/page.h | 2
arch/mips/kernel/traps.c | 4
arch/mips/mm/init.c | 7
arch/nds32/include/asm/memory.h | 6
arch/openrisc/include/asm/tlbflush.h | 2
arch/powerpc/Kconfig | 2
arch/powerpc/include/asm/mmzone.h | 4
arch/powerpc/kernel/setup_64.c | 2
arch/powerpc/kernel/smp.c | 2
arch/powerpc/kexec/core.c | 4
arch/powerpc/kvm/book3s_hv.c | 4
arch/powerpc/kvm/book3s_hv_uvmem.c | 2
arch/powerpc/mm/Makefile | 2
arch/powerpc/mm/mem.c | 4
arch/riscv/Kconfig | 2
arch/s390/Kconfig | 2
arch/s390/include/asm/pgtable.h | 2
arch/sh/include/asm/mmzone.h | 4
arch/sh/kernel/topology.c | 2
arch/sh/mm/Kconfig | 2
arch/sh/mm/init.c | 2
arch/sparc/Kconfig | 2
arch/sparc/include/asm/mmzone.h | 4
arch/sparc/kernel/smp_64.c | 2
arch/sparc/mm/init_64.c | 12
arch/x86/Kconfig | 2
arch/x86/ia32/ia32_aout.c | 4
arch/x86/kernel/cpu/mce/core.c | 13
arch/x86/kernel/cpu/sgx/encl.h | 4
arch/x86/kernel/setup_percpu.c | 6
arch/x86/mm/init_32.c | 4
arch/xtensa/include/asm/page.h | 4
arch/xtensa/include/asm/tlbflush.h | 4
drivers/base/node.c | 18
drivers/block/loop.c | 270 ++++-
drivers/block/loop.h | 15
drivers/dax/device.c | 2
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c | 4
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c | 2
drivers/media/common/videobuf2/frame_vector.c | 2
drivers/misc/sgi-gru/grufault.c | 4
drivers/vfio/vfio_iommu_type1.c | 2
drivers/virtio/virtio_balloon.c | 17
fs/adfs/inode.c | 1
fs/affs/file.c | 2
fs/bfs/file.c | 1
fs/binfmt_aout.c | 4
fs/binfmt_elf.c | 2
fs/binfmt_elf_fdpic.c | 11
fs/binfmt_flat.c | 2
fs/block_dev.c | 1
fs/buffer.c | 25
fs/configfs/inode.c | 8
fs/dax.c | 3
fs/ecryptfs/mmap.c | 13
fs/exfat/inode.c | 1
fs/ext2/inode.c | 4
fs/ext4/inode.c | 2
fs/fat/inode.c | 1
fs/fs-writeback.c | 366 +++++---
fs/fuse/dax.c | 3
fs/gfs2/aops.c | 2
fs/gfs2/meta_io.c | 2
fs/hfs/inode.c | 2
fs/hfsplus/inode.c | 2
fs/hpfs/file.c | 1
fs/iomap/buffered-io.c | 27
fs/jfs/inode.c | 1
fs/kernfs/inode.c | 8
fs/libfs.c | 44
fs/minix/inode.c | 1
fs/nilfs2/mdt.c | 1
fs/ntfs/inode.c | 2
fs/ocfs2/aops.c | 4
fs/ocfs2/cluster/heartbeat.c | 7
fs/ocfs2/cluster/nodemanager.c | 2
fs/ocfs2/dlm/dlmmaster.c | 2
fs/ocfs2/filecheck.c | 6
fs/ocfs2/stackglue.c | 8
fs/omfs/file.c | 1
fs/proc/task_mmu.c | 2
fs/ramfs/inode.c | 9
fs/squashfs/block.c | 5
fs/squashfs/squashfs_fs_sb.h | 1
fs/squashfs/super.c | 86 +
fs/sysv/itree.c | 1
fs/udf/file.c | 1
fs/udf/inode.c | 1
fs/ufs/inode.c | 1
fs/xfs/xfs_aops.c | 4
fs/zonefs/super.c | 4
include/asm-generic/memory_model.h | 37
include/asm-generic/pgtable-nop4d.h | 1
include/asm-generic/topology.h | 2
include/kunit/test.h | 5
include/linux/backing-dev-defs.h | 20
include/linux/cpuhotplug.h | 2
include/linux/fs.h | 6
include/linux/gfp.h | 13
include/linux/iomap.h | 1
include/linux/kasan.h | 7
include/linux/kernel.h | 2
include/linux/kthread.h | 2
include/linux/memblock.h | 6
include/linux/memcontrol.h | 60 -
include/linux/mm.h | 53 -
include/linux/mm_types.h | 10
include/linux/mman.h | 2
include/linux/mmdebug.h | 3
include/linux/mmzone.h | 96 +-
include/linux/page-flags.h | 10
include/linux/page_owner.h | 6
include/linux/page_ref.h | 4
include/linux/page_reporting.h | 3
include/linux/pageblock-flags.h | 2
include/linux/pagemap.h | 4
include/linux/pgtable.h | 22
include/linux/printk.h | 5
include/linux/sched/coredump.h | 8
include/linux/slab.h | 59 +
include/linux/swap.h | 19
include/linux/swapops.h | 5
include/linux/vmstat.h | 69 -
include/linux/writeback.h | 1
include/trace/events/cma.h | 4
include/trace/events/filemap.h | 2
include/trace/events/kmem.h | 12
include/trace/events/page_pool.h | 4
include/trace/events/pagemap.h | 4
include/trace/events/vmscan.h | 2
kernel/cgroup/cgroup.c | 1
kernel/crash_core.c | 4
kernel/events/core.c | 2
kernel/events/uprobes.c | 4
kernel/fork.c | 1
kernel/kthread.c | 19
kernel/sysctl.c | 16
kernel/watchdog.c | 12
lib/Kconfig.debug | 15
lib/Kconfig.kasan | 16
lib/Makefile | 1
lib/dump_stack.c | 20
lib/kunit/test.c | 18
lib/slub_kunit.c | 152 +++
lib/test_hmm.c | 5
lib/test_kasan.c | 11
lib/vsprintf.c | 2
mm/Kconfig | 38
mm/backing-dev.c | 66 +
mm/compaction.c | 2
mm/debug.c | 27
mm/debug_vm_pgtable.c | 63 +
mm/dmapool.c | 5
mm/filemap.c | 2
mm/gup.c | 81 +
mm/hugetlb.c | 2
mm/internal.h | 9
mm/kasan/Makefile | 4
mm/kasan/common.c | 6
mm/kasan/generic.c | 3
mm/kasan/hw_tags.c | 22
mm/kasan/init.c | 6
mm/kasan/kasan.h | 12
mm/kasan/report.c | 6
mm/kasan/report_hw_tags.c | 5
mm/kasan/report_sw_tags.c | 45
mm/kasan/report_tags.c | 51 +
mm/kasan/shadow.c | 6
mm/kasan/sw_tags.c | 45
mm/kasan/tags.c | 59 +
mm/kfence/kfence_test.c | 5
mm/kmemleak.c | 18
mm/ksm.c | 6
mm/memblock.c | 8
mm/memcontrol.c | 385 ++++++--
mm/memory-failure.c | 344 +++++--
mm/memory.c | 22
mm/memory_hotplug.c | 6
mm/mempolicy.c | 4
mm/migrate.c | 4
mm/mmap.c | 54 -
mm/mmap_lock.c | 33
mm/mprotect.c | 52 +
mm/mremap.c | 5
mm/nommu.c | 2
mm/page-writeback.c | 89 +
mm/page_alloc.c | 950 +++++++++++++--------
mm/page_ext.c | 2
mm/page_owner.c | 2
mm/page_reporting.c | 19
mm/page_reporting.h | 5
mm/pagewalk.c | 58 +
mm/shmem.c | 18
mm/slab.h | 24
mm/slab_common.c | 60 -
mm/slub.c | 420 +++++----
mm/sparse.c | 2
mm/swap.c | 4
mm/swap_slots.c | 2
mm/swap_state.c | 20
mm/swapfile.c | 177 +--
mm/vmalloc.c | 181 ++--
mm/vmscan.c | 43
mm/vmstat.c | 282 ++----
mm/workingset.c | 2
net/ipv4/tcp.c | 4
scripts/kconfig/streamline_config.pl | 76 -
scripts/link-vmlinux.sh | 4
scripts/spelling.txt | 16
tools/testing/selftests/vm/gup_test.c | 96 +-
tools/vm/page_owner_sort.c | 4
virt/kvm/kvm_main.c | 2
260 files changed, 3989 insertions(+), 2996 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 001/192] mm/gup: fix try_grab_compound_head() race with split_huge_page()
2021-06-29 2:32 incoming Andrew Morton
@ 2021-06-29 2:33 ` Andrew Morton
2021-06-29 2:33 ` [patch 002/192] mm/page_alloc: fix memory map initialization for descending nodes Andrew Morton
` (190 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:33 UTC (permalink / raw)
To: akpm, jack, jannh, jhubbard, kirill, linux-mm, mm-commits,
stable, torvalds, willy
From: Jann Horn <jannh@google.com>
Subject: mm/gup: fix try_grab_compound_head() race with split_huge_page()
try_grab_compound_head() is used to grab a reference to a page from
get_user_pages_fast(), which is only protected against concurrent freeing
of page tables (via local_irq_save()), but not against concurrent TLB
flushes, freeing of data pages, or splitting of compound pages.
Because no reference is held to the page when try_grab_compound_head() is
called, the page may have been freed and reallocated by the time its
refcount has been elevated; therefore, once we're holding a stable
reference to the page, the caller re-checks whether the PTE still points
to the same page (with the same access rights).
The problem is that try_grab_compound_head() has to grab a reference on
the head page; but between the time we look up what the head page is and
the time we actually grab a reference on the head page, the compound page
may have been split up (either explicitly through split_huge_page() or by
freeing the compound page to the buddy allocator and then allocating its
individual order-0 pages). If that happens, get_user_pages_fast() may end
up returning the right page but lifting the refcount on a now-unrelated
page, leading to use-after-free of pages.
To fix it: Re-check whether the pages still belong together after lifting
the refcount on the head page. Move anything else that checks
compound_head(page) below the refcount increment.
This can't actually happen on bare-metal x86 (because there, disabling
IRQs locks out remote TLB flushes), but it can happen on virtualized x86
(e.g. under KVM) and probably also on arm64. The race window is pretty
narrow, and constantly allocating and shattering hugepages isn't exactly
fast; for now I've only managed to reproduce this in an x86 KVM guest with
an artificially widened timing window (by adding a loop that repeatedly
calls `inl(0x3f8 + 5)` in `try_get_compound_head()` to force VM exits, so
that PV TLB flushes are used instead of IPIs).
As requested on the list, also replace the existing VM_BUG_ON_PAGE() with
a warning and bailout. Since the existing code only performed the BUG_ON
check on DEBUG_VM kernels, ensure that the new code also only performs the
check under that configuration - I don't want to mix two logically
separate changes together too much. The macro VM_WARN_ON_ONCE_PAGE()
doesn't return a value on !DEBUG_VM, so wrap the whole check in an #ifdef
block. An alternative would be to change the VM_WARN_ON_ONCE_PAGE()
definition for !DEBUG_VM such that it always returns false, but since that
would differ from the behavior of the normal WARN macros, it might be too
confusing for readers.
Link: https://lkml.kernel.org/r/20210615012014.1100672-1-jannh@google.com
Fixes: 7aef4172c795 ("mm: handle PTE-mapped tail pages in gerneric fast gup implementaiton")
Signed-off-by: Jann Horn <jannh@google.com>
Reviewed-by: John Hubbard <jhubbard@nvidia.com>
Cc: Matthew Wilcox <willy@infradead.org>
Cc: Kirill A. Shutemov <kirill@shutemov.name>
Cc: Jan Kara <jack@suse.cz>
Cc: <stable@vger.kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/gup.c | 58 +++++++++++++++++++++++++++++++++++++++--------------
1 file changed, 43 insertions(+), 15 deletions(-)
--- a/mm/gup.c~mm-gup-fix-try_grab_compound_head-race-with-split_huge_page
+++ a/mm/gup.c
@@ -44,6 +44,23 @@ static void hpage_pincount_sub(struct pa
atomic_sub(refs, compound_pincount_ptr(page));
}
+/* Equivalent to calling put_page() @refs times. */
+static void put_page_refs(struct page *page, int refs)
+{
+#ifdef CONFIG_DEBUG_VM
+ if (VM_WARN_ON_ONCE_PAGE(page_ref_count(page) < refs, page))
+ return;
+#endif
+
+ /*
+ * Calling put_page() for each ref is unnecessarily slow. Only the last
+ * ref needs a put_page().
+ */
+ if (refs > 1)
+ page_ref_sub(page, refs - 1);
+ put_page(page);
+}
+
/*
* Return the compound head page with ref appropriately incremented,
* or NULL if that failed.
@@ -56,6 +73,21 @@ static inline struct page *try_get_compo
return NULL;
if (unlikely(!page_cache_add_speculative(head, refs)))
return NULL;
+
+ /*
+ * At this point we have a stable reference to the head page; but it
+ * could be that between the compound_head() lookup and the refcount
+ * increment, the compound page was split, in which case we'd end up
+ * holding a reference on a page that has nothing to do with the page
+ * we were given anymore.
+ * So now that the head page is stable, recheck that the pages still
+ * belong together.
+ */
+ if (unlikely(compound_head(page) != head)) {
+ put_page_refs(head, refs);
+ return NULL;
+ }
+
return head;
}
@@ -96,6 +128,14 @@ __maybe_unused struct page *try_grab_com
return NULL;
/*
+ * CAUTION: Don't use compound_head() on the page before this
+ * point, the result won't be stable.
+ */
+ page = try_get_compound_head(page, refs);
+ if (!page)
+ return NULL;
+
+ /*
* When pinning a compound page of order > 1 (which is what
* hpage_pincount_available() checks for), use an exact count to
* track it, via hpage_pincount_add/_sub().
@@ -103,15 +143,10 @@ __maybe_unused struct page *try_grab_com
* However, be sure to *also* increment the normal page refcount
* field at least once, so that the page really is pinned.
*/
- if (!hpage_pincount_available(page))
- refs *= GUP_PIN_COUNTING_BIAS;
-
- page = try_get_compound_head(page, refs);
- if (!page)
- return NULL;
-
if (hpage_pincount_available(page))
hpage_pincount_add(page, refs);
+ else
+ page_ref_add(page, refs * (GUP_PIN_COUNTING_BIAS - 1));
mod_node_page_state(page_pgdat(page), NR_FOLL_PIN_ACQUIRED,
orig_refs);
@@ -135,14 +170,7 @@ static void put_compound_head(struct pag
refs *= GUP_PIN_COUNTING_BIAS;
}
- VM_BUG_ON_PAGE(page_ref_count(page) < refs, page);
- /*
- * Calling put_page() for each ref is unnecessarily slow. Only the last
- * ref needs a put_page().
- */
- if (refs > 1)
- page_ref_sub(page, refs - 1);
- put_page(page);
+ put_page_refs(page, refs);
}
/**
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 002/192] mm/page_alloc: fix memory map initialization for descending nodes
2021-06-29 2:32 incoming Andrew Morton
2021-06-29 2:33 ` [patch 001/192] mm/gup: fix try_grab_compound_head() race with split_huge_page() Andrew Morton
@ 2021-06-29 2:33 ` Andrew Morton
2021-06-29 2:33 ` [patch 003/192] mm/page_alloc: correct return value of populated elements if bulk array is populated Andrew Morton
` (189 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:33 UTC (permalink / raw)
To: akpm, bhe, bp, david, linux-mm, mm-commits, robert.shteynfeld,
rppt, stable, torvalds, vbabka
From: Mike Rapoport <rppt@linux.ibm.com>
Subject: mm/page_alloc: fix memory map initialization for descending nodes
On systems with memory nodes sorted in descending order, for instance Dell
Precision WorkStation T5500, the struct pages for higher PFNs and
respectively lower nodes, could be overwritten by the initialization of
struct pages corresponding to the holes in the memory sections.
For example for the below memory layout
[ 0.245624] Early memory node ranges
[ 0.248496] node 1: [mem 0x0000000000001000-0x0000000000090fff]
[ 0.251376] node 1: [mem 0x0000000000100000-0x00000000dbdf8fff]
[ 0.254256] node 1: [mem 0x0000000100000000-0x0000001423ffffff]
[ 0.257144] node 0: [mem 0x0000001424000000-0x0000002023ffffff]
the range 0x1424000000 - 0x1428000000 in the beginning of node 0 starts in
the middle of a section and will be considered as a hole during the
initialization of the last section in node 1.
The wrong initialization of the memory map causes panic on boot when
CONFIG_DEBUG_VM is enabled.
Reorder loop order of the memory map initialization so that the outer loop
will always iterate over populated memory regions in the ascending order
and the inner loop will select the zone corresponding to the PFN range.
This way initialization of the struct pages for the memory holes will be
always done for the ranges that are actually not populated.
[akpm@linux-foundation.org: coding style fixes]
Link: https://lkml.kernel.org/r/YNXlMqBbL+tBG7yq@kernel.org
Link: https://bugzilla.kernel.org/show_bug.cgi?id=213073
Link: https://lkml.kernel.org/r/20210624062305.10940-1-rppt@kernel.org
Fixes: 0740a50b9baa ("mm/page_alloc.c: refactor initialization of struct page for holes in memory layout")
Signed-off-by: Mike Rapoport <rppt@linux.ibm.com>
Cc: Boris Petkov <bp@alien8.de>
Cc: Robert Shteynfeld <robert.shteynfeld@gmail.com>
Cc: Baoquan He <bhe@redhat.com>
Cc: Vlastimil Babka <vbabka@suse.cz>
Cc: David Hildenbrand <david@redhat.com>
Cc: <stable@vger.kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/linux/mm.h | 1
mm/page_alloc.c | 96 ++++++++++++++++++++++++++-----------------
2 files changed, 59 insertions(+), 38 deletions(-)
--- a/include/linux/mm.h~mm-page_alloc-fix-memory-map-initialization-for-descending-nodes
+++ a/include/linux/mm.h
@@ -2474,7 +2474,6 @@ extern void set_dma_reserve(unsigned lon
extern void memmap_init_range(unsigned long, int, unsigned long,
unsigned long, unsigned long, enum meminit_context,
struct vmem_altmap *, int migratetype);
-extern void memmap_init_zone(struct zone *zone);
extern void setup_per_zone_wmarks(void);
extern int __meminit init_per_zone_wmark_min(void);
extern void mem_init(void);
--- a/mm/page_alloc.c~mm-page_alloc-fix-memory-map-initialization-for-descending-nodes
+++ a/mm/page_alloc.c
@@ -6400,7 +6400,7 @@ void __ref memmap_init_zone_device(struc
return;
/*
- * The call to memmap_init_zone should have already taken care
+ * The call to memmap_init should have already taken care
* of the pages reserved for the memmap, so we can just jump to
* the end of that region and start processing the device pages.
*/
@@ -6465,7 +6465,7 @@ static void __meminit zone_init_free_lis
/*
* Only struct pages that correspond to ranges defined by memblock.memory
* are zeroed and initialized by going through __init_single_page() during
- * memmap_init_zone().
+ * memmap_init_zone_range().
*
* But, there could be struct pages that correspond to holes in
* memblock.memory. This can happen because of the following reasons:
@@ -6484,9 +6484,9 @@ static void __meminit zone_init_free_lis
* zone/node above the hole except for the trailing pages in the last
* section that will be appended to the zone/node below.
*/
-static u64 __meminit init_unavailable_range(unsigned long spfn,
- unsigned long epfn,
- int zone, int node)
+static void __init init_unavailable_range(unsigned long spfn,
+ unsigned long epfn,
+ int zone, int node)
{
unsigned long pfn;
u64 pgcnt = 0;
@@ -6502,56 +6502,77 @@ static u64 __meminit init_unavailable_ra
pgcnt++;
}
- return pgcnt;
+ if (pgcnt)
+ pr_info("On node %d, zone %s: %lld pages in unavailable ranges",
+ node, zone_names[zone], pgcnt);
}
#else
-static inline u64 init_unavailable_range(unsigned long spfn, unsigned long epfn,
- int zone, int node)
+static inline void init_unavailable_range(unsigned long spfn,
+ unsigned long epfn,
+ int zone, int node)
{
- return 0;
}
#endif
-void __meminit __weak memmap_init_zone(struct zone *zone)
+static void __init memmap_init_zone_range(struct zone *zone,
+ unsigned long start_pfn,
+ unsigned long end_pfn,
+ unsigned long *hole_pfn)
{
unsigned long zone_start_pfn = zone->zone_start_pfn;
unsigned long zone_end_pfn = zone_start_pfn + zone->spanned_pages;
- int i, nid = zone_to_nid(zone), zone_id = zone_idx(zone);
- static unsigned long hole_pfn;
+ int nid = zone_to_nid(zone), zone_id = zone_idx(zone);
+
+ start_pfn = clamp(start_pfn, zone_start_pfn, zone_end_pfn);
+ end_pfn = clamp(end_pfn, zone_start_pfn, zone_end_pfn);
+
+ if (start_pfn >= end_pfn)
+ return;
+
+ memmap_init_range(end_pfn - start_pfn, nid, zone_id, start_pfn,
+ zone_end_pfn, MEMINIT_EARLY, NULL, MIGRATE_MOVABLE);
+
+ if (*hole_pfn < start_pfn)
+ init_unavailable_range(*hole_pfn, start_pfn, zone_id, nid);
+
+ *hole_pfn = end_pfn;
+}
+
+static void __init memmap_init(void)
+{
unsigned long start_pfn, end_pfn;
- u64 pgcnt = 0;
+ unsigned long hole_pfn = 0;
+ int i, j, zone_id, nid;
- for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, NULL) {
- start_pfn = clamp(start_pfn, zone_start_pfn, zone_end_pfn);
- end_pfn = clamp(end_pfn, zone_start_pfn, zone_end_pfn);
+ for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) {
+ struct pglist_data *node = NODE_DATA(nid);
+
+ for (j = 0; j < MAX_NR_ZONES; j++) {
+ struct zone *zone = node->node_zones + j;
+
+ if (!populated_zone(zone))
+ continue;
- if (end_pfn > start_pfn)
- memmap_init_range(end_pfn - start_pfn, nid,
- zone_id, start_pfn, zone_end_pfn,
- MEMINIT_EARLY, NULL, MIGRATE_MOVABLE);
-
- if (hole_pfn < start_pfn)
- pgcnt += init_unavailable_range(hole_pfn, start_pfn,
- zone_id, nid);
- hole_pfn = end_pfn;
+ memmap_init_zone_range(zone, start_pfn, end_pfn,
+ &hole_pfn);
+ zone_id = j;
+ }
}
#ifdef CONFIG_SPARSEMEM
/*
- * Initialize the hole in the range [zone_end_pfn, section_end].
- * If zone boundary falls in the middle of a section, this hole
- * will be re-initialized during the call to this function for the
- * higher zone.
+ * Initialize the memory map for hole in the range [memory_end,
+ * section_end].
+ * Append the pages in this hole to the highest zone in the last
+ * node.
+ * The call to init_unavailable_range() is outside the ifdef to
+ * silence the compiler warining about zone_id set but not used;
+ * for FLATMEM it is a nop anyway
*/
- end_pfn = round_up(zone_end_pfn, PAGES_PER_SECTION);
+ end_pfn = round_up(end_pfn, PAGES_PER_SECTION);
if (hole_pfn < end_pfn)
- pgcnt += init_unavailable_range(hole_pfn, end_pfn,
- zone_id, nid);
#endif
-
- if (pgcnt)
- pr_info(" %s zone: %llu pages in unavailable ranges\n",
- zone->name, pgcnt);
+ init_unavailable_range(hole_pfn, end_pfn, zone_id, nid);
}
static int zone_batchsize(struct zone *zone)
@@ -7254,7 +7275,6 @@ static void __init free_area_init_core(s
set_pageblock_order();
setup_usemap(zone);
init_currently_empty_zone(zone, zone->zone_start_pfn, size);
- memmap_init_zone(zone);
}
}
@@ -7780,6 +7800,8 @@ void __init free_area_init(unsigned long
node_set_state(nid, N_MEMORY);
check_for_memory(pgdat, nid);
}
+
+ memmap_init();
}
static int __init cmdline_parse_core(char *p, unsigned long *core,
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 003/192] mm/page_alloc: correct return value of populated elements if bulk array is populated
2021-06-29 2:32 incoming Andrew Morton
2021-06-29 2:33 ` [patch 001/192] mm/gup: fix try_grab_compound_head() race with split_huge_page() Andrew Morton
2021-06-29 2:33 ` [patch 002/192] mm/page_alloc: fix memory map initialization for descending nodes Andrew Morton
@ 2021-06-29 2:33 ` Andrew Morton
2021-06-29 2:33 ` [patch 004/192] kthread: switch to new kerneldoc syntax for named variable macro argument Andrew Morton
` (188 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:33 UTC (permalink / raw)
To: akpm, brouer, dan.carpenter, davej, linux-mm, mgorman,
mm-commits, stable, torvalds, vbabka
From: Mel Gorman <mgorman@techsingularity.net>
Subject: mm/page_alloc: correct return value of populated elements if bulk array is populated
Dave Jones reported the following
This made it into 5.13 final, and completely breaks NFSD for me
(Serving tcp v3 mounts). Existing mounts on clients hang, as do
new mounts from new clients. Rebooting the server back to rc7
everything recovers.
The commit b3b64ebd3822 ("mm/page_alloc: do bulk array bounds check after
checking populated elements") returns the wrong value if the array is
already populated which is interpreted as an allocation failure. Dave
reported this fixes his problem and it also passed a test running dbench
over NFS.
Link: https://lkml.kernel.org/r/20210628150219.GC3840@techsingularity.net
Fixes: b3b64ebd3822 ("mm/page_alloc: do bulk array bounds check after checking populated elements")
Signed-off-by: Mel Gorman <mgorman@techsingularity.net>
Reported-by: Dave Jones <davej@codemonkey.org.uk>
Tested-by: Dave Jones <davej@codemonkey.org.uk>
Cc: Dan Carpenter <dan.carpenter@oracle.com>
Cc: Jesper Dangaard Brouer <brouer@redhat.com>
Cc: Vlastimil Babka <vbabka@suse.cz>
Cc: <stable@vger.kernel.org> [5.13+]
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/page_alloc.c | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
--- a/mm/page_alloc.c~mm-page_alloc-correct-return-value-of-populated-elements-if-bulk-array-is-populated
+++ a/mm/page_alloc.c
@@ -5058,7 +5058,7 @@ unsigned long __alloc_pages_bulk(gfp_t g
/* Already populated array? */
if (unlikely(page_array && nr_pages - nr_populated == 0))
- return 0;
+ return nr_populated;
/* Use the single page allocator for one page. */
if (nr_pages - nr_populated == 1)
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 004/192] kthread: switch to new kerneldoc syntax for named variable macro argument
2021-06-29 2:32 incoming Andrew Morton
` (2 preceding siblings ...)
2021-06-29 2:33 ` [patch 003/192] mm/page_alloc: correct return value of populated elements if bulk array is populated Andrew Morton
@ 2021-06-29 2:33 ` Andrew Morton
2021-06-29 2:33 ` [patch 005/192] kthread_worker: fix return value when kthread_mod_delayed_work() races with kthread_cancel_delayed_work_sync() Andrew Morton
` (187 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:33 UTC (permalink / raw)
To: akpm, axboe, Felix.Kuehling, j.neuschaefer, linux-mm, mm-commits,
peterz, torvalds, valentin.schneider
From: Jonathan Neuschäfer <j.neuschaefer@gmx.net>
Subject: kthread: switch to new kerneldoc syntax for named variable macro argument
The syntax without dots is available since commit 43756e347f21
("scripts/kernel-doc: Add support for named variable macro arguments").
The same HTML output is produced with and without this patch.
Link: https://lkml.kernel.org/r/20210513161702.1721039-1-j.neuschaefer@gmx.net
Signed-off-by: Jonathan Neuschäfer <j.neuschaefer@gmx.net>
Cc: Jens Axboe <axboe@kernel.dk>
Cc: Felix Kuehling <Felix.Kuehling@amd.com>
Cc: Valentin Schneider <valentin.schneider@arm.com>
Cc: Peter Zijlstra <peterz@infradead.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/linux/kthread.h | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
--- a/include/linux/kthread.h~kthread-switch-to-new-kerneldoc-syntax-for-named-variable-macro-argument
+++ a/include/linux/kthread.h
@@ -18,7 +18,7 @@ struct task_struct *kthread_create_on_no
* @threadfn: the function to run in the thread
* @data: data pointer for @threadfn()
* @namefmt: printf-style format string for the thread name
- * @arg...: arguments for @namefmt.
+ * @arg: arguments for @namefmt.
*
* This macro will create a kthread on the current node, leaving it in
* the stopped state. This is just a helper for kthread_create_on_node();
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 005/192] kthread_worker: fix return value when kthread_mod_delayed_work() races with kthread_cancel_delayed_work_sync()
2021-06-29 2:32 incoming Andrew Morton
` (3 preceding siblings ...)
2021-06-29 2:33 ` [patch 004/192] kthread: switch to new kerneldoc syntax for named variable macro argument Andrew Morton
@ 2021-06-29 2:33 ` Andrew Morton
2021-06-29 2:33 ` [patch 006/192] ia64: headers: drop duplicated words Andrew Morton
` (186 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:33 UTC (permalink / raw)
To: akpm, jenhaochen, linux-mm, liumartin, minchan, mm-commits,
nathan, ndesaulniers, oleg, pmladek, tj, torvalds
From: Petr Mladek <pmladek@suse.com>
Subject: kthread_worker: fix return value when kthread_mod_delayed_work() races with kthread_cancel_delayed_work_sync()
kthread_mod_delayed_work() might race with
kthread_cancel_delayed_work_sync() or another kthread_mod_delayed_work()
call. The function lets the other operation win when it sees
work->canceling counter set. And it returns @false.
But it should return @true as it is done by the related workqueue API, see
mod_delayed_work_on().
The reason is that the return value might be used for reference counting.
It has to distinguish the case when the number of queued works has changed
or stayed the same.
The change is safe. kthread_mod_delayed_work() return value is not
checked anywhere at the moment.
Link: https://lore.kernel.org/r/20210521163526.GA17916@redhat.com
Link: https://lkml.kernel.org/r/20210610133051.15337-4-pmladek@suse.com
Signed-off-by: Petr Mladek <pmladek@suse.com>
Reported-by: Oleg Nesterov <oleg@redhat.com>
Cc: Nathan Chancellor <nathan@kernel.org>
Cc: Nick Desaulniers <ndesaulniers@google.com>
Cc: Tejun Heo <tj@kernel.org>
Cc: Minchan Kim <minchan@google.com>
Cc: <jenhaochen@google.com>
Cc: Martin Liu <liumartin@google.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
kernel/kthread.c | 19 ++++++++++++-------
1 file changed, 12 insertions(+), 7 deletions(-)
--- a/kernel/kthread.c~kthread_worker-fix-return-value-when-kthread_mod_delayed_work-races-with-kthread_cancel_delayed_work_sync
+++ a/kernel/kthread.c
@@ -1156,14 +1156,14 @@ static bool __kthread_cancel_work(struct
* modify @dwork's timer so that it expires after @delay. If @delay is zero,
* @work is guaranteed to be queued immediately.
*
- * Return: %true if @dwork was pending and its timer was modified,
- * %false otherwise.
+ * Return: %false if @dwork was idle and queued, %true otherwise.
*
* A special case is when the work is being canceled in parallel.
* It might be caused either by the real kthread_cancel_delayed_work_sync()
* or yet another kthread_mod_delayed_work() call. We let the other command
- * win and return %false here. The caller is supposed to synchronize these
- * operations a reasonable way.
+ * win and return %true here. The return value can be used for reference
+ * counting and the number of queued works stays the same. Anyway, the caller
+ * is supposed to synchronize these operations a reasonable way.
*
* This function is safe to call from any context including IRQ handler.
* See __kthread_cancel_work() and kthread_delayed_work_timer_fn()
@@ -1175,13 +1175,15 @@ bool kthread_mod_delayed_work(struct kth
{
struct kthread_work *work = &dwork->work;
unsigned long flags;
- int ret = false;
+ int ret;
raw_spin_lock_irqsave(&worker->lock, flags);
/* Do not bother with canceling when never queued. */
- if (!work->worker)
+ if (!work->worker) {
+ ret = false;
goto fast_queue;
+ }
/* Work must not be used with >1 worker, see kthread_queue_work() */
WARN_ON_ONCE(work->worker != worker);
@@ -1199,8 +1201,11 @@ bool kthread_mod_delayed_work(struct kth
* be used for reference counting.
*/
kthread_cancel_delayed_work_timer(work, &flags);
- if (work->canceling)
+ if (work->canceling) {
+ /* The number of works in the queue does not change. */
+ ret = true;
goto out;
+ }
ret = __kthread_cancel_work(work);
fast_queue:
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 006/192] ia64: headers: drop duplicated words
2021-06-29 2:32 incoming Andrew Morton
` (4 preceding siblings ...)
2021-06-29 2:33 ` [patch 005/192] kthread_worker: fix return value when kthread_mod_delayed_work() races with kthread_cancel_delayed_work_sync() Andrew Morton
@ 2021-06-29 2:33 ` Andrew Morton
2021-06-29 2:33 ` [patch 007/192] ia64: mca_drv: fix incorrect array size calculation Andrew Morton
` (185 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:33 UTC (permalink / raw)
To: akpm, fenghua.yu, linux-mm, mm-commits, rdunlap, torvalds
From: Randy Dunlap <rdunlap@infradead.org>
Subject: ia64: headers: drop duplicated words
Delete the repeated words "to" and "the".
Link: https://lkml.kernel.org/r/20210507184837.10754-1-rdunlap@infradead.org
Signed-off-by: Randy Dunlap <rdunlap@infradead.org>
Cc: Fenghua Yu <fenghua.yu@intel.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
arch/ia64/include/asm/pal.h | 2 +-
arch/ia64/include/asm/spinlock.h | 2 +-
arch/ia64/include/asm/uv/uv_hub.h | 2 +-
3 files changed, 3 insertions(+), 3 deletions(-)
--- a/arch/ia64/include/asm/pal.h~ia64-headers-drop-duplicated-words
+++ a/arch/ia64/include/asm/pal.h
@@ -1086,7 +1086,7 @@ static inline long ia64_pal_freq_base(un
/*
* Get the ratios for processor frequency, bus frequency and interval timer to
- * to base frequency of the platform
+ * the base frequency of the platform
*/
static inline s64
ia64_pal_freq_ratios (struct pal_freq_ratio *proc_ratio, struct pal_freq_ratio *bus_ratio,
--- a/arch/ia64/include/asm/spinlock.h~ia64-headers-drop-duplicated-words
+++ a/arch/ia64/include/asm/spinlock.h
@@ -26,7 +26,7 @@
* the queue, and the other indicating the current tail. The lock is acquired
* by atomically noting the tail and incrementing it by one (thus adding
* ourself to the queue and noting our position), then waiting until the head
- * becomes equal to the the initial value of the tail.
+ * becomes equal to the initial value of the tail.
* The pad bits in the middle are used to prevent the next_ticket number
* overflowing into the now_serving number.
*
--- a/arch/ia64/include/asm/uv/uv_hub.h~ia64-headers-drop-duplicated-words
+++ a/arch/ia64/include/asm/uv/uv_hub.h
@@ -257,7 +257,7 @@ static inline int uv_numa_blade_id(void)
return 0;
}
-/* Convert a cpu number to the the UV blade number */
+/* Convert a cpu number to the UV blade number */
static inline int uv_cpu_to_blade_id(int cpu)
{
return 0;
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 007/192] ia64: mca_drv: fix incorrect array size calculation
2021-06-29 2:32 incoming Andrew Morton
` (5 preceding siblings ...)
2021-06-29 2:33 ` [patch 006/192] ia64: headers: drop duplicated words Andrew Morton
@ 2021-06-29 2:33 ` Andrew Morton
2021-06-29 2:33 ` [patch 008/192] streamline_config.pl: make spacing consistent Andrew Morton
` (184 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:33 UTC (permalink / raw)
To: akpm, arnd, linux-mm, masahiroy, mm-commits, rdunlap, torvalds
From: Arnd Bergmann <arnd@arndb.de>
Subject: ia64: mca_drv: fix incorrect array size calculation
gcc points out a mistake in the mca driver that goes back to before the
git history:
arch/ia64/kernel/mca_drv.c: In function 'init_record_index_pools':
arch/ia64/kernel/mca_drv.c:346:54: error: expression does not compute the number of elements in this array; element typ
e is 'int', not 'size_t' {aka 'long unsigned int'} [-Werror=sizeof-array-div]
346 | for (i = 1; i < sizeof sal_log_sect_min_sizes/sizeof(size_t); i++)
| ^
This is the same as sizeof(size_t), which is two shorter than the actual
array. Use the ARRAY_SIZE() macro to get the correct calculation instead.
Link: https://lkml.kernel.org/r/20210514214123.875971-1-arnd@kernel.org
Signed-off-by: Arnd Bergmann <arnd@arndb.de>
Cc: Masahiro Yamada <masahiroy@kernel.org>
Cc: Randy Dunlap <rdunlap@infradead.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
arch/ia64/kernel/mca_drv.c | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
--- a/arch/ia64/kernel/mca_drv.c~ia64-mca_drv-fix-incorrect-array-size-calculation
+++ a/arch/ia64/kernel/mca_drv.c
@@ -343,7 +343,7 @@ init_record_index_pools(void)
/* - 2 - */
sect_min_size = sal_log_sect_min_sizes[0];
- for (i = 1; i < sizeof sal_log_sect_min_sizes/sizeof(size_t); i++)
+ for (i = 1; i < ARRAY_SIZE(sal_log_sect_min_sizes); i++)
if (sect_min_size > sal_log_sect_min_sizes[i])
sect_min_size = sal_log_sect_min_sizes[i];
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 008/192] streamline_config.pl: make spacing consistent
2021-06-29 2:32 incoming Andrew Morton
` (6 preceding siblings ...)
2021-06-29 2:33 ` [patch 007/192] ia64: mca_drv: fix incorrect array size calculation Andrew Morton
@ 2021-06-29 2:33 ` Andrew Morton
2021-06-29 2:33 ` [patch 009/192] streamline_config.pl: add softtabstop=4 for vim users Andrew Morton
` (183 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:33 UTC (permalink / raw)
To: akpm, linux-mm, masahiroy, mm-commits, rostedt, torvalds, warthog9
From: "Steven Rostedt (VMware)" <rostedt@goodmis.org>
Subject: streamline_config.pl: make spacing consistent
Patch series "streamline_config.pl: Fix Perl spacing".
Talking with John Hawley about how vim and emacs deal with Perl files with
respect to tabs and spaces, I found that some of my Perl code in the
kernel had inconsistent spacing. The way emacs handles Perl by default is
to use 4 spaces per indent, but make all 8 spaces into a single tab. Vim
does not do this by default. But if you add the vim variable control:
# vim: softtabstop=4
to a perl file, it makes vim behave the same way as emacs.
The first patch is to change all 8 spaces into a single tab (mostly from
people editing the file with vim). The next patch adds the softtabstop
variable to make vim act like emacs by default.
This patch (of 2):
As Perl code tends to have 4 space indentation, but uses tabs for every 8
spaces, make that consistent in the streamline_config.pl code. Replace
all 8 spaces with a single tab.
Link: https://lkml.kernel.org/r/20210322214032.133596267@goodmis.org
Signed-off-by: Steven Rostedt (VMware) <rostedt@goodmis.org>
Cc: "John (Warthog9) Hawley" <warthog9@kernel.org>
Cc: Masahiro Yamada <masahiroy@kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
scripts/kconfig/streamline_config.pl | 74 ++++++++++++-------------
1 file changed, 37 insertions(+), 37 deletions(-)
--- a/scripts/kconfig/streamline_config.pl~streamline_configpl-make-spacing-consistent
+++ a/scripts/kconfig/streamline_config.pl
@@ -601,12 +601,12 @@ if (defined($ENV{'LMC_KEEP'})) {
sub in_preserved_kconfigs {
my $kconfig = $config2kfile{$_[0]};
if (!defined($kconfig)) {
- return 0;
+ return 0;
}
foreach my $excl (@preserved_kconfigs) {
- if($kconfig =~ /^$excl/) {
- return 1;
- }
+ if($kconfig =~ /^$excl/) {
+ return 1;
+ }
}
return 0;
}
@@ -629,52 +629,52 @@ foreach my $line (@config_file) {
}
if (/CONFIG_MODULE_SIG_KEY="(.+)"/) {
- my $orig_cert = $1;
- my $default_cert = "certs/signing_key.pem";
+ my $orig_cert = $1;
+ my $default_cert = "certs/signing_key.pem";
- # Check that the logic in this script still matches the one in Kconfig
- if (!defined($depends{"MODULE_SIG_KEY"}) ||
- $depends{"MODULE_SIG_KEY"} !~ /"\Q$default_cert\E"/) {
- print STDERR "WARNING: MODULE_SIG_KEY assertion failure, ",
- "update needed to ", __FILE__, " line ", __LINE__, "\n";
- print;
- } elsif ($orig_cert ne $default_cert && ! -f $orig_cert) {
- print STDERR "Module signature verification enabled but ",
- "module signing key \"$orig_cert\" not found. Resetting ",
- "signing key to default value.\n";
- print "CONFIG_MODULE_SIG_KEY=\"$default_cert\"\n";
- } else {
- print;
- }
- next;
+ # Check that the logic in this script still matches the one in Kconfig
+ if (!defined($depends{"MODULE_SIG_KEY"}) ||
+ $depends{"MODULE_SIG_KEY"} !~ /"\Q$default_cert\E"/) {
+ print STDERR "WARNING: MODULE_SIG_KEY assertion failure, ",
+ "update needed to ", __FILE__, " line ", __LINE__, "\n";
+ print;
+ } elsif ($orig_cert ne $default_cert && ! -f $orig_cert) {
+ print STDERR "Module signature verification enabled but ",
+ "module signing key \"$orig_cert\" not found. Resetting ",
+ "signing key to default value.\n";
+ print "CONFIG_MODULE_SIG_KEY=\"$default_cert\"\n";
+ } else {
+ print;
+ }
+ next;
}
if (/CONFIG_SYSTEM_TRUSTED_KEYS="(.+)"/) {
- my $orig_keys = $1;
+ my $orig_keys = $1;
- if (! -f $orig_keys) {
- print STDERR "System keyring enabled but keys \"$orig_keys\" ",
- "not found. Resetting keys to default value.\n";
- print "CONFIG_SYSTEM_TRUSTED_KEYS=\"\"\n";
- } else {
- print;
- }
- next;
+ if (! -f $orig_keys) {
+ print STDERR "System keyring enabled but keys \"$orig_keys\" ",
+ "not found. Resetting keys to default value.\n";
+ print "CONFIG_SYSTEM_TRUSTED_KEYS=\"\"\n";
+ } else {
+ print;
+ }
+ next;
}
if (/^(CONFIG.*)=(m|y)/) {
- if (in_preserved_kconfigs($1)) {
- dprint "Preserve config $1";
- print;
- next;
- }
+ if (in_preserved_kconfigs($1)) {
+ dprint "Preserve config $1";
+ print;
+ next;
+ }
if (defined($configs{$1})) {
if ($localyesconfig) {
- $setconfigs{$1} = 'y';
+ $setconfigs{$1} = 'y';
print "$1=y\n";
next;
} else {
- $setconfigs{$1} = $2;
+ $setconfigs{$1} = $2;
}
} elsif ($2 eq "m") {
print "# $1 is not set\n";
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 009/192] streamline_config.pl: add softtabstop=4 for vim users
2021-06-29 2:32 incoming Andrew Morton
` (7 preceding siblings ...)
2021-06-29 2:33 ` [patch 008/192] streamline_config.pl: make spacing consistent Andrew Morton
@ 2021-06-29 2:33 ` Andrew Morton
2021-06-29 2:33 ` [patch 010/192] scripts/spelling.txt: add more spellings to spelling.txt Andrew Morton
` (182 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:33 UTC (permalink / raw)
To: akpm, linux-mm, masahiroy, mm-commits, rostedt, torvalds, warthog9
From: "Steven Rostedt (VMware)" <rostedt@goodmis.org>
Subject: streamline_config.pl: add softtabstop=4 for vim users
The tab stop for Perl files is by default (at least in emacs) to be 4
spaces, where a tab is used for all 8 spaces. Add a local variable
comment to make vim do the same by default, and this will help keep the
file consistent in the future when others edit it via vim and not emacs.
Link: https://lkml.kernel.org/r/20210322214032.293992979@goodmis.org
Signed-off-by: Steven Rostedt (VMware) <rostedt@goodmis.org>
Cc: Masahiro Yamada <masahiroy@kernel.org>
Cc: "John (Warthog9) Hawley" <warthog9@kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
scripts/kconfig/streamline_config.pl | 2 ++
1 file changed, 2 insertions(+)
--- a/scripts/kconfig/streamline_config.pl~streamline_configpl-add-softtabstop=4-for-vim-users
+++ a/scripts/kconfig/streamline_config.pl
@@ -702,3 +702,5 @@ foreach my $module (keys(%modules)) {
print STDERR "\n";
}
}
+
+# vim: softtabstop=4
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 010/192] scripts/spelling.txt: add more spellings to spelling.txt
2021-06-29 2:32 incoming Andrew Morton
` (8 preceding siblings ...)
2021-06-29 2:33 ` [patch 009/192] streamline_config.pl: add softtabstop=4 for vim users Andrew Morton
@ 2021-06-29 2:33 ` Andrew Morton
2021-06-29 2:33 ` [patch 011/192] ntfs: fix validity check for file name attribute Andrew Morton
` (181 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:33 UTC (permalink / raw)
To: akpm, colin.king, linux-mm, mm-commits, torvalds
From: Colin Ian King <colin.king@canonical.com>
Subject: scripts/spelling.txt: add more spellings to spelling.txt
Here are some of the more common spelling mistakes and typos that I've
found while fixing up spelling mistakes in the kernel in the past few
months.
Link: https://lkml.kernel.org/r/20210514093655.8829-1-colin.king@canonical.com
Signed-off-by: Colin Ian King <colin.king@canonical.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
scripts/spelling.txt | 16 ++++++++++++++++
1 file changed, 16 insertions(+)
--- a/scripts/spelling.txt~scripts-spellingtxt-add-more-spellings-to-spellingtxt
+++ a/scripts/spelling.txt
@@ -22,6 +22,7 @@ absolut||absolute
absoulte||absolute
acccess||access
acceess||access
+accelaration||acceleration
acceleratoin||acceleration
accelleration||acceleration
accesing||accessing
@@ -264,6 +265,7 @@ calucate||calculate
calulate||calculate
cancelation||cancellation
cancle||cancel
+canot||cannot
capabilites||capabilities
capabilties||capabilities
capabilty||capability
@@ -494,7 +496,10 @@ digial||digital
dimention||dimension
dimesions||dimensions
diconnected||disconnected
+disabed||disabled
+disble||disable
disgest||digest
+disired||desired
dispalying||displaying
diplay||display
directon||direction
@@ -710,6 +715,7 @@ havind||having
heirarchically||hierarchically
heirarchy||hierarchy
helpfull||helpful
+hearbeat||heartbeat
heterogenous||heterogeneous
hexdecimal||hexadecimal
hybernate||hibernate
@@ -989,6 +995,7 @@ notications||notifications
notifcations||notifications
notifed||notified
notity||notify
+nubmer||number
numebr||number
numner||number
obtaion||obtain
@@ -1014,8 +1021,10 @@ ommiting||omitting
ommitted||omitted
onself||oneself
ony||only
+openning||opening
operatione||operation
opertaions||operations
+opportunies||opportunities
optionnal||optional
optmizations||optimizations
orientatied||orientated
@@ -1111,6 +1120,7 @@ prefitler||prefilter
preform||perform
premption||preemption
prepaired||prepared
+prepate||prepare
preperation||preparation
preprare||prepare
pressre||pressure
@@ -1123,6 +1133,7 @@ privilaged||privileged
privilage||privilege
priviledge||privilege
priviledges||privileges
+privleges||privileges
probaly||probably
procceed||proceed
proccesors||processors
@@ -1167,6 +1178,7 @@ promixity||proximity
psudo||pseudo
psuedo||pseudo
psychadelic||psychedelic
+purgable||purgeable
pwoer||power
queing||queuing
quering||querying
@@ -1180,6 +1192,7 @@ receieve||receive
recepient||recipient
recevied||received
receving||receiving
+recievd||received
recieved||received
recieve||receive
reciever||receiver
@@ -1228,6 +1241,7 @@ reponse||response
representaion||representation
reqeust||request
reqister||register
+requed||requeued
requestied||requested
requiere||require
requirment||requirement
@@ -1332,6 +1346,7 @@ singal||signal
singed||signed
sleeped||slept
sliped||slipped
+softwade||software
softwares||software
soley||solely
souce||source
@@ -1510,6 +1525,7 @@ unintialized||uninitialized
unitialized||uninitialized
unkmown||unknown
unknonw||unknown
+unknouwn||unknown
unknow||unknown
unkown||unknown
unamed||unnamed
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 011/192] ntfs: fix validity check for file name attribute
2021-06-29 2:32 incoming Andrew Morton
` (9 preceding siblings ...)
2021-06-29 2:33 ` [patch 010/192] scripts/spelling.txt: add more spellings to spelling.txt Andrew Morton
@ 2021-06-29 2:33 ` Andrew Morton
2021-06-29 2:33 ` [patch 012/192] squashfs: add option to panic on errors Andrew Morton
` (180 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:33 UTC (permalink / raw)
To: akpm, anton, desmondcheongzx, gregkh, linux-mm, mm-commits,
skhan, stable, torvalds
From: Desmond Cheong Zhi Xi <desmondcheongzx@gmail.com>
Subject: ntfs: fix validity check for file name attribute
When checking the file name attribute, we want to ensure that it fits
within the bounds of ATTR_RECORD. To do this, we should check that (attr
record + file name offset + file name length) < (attr record + attr record
length).
However, the original check did not include the file name offset in the
calculation. This means that corrupted on-disk metadata might not caught
by the incorrect file name check, and lead to an invalid memory access.
An example can be seen in the crash report of a memory corruption error
found by Syzbot:
https://syzkaller.appspot.com/bug?id=a1a1e379b225812688566745c3e2f7242bffc246
Adding the file name offset to the validity check fixes this error and
passes the Syzbot reproducer test.
Link: https://lkml.kernel.org/r/20210614050540.289494-1-desmondcheongzx@gmail.com
Signed-off-by: Desmond Cheong Zhi Xi <desmondcheongzx@gmail.com>
Reported-by: syzbot+213ac8bb98f7f4420840@syzkaller.appspotmail.com
Tested-by: syzbot+213ac8bb98f7f4420840@syzkaller.appspotmail.com
Acked-by: Anton Altaparmakov <anton@tuxera.com>
Cc: Shuah Khan <skhan@linuxfoundation.org>
Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Cc: <stable@vger.kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
fs/ntfs/inode.c | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
--- a/fs/ntfs/inode.c~ntfs-fix-validity-check-for-file-name-attribute
+++ a/fs/ntfs/inode.c
@@ -477,7 +477,7 @@ err_corrupt_attr:
}
file_name_attr = (FILE_NAME_ATTR*)((u8*)attr +
le16_to_cpu(attr->data.resident.value_offset));
- p2 = (u8*)attr + le32_to_cpu(attr->data.resident.value_length);
+ p2 = (u8 *)file_name_attr + le32_to_cpu(attr->data.resident.value_length);
if (p2 < (u8*)attr || p2 > p)
goto err_corrupt_attr;
/* This attribute is ok, but is it in the $Extend directory? */
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 012/192] squashfs: add option to panic on errors
2021-06-29 2:32 incoming Andrew Morton
` (10 preceding siblings ...)
2021-06-29 2:33 ` [patch 011/192] ntfs: fix validity check for file name attribute Andrew Morton
@ 2021-06-29 2:33 ` Andrew Morton
2021-06-29 2:33 ` [patch 013/192] ocfs2: remove unnecessary INIT_LIST_HEAD() Andrew Morton
` (179 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:33 UTC (permalink / raw)
To: akpm, linux-mm, mm-commits, phillip, torvalds, vincent.whitchurch
From: Vincent Whitchurch <vincent.whitchurch@axis.com>
Subject: squashfs: add option to panic on errors
Add an errors=panic mount option to make squashfs trigger a panic when
errors are encountered, similar to several other filesystems. This allows
a kernel dump to be saved using which the corruption can be analysed and
debugged.
Inspired by a pre-fs_context patch by Anton Eliasson.
Link: https://lkml.kernel.org/r/20210527125019.14511-1-vincent.whitchurch@axis.com
Signed-off-by: Vincent Whitchurch <vincent.whitchurch@axis.com>
Signed-off-by: Phillip Lougher <phillip@squashfs.org.uk>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
fs/squashfs/block.c | 5 +
fs/squashfs/squashfs_fs_sb.h | 1
fs/squashfs/super.c | 86 +++++++++++++++++++++++++++++++++
3 files changed, 91 insertions(+), 1 deletion(-)
--- a/fs/squashfs/block.c~squashfs-add-option-to-panic-on-errors
+++ a/fs/squashfs/block.c
@@ -226,8 +226,11 @@ out_free_bio:
bio_free_pages(bio);
bio_put(bio);
out:
- if (res < 0)
+ if (res < 0) {
ERROR("Failed to read block 0x%llx: %d\n", index, res);
+ if (msblk->panic_on_errors)
+ panic("squashfs read failed");
+ }
return res;
}
--- a/fs/squashfs/squashfs_fs_sb.h~squashfs-add-option-to-panic-on-errors
+++ a/fs/squashfs/squashfs_fs_sb.h
@@ -65,5 +65,6 @@ struct squashfs_sb_info {
unsigned int fragments;
int xattr_ids;
unsigned int ids;
+ bool panic_on_errors;
};
#endif
--- a/fs/squashfs/super.c~squashfs-add-option-to-panic-on-errors
+++ a/fs/squashfs/super.c
@@ -18,9 +18,11 @@
#include <linux/fs.h>
#include <linux/fs_context.h>
+#include <linux/fs_parser.h>
#include <linux/vfs.h>
#include <linux/slab.h>
#include <linux/mutex.h>
+#include <linux/seq_file.h>
#include <linux/pagemap.h>
#include <linux/init.h>
#include <linux/module.h>
@@ -37,6 +39,51 @@
static struct file_system_type squashfs_fs_type;
static const struct super_operations squashfs_super_ops;
+enum Opt_errors {
+ Opt_errors_continue,
+ Opt_errors_panic,
+};
+
+enum squashfs_param {
+ Opt_errors,
+};
+
+struct squashfs_mount_opts {
+ enum Opt_errors errors;
+};
+
+static const struct constant_table squashfs_param_errors[] = {
+ {"continue", Opt_errors_continue },
+ {"panic", Opt_errors_panic },
+ {}
+};
+
+static const struct fs_parameter_spec squashfs_fs_parameters[] = {
+ fsparam_enum("errors", Opt_errors, squashfs_param_errors),
+ {}
+};
+
+static int squashfs_parse_param(struct fs_context *fc, struct fs_parameter *param)
+{
+ struct squashfs_mount_opts *opts = fc->fs_private;
+ struct fs_parse_result result;
+ int opt;
+
+ opt = fs_parse(fc, squashfs_fs_parameters, param, &result);
+ if (opt < 0)
+ return opt;
+
+ switch (opt) {
+ case Opt_errors:
+ opts->errors = result.uint_32;
+ break;
+ default:
+ return -EINVAL;
+ }
+
+ return 0;
+}
+
static const struct squashfs_decompressor *supported_squashfs_filesystem(
struct fs_context *fc,
short major, short minor, short id)
@@ -67,6 +114,7 @@ static const struct squashfs_decompresso
static int squashfs_fill_super(struct super_block *sb, struct fs_context *fc)
{
+ struct squashfs_mount_opts *opts = fc->fs_private;
struct squashfs_sb_info *msblk;
struct squashfs_super_block *sblk = NULL;
struct inode *root;
@@ -85,6 +133,8 @@ static int squashfs_fill_super(struct su
}
msblk = sb->s_fs_info;
+ msblk->panic_on_errors = (opts->errors == Opt_errors_panic);
+
msblk->devblksize = sb_min_blocksize(sb, SQUASHFS_DEVBLK_SIZE);
msblk->devblksize_log2 = ffz(~msblk->devblksize);
@@ -350,18 +400,52 @@ static int squashfs_get_tree(struct fs_c
static int squashfs_reconfigure(struct fs_context *fc)
{
+ struct super_block *sb = fc->root->d_sb;
+ struct squashfs_sb_info *msblk = sb->s_fs_info;
+ struct squashfs_mount_opts *opts = fc->fs_private;
+
sync_filesystem(fc->root->d_sb);
fc->sb_flags |= SB_RDONLY;
+
+ msblk->panic_on_errors = (opts->errors == Opt_errors_panic);
+
return 0;
}
+static void squashfs_free_fs_context(struct fs_context *fc)
+{
+ kfree(fc->fs_private);
+}
+
static const struct fs_context_operations squashfs_context_ops = {
.get_tree = squashfs_get_tree,
+ .free = squashfs_free_fs_context,
+ .parse_param = squashfs_parse_param,
.reconfigure = squashfs_reconfigure,
};
+static int squashfs_show_options(struct seq_file *s, struct dentry *root)
+{
+ struct super_block *sb = root->d_sb;
+ struct squashfs_sb_info *msblk = sb->s_fs_info;
+
+ if (msblk->panic_on_errors)
+ seq_puts(s, ",errors=panic");
+ else
+ seq_puts(s, ",errors=continue");
+
+ return 0;
+}
+
static int squashfs_init_fs_context(struct fs_context *fc)
{
+ struct squashfs_mount_opts *opts;
+
+ opts = kzalloc(sizeof(*opts), GFP_KERNEL);
+ if (!opts)
+ return -ENOMEM;
+
+ fc->fs_private = opts;
fc->ops = &squashfs_context_ops;
return 0;
}
@@ -481,6 +565,7 @@ static struct file_system_type squashfs_
.owner = THIS_MODULE,
.name = "squashfs",
.init_fs_context = squashfs_init_fs_context,
+ .parameters = squashfs_fs_parameters,
.kill_sb = kill_block_super,
.fs_flags = FS_REQUIRES_DEV
};
@@ -491,6 +576,7 @@ static const struct super_operations squ
.free_inode = squashfs_free_inode,
.statfs = squashfs_statfs,
.put_super = squashfs_put_super,
+ .show_options = squashfs_show_options,
};
module_init(init_squashfs_fs);
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 013/192] ocfs2: remove unnecessary INIT_LIST_HEAD()
2021-06-29 2:32 incoming Andrew Morton
` (11 preceding siblings ...)
2021-06-29 2:33 ` [patch 012/192] squashfs: add option to panic on errors Andrew Morton
@ 2021-06-29 2:33 ` Andrew Morton
2021-06-29 2:34 ` [patch 014/192] ocfs2: fix snprintf() checking Andrew Morton
` (178 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:33 UTC (permalink / raw)
To: akpm, gechangwei, ghe, hulkci, jiangqi903, jlbec, junxiao.bi,
linux-mm, mark, mm-commits, piaojun, torvalds, yangyingliang
From: Yang Yingliang <yangyingliang@huawei.com>
Subject: ocfs2: remove unnecessary INIT_LIST_HEAD()
The list_head o2hb_node_events is initialized statically. It is
unnecessary to initialize by INIT_LIST_HEAD().
Link: https://lkml.kernel.org/r/20210511115847.3817395-1-yangyingliang@huawei.com
Signed-off-by: Yang Yingliang <yangyingliang@huawei.com>
Reported-by: Hulk Robot <hulkci@huawei.com>
Cc: Mark Fasheh <mark@fasheh.com>
Cc: Joel Becker <jlbec@evilplan.org>
Cc: Junxiao Bi <junxiao.bi@oracle.com>
Cc: Joseph Qi <jiangqi903@gmail.com>
Cc: Changwei Ge <gechangwei@live.cn>
Cc: Gang He <ghe@suse.com>
Cc: Jun Piao <piaojun@huawei.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
fs/ocfs2/cluster/heartbeat.c | 2 --
1 file changed, 2 deletions(-)
--- a/fs/ocfs2/cluster/heartbeat.c~ocfs2-remove-unnecessary-init_list_head
+++ a/fs/ocfs2/cluster/heartbeat.c
@@ -1442,8 +1442,6 @@ void o2hb_init(void)
for (i = 0; i < ARRAY_SIZE(o2hb_live_slots); i++)
INIT_LIST_HEAD(&o2hb_live_slots[i]);
- INIT_LIST_HEAD(&o2hb_node_events);
-
memset(o2hb_live_node_bitmap, 0, sizeof(o2hb_live_node_bitmap));
memset(o2hb_region_bitmap, 0, sizeof(o2hb_region_bitmap));
memset(o2hb_live_region_bitmap, 0, sizeof(o2hb_live_region_bitmap));
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 014/192] ocfs2: fix snprintf() checking
2021-06-29 2:32 incoming Andrew Morton
` (12 preceding siblings ...)
2021-06-29 2:33 ` [patch 013/192] ocfs2: remove unnecessary INIT_LIST_HEAD() Andrew Morton
@ 2021-06-29 2:34 ` Andrew Morton
2021-06-29 2:34 ` [patch 015/192] ocfs2: remove redundant assignment to pointer queue Andrew Morton
` (177 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:34 UTC (permalink / raw)
To: akpm, dan.carpenter, gechangwei, ghe, jlbec, joseph.qi,
junxiao.bi, linux-mm, mark, mm-commits, piaojun, torvalds
From: Dan Carpenter <dan.carpenter@oracle.com>
Subject: ocfs2: fix snprintf() checking
The snprintf() function returns the number of bytes which would have been
printed if the buffer was large enough. In other words it can return ">=
remain" but this code assumes it returns "== remain".
The run time impact of this bug is not very severe. The next iteration
through the loop would trigger a WARN() when we pass a negative limit to
snprintf(). We would then return success instead of -E2BIG.
The kernel implementation of snprintf() will never return negatives so
there is no need to check and I have deleted that dead code.
Link: https://lkml.kernel.org/r/20210511135350.GV1955@kadam
Fixes: a860f6eb4c6a ("ocfs2: sysfile interfaces for online file check")
Fixes: 74ae4e104dfc ("ocfs2: Create stack glue sysfs files.")
Signed-off-by: Dan Carpenter <dan.carpenter@oracle.com>
Reviewed-by: Joseph Qi <joseph.qi@linux.alibaba.com>
Cc: Mark Fasheh <mark@fasheh.com>
Cc: Joel Becker <jlbec@evilplan.org>
Cc: Junxiao Bi <junxiao.bi@oracle.com>
Cc: Changwei Ge <gechangwei@live.cn>
Cc: Gang He <ghe@suse.com>
Cc: Jun Piao <piaojun@huawei.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
fs/ocfs2/filecheck.c | 6 +-----
fs/ocfs2/stackglue.c | 8 ++------
2 files changed, 3 insertions(+), 11 deletions(-)
--- a/fs/ocfs2/filecheck.c~ocfs2-fix-snprintf-checking
+++ a/fs/ocfs2/filecheck.c
@@ -326,11 +326,7 @@ static ssize_t ocfs2_filecheck_attr_show
ret = snprintf(buf + total, remain, "%lu\t\t%u\t%s\n",
p->fe_ino, p->fe_done,
ocfs2_filecheck_error(p->fe_status));
- if (ret < 0) {
- total = ret;
- break;
- }
- if (ret == remain) {
+ if (ret >= remain) {
/* snprintf() didn't fit */
total = -E2BIG;
break;
--- a/fs/ocfs2/stackglue.c~ocfs2-fix-snprintf-checking
+++ a/fs/ocfs2/stackglue.c
@@ -500,11 +500,7 @@ static ssize_t ocfs2_loaded_cluster_plug
list_for_each_entry(p, &ocfs2_stack_list, sp_list) {
ret = snprintf(buf, remain, "%s\n",
p->sp_name);
- if (ret < 0) {
- total = ret;
- break;
- }
- if (ret == remain) {
+ if (ret >= remain) {
/* snprintf() didn't fit */
total = -E2BIG;
break;
@@ -531,7 +527,7 @@ static ssize_t ocfs2_active_cluster_plug
if (active_stack) {
ret = snprintf(buf, PAGE_SIZE, "%s\n",
active_stack->sp_name);
- if (ret == PAGE_SIZE)
+ if (ret >= PAGE_SIZE)
ret = -E2BIG;
}
spin_unlock(&ocfs2_stack_lock);
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 015/192] ocfs2: remove redundant assignment to pointer queue
2021-06-29 2:32 incoming Andrew Morton
` (13 preceding siblings ...)
2021-06-29 2:34 ` [patch 014/192] ocfs2: fix snprintf() checking Andrew Morton
@ 2021-06-29 2:34 ` Andrew Morton
2021-06-29 2:34 ` [patch 016/192] ocfs2: remove repeated uptodate check for buffer Andrew Morton
` (176 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:34 UTC (permalink / raw)
To: akpm, colin.king, gechangwei, ghe, jlbec, joseph.qi, junxiao.bi,
linux-mm, mark, mm-commits, piaojun, torvalds
From: Colin Ian King <colin.king@canonical.com>
Subject: ocfs2: remove redundant assignment to pointer queue
The pointer queue is being initialized with a value that is never read and
it is being updated later with a new value. The initialization is
redundant and can be removed.
Addresses-Coverity: ("Unused value")
Link: https://lkml.kernel.org/r/20210513113957.57539-1-colin.king@canonical.com
Signed-off-by: Colin Ian King <colin.king@canonical.com>
Acked-by: Joseph Qi <joseph.qi@linux.alibaba.com>
Cc: Mark Fasheh <mark@fasheh.com>
Cc: Joel Becker <jlbec@evilplan.org>
Cc: Junxiao Bi <junxiao.bi@oracle.com>
Cc: Changwei Ge <gechangwei@live.cn>
Cc: Gang He <ghe@suse.com>
Cc: Jun Piao <piaojun@huawei.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
fs/ocfs2/dlm/dlmmaster.c | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
--- a/fs/ocfs2/dlm/dlmmaster.c~ocfs2-remove-redundant-assignment-to-pointer-queue
+++ a/fs/ocfs2/dlm/dlmmaster.c
@@ -2977,7 +2977,7 @@ static u8 dlm_pick_migration_target(stru
struct dlm_lock_resource *res)
{
enum dlm_lockres_list idx;
- struct list_head *queue = &res->granted;
+ struct list_head *queue;
struct dlm_lock *lock;
int noderef;
u8 nodenum = O2NM_MAX_NODES;
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 016/192] ocfs2: remove repeated uptodate check for buffer
2021-06-29 2:32 incoming Andrew Morton
` (14 preceding siblings ...)
2021-06-29 2:34 ` [patch 015/192] ocfs2: remove redundant assignment to pointer queue Andrew Morton
@ 2021-06-29 2:34 ` Andrew Morton
2021-06-29 2:34 ` [patch 017/192] ocfs2: replace simple_strtoull() with kstrtoull() Andrew Morton
` (175 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:34 UTC (permalink / raw)
To: akpm, gechangwei, ghe, jlbec, joseph.qi, junxiao.bi, linux-mm,
mark, mm-commits, piaojun, torvalds, wanjiabing
From: Wan Jiabing <wanjiabing@vivo.com>
Subject: ocfs2: remove repeated uptodate check for buffer
In commit 60f91826ca62 ("buffer: Avoid setting buffer bits that are
already set"), function set_buffer_##name was added a test_bit() to check
buffer, which is the same as function buffer_##name. The
!buffer_uptodate(bh) here is a repeated check. Remove it.
Link: https://lkml.kernel.org/r/20210425025702.13628-1-wanjiabing@vivo.com
Signed-off-by: Wan Jiabing <wanjiabing@vivo.com>
Reviewed-by: Joseph Qi <joseph.qi@linux.alibaba.com>
Cc: Mark Fasheh <mark@fasheh.com>
Cc: Joel Becker <jlbec@evilplan.org>
Cc: Junxiao Bi <junxiao.bi@oracle.com>
Cc: Changwei Ge <gechangwei@live.cn>
Cc: Gang He <ghe@suse.com>
Cc: Jun Piao <piaojun@huawei.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
fs/ocfs2/aops.c | 3 +--
1 file changed, 1 insertion(+), 2 deletions(-)
--- a/fs/ocfs2/aops.c~ocfs2-remove-repeated-uptodate-check-for-buffer
+++ a/fs/ocfs2/aops.c
@@ -632,8 +632,7 @@ int ocfs2_map_page_blocks(struct page *p
}
if (PageUptodate(page)) {
- if (!buffer_uptodate(bh))
- set_buffer_uptodate(bh);
+ set_buffer_uptodate(bh);
} else if (!buffer_uptodate(bh) && !buffer_delay(bh) &&
!buffer_new(bh) &&
ocfs2_should_read_blk(inode, page, block_start) &&
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 017/192] ocfs2: replace simple_strtoull() with kstrtoull()
2021-06-29 2:32 incoming Andrew Morton
` (15 preceding siblings ...)
2021-06-29 2:34 ` [patch 016/192] ocfs2: remove repeated uptodate check for buffer Andrew Morton
@ 2021-06-29 2:34 ` Andrew Morton
2021-06-29 2:34 ` [patch 018/192] ocfs2: remove redundant initialization of variable ret Andrew Morton
` (174 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:34 UTC (permalink / raw)
To: akpm, chenhuang5, gechangwei, ghe, jiangqi903, jlbec, junxiao.bi,
linux-mm, mark, mm-commits, piaojun, torvalds
From: Chen Huang <chenhuang5@huawei.com>
Subject: ocfs2: replace simple_strtoull() with kstrtoull()
simple_strtoull() is deprecated in some situation since it does not check
for the range overflow, use kstrtoull() instead.
Link: https://lkml.kernel.org/r/20210526092020.554341-3-chenhuang5@huawei.com
Signed-off-by: Chen Huang <chenhuang5@huawei.com>
Cc: Mark Fasheh <mark@fasheh.com>
Cc: Joel Becker <jlbec@evilplan.org>
Cc: Junxiao Bi <junxiao.bi@oracle.com>
Cc: Joseph Qi <jiangqi903@gmail.com>
Cc: Changwei Ge <gechangwei@live.cn>
Cc: Gang He <ghe@suse.com>
Cc: Jun Piao <piaojun@huawei.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
fs/ocfs2/cluster/heartbeat.c | 5 +++--
1 file changed, 3 insertions(+), 2 deletions(-)
--- a/fs/ocfs2/cluster/heartbeat.c~ocfs2-replaced-simple_strtoull-with-kstrtoull
+++ a/fs/ocfs2/cluster/heartbeat.c
@@ -1596,12 +1596,13 @@ static ssize_t o2hb_region_start_block_s
struct o2hb_region *reg = to_o2hb_region(item);
unsigned long long tmp;
char *p = (char *)page;
+ ssize_t ret;
if (reg->hr_bdev)
return -EINVAL;
- tmp = simple_strtoull(p, &p, 0);
- if (!p || (*p && (*p != '\n')))
+ ret = kstrtoull(p, 0, &tmp);
+ if (ret)
return -EINVAL;
reg->hr_start_block = tmp;
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 018/192] ocfs2: remove redundant initialization of variable ret
2021-06-29 2:32 incoming Andrew Morton
` (16 preceding siblings ...)
2021-06-29 2:34 ` [patch 017/192] ocfs2: replace simple_strtoull() with kstrtoull() Andrew Morton
@ 2021-06-29 2:34 ` Andrew Morton
2021-06-29 2:34 ` [patch 019/192] kernel: watchdog: modify the explanation related to watchdog thread Andrew Morton
` (173 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:34 UTC (permalink / raw)
To: akpm, colin.king, gechangwei, ghe, jlbec, joseph.qi, junxiao.bi,
linux-mm, mark, mm-commits, piaojun, torvalds
From: Colin Ian King <colin.king@canonical.com>
Subject: ocfs2: remove redundant initialization of variable ret
The variable ret is being initialized with a value that is never read, the
assignment is redundant and can be removed.
Addresses-Coverity: ("Unused value")
Link: https://lkml.kernel.org/r/20210613135148.74658-1-colin.king@canonical.com
Signed-off-by: Colin Ian King <colin.king@canonical.com>
Acked-by: Joseph Qi <joseph.qi@linux.alibaba.com>
Cc: Mark Fasheh <mark@fasheh.com>
Cc: Joel Becker <jlbec@evilplan.org>
Cc: Junxiao Bi <junxiao.bi@oracle.com>
Cc: Changwei Ge <gechangwei@live.cn>
Cc: Gang He <ghe@suse.com>
Cc: Jun Piao <piaojun@huawei.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
fs/ocfs2/cluster/nodemanager.c | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
--- a/fs/ocfs2/cluster/nodemanager.c~ocfs2-remove-redundant-initialization-of-variable-ret
+++ a/fs/ocfs2/cluster/nodemanager.c
@@ -824,7 +824,7 @@ static void __exit exit_o2nm(void)
static int __init init_o2nm(void)
{
- int ret = -1;
+ int ret;
o2hb_init();
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 019/192] kernel: watchdog: modify the explanation related to watchdog thread
2021-06-29 2:32 incoming Andrew Morton
` (17 preceding siblings ...)
2021-06-29 2:34 ` [patch 018/192] ocfs2: remove redundant initialization of variable ret Andrew Morton
@ 2021-06-29 2:34 ` Andrew Morton
2021-06-29 2:34 ` [patch 020/192] doc: " Andrew Morton
` (172 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:34 UTC (permalink / raw)
To: akpm, corbet, gpiccoli, joe, keescook, linux-mm, mchehab+huawei,
mm-commits, pmladek, qais.yousef, rdunlap, santosh, steve,
torvalds, vbabka, wangqing
From: Wang Qing <wangqing@vivo.com>
Subject: kernel: watchdog: modify the explanation related to watchdog thread
The watchdog thread has been replaced by cpu_stop_work, modify the
explanation related.
Link: https://lkml.kernel.org/r/1619687073-24686-2-git-send-email-wangqing@vivo.com
Signed-off-by: Wang Qing <wangqing@vivo.com>
Reviewed-by: Petr Mladek <pmladek@suse.com>
Cc: Jonathan Corbet <corbet@lwn.net>
Cc: Mauro Carvalho Chehab <mchehab+huawei@kernel.org>
Cc: Joe Perches <joe@perches.com>
Cc: Stephen Kitt <steve@sk2.org>
Cc: Kees Cook <keescook@chromium.org>
Cc: Randy Dunlap <rdunlap@infradead.org>
Cc: "Guilherme G. Piccoli" <gpiccoli@canonical.com>
Cc: Qais Yousef <qais.yousef@arm.com>
Cc: Santosh Sivaraj <santosh@fossix.org>
Cc: Vlastimil Babka <vbabka@suse.cz>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
kernel/watchdog.c | 12 ++++--------
1 file changed, 4 insertions(+), 8 deletions(-)
--- a/kernel/watchdog.c~kernel-watchdog-modify-the-explanation-related-to-watchdog-thread
+++ a/kernel/watchdog.c
@@ -92,7 +92,7 @@ __setup("nmi_watchdog=", hardlockup_pani
* own hardlockup detector.
*
* watchdog_nmi_enable/disable can be implemented to start and stop when
- * softlockup watchdog threads start and stop. The arch must select the
+ * softlockup watchdog start and stop. The arch must select the
* SOFTLOCKUP_DETECTOR Kconfig.
*/
int __weak watchdog_nmi_enable(unsigned int cpu)
@@ -335,7 +335,7 @@ static DEFINE_PER_CPU(struct completion,
static DEFINE_PER_CPU(struct cpu_stop_work, softlockup_stop_work);
/*
- * The watchdog thread function - touches the timestamp.
+ * The watchdog feed function - touches the timestamp.
*
* It only runs once every sample_period seconds (4 seconds by
* default) to reset the softlockup timestamp. If this gets delayed
@@ -558,11 +558,7 @@ static void lockup_detector_reconfigure(
}
/*
- * Create the watchdog thread infrastructure and configure the detector(s).
- *
- * The threads are not unparked as watchdog_allowed_mask is empty. When
- * the threads are successfully initialized, take the proper locks and
- * unpark the threads in the watchdog_cpumask if the watchdog is enabled.
+ * Create the watchdog infrastructure and configure the detector(s).
*/
static __init void lockup_detector_setup(void)
{
@@ -628,7 +624,7 @@ void lockup_detector_soft_poweroff(void)
#ifdef CONFIG_SYSCTL
-/* Propagate any changes to the watchdog threads */
+/* Propagate any changes to the watchdog infrastructure */
static void proc_watchdog_update(void)
{
/* Remove impossible cpus to keep sysctl output clean. */
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 020/192] doc: watchdog: modify the explanation related to watchdog thread
2021-06-29 2:32 incoming Andrew Morton
` (18 preceding siblings ...)
2021-06-29 2:34 ` [patch 019/192] kernel: watchdog: modify the explanation related to watchdog thread Andrew Morton
@ 2021-06-29 2:34 ` Andrew Morton
2021-06-29 2:34 ` [patch 021/192] doc: watchdog: modify the doc related to "watchdog/%u" Andrew Morton
` (171 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:34 UTC (permalink / raw)
To: akpm, corbet, gpiccoli, joe, keescook, linux-mm, mchehab+huawei,
mm-commits, pmladek, qais.yousef, rdunlap, santosh, steve,
torvalds, vbabka, wangqing
From: Wang Qing <wangqing@vivo.com>
Subject: doc: watchdog: modify the explanation related to watchdog thread
"watchdog/%u" threads has be replaced by cpu_stop_work. The current
description is extremely misleading.
Link: https://lkml.kernel.org/r/1619687073-24686-4-git-send-email-wangqing@vivo.com
Signed-off-by: Wang Qing <wangqing@vivo.com>
Reviewed-by: Petr Mladek <pmladek@suse.com>
Cc: "Guilherme G. Piccoli" <gpiccoli@canonical.com>
Cc: Joe Perches <joe@perches.com>
Cc: Jonathan Corbet <corbet@lwn.net>
Cc: Kees Cook <keescook@chromium.org>
Cc: Mauro Carvalho Chehab <mchehab+huawei@kernel.org>
Cc: Qais Yousef <qais.yousef@arm.com>
Cc: Randy Dunlap <rdunlap@infradead.org>
Cc: Santosh Sivaraj <santosh@fossix.org>
Cc: Stephen Kitt <steve@sk2.org>
Cc: Vlastimil Babka <vbabka@suse.cz>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
Documentation/admin-guide/lockup-watchdogs.rst | 4 ++--
1 file changed, 2 insertions(+), 2 deletions(-)
--- a/Documentation/admin-guide/lockup-watchdogs.rst~doc-watchdog-modify-the-explanation-related-to-watchdog-thread
+++ a/Documentation/admin-guide/lockup-watchdogs.rst
@@ -39,7 +39,7 @@ in principle, they should work in any ar
subsystems are present.
A periodic hrtimer runs to generate interrupts and kick the watchdog
-task. An NMI perf event is generated every "watchdog_thresh"
+job. An NMI perf event is generated every "watchdog_thresh"
(compile-time initialized to 10 and configurable through sysctl of the
same name) seconds to check for hardlockups. If any CPU in the system
does not receive any hrtimer interrupt during that time the
@@ -47,7 +47,7 @@ does not receive any hrtimer interrupt d
generate a kernel warning or call panic, depending on the
configuration.
-The watchdog task is a high priority kernel thread that updates a
+The watchdog job runs in a stop scheduling thread that updates a
timestamp every time it is scheduled. If that timestamp is not updated
for 2*watchdog_thresh seconds (the softlockup threshold) the
'softlockup detector' (coded inside the hrtimer callback function)
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 021/192] doc: watchdog: modify the doc related to "watchdog/%u"
2021-06-29 2:32 incoming Andrew Morton
` (19 preceding siblings ...)
2021-06-29 2:34 ` [patch 020/192] doc: " Andrew Morton
@ 2021-06-29 2:34 ` Andrew Morton
2021-06-29 2:34 ` [patch 022/192] slab: use __func__ to trace function name Andrew Morton
` (170 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:34 UTC (permalink / raw)
To: akpm, corbet, gpiccoli, joe, keescook, linux-mm, mchehab+huawei,
mm-commits, pmladek, qais.yousef, rdunlap, santosh, steve,
torvalds, vbabka, wangqing
From: Wang Qing <wangqing@vivo.com>
Subject: doc: watchdog: modify the doc related to "watchdog/%u"
"watchdog/%u" threads has be replaced by cpu_stop_work. The current
description is extremely misleading.
Link: https://lkml.kernel.org/r/1619687073-24686-5-git-send-email-wangqing@vivo.com
Signed-off-by: Wang Qing <wangqing@vivo.com>
Reviewed-by: Petr Mladek <pmladek@suse.com>
Cc: "Guilherme G. Piccoli" <gpiccoli@canonical.com>
Cc: Joe Perches <joe@perches.com>
Cc: Jonathan Corbet <corbet@lwn.net>
Cc: Kees Cook <keescook@chromium.org>
Cc: Mauro Carvalho Chehab <mchehab+huawei@kernel.org>
Cc: Qais Yousef <qais.yousef@arm.com>
Cc: Randy Dunlap <rdunlap@infradead.org>
Cc: Santosh Sivaraj <santosh@fossix.org>
Cc: Stephen Kitt <steve@sk2.org>
Cc: Vlastimil Babka <vbabka@suse.cz>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
Documentation/admin-guide/sysctl/kernel.rst | 10 +++++-----
1 file changed, 5 insertions(+), 5 deletions(-)
--- a/Documentation/admin-guide/sysctl/kernel.rst~doc-watchdog-modify-the-doc-related-to-watchdog-%u
+++ a/Documentation/admin-guide/sysctl/kernel.rst
@@ -1283,11 +1283,11 @@ This parameter can be used to control th
= =================================
The soft lockup detector monitors CPUs for threads that are hogging the CPUs
-without rescheduling voluntarily, and thus prevent the 'watchdog/N' threads
-from running. The mechanism depends on the CPUs ability to respond to timer
-interrupts which are needed for the 'watchdog/N' threads to be woken up by
-the watchdog timer function, otherwise the NMI watchdog — if enabled — can
-detect a hard lockup condition.
+without rescheduling voluntarily, and thus prevent the 'migration/N' threads
+from running, causing the watchdog work fail to execute. The mechanism depends
+on the CPUs ability to respond to timer interrupts which are needed for the
+watchdog work to be queued by the watchdog timer function, otherwise the NMI
+watchdog — if enabled — can detect a hard lockup condition.
stack_erasing
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 022/192] slab: use __func__ to trace function name
2021-06-29 2:32 incoming Andrew Morton
` (20 preceding siblings ...)
2021-06-29 2:34 ` [patch 021/192] doc: watchdog: modify the doc related to "watchdog/%u" Andrew Morton
@ 2021-06-29 2:34 ` Andrew Morton
2021-06-29 2:34 ` [patch 023/192] kunit: make test->lock irq safe Andrew Morton
` (169 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:34 UTC (permalink / raw)
To: akpm, atomlin, cl, gumingtao1225, gumingtao, iamjoonsoo.kim,
linux-mm, mm-commits, penberg, rientjes, torvalds, vbabka
From: gumingtao <gumingtao1225@gmail.com>
Subject: slab: use __func__ to trace function name
It is better to use __func__ to trace function name.
Link: https://lkml.kernel.org/r/31fdbad5c45cd1e26be9ff37be321b8586b80fee.1624355507.git.gumingtao@xiaomi.com
Signed-off-by: gumingtao <gumingtao@xiaomi.com>
Acked-by: Christoph Lameter <cl@linux.com>
Acked-by: David Rientjes <rientjes@google.com>
Reviewed-by: Aaron Tomlin <atomlin@redhat.com>
Cc: Pekka Enberg <penberg@kernel.org>
Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Cc: Vlastimil Babka <vbabka@suse.cz>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/slab_common.c | 12 ++++++------
1 file changed, 6 insertions(+), 6 deletions(-)
--- a/mm/slab_common.c~slab-use-__func__-to-trace-function-name
+++ a/mm/slab_common.c
@@ -377,11 +377,11 @@ out_unlock:
if (err) {
if (flags & SLAB_PANIC)
- panic("kmem_cache_create: Failed to create slab '%s'. Error %d\n",
- name, err);
+ panic("%s: Failed to create slab '%s'. Error %d\n",
+ __func__, name, err);
else {
- pr_warn("kmem_cache_create(%s) failed with error %d\n",
- name, err);
+ pr_warn("%s(%s) failed with error %d\n",
+ __func__, name, err);
dump_stack();
}
return NULL;
@@ -508,8 +508,8 @@ void kmem_cache_destroy(struct kmem_cach
err = shutdown_cache(s);
if (err) {
- pr_err("kmem_cache_destroy %s: Slab cache still has objects\n",
- s->name);
+ pr_err("%s %s: Slab cache still has objects\n",
+ __func__, s->name);
dump_stack();
}
out_unlock:
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 023/192] kunit: make test->lock irq safe
2021-06-29 2:32 incoming Andrew Morton
` (21 preceding siblings ...)
2021-06-29 2:34 ` [patch 022/192] slab: use __func__ to trace function name Andrew Morton
@ 2021-06-29 2:34 ` Andrew Morton
2021-06-29 2:34 ` [patch 024/192] mm/slub, kunit: add a KUnit test for SLUB debugging functionality Andrew Morton
` (168 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:34 UTC (permalink / raw)
To: akpm, brendanhiggins, cl, dlatypov, elver, glittao,
iamjoonsoo.kim, linux-mm, mm-commits, penberg, rientjes,
torvalds, vbabka
From: Vlastimil Babka <vbabka@suse.cz>
Subject: kunit: make test->lock irq safe
The upcoming SLUB kunit test will be calling kunit_find_named_resource()
from a context with disabled interrupts. That means kunit's test->lock
needs to be IRQ safe to avoid potential deadlocks and lockdep splats.
This patch therefore changes the test->lock usage to spin_lock_irqsave()
and spin_unlock_irqrestore().
Link: https://lkml.kernel.org/r/20210511150734.3492-1-glittao@gmail.com
Signed-off-by: Vlastimil Babka <vbabka@suse.cz>
Signed-off-by: Oliver Glitta <glittao@gmail.com>
Reviewed-by: Brendan Higgins <brendanhiggins@google.com>
Cc: Christoph Lameter <cl@linux.com>
Cc: Daniel Latypov <dlatypov@google.com>
Cc: David Rientjes <rientjes@google.com>
Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Cc: Marco Elver <elver@google.com>
Cc: Pekka Enberg <penberg@kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/kunit/test.h | 5 +++--
lib/kunit/test.c | 18 +++++++++++-------
2 files changed, 14 insertions(+), 9 deletions(-)
--- a/include/kunit/test.h~kunit-make-test-lock-irq-safe
+++ a/include/kunit/test.h
@@ -515,8 +515,9 @@ kunit_find_resource(struct kunit *test,
void *match_data)
{
struct kunit_resource *res, *found = NULL;
+ unsigned long flags;
- spin_lock(&test->lock);
+ spin_lock_irqsave(&test->lock, flags);
list_for_each_entry_reverse(res, &test->resources, node) {
if (match(test, res, (void *)match_data)) {
@@ -526,7 +527,7 @@ kunit_find_resource(struct kunit *test,
}
}
- spin_unlock(&test->lock);
+ spin_unlock_irqrestore(&test->lock, flags);
return found;
}
--- a/lib/kunit/test.c~kunit-make-test-lock-irq-safe
+++ a/lib/kunit/test.c
@@ -475,6 +475,7 @@ int kunit_add_resource(struct kunit *tes
void *data)
{
int ret = 0;
+ unsigned long flags;
res->free = free;
kref_init(&res->refcount);
@@ -487,10 +488,10 @@ int kunit_add_resource(struct kunit *tes
res->data = data;
}
- spin_lock(&test->lock);
+ spin_lock_irqsave(&test->lock, flags);
list_add_tail(&res->node, &test->resources);
/* refcount for list is established by kref_init() */
- spin_unlock(&test->lock);
+ spin_unlock_irqrestore(&test->lock, flags);
return ret;
}
@@ -548,9 +549,11 @@ EXPORT_SYMBOL_GPL(kunit_alloc_and_get_re
void kunit_remove_resource(struct kunit *test, struct kunit_resource *res)
{
- spin_lock(&test->lock);
+ unsigned long flags;
+
+ spin_lock_irqsave(&test->lock, flags);
list_del(&res->node);
- spin_unlock(&test->lock);
+ spin_unlock_irqrestore(&test->lock, flags);
kunit_put_resource(res);
}
EXPORT_SYMBOL_GPL(kunit_remove_resource);
@@ -630,6 +633,7 @@ EXPORT_SYMBOL_GPL(kunit_kfree);
void kunit_cleanup(struct kunit *test)
{
struct kunit_resource *res;
+ unsigned long flags;
/*
* test->resources is a stack - each allocation must be freed in the
@@ -641,9 +645,9 @@ void kunit_cleanup(struct kunit *test)
* protect against the current node being deleted, not the next.
*/
while (true) {
- spin_lock(&test->lock);
+ spin_lock_irqsave(&test->lock, flags);
if (list_empty(&test->resources)) {
- spin_unlock(&test->lock);
+ spin_unlock_irqrestore(&test->lock, flags);
break;
}
res = list_last_entry(&test->resources,
@@ -654,7 +658,7 @@ void kunit_cleanup(struct kunit *test)
* resource, and this can't happen if the test->lock
* is held.
*/
- spin_unlock(&test->lock);
+ spin_unlock_irqrestore(&test->lock, flags);
kunit_remove_resource(test, res);
}
current->kunit_test = NULL;
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 024/192] mm/slub, kunit: add a KUnit test for SLUB debugging functionality
2021-06-29 2:32 incoming Andrew Morton
` (22 preceding siblings ...)
2021-06-29 2:34 ` [patch 023/192] kunit: make test->lock irq safe Andrew Morton
@ 2021-06-29 2:34 ` Andrew Morton
2021-06-29 2:34 ` [patch 025/192] slub: remove resiliency_test() function Andrew Morton
` (167 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:34 UTC (permalink / raw)
To: akpm, brendanhiggins, cl, dlatypov, elver, glittao,
iamjoonsoo.kim, linux-mm, mm-commits, penberg, rientjes,
torvalds, vbabka
From: Oliver Glitta <glittao@gmail.com>
Subject: mm/slub, kunit: add a KUnit test for SLUB debugging functionality
SLUB has resiliency_test() function which is hidden behind #ifdef
SLUB_RESILIENCY_TEST that is not part of Kconfig, so nobody runs it.
KUnit should be a proper replacement for it.
Try changing byte in redzone after allocation and changing pointer to next
free node, first byte, 50th byte and redzone byte. Check if validation
finds errors.
There are several differences from the original resiliency test: Tests
create own caches with known state instead of corrupting shared kmalloc
caches.
The corruption of freepointer uses correct offset, the original resiliency
test got broken with freepointer changes.
Scratch changing random byte test, because it does not have meaning in
this form where we need deterministic results.
Add new option CONFIG_SLUB_KUNIT_TEST in Kconfig. Tests next_pointer,
first_word and clobber_50th_byte do not run with KASAN option on. Because
the test deliberately modifies non-allocated objects.
Use kunit_resource to count errors in cache and silence bug reports.
Count error whenever slab_bug() or slab_fix() is called or when the count
of pages is wrong.
[glittao@gmail.com: remove unused function test_exit(), from SLUB KUnit test]
Link: https://lkml.kernel.org/r/20210512140656.12083-1-glittao@gmail.com
[akpm@linux-foundation.org: export kasan_enable/disable_current to modules]
Link: https://lkml.kernel.org/r/20210511150734.3492-2-glittao@gmail.com
Signed-off-by: Oliver Glitta <glittao@gmail.com>
Reviewed-by: Vlastimil Babka <vbabka@suse.cz>
Acked-by: Daniel Latypov <dlatypov@google.com>
Acked-by: Marco Elver <elver@google.com>
Cc: Brendan Higgins <brendanhiggins@google.com>
Cc: Christoph Lameter <cl@linux.com>
Cc: David Rientjes <rientjes@google.com>
Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Cc: Pekka Enberg <penberg@kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
lib/Kconfig.debug | 12 +++
lib/Makefile | 1
lib/slub_kunit.c | 152 ++++++++++++++++++++++++++++++++++++++++++++
mm/kasan/common.c | 3
mm/slab.h | 1
mm/slub.c | 46 ++++++++++++-
6 files changed, 212 insertions(+), 3 deletions(-)
--- a/lib/Kconfig.debug~mm-slub-kunit-add-a-kunit-test-for-slub-debugging-functionality
+++ a/lib/Kconfig.debug
@@ -2429,6 +2429,18 @@ config BITS_TEST
If unsure, say N.
+config SLUB_KUNIT_TEST
+ tristate "KUnit test for SLUB cache error detection" if !KUNIT_ALL_TESTS
+ depends on SLUB_DEBUG && KUNIT
+ default KUNIT_ALL_TESTS
+ help
+ This builds SLUB allocator unit test.
+ Tests SLUB cache debugging functionality.
+ For more information on KUnit and unit tests in general please refer
+ to the KUnit documentation in Documentation/dev-tools/kunit/.
+
+ If unsure, say N.
+
config TEST_UDELAY
tristate "udelay test driver"
help
--- a/lib/Makefile~mm-slub-kunit-add-a-kunit-test-for-slub-debugging-functionality
+++ a/lib/Makefile
@@ -354,5 +354,6 @@ obj-$(CONFIG_LIST_KUNIT_TEST) += list-te
obj-$(CONFIG_LINEAR_RANGES_TEST) += test_linear_ranges.o
obj-$(CONFIG_BITS_TEST) += test_bits.o
obj-$(CONFIG_CMDLINE_KUNIT_TEST) += cmdline_kunit.o
+obj-$(CONFIG_SLUB_KUNIT_TEST) += slub_kunit.o
obj-$(CONFIG_GENERIC_LIB_DEVMEM_IS_ALLOWED) += devmem_is_allowed.o
--- /dev/null
+++ a/lib/slub_kunit.c
@@ -0,0 +1,152 @@
+// SPDX-License-Identifier: GPL-2.0
+#include <kunit/test.h>
+#include <linux/mm.h>
+#include <linux/slab.h>
+#include <linux/module.h>
+#include <linux/kernel.h>
+#include "../mm/slab.h"
+
+static struct kunit_resource resource;
+static int slab_errors;
+
+static void test_clobber_zone(struct kunit *test)
+{
+ struct kmem_cache *s = kmem_cache_create("TestSlub_RZ_alloc", 64, 0,
+ SLAB_RED_ZONE, NULL);
+ u8 *p = kmem_cache_alloc(s, GFP_KERNEL);
+
+ kasan_disable_current();
+ p[64] = 0x12;
+
+ validate_slab_cache(s);
+ KUNIT_EXPECT_EQ(test, 2, slab_errors);
+
+ kasan_enable_current();
+ kmem_cache_free(s, p);
+ kmem_cache_destroy(s);
+}
+
+#ifndef CONFIG_KASAN
+static void test_next_pointer(struct kunit *test)
+{
+ struct kmem_cache *s = kmem_cache_create("TestSlub_next_ptr_free", 64, 0,
+ SLAB_POISON, NULL);
+ u8 *p = kmem_cache_alloc(s, GFP_KERNEL);
+ unsigned long tmp;
+ unsigned long *ptr_addr;
+
+ kmem_cache_free(s, p);
+
+ ptr_addr = (unsigned long *)(p + s->offset);
+ tmp = *ptr_addr;
+ p[s->offset] = 0x12;
+
+ /*
+ * Expecting three errors.
+ * One for the corrupted freechain and the other one for the wrong
+ * count of objects in use. The third error is fixing broken cache.
+ */
+ validate_slab_cache(s);
+ KUNIT_EXPECT_EQ(test, 3, slab_errors);
+
+ /*
+ * Try to repair corrupted freepointer.
+ * Still expecting two errors. The first for the wrong count
+ * of objects in use.
+ * The second error is for fixing broken cache.
+ */
+ *ptr_addr = tmp;
+ slab_errors = 0;
+
+ validate_slab_cache(s);
+ KUNIT_EXPECT_EQ(test, 2, slab_errors);
+
+ /*
+ * Previous validation repaired the count of objects in use.
+ * Now expecting no error.
+ */
+ slab_errors = 0;
+ validate_slab_cache(s);
+ KUNIT_EXPECT_EQ(test, 0, slab_errors);
+
+ kmem_cache_destroy(s);
+}
+
+static void test_first_word(struct kunit *test)
+{
+ struct kmem_cache *s = kmem_cache_create("TestSlub_1th_word_free", 64, 0,
+ SLAB_POISON, NULL);
+ u8 *p = kmem_cache_alloc(s, GFP_KERNEL);
+
+ kmem_cache_free(s, p);
+ *p = 0x78;
+
+ validate_slab_cache(s);
+ KUNIT_EXPECT_EQ(test, 2, slab_errors);
+
+ kmem_cache_destroy(s);
+}
+
+static void test_clobber_50th_byte(struct kunit *test)
+{
+ struct kmem_cache *s = kmem_cache_create("TestSlub_50th_word_free", 64, 0,
+ SLAB_POISON, NULL);
+ u8 *p = kmem_cache_alloc(s, GFP_KERNEL);
+
+ kmem_cache_free(s, p);
+ p[50] = 0x9a;
+
+ validate_slab_cache(s);
+ KUNIT_EXPECT_EQ(test, 2, slab_errors);
+
+ kmem_cache_destroy(s);
+}
+#endif
+
+static void test_clobber_redzone_free(struct kunit *test)
+{
+ struct kmem_cache *s = kmem_cache_create("TestSlub_RZ_free", 64, 0,
+ SLAB_RED_ZONE, NULL);
+ u8 *p = kmem_cache_alloc(s, GFP_KERNEL);
+
+ kasan_disable_current();
+ kmem_cache_free(s, p);
+ p[64] = 0xab;
+
+ validate_slab_cache(s);
+ KUNIT_EXPECT_EQ(test, 2, slab_errors);
+
+ kasan_enable_current();
+ kmem_cache_destroy(s);
+}
+
+static int test_init(struct kunit *test)
+{
+ slab_errors = 0;
+
+ kunit_add_named_resource(test, NULL, NULL, &resource,
+ "slab_errors", &slab_errors);
+ return 0;
+}
+
+static struct kunit_case test_cases[] = {
+ KUNIT_CASE(test_clobber_zone),
+
+#ifndef CONFIG_KASAN
+ KUNIT_CASE(test_next_pointer),
+ KUNIT_CASE(test_first_word),
+ KUNIT_CASE(test_clobber_50th_byte),
+#endif
+
+ KUNIT_CASE(test_clobber_redzone_free),
+ {}
+};
+
+static struct kunit_suite test_suite = {
+ .name = "slub_test",
+ .init = test_init,
+ .test_cases = test_cases,
+};
+kunit_test_suite(test_suite);
+
+MODULE_LICENSE("GPL");
--- a/mm/kasan/common.c~mm-slub-kunit-add-a-kunit-test-for-slub-debugging-functionality
+++ a/mm/kasan/common.c
@@ -51,11 +51,14 @@ void kasan_enable_current(void)
{
current->kasan_depth++;
}
+EXPORT_SYMBOL(kasan_enable_current);
void kasan_disable_current(void)
{
current->kasan_depth--;
}
+EXPORT_SYMBOL(kasan_disable_current);
+
#endif /* CONFIG_KASAN_GENERIC || CONFIG_KASAN_SW_TAGS */
void __kasan_unpoison_range(const void *address, size_t size)
--- a/mm/slab.h~mm-slub-kunit-add-a-kunit-test-for-slub-debugging-functionality
+++ a/mm/slab.h
@@ -215,6 +215,7 @@ DECLARE_STATIC_KEY_TRUE(slub_debug_enabl
DECLARE_STATIC_KEY_FALSE(slub_debug_enabled);
#endif
extern void print_tracking(struct kmem_cache *s, void *object);
+long validate_slab_cache(struct kmem_cache *s);
#else
static inline void print_tracking(struct kmem_cache *s, void *object)
{
--- a/mm/slub.c~mm-slub-kunit-add-a-kunit-test-for-slub-debugging-functionality
+++ a/mm/slub.c
@@ -36,6 +36,7 @@
#include <linux/prefetch.h>
#include <linux/memcontrol.h>
#include <linux/random.h>
+#include <kunit/test.h>
#include <trace/events/kmem.h>
@@ -449,6 +450,26 @@ static inline bool cmpxchg_double_slab(s
static unsigned long object_map[BITS_TO_LONGS(MAX_OBJS_PER_PAGE)];
static DEFINE_SPINLOCK(object_map_lock);
+#if IS_ENABLED(CONFIG_KUNIT)
+static bool slab_add_kunit_errors(void)
+{
+ struct kunit_resource *resource;
+
+ if (likely(!current->kunit_test))
+ return false;
+
+ resource = kunit_find_named_resource(current->kunit_test, "slab_errors");
+ if (!resource)
+ return false;
+
+ (*(int *)resource->data)++;
+ kunit_put_resource(resource);
+ return true;
+}
+#else
+static inline bool slab_add_kunit_errors(void) { return false; }
+#endif
+
/*
* Determine a map of object in use on a page.
*
@@ -679,6 +700,9 @@ static void slab_fix(struct kmem_cache *
struct va_format vaf;
va_list args;
+ if (slab_add_kunit_errors())
+ return;
+
va_start(args, fmt);
vaf.fmt = fmt;
vaf.va = &args;
@@ -742,6 +766,9 @@ static void print_trailer(struct kmem_ca
void object_err(struct kmem_cache *s, struct page *page,
u8 *object, char *reason)
{
+ if (slab_add_kunit_errors())
+ return;
+
slab_bug(s, "%s", reason);
print_trailer(s, page, object);
}
@@ -752,6 +779,9 @@ static __printf(3, 4) void slab_err(stru
va_list args;
char buf[100];
+ if (slab_add_kunit_errors())
+ return;
+
va_start(args, fmt);
vsnprintf(buf, sizeof(buf), fmt, args);
va_end(args);
@@ -801,12 +831,16 @@ static int check_bytes_and_report(struct
while (end > fault && end[-1] == value)
end--;
+ if (slab_add_kunit_errors())
+ goto skip_bug_print;
+
slab_bug(s, "%s overwritten", what);
pr_err("0x%p-0x%p @offset=%tu. First byte 0x%x instead of 0x%x\n",
fault, end - 1, fault - addr,
fault[0], value);
print_trailer(s, page, object);
+skip_bug_print:
restore_bytes(s, what, value, fault, end);
return 0;
}
@@ -4649,9 +4683,11 @@ static int validate_slab_node(struct kme
validate_slab(s, page);
count++;
}
- if (count != n->nr_partial)
+ if (count != n->nr_partial) {
pr_err("SLUB %s: %ld partial slabs counted but counter=%ld\n",
s->name, count, n->nr_partial);
+ slab_add_kunit_errors();
+ }
if (!(s->flags & SLAB_STORE_USER))
goto out;
@@ -4660,16 +4696,18 @@ static int validate_slab_node(struct kme
validate_slab(s, page);
count++;
}
- if (count != atomic_long_read(&n->nr_slabs))
+ if (count != atomic_long_read(&n->nr_slabs)) {
pr_err("SLUB: %s %ld slabs counted but counter=%ld\n",
s->name, count, atomic_long_read(&n->nr_slabs));
+ slab_add_kunit_errors();
+ }
out:
spin_unlock_irqrestore(&n->list_lock, flags);
return count;
}
-static long validate_slab_cache(struct kmem_cache *s)
+long validate_slab_cache(struct kmem_cache *s)
{
int node;
unsigned long count = 0;
@@ -4681,6 +4719,8 @@ static long validate_slab_cache(struct k
return count;
}
+EXPORT_SYMBOL(validate_slab_cache);
+
/*
* Generate lists of code addresses where slabcache objects are allocated
* and freed.
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 025/192] slub: remove resiliency_test() function
2021-06-29 2:32 incoming Andrew Morton
` (23 preceding siblings ...)
2021-06-29 2:34 ` [patch 024/192] mm/slub, kunit: add a KUnit test for SLUB debugging functionality Andrew Morton
@ 2021-06-29 2:34 ` Andrew Morton
2021-06-29 2:34 ` [patch 026/192] mm, slub: change run-time assertion in kmalloc_index() to compile-time Andrew Morton
` (166 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:34 UTC (permalink / raw)
To: akpm, brendanhiggins, cl, dlatypov, elver, glittao,
iamjoonsoo.kim, linux-mm, mm-commits, penberg, rientjes,
torvalds, vbabka
From: Oliver Glitta <glittao@gmail.com>
Subject: slub: remove resiliency_test() function
Function resiliency_test() is hidden behind #ifdef SLUB_RESILIENCY_TEST
that is not part of Kconfig, so nobody runs it.
This function is replaced with KUnit test for SLUB added by the previous
patch "selftests: add a KUnit test for SLUB debugging functionality".
Link: https://lkml.kernel.org/r/20210511150734.3492-3-glittao@gmail.com
Signed-off-by: Oliver Glitta <glittao@gmail.com>
Reviewed-by: Marco Elver <elver@google.com>
Acked-by: Vlastimil Babka <vbabka@suse.cz>
Acked-by: David Rientjes <rientjes@google.com>
Cc: Christoph Lameter <cl@linux.com>
Cc: Pekka Enberg <penberg@kernel.org>
Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Cc: Oliver Glitta <glittao@gmail.com>
Cc: Brendan Higgins <brendanhiggins@google.com>
Cc: Daniel Latypov <dlatypov@google.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/slub.c | 64 ----------------------------------------------------
1 file changed, 64 deletions(-)
--- a/mm/slub.c~slub-remove-resiliency_test-function
+++ a/mm/slub.c
@@ -155,9 +155,6 @@ static inline bool kmem_cache_has_cpu_pa
* - Variable sizing of the per node arrays
*/
-/* Enable to test recovery from slab corruption on boot */
-#undef SLUB_RESILIENCY_TEST
-
/* Enable to log cmpxchg failures */
#undef SLUB_DEBUG_CMPXCHG
@@ -4938,66 +4935,6 @@ static int list_locations(struct kmem_ca
}
#endif /* CONFIG_SLUB_DEBUG */
-#ifdef SLUB_RESILIENCY_TEST
-static void __init resiliency_test(void)
-{
- u8 *p;
- int type = KMALLOC_NORMAL;
-
- BUILD_BUG_ON(KMALLOC_MIN_SIZE > 16 || KMALLOC_SHIFT_HIGH < 10);
-
- pr_err("SLUB resiliency testing\n");
- pr_err("-----------------------\n");
- pr_err("A. Corruption after allocation\n");
-
- p = kzalloc(16, GFP_KERNEL);
- p[16] = 0x12;
- pr_err("\n1. kmalloc-16: Clobber Redzone/next pointer 0x12->0x%p\n\n",
- p + 16);
-
- validate_slab_cache(kmalloc_caches[type][4]);
-
- /* Hmmm... The next two are dangerous */
- p = kzalloc(32, GFP_KERNEL);
- p[32 + sizeof(void *)] = 0x34;
- pr_err("\n2. kmalloc-32: Clobber next pointer/next slab 0x34 -> -0x%p\n",
- p);
- pr_err("If allocated object is overwritten then not detectable\n\n");
-
- validate_slab_cache(kmalloc_caches[type][5]);
- p = kzalloc(64, GFP_KERNEL);
- p += 64 + (get_cycles() & 0xff) * sizeof(void *);
- *p = 0x56;
- pr_err("\n3. kmalloc-64: corrupting random byte 0x56->0x%p\n",
- p);
- pr_err("If allocated object is overwritten then not detectable\n\n");
- validate_slab_cache(kmalloc_caches[type][6]);
-
- pr_err("\nB. Corruption after free\n");
- p = kzalloc(128, GFP_KERNEL);
- kfree(p);
- *p = 0x78;
- pr_err("1. kmalloc-128: Clobber first word 0x78->0x%p\n\n", p);
- validate_slab_cache(kmalloc_caches[type][7]);
-
- p = kzalloc(256, GFP_KERNEL);
- kfree(p);
- p[50] = 0x9a;
- pr_err("\n2. kmalloc-256: Clobber 50th byte 0x9a->0x%p\n\n", p);
- validate_slab_cache(kmalloc_caches[type][8]);
-
- p = kzalloc(512, GFP_KERNEL);
- kfree(p);
- p[512] = 0xab;
- pr_err("\n3. kmalloc-512: Clobber redzone 0xab->0x%p\n\n", p);
- validate_slab_cache(kmalloc_caches[type][9]);
-}
-#else
-#ifdef CONFIG_SYSFS
-static void resiliency_test(void) {};
-#endif
-#endif /* SLUB_RESILIENCY_TEST */
-
#ifdef CONFIG_SYSFS
enum slab_stat_type {
SL_ALL, /* All slabs */
@@ -5846,7 +5783,6 @@ static int __init slab_sysfs_init(void)
}
mutex_unlock(&slab_mutex);
- resiliency_test();
return 0;
}
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 026/192] mm, slub: change run-time assertion in kmalloc_index() to compile-time
2021-06-29 2:32 incoming Andrew Morton
` (24 preceding siblings ...)
2021-06-29 2:34 ` [patch 025/192] slub: remove resiliency_test() function Andrew Morton
@ 2021-06-29 2:34 ` Andrew Morton
2021-06-29 2:34 ` [patch 027/192] slub: restore slub_debug=- behavior Andrew Morton
` (165 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:34 UTC (permalink / raw)
To: 42.hyeyoo, akpm, cl, elver, iamjoonsoo.kim, linux-mm, mm-commits,
penberg, rientjes, torvalds, vbabka
From: Hyeonggon Yoo <42.hyeyoo@gmail.com>
Subject: mm, slub: change run-time assertion in kmalloc_index() to compile-time
Currently when size is not supported by kmalloc_index, compiler will
generate a run-time BUG() while compile-time error is also possible, and
better. So change BUG to BUILD_BUG_ON_MSG to make compile-time check
possible.
Also remove code that allocates more than 32MB because current
implementation supports only up to 32MB.
[42.hyeyoo@gmail.com: fix support for clang 10]
Link: https://lkml.kernel.org/r/20210518181247.GA10062@hyeyoo
[vbabka@suse.cz: fix false-positive assert in kernel/bpf/local_storage.c]
Link: https://lkml.kernel.org/r/bea97388-01df-8eac-091b-a3c89b4a4a09@suse.czLink: https://lkml.kernel.org/r/20210511173448.GA54466@hyeyoo
[elver@google.com: kfence fix]
Link: https://lkml.kernel.org/r/20210512195227.245000695c9014242e9a00e5@linux-foundation.org
Signed-off-by: Hyeonggon Yoo <42.hyeyoo@gmail.com>
Signed-off-by: Vlastimil Babka <vbabka@suse.cz>
Reviewed-by: Vlastimil Babka <vbabka@suse.cz>
Signed-off-by: Marco Elver <elver@google.com>
Cc: Christoph Lameter <cl@linux.com>
Cc: Pekka Enberg <penberg@kernel.org>
Cc: David Rientjes <rientjes@google.com>
Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Cc: Marco Elver <elver@google.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/linux/slab.h | 17 ++++++++++++++---
mm/kfence/kfence_test.c | 5 +++--
mm/slab_common.c | 7 +++----
3 files changed, 20 insertions(+), 9 deletions(-)
--- a/include/linux/slab.h~mm-slub-change-run-time-assertion-in-kmalloc_index-to-compile-time
+++ a/include/linux/slab.h
@@ -346,8 +346,14 @@ static __always_inline enum kmalloc_cach
* 1 = 65 .. 96 bytes
* 2 = 129 .. 192 bytes
* n = 2^(n-1)+1 .. 2^n
+ *
+ * Note: __kmalloc_index() is compile-time optimized, and not runtime optimized;
+ * typical usage is via kmalloc_index() and therefore evaluated at compile-time.
+ * Callers where !size_is_constant should only be test modules, where runtime
+ * overheads of __kmalloc_index() can be tolerated. Also see kmalloc_slab().
*/
-static __always_inline unsigned int kmalloc_index(size_t size)
+static __always_inline unsigned int __kmalloc_index(size_t size,
+ bool size_is_constant)
{
if (!size)
return 0;
@@ -382,12 +388,17 @@ static __always_inline unsigned int kmal
if (size <= 8 * 1024 * 1024) return 23;
if (size <= 16 * 1024 * 1024) return 24;
if (size <= 32 * 1024 * 1024) return 25;
- if (size <= 64 * 1024 * 1024) return 26;
- BUG();
+
+ if ((IS_ENABLED(CONFIG_CC_IS_GCC) || CONFIG_CLANG_VERSION >= 110000)
+ && !IS_ENABLED(CONFIG_PROFILE_ALL_BRANCHES) && size_is_constant)
+ BUILD_BUG_ON_MSG(1, "unexpected size in kmalloc_index()");
+ else
+ BUG();
/* Will never be reached. Needed because the compiler may complain */
return -1;
}
+#define kmalloc_index(s) __kmalloc_index(s, true)
#endif /* !CONFIG_SLOB */
void *__kmalloc(size_t size, gfp_t flags) __assume_kmalloc_alignment __malloc;
--- a/mm/kfence/kfence_test.c~mm-slub-change-run-time-assertion-in-kmalloc_index-to-compile-time
+++ a/mm/kfence/kfence_test.c
@@ -197,7 +197,7 @@ static void test_cache_destroy(void)
static inline size_t kmalloc_cache_alignment(size_t size)
{
- return kmalloc_caches[kmalloc_type(GFP_KERNEL)][kmalloc_index(size)]->align;
+ return kmalloc_caches[kmalloc_type(GFP_KERNEL)][__kmalloc_index(size, false)]->align;
}
/* Must always inline to match stack trace against caller. */
@@ -267,7 +267,8 @@ static void *test_alloc(struct kunit *te
if (is_kfence_address(alloc)) {
struct page *page = virt_to_head_page(alloc);
- struct kmem_cache *s = test_cache ?: kmalloc_caches[kmalloc_type(GFP_KERNEL)][kmalloc_index(size)];
+ struct kmem_cache *s = test_cache ?:
+ kmalloc_caches[kmalloc_type(GFP_KERNEL)][__kmalloc_index(size, false)];
/*
* Verify that various helpers return the right values
--- a/mm/slab_common.c~mm-slub-change-run-time-assertion-in-kmalloc_index-to-compile-time
+++ a/mm/slab_common.c
@@ -754,8 +754,8 @@ struct kmem_cache *kmalloc_slab(size_t s
/*
* kmalloc_info[] is to make slub_debug=,kmalloc-xx option work at boot time.
- * kmalloc_index() supports up to 2^26=64MB, so the final entry of the table is
- * kmalloc-67108864.
+ * kmalloc_index() supports up to 2^25=32MB, so the final entry of the table is
+ * kmalloc-32M.
*/
const struct kmalloc_info_struct kmalloc_info[] __initconst = {
INIT_KMALLOC_INFO(0, 0),
@@ -783,8 +783,7 @@ const struct kmalloc_info_struct kmalloc
INIT_KMALLOC_INFO(4194304, 4M),
INIT_KMALLOC_INFO(8388608, 8M),
INIT_KMALLOC_INFO(16777216, 16M),
- INIT_KMALLOC_INFO(33554432, 32M),
- INIT_KMALLOC_INFO(67108864, 64M)
+ INIT_KMALLOC_INFO(33554432, 32M)
};
/*
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 027/192] slub: restore slub_debug=- behavior
2021-06-29 2:32 incoming Andrew Morton
` (25 preceding siblings ...)
2021-06-29 2:34 ` [patch 026/192] mm, slub: change run-time assertion in kmalloc_index() to compile-time Andrew Morton
@ 2021-06-29 2:34 ` Andrew Morton
2021-06-29 2:34 ` [patch 028/192] slub: actually use 'message' in restore_bytes() Andrew Morton
` (164 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:34 UTC (permalink / raw)
To: akpm, cl, iamjoonsoo.kim, joe, linux-mm, mm-commits, penberg,
pmladek, rientjes, songmuchun, swboyd, torvalds, vbabka
From: Stephen Boyd <swboyd@chromium.org>
Subject: slub: restore slub_debug=- behavior
Petch series "slub: Print non-hashed pointers in slub debugging", v3.
I was doing some debugging recently and noticed that my pointers were
being hashed while slub_debug was on the kernel commandline. Let's force
on the no hash pointer option when slub_debug is on the kernel commandline
so that the prints are more meaningful.
The first two patches are something else I noticed while looking at the
code. The message argument is never used so the debugging messages are
not as clear as they could be and the slub_debug=- behavior seems to be
busted. Then there's a printf fixup from Joe and the final patch is the
one that force disables pointer hashing.
This patch (of 4):
Passing slub_debug=- on the kernel commandline is supposed to disable slub
debugging. This is especially useful with CONFIG_SLUB_DEBUG_ON where the
default is to have slub debugging enabled in the build. Due to some code
reorganization this behavior was dropped, but the code to make it work
mostly stuck around. Restore the previous behavior by disabling the
static key when we parse the commandline and see that we're trying to
disable slub debugging.
Link: https://lkml.kernel.org/r/20210601182202.3011020-1-swboyd@chromium.org
Link: https://lkml.kernel.org/r/20210601182202.3011020-2-swboyd@chromium.org
Fixes: ca0cab65ea2b ("mm, slub: introduce static key for slub_debug()")
Signed-off-by: Stephen Boyd <swboyd@chromium.org>
Acked-by: Vlastimil Babka <vbabka@suse.cz>
Reviewed-by: Muchun Song <songmuchun@bytedance.com>
Cc: Christoph Lameter <cl@linux.com>
Cc: David Rientjes <rientjes@google.com>
Cc: Joe Perches <joe@perches.com>
Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Cc: Pekka Enberg <penberg@kernel.org>
Cc: Petr Mladek <pmladek@suse.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/slub.c | 2 ++
1 file changed, 2 insertions(+)
--- a/mm/slub.c~slub-restore-slub_debug=-behavior
+++ a/mm/slub.c
@@ -1429,6 +1429,8 @@ static int __init setup_slub_debug(char
out:
if (slub_debug != 0 || slub_debug_string)
static_branch_enable(&slub_debug_enabled);
+ else
+ static_branch_disable(&slub_debug_enabled);
if ((static_branch_unlikely(&init_on_alloc) ||
static_branch_unlikely(&init_on_free)) &&
(slub_debug & SLAB_POISON))
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 028/192] slub: actually use 'message' in restore_bytes()
2021-06-29 2:32 incoming Andrew Morton
` (26 preceding siblings ...)
2021-06-29 2:34 ` [patch 027/192] slub: restore slub_debug=- behavior Andrew Morton
@ 2021-06-29 2:34 ` Andrew Morton
2021-06-29 2:34 ` [patch 029/192] slub: indicate slab_fix() uses printf formats Andrew Morton
` (163 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:34 UTC (permalink / raw)
To: akpm, cl, iamjoonsoo.kim, joe, linux-mm, mm-commits, penberg,
pmladek, rientjes, songmuchun, swboyd, torvalds, vbabka
From: Stephen Boyd <swboyd@chromium.org>
Subject: slub: actually use 'message' in restore_bytes()
The message argument isn't used here. Let's pass the string to the printk
message so that the developer can figure out what's happening, instead of
guessing that a redzone is being restored, etc.
Link: https://lkml.kernel.org/r/20210601182202.3011020-3-swboyd@chromium.org
Signed-off-by: Stephen Boyd <swboyd@chromium.org>
Reviewed-by: Vlastimil Babka <vbabka@suse.cz>
Acked-by: David Rientjes <rientjes@google.com>
Reviewed-by: Muchun Song <songmuchun@bytedance.com>
Cc: Christoph Lameter <cl@linux.com>
Cc: Joe Perches <joe@perches.com>
Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Cc: Pekka Enberg <penberg@kernel.org>
Cc: Petr Mladek <pmladek@suse.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/slub.c | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
--- a/mm/slub.c~slub-actually-use-message-in-restore_bytes
+++ a/mm/slub.c
@@ -806,7 +806,7 @@ static void init_object(struct kmem_cach
static void restore_bytes(struct kmem_cache *s, char *message, u8 data,
void *from, void *to)
{
- slab_fix(s, "Restoring 0x%p-0x%p=0x%x\n", from, to - 1, data);
+ slab_fix(s, "Restoring %s 0x%p-0x%p=0x%x\n", message, from, to - 1, data);
memset(from, data, to - from);
}
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 029/192] slub: indicate slab_fix() uses printf formats
2021-06-29 2:32 incoming Andrew Morton
` (27 preceding siblings ...)
2021-06-29 2:34 ` [patch 028/192] slub: actually use 'message' in restore_bytes() Andrew Morton
@ 2021-06-29 2:34 ` Andrew Morton
2021-06-29 2:34 ` [patch 030/192] slub: force on no_hash_pointers when slub_debug is enabled Andrew Morton
` (162 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:34 UTC (permalink / raw)
To: akpm, cl, iamjoonsoo.kim, joe, linux-mm, mm-commits, penberg,
pmladek, rientjes, songmuchun, swboyd, torvalds, vbabka
From: Joe Perches <joe@perches.com>
Subject: slub: indicate slab_fix() uses printf formats
Ideally, slab_fix() would be marked with __printf and the format here
would not use \n as that's emitted by the slab_fix(). Make these changes.
Link: https://lkml.kernel.org/r/20210601182202.3011020-4-swboyd@chromium.org
Signed-off-by: Joe Perches <joe@perches.com>
Acked-by: Vlastimil Babka <vbabka@suse.cz>
Signed-off-by: Stephen Boyd <swboyd@chromium.org>
Acked-by: David Rientjes <rientjes@google.com>
Cc: Christoph Lameter <cl@linux.com>
Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Cc: Muchun Song <songmuchun@bytedance.com>
Cc: Pekka Enberg <penberg@kernel.org>
Cc: Petr Mladek <pmladek@suse.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/slub.c | 7 ++++---
1 file changed, 4 insertions(+), 3 deletions(-)
--- a/mm/slub.c~slub-indicate-slab_fix-uses-printf-formats
+++ a/mm/slub.c
@@ -692,6 +692,7 @@ static void slab_bug(struct kmem_cache *
va_end(args);
}
+__printf(2, 3)
static void slab_fix(struct kmem_cache *s, char *fmt, ...)
{
struct va_format vaf;
@@ -806,7 +807,7 @@ static void init_object(struct kmem_cach
static void restore_bytes(struct kmem_cache *s, char *message, u8 data,
void *from, void *to)
{
- slab_fix(s, "Restoring %s 0x%p-0x%p=0x%x\n", message, from, to - 1, data);
+ slab_fix(s, "Restoring %s 0x%p-0x%p=0x%x", message, from, to - 1, data);
memset(from, data, to - from);
}
@@ -1059,13 +1060,13 @@ static int on_freelist(struct kmem_cache
slab_err(s, page, "Wrong number of objects. Found %d but should be %d",
page->objects, max_objects);
page->objects = max_objects;
- slab_fix(s, "Number of objects adjusted.");
+ slab_fix(s, "Number of objects adjusted");
}
if (page->inuse != page->objects - nr) {
slab_err(s, page, "Wrong object count. Counter is %d but counted were %d",
page->inuse, page->objects - nr);
page->inuse = page->objects - nr;
- slab_fix(s, "Object count adjusted.");
+ slab_fix(s, "Object count adjusted");
}
return search == NULL;
}
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 030/192] slub: force on no_hash_pointers when slub_debug is enabled
2021-06-29 2:32 incoming Andrew Morton
` (28 preceding siblings ...)
2021-06-29 2:34 ` [patch 029/192] slub: indicate slab_fix() uses printf formats Andrew Morton
@ 2021-06-29 2:34 ` Andrew Morton
2021-06-29 2:34 ` [patch 031/192] mm: slub: move sysfs slab alloc/free interfaces to debugfs Andrew Morton
` (161 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:34 UTC (permalink / raw)
To: akpm, cl, iamjoonsoo.kim, joe, linux-mm, mm-commits, penberg,
pmladek, rientjes, songmuchun, swboyd, torvalds, vbabka
From: Stephen Boyd <swboyd@chromium.org>
Subject: slub: force on no_hash_pointers when slub_debug is enabled
Obscuring the pointers that slub shows when debugging makes for some
confusing slub debug messages:
Padding overwritten. 0x0000000079f0674a-0x000000000d4dce17
Those addresses are hashed for kernel security reasons. If we're trying
to be secure with slub_debug on the commandline we have some big problems
given that we dump whole chunks of kernel memory to the kernel logs.
Let's force on the no_hash_pointers commandline flag when slub_debug is on
the commandline. This makes slub debug messages more meaningful and if by
chance a kernel address is in some slub debug object dump we will have a
better chance of figuring out what went wrong.
Note that we don't use %px in the slub code because we want to reduce the
number of places that %px is used in the kernel. This also nicely prints
a big fat warning at kernel boot if slub_debug is on the commandline so
that we know that this kernel shouldn't be used on production systems.
[akpm@linux-foundation.org: fix build with CONFIG_SLUB_DEBUG=n]
Link: https://lkml.kernel.org/r/20210601182202.3011020-5-swboyd@chromium.org
Signed-off-by: Stephen Boyd <swboyd@chromium.org>
Acked-by: Vlastimil Babka <vbabka@suse.cz>
Acked-by: Petr Mladek <pmladek@suse.com>
Cc: Joe Perches <joe@perches.com>
Cc: Christoph Lameter <cl@linux.com>
Cc: Pekka Enberg <penberg@kernel.org>
Cc: David Rientjes <rientjes@google.com>
Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Cc: Muchun Song <songmuchun@bytedance.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/linux/kernel.h | 2 ++
lib/vsprintf.c | 2 +-
mm/slub.c | 20 +++++++++++++++++++-
3 files changed, 22 insertions(+), 2 deletions(-)
--- a/include/linux/kernel.h~slub-force-on-no_hash_pointers-when-slub_debug-is-enabled
+++ a/include/linux/kernel.h
@@ -357,6 +357,8 @@ int sscanf(const char *, const char *, .
extern __scanf(2, 0)
int vsscanf(const char *, const char *, va_list);
+extern int no_hash_pointers_enable(char *str);
+
extern int get_option(char **str, int *pint);
extern char *get_options(const char *str, int nints, int *ints);
extern unsigned long long memparse(const char *ptr, char **retptr);
--- a/lib/vsprintf.c~slub-force-on-no_hash_pointers-when-slub_debug-is-enabled
+++ a/lib/vsprintf.c
@@ -2186,7 +2186,7 @@ char *fwnode_string(char *buf, char *end
bool no_hash_pointers __ro_after_init;
EXPORT_SYMBOL_GPL(no_hash_pointers);
-static int __init no_hash_pointers_enable(char *str)
+int __init no_hash_pointers_enable(char *str)
{
if (no_hash_pointers)
return 0;
--- a/mm/slub.c~slub-force-on-no_hash_pointers-when-slub_debug-is-enabled
+++ a/mm/slub.c
@@ -118,12 +118,26 @@
*/
#ifdef CONFIG_SLUB_DEBUG
+
#ifdef CONFIG_SLUB_DEBUG_ON
DEFINE_STATIC_KEY_TRUE(slub_debug_enabled);
#else
DEFINE_STATIC_KEY_FALSE(slub_debug_enabled);
#endif
-#endif
+
+static inline bool __slub_debug_enabled(void)
+{
+ return static_branch_unlikely(&slub_debug_enabled);
+}
+
+#else /* CONFIG_SLUB_DEBUG */
+
+static inline bool __slub_debug_enabled(void)
+{
+ return false;
+}
+
+#endif /* CONFIG_SLUB_DEBUG */
static inline bool kmem_cache_debug(struct kmem_cache *s)
{
@@ -4487,6 +4501,10 @@ void __init kmem_cache_init(void)
if (debug_guardpage_minorder())
slub_max_order = 0;
+ /* Print slub debugging pointers without hashing */
+ if (__slub_debug_enabled())
+ no_hash_pointers_enable(NULL);
+
kmem_cache_node = &boot_kmem_cache_node;
kmem_cache = &boot_kmem_cache;
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 031/192] mm: slub: move sysfs slab alloc/free interfaces to debugfs
2021-06-29 2:32 incoming Andrew Morton
` (29 preceding siblings ...)
2021-06-29 2:34 ` [patch 030/192] slub: force on no_hash_pointers when slub_debug is enabled Andrew Morton
@ 2021-06-29 2:34 ` Andrew Morton
2021-06-29 2:34 ` [patch 032/192] mm/slub: add taint after the errors are printed Andrew Morton
` (160 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:34 UTC (permalink / raw)
To: akpm, cl, faiyazm, gregkh, iamjoonsoo.kim, linux-mm, mm-commits,
penberg, rientjes, torvalds, vbabka
From: Faiyaz Mohammed <faiyazm@codeaurora.org>
Subject: mm: slub: move sysfs slab alloc/free interfaces to debugfs
alloc_calls and free_calls implementation in sysfs have two issues, one is
PAGE_SIZE limitation of sysfs and other is it does not adhere to "one
value per file" rule.
To overcome this issues, move the alloc_calls and free_calls
implementation to debugfs.
Debugfs cache will be created if SLAB_STORE_USER flag is set.
Rename the alloc_calls/free_calls to alloc_traces/free_traces, to be
inline with what it does.
[faiyazm@codeaurora.org: fix the leak of alloc/free traces debugfs interface]
Link: https://lkml.kernel.org/r/1624248060-30286-1-git-send-email-faiyazm@codeaurora.org
Link: https://lkml.kernel.org/r/1623438200-19361-1-git-send-email-faiyazm@codeaurora.org
Signed-off-by: Faiyaz Mohammed <faiyazm@codeaurora.org>
Reviewed-by: Vlastimil Babka <vbabka@suse.cz>
Reviewed-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Cc: Christoph Lameter <cl@linux.com>
Cc: Pekka Enberg <penberg@kernel.org>
Cc: David Rientjes <rientjes@google.com>
Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/slab.h | 6
mm/slab_common.c | 2
mm/slub.c | 274 +++++++++++++++++++++++++++++----------------
3 files changed, 189 insertions(+), 93 deletions(-)
--- a/mm/slab_common.c~mm-slub-move-sysfs-slab-alloc-free-interfaces-to-debugfs
+++ a/mm/slab_common.c
@@ -448,6 +448,7 @@ static void slab_caches_to_rcu_destroy_w
rcu_barrier();
list_for_each_entry_safe(s, s2, &to_destroy, list) {
+ debugfs_slab_release(s);
kfence_shutdown_cache(s);
#ifdef SLAB_SUPPORTS_SYSFS
sysfs_slab_release(s);
@@ -475,6 +476,7 @@ static int shutdown_cache(struct kmem_ca
schedule_work(&slab_caches_to_rcu_destroy_work);
} else {
kfence_shutdown_cache(s);
+ debugfs_slab_release(s);
#ifdef SLAB_SUPPORTS_SYSFS
sysfs_slab_unlink(s);
sysfs_slab_release(s);
--- a/mm/slab.h~mm-slub-move-sysfs-slab-alloc-free-interfaces-to-debugfs
+++ a/mm/slab.h
@@ -631,6 +631,12 @@ static inline bool slab_want_init_on_fre
return false;
}
+#if defined(CONFIG_DEBUG_FS) && defined(CONFIG_SLUB_DEBUG)
+void debugfs_slab_release(struct kmem_cache *);
+#else
+static inline void debugfs_slab_release(struct kmem_cache *s) { }
+#endif
+
#ifdef CONFIG_PRINTK
#define KS_ADDRS_COUNT 16
struct kmem_obj_info {
--- a/mm/slub.c~mm-slub-move-sysfs-slab-alloc-free-interfaces-to-debugfs
+++ a/mm/slub.c
@@ -38,6 +38,7 @@
#include <linux/random.h>
#include <kunit/test.h>
+#include <linux/debugfs.h>
#include <trace/events/kmem.h>
#include "internal.h"
@@ -238,6 +239,12 @@ static inline int sysfs_slab_alias(struc
{ return 0; }
#endif
+#if defined(CONFIG_DEBUG_FS) && defined(CONFIG_SLUB_DEBUG)
+static void debugfs_slab_add(struct kmem_cache *);
+#else
+static inline void debugfs_slab_add(struct kmem_cache *s) { }
+#endif
+
static inline void stat(const struct kmem_cache *s, enum stat_item si)
{
#ifdef CONFIG_SLUB_STATS
@@ -4593,6 +4600,9 @@ int __kmem_cache_create(struct kmem_cach
if (err)
__kmem_cache_release(s);
+ if (s->flags & SLAB_STORE_USER)
+ debugfs_slab_add(s);
+
return err;
}
@@ -4739,6 +4749,7 @@ long validate_slab_cache(struct kmem_cac
}
EXPORT_SYMBOL(validate_slab_cache);
+#ifdef CONFIG_DEBUG_FS
/*
* Generate lists of code addresses where slabcache objects are allocated
* and freed.
@@ -4762,6 +4773,8 @@ struct loc_track {
struct location *loc;
};
+static struct dentry *slab_debugfs_root;
+
static void free_loc_track(struct loc_track *t)
{
if (t->max)
@@ -4878,82 +4891,7 @@ static void process_slab(struct loc_trac
add_location(t, s, get_track(s, p, alloc));
put_map(map);
}
-
-static int list_locations(struct kmem_cache *s, char *buf,
- enum track_item alloc)
-{
- int len = 0;
- unsigned long i;
- struct loc_track t = { 0, 0, NULL };
- int node;
- struct kmem_cache_node *n;
-
- if (!alloc_loc_track(&t, PAGE_SIZE / sizeof(struct location),
- GFP_KERNEL)) {
- return sysfs_emit(buf, "Out of memory\n");
- }
- /* Push back cpu slabs */
- flush_all(s);
-
- for_each_kmem_cache_node(s, node, n) {
- unsigned long flags;
- struct page *page;
-
- if (!atomic_long_read(&n->nr_slabs))
- continue;
-
- spin_lock_irqsave(&n->list_lock, flags);
- list_for_each_entry(page, &n->partial, slab_list)
- process_slab(&t, s, page, alloc);
- list_for_each_entry(page, &n->full, slab_list)
- process_slab(&t, s, page, alloc);
- spin_unlock_irqrestore(&n->list_lock, flags);
- }
-
- for (i = 0; i < t.count; i++) {
- struct location *l = &t.loc[i];
-
- len += sysfs_emit_at(buf, len, "%7ld ", l->count);
-
- if (l->addr)
- len += sysfs_emit_at(buf, len, "%pS", (void *)l->addr);
- else
- len += sysfs_emit_at(buf, len, "<not-available>");
-
- if (l->sum_time != l->min_time)
- len += sysfs_emit_at(buf, len, " age=%ld/%ld/%ld",
- l->min_time,
- (long)div_u64(l->sum_time,
- l->count),
- l->max_time);
- else
- len += sysfs_emit_at(buf, len, " age=%ld", l->min_time);
-
- if (l->min_pid != l->max_pid)
- len += sysfs_emit_at(buf, len, " pid=%ld-%ld",
- l->min_pid, l->max_pid);
- else
- len += sysfs_emit_at(buf, len, " pid=%ld",
- l->min_pid);
-
- if (num_online_cpus() > 1 &&
- !cpumask_empty(to_cpumask(l->cpus)))
- len += sysfs_emit_at(buf, len, " cpus=%*pbl",
- cpumask_pr_args(to_cpumask(l->cpus)));
-
- if (nr_online_nodes > 1 && !nodes_empty(l->nodes))
- len += sysfs_emit_at(buf, len, " nodes=%*pbl",
- nodemask_pr_args(&l->nodes));
-
- len += sysfs_emit_at(buf, len, "\n");
- }
-
- free_loc_track(&t);
- if (!t.count)
- len += sysfs_emit_at(buf, len, "No data\n");
-
- return len;
-}
+#endif /* CONFIG_DEBUG_FS */
#endif /* CONFIG_SLUB_DEBUG */
#ifdef CONFIG_SYSFS
@@ -5343,21 +5281,6 @@ static ssize_t validate_store(struct kme
}
SLAB_ATTR(validate);
-static ssize_t alloc_calls_show(struct kmem_cache *s, char *buf)
-{
- if (!(s->flags & SLAB_STORE_USER))
- return -ENOSYS;
- return list_locations(s, buf, TRACK_ALLOC);
-}
-SLAB_ATTR_RO(alloc_calls);
-
-static ssize_t free_calls_show(struct kmem_cache *s, char *buf)
-{
- if (!(s->flags & SLAB_STORE_USER))
- return -ENOSYS;
- return list_locations(s, buf, TRACK_FREE);
-}
-SLAB_ATTR_RO(free_calls);
#endif /* CONFIG_SLUB_DEBUG */
#ifdef CONFIG_FAILSLAB
@@ -5521,8 +5444,6 @@ static struct attribute *slab_attrs[] =
&poison_attr.attr,
&store_user_attr.attr,
&validate_attr.attr,
- &alloc_calls_attr.attr,
- &free_calls_attr.attr,
#endif
#ifdef CONFIG_ZONE_DMA
&cache_dma_attr.attr,
@@ -5810,6 +5731,173 @@ static int __init slab_sysfs_init(void)
__initcall(slab_sysfs_init);
#endif /* CONFIG_SYSFS */
+#if defined(CONFIG_SLUB_DEBUG) && defined(CONFIG_DEBUG_FS)
+static int slab_debugfs_show(struct seq_file *seq, void *v)
+{
+
+ struct location *l;
+ unsigned int idx = *(unsigned int *)v;
+ struct loc_track *t = seq->private;
+
+ if (idx < t->count) {
+ l = &t->loc[idx];
+
+ seq_printf(seq, "%7ld ", l->count);
+
+ if (l->addr)
+ seq_printf(seq, "%pS", (void *)l->addr);
+ else
+ seq_puts(seq, "<not-available>");
+
+ if (l->sum_time != l->min_time) {
+ seq_printf(seq, " age=%ld/%llu/%ld",
+ l->min_time, div_u64(l->sum_time, l->count),
+ l->max_time);
+ } else
+ seq_printf(seq, " age=%ld", l->min_time);
+
+ if (l->min_pid != l->max_pid)
+ seq_printf(seq, " pid=%ld-%ld", l->min_pid, l->max_pid);
+ else
+ seq_printf(seq, " pid=%ld",
+ l->min_pid);
+
+ if (num_online_cpus() > 1 && !cpumask_empty(to_cpumask(l->cpus)))
+ seq_printf(seq, " cpus=%*pbl",
+ cpumask_pr_args(to_cpumask(l->cpus)));
+
+ if (nr_online_nodes > 1 && !nodes_empty(l->nodes))
+ seq_printf(seq, " nodes=%*pbl",
+ nodemask_pr_args(&l->nodes));
+
+ seq_puts(seq, "\n");
+ }
+
+ if (!idx && !t->count)
+ seq_puts(seq, "No data\n");
+
+ return 0;
+}
+
+static void slab_debugfs_stop(struct seq_file *seq, void *v)
+{
+}
+
+static void *slab_debugfs_next(struct seq_file *seq, void *v, loff_t *ppos)
+{
+ struct loc_track *t = seq->private;
+
+ v = ppos;
+ ++*ppos;
+ if (*ppos <= t->count)
+ return v;
+
+ return NULL;
+}
+
+static void *slab_debugfs_start(struct seq_file *seq, loff_t *ppos)
+{
+ return ppos;
+}
+
+static const struct seq_operations slab_debugfs_sops = {
+ .start = slab_debugfs_start,
+ .next = slab_debugfs_next,
+ .stop = slab_debugfs_stop,
+ .show = slab_debugfs_show,
+};
+
+static int slab_debug_trace_open(struct inode *inode, struct file *filep)
+{
+
+ struct kmem_cache_node *n;
+ enum track_item alloc;
+ int node;
+ struct loc_track *t = __seq_open_private(filep, &slab_debugfs_sops,
+ sizeof(struct loc_track));
+ struct kmem_cache *s = file_inode(filep)->i_private;
+
+ if (strcmp(filep->f_path.dentry->d_name.name, "alloc_traces") == 0)
+ alloc = TRACK_ALLOC;
+ else
+ alloc = TRACK_FREE;
+
+ if (!alloc_loc_track(t, PAGE_SIZE / sizeof(struct location), GFP_KERNEL))
+ return -ENOMEM;
+
+ /* Push back cpu slabs */
+ flush_all(s);
+
+ for_each_kmem_cache_node(s, node, n) {
+ unsigned long flags;
+ struct page *page;
+
+ if (!atomic_long_read(&n->nr_slabs))
+ continue;
+
+ spin_lock_irqsave(&n->list_lock, flags);
+ list_for_each_entry(page, &n->partial, slab_list)
+ process_slab(t, s, page, alloc);
+ list_for_each_entry(page, &n->full, slab_list)
+ process_slab(t, s, page, alloc);
+ spin_unlock_irqrestore(&n->list_lock, flags);
+ }
+
+ return 0;
+}
+
+static int slab_debug_trace_release(struct inode *inode, struct file *file)
+{
+ struct seq_file *seq = file->private_data;
+ struct loc_track *t = seq->private;
+
+ free_loc_track(t);
+ return seq_release_private(inode, file);
+}
+
+static const struct file_operations slab_debugfs_fops = {
+ .open = slab_debug_trace_open,
+ .read = seq_read,
+ .llseek = seq_lseek,
+ .release = slab_debug_trace_release,
+};
+
+static void debugfs_slab_add(struct kmem_cache *s)
+{
+ struct dentry *slab_cache_dir;
+
+ if (unlikely(!slab_debugfs_root))
+ return;
+
+ slab_cache_dir = debugfs_create_dir(s->name, slab_debugfs_root);
+
+ debugfs_create_file("alloc_traces", 0400,
+ slab_cache_dir, s, &slab_debugfs_fops);
+
+ debugfs_create_file("free_traces", 0400,
+ slab_cache_dir, s, &slab_debugfs_fops);
+}
+
+void debugfs_slab_release(struct kmem_cache *s)
+{
+ debugfs_remove_recursive(debugfs_lookup(s->name, slab_debugfs_root));
+}
+
+static int __init slab_debugfs_init(void)
+{
+ struct kmem_cache *s;
+
+ slab_debugfs_root = debugfs_create_dir("slab", NULL);
+
+ list_for_each_entry(s, &slab_caches, list)
+ if (s->flags & SLAB_STORE_USER)
+ debugfs_slab_add(s);
+
+ return 0;
+
+}
+__initcall(slab_debugfs_init);
+#endif
/*
* The /proc/slabinfo ABI
*/
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 032/192] mm/slub: add taint after the errors are printed
2021-06-29 2:32 incoming Andrew Morton
` (30 preceding siblings ...)
2021-06-29 2:34 ` [patch 031/192] mm: slub: move sysfs slab alloc/free interfaces to debugfs Andrew Morton
@ 2021-06-29 2:34 ` Andrew Morton
2021-06-29 2:35 ` [patch 033/192] mm/kmemleak: fix possible wrong memory scanning period Andrew Morton
` (159 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:34 UTC (permalink / raw)
To: akpm, aquini, atomlin, cl, iamjoonsoo.kim, linux-mm, mm-commits,
penberg, quic_c_gdjako, rientjes, torvalds, vbabka
From: Georgi Djakov <quic_c_gdjako@quicinc.com>
Subject: mm/slub: add taint after the errors are printed
When running the kernel with panic_on_taint, the usual slub debug error
messages are not being printed when object corruption happens. That's
because we panic in add_taint(), which is called before printing the
additional information. This is a bit unfortunate as the error messages
are actually very useful, especially before a panic. Let's fix this by
moving add_taint() after the errors are printed on the console.
Link: https://lkml.kernel.org/r/1623860738-146761-1-git-send-email-quic_c_gdjako@quicinc.com
Signed-off-by: Georgi Djakov <quic_c_gdjako@quicinc.com>
Acked-by: Rafael Aquini <aquini@redhat.com>
Acked-by: David Rientjes <rientjes@google.com>
Acked-by: Vlastimil Babka <vbabka@suse.cz>
Reviewed-by: Aaron Tomlin <atomlin@redhat.com>
Cc: Christoph Lameter <cl@linux.com>
Cc: Pekka Enberg <penberg@kernel.org>
Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/slub.c | 5 +++--
1 file changed, 3 insertions(+), 2 deletions(-)
--- a/mm/slub.c~mm-slub-add-taint-after-the-errors-are-printed
+++ a/mm/slub.c
@@ -708,8 +708,6 @@ static void slab_bug(struct kmem_cache *
pr_err("=============================================================================\n");
pr_err("BUG %s (%s): %pV\n", s->name, print_tainted(), &vaf);
pr_err("-----------------------------------------------------------------------------\n\n");
-
- add_taint(TAINT_BAD_PAGE, LOCKDEP_NOW_UNRELIABLE);
va_end(args);
}
@@ -790,6 +788,7 @@ void object_err(struct kmem_cache *s, st
slab_bug(s, "%s", reason);
print_trailer(s, page, object);
+ add_taint(TAINT_BAD_PAGE, LOCKDEP_NOW_UNRELIABLE);
}
static __printf(3, 4) void slab_err(struct kmem_cache *s, struct page *page,
@@ -807,6 +806,7 @@ static __printf(3, 4) void slab_err(stru
slab_bug(s, "%s", buf);
print_page_info(page);
dump_stack();
+ add_taint(TAINT_BAD_PAGE, LOCKDEP_NOW_UNRELIABLE);
}
static void init_object(struct kmem_cache *s, void *object, u8 val)
@@ -858,6 +858,7 @@ static int check_bytes_and_report(struct
fault, end - 1, fault - addr,
fault[0], value);
print_trailer(s, page, object);
+ add_taint(TAINT_BAD_PAGE, LOCKDEP_NOW_UNRELIABLE);
skip_bug_print:
restore_bytes(s, what, value, fault, end);
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 033/192] mm/kmemleak: fix possible wrong memory scanning period
2021-06-29 2:32 incoming Andrew Morton
` (31 preceding siblings ...)
2021-06-29 2:34 ` [patch 032/192] mm/slub: add taint after the errors are printed Andrew Morton
@ 2021-06-29 2:35 ` Andrew Morton
2021-06-29 2:35 ` [patch 034/192] dax: fix ENOMEM handling in grab_mapping_entry() Andrew Morton
` (158 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:35 UTC (permalink / raw)
To: akpm, catalin.marinas, linux-mm, mm-commits, torvalds, yanfei.xu
From: Yanfei Xu <yanfei.xu@windriver.com>
Subject: mm/kmemleak: fix possible wrong memory scanning period
This commit contains 3 modifications:
1. Convert the type of jiffies_scan_wait to "unsigned long".
2. Use READ/WRITE_ONCE() for accessing "jiffies_scan_wait".
3. Fix the possible wrong memory scanning period. If you set a large
memory scanning period like blow, then the "secs" variable will be
non-zero, however the value of "jiffies_scan_wait" will be zero.
echo "scan=0x10000000" > /sys/kernel/debug/kmemleak
It is because the type of the msecs_to_jiffies()'s parameter is "unsigned
int", and the "secs * 1000" is larger than its max value. This in turn
leads a unexpected jiffies_scan_wait, maybe zero. We corret it by
replacing kstrtoul() with kstrtouint(), and check the msecs to prevent it
larger than UINT_MAX.
Link: https://lkml.kernel.org/r/20210613174022.23044-1-yanfei.xu@windriver.com
Signed-off-by: Yanfei Xu <yanfei.xu@windriver.com>
Acked-by: Catalin Marinas <catalin.marinas@arm.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/kmemleak.c | 18 ++++++++++++------
1 file changed, 12 insertions(+), 6 deletions(-)
--- a/mm/kmemleak.c~mm-kmemleak-fix-the-possible-wrong-memory-scanning-period
+++ a/mm/kmemleak.c
@@ -219,7 +219,7 @@ static struct task_struct *scan_thread;
static unsigned long jiffies_min_age;
static unsigned long jiffies_last_scan;
/* delay between automatic memory scannings */
-static signed long jiffies_scan_wait;
+static unsigned long jiffies_scan_wait;
/* enables or disables the task stacks scanning */
static int kmemleak_stack_scan = 1;
/* protects the memory scanning, parameters and debug/kmemleak file access */
@@ -1567,7 +1567,7 @@ static int kmemleak_scan_thread(void *ar
}
while (!kthread_should_stop()) {
- signed long timeout = jiffies_scan_wait;
+ signed long timeout = READ_ONCE(jiffies_scan_wait);
mutex_lock(&scan_mutex);
kmemleak_scan();
@@ -1807,14 +1807,20 @@ static ssize_t kmemleak_write(struct fil
else if (strncmp(buf, "scan=off", 8) == 0)
stop_scan_thread();
else if (strncmp(buf, "scan=", 5) == 0) {
- unsigned long secs;
+ unsigned secs;
+ unsigned long msecs;
- ret = kstrtoul(buf + 5, 0, &secs);
+ ret = kstrtouint(buf + 5, 0, &secs);
if (ret < 0)
goto out;
+
+ msecs = secs * MSEC_PER_SEC;
+ if (msecs > UINT_MAX)
+ msecs = UINT_MAX;
+
stop_scan_thread();
- if (secs) {
- jiffies_scan_wait = msecs_to_jiffies(secs * 1000);
+ if (msecs) {
+ WRITE_ONCE(jiffies_scan_wait, msecs_to_jiffies(msecs));
start_scan_thread();
}
} else if (strncmp(buf, "scan", 4) == 0)
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 034/192] dax: fix ENOMEM handling in grab_mapping_entry()
2021-06-29 2:32 incoming Andrew Morton
` (32 preceding siblings ...)
2021-06-29 2:35 ` [patch 033/192] mm/kmemleak: fix possible wrong memory scanning period Andrew Morton
@ 2021-06-29 2:35 ` Andrew Morton
2021-06-29 2:35 ` [patch 035/192] tools/vm/page_owner_sort.c: check malloc() return Andrew Morton
` (157 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:35 UTC (permalink / raw)
To: akpm, aneesh.kumar, dan.j.williams, jack, linux-mm, mm-commits,
torvalds, willy
From: Jan Kara <jack@suse.cz>
Subject: dax: fix ENOMEM handling in grab_mapping_entry()
grab_mapping_entry() has a bug in handling of ENOMEM condition. Suppose
we have a PMD entry at index i which we are downgrading to a PTE entry.
grab_mapping_entry() will set pmd_downgrade to true, lock the entry, clear
the entry in xarray, and decrement mapping->nrpages. The it will call:
entry = dax_make_entry(pfn_to_pfn_t(0), flags);
dax_lock_entry(xas, entry);
which inserts new PTE entry into xarray. However this may fail allocating
the new node. We handle this by:
if (xas_nomem(xas, mapping_gfp_mask(mapping) & ~__GFP_HIGHMEM))
goto retry;
however pmd_downgrade stays set to true even though 'entry' returned from
get_unlocked_entry() will be NULL now. And we will go again through the
downgrade branch. This is mostly harmless except that mapping->nrpages is
decremented again and we temporarily have an invalid entry stored in
xarray. Fix the problem by setting pmd_downgrade to false each time we
lookup the entry we work with so that it matches the entry we found.
Link: https://lkml.kernel.org/r/20210622160015.18004-1-jack@suse.cz
Fixes: b15cd800682f ("dax: Convert page fault handlers to XArray")
Signed-off-by: Jan Kara <jack@suse.cz>
Reviewed-by: Dan Williams <dan.j.williams@intel.com>
Cc: Matthew Wilcox <willy@infradead.org>
Cc: "Aneesh Kumar K.V" <aneesh.kumar@linux.ibm.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
fs/dax.c | 3 ++-
1 file changed, 2 insertions(+), 1 deletion(-)
--- a/fs/dax.c~dax-fix-enomem-handling-in-grab_mapping_entry
+++ a/fs/dax.c
@@ -488,10 +488,11 @@ static void *grab_mapping_entry(struct x
struct address_space *mapping, unsigned int order)
{
unsigned long index = xas->xa_index;
- bool pmd_downgrade = false; /* splitting PMD entry into PTE entries? */
+ bool pmd_downgrade; /* splitting PMD entry into PTE entries? */
void *entry;
retry:
+ pmd_downgrade = false;
xas_lock_irq(xas);
entry = get_unlocked_entry(xas, order);
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 035/192] tools/vm/page_owner_sort.c: check malloc() return
2021-06-29 2:32 incoming Andrew Morton
` (33 preceding siblings ...)
2021-06-29 2:35 ` [patch 034/192] dax: fix ENOMEM handling in grab_mapping_entry() Andrew Morton
@ 2021-06-29 2:35 ` Andrew Morton
2021-06-29 2:35 ` [patch 036/192] mm/debug_vm_pgtable: ensure THP availability via has_transparent_hugepage() Andrew Morton
` (156 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:35 UTC (permalink / raw)
To: akpm, linux-mm, mm-commits, tangbin, torvalds, zhangshengju
From: Tang Bin <tangbin@cmss.chinamobile.com>
Subject: tools/vm/page_owner_sort.c: check malloc() return
Link: https://lkml.kernel.org/r/20210506131402.10416-1-tangbin@cmss.chinamobile.com
Signed-off-by: Zhang Shengju <zhangshengju@cmss.chinamobile.com>
Signed-off-by: Tang Bin <tangbin@cmss.chinamobile.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
tools/vm/page_owner_sort.c | 4 ++++
1 file changed, 4 insertions(+)
--- a/tools/vm/page_owner_sort.c~tools-vm-page_owner_sortc-fix-the-potential-stack-overflow-risk
+++ a/tools/vm/page_owner_sort.c
@@ -132,6 +132,10 @@ int main(int argc, char **argv)
qsort(list, list_size, sizeof(list[0]), compare_txt);
list2 = malloc(sizeof(*list) * list_size);
+ if (!list2) {
+ printf("Out of memory\n");
+ exit(1);
+ }
printf("culling\n");
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 036/192] mm/debug_vm_pgtable: ensure THP availability via has_transparent_hugepage()
2021-06-29 2:32 incoming Andrew Morton
` (34 preceding siblings ...)
2021-06-29 2:35 ` [patch 035/192] tools/vm/page_owner_sort.c: check malloc() return Andrew Morton
@ 2021-06-29 2:35 ` Andrew Morton
2021-06-29 2:35 ` [patch 037/192] mm: mmap_lock: use local locks instead of disabling preemption Andrew Morton
` (155 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:35 UTC (permalink / raw)
To: akpm, aneesh.kumar, anshuman.khandual, christophe.leroy,
linux-mm, mm-commits, torvalds
From: Anshuman Khandual <anshuman.khandual@arm.com>
Subject: mm/debug_vm_pgtable: ensure THP availability via has_transparent_hugepage()
On certain platforms, THP support could not just be validated via the
build option CONFIG_TRANSPARENT_HUGEPAGE. Instead
has_transparent_hugepage() also needs to be called upon to verify THP
runtime support. Otherwise the debug test will just run into unusable THP
helpers like in the case of a 4K hash config on powerpc platform [1].
This just moves all pfn_pmd() and pfn_pud() after THP runtime validation
with has_transparent_hugepage() which prevents the mentioned problem.
[1] https://bugzilla.kernel.org/show_bug.cgi?id=213069
Link: https://lkml.kernel.org/r/1621397588-19211-1-git-send-email-anshuman.khandual@arm.com
Fixes: 787d563b8642 ("mm/debug_vm_pgtable: fix kernel crash by checking for THP support")
Signed-off-by: Anshuman Khandual <anshuman.khandual@arm.com>
Cc: Aneesh Kumar K.V <aneesh.kumar@linux.ibm.com>
Cc: Christophe Leroy <christophe.leroy@csgroup.eu>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/debug_vm_pgtable.c | 63 ++++++++++++++++++++++++++++++++--------
1 file changed, 51 insertions(+), 12 deletions(-)
--- a/mm/debug_vm_pgtable.c~mm-debug_vm_pgtable-ensure-thp-availability-via-has_transparent_hugepage
+++ a/mm/debug_vm_pgtable.c
@@ -146,13 +146,14 @@ static void __init pte_savedwrite_tests(
static void __init pmd_basic_tests(unsigned long pfn, int idx)
{
pgprot_t prot = protection_map[idx];
- pmd_t pmd = pfn_pmd(pfn, prot);
unsigned long val = idx, *ptr = &val;
+ pmd_t pmd;
if (!has_transparent_hugepage())
return;
pr_debug("Validating PMD basic (%pGv)\n", ptr);
+ pmd = pfn_pmd(pfn, prot);
/*
* This test needs to be executed after the given page table entry
@@ -185,7 +186,7 @@ static void __init pmd_advanced_tests(st
unsigned long pfn, unsigned long vaddr,
pgprot_t prot, pgtable_t pgtable)
{
- pmd_t pmd = pfn_pmd(pfn, prot);
+ pmd_t pmd;
if (!has_transparent_hugepage())
return;
@@ -232,9 +233,14 @@ static void __init pmd_advanced_tests(st
static void __init pmd_leaf_tests(unsigned long pfn, pgprot_t prot)
{
- pmd_t pmd = pfn_pmd(pfn, prot);
+ pmd_t pmd;
+
+ if (!has_transparent_hugepage())
+ return;
pr_debug("Validating PMD leaf\n");
+ pmd = pfn_pmd(pfn, prot);
+
/*
* PMD based THP is a leaf entry.
*/
@@ -267,12 +273,16 @@ static void __init pmd_huge_tests(pmd_t
static void __init pmd_savedwrite_tests(unsigned long pfn, pgprot_t prot)
{
- pmd_t pmd = pfn_pmd(pfn, prot);
+ pmd_t pmd;
if (!IS_ENABLED(CONFIG_NUMA_BALANCING))
return;
+ if (!has_transparent_hugepage())
+ return;
+
pr_debug("Validating PMD saved write\n");
+ pmd = pfn_pmd(pfn, prot);
WARN_ON(!pmd_savedwrite(pmd_mk_savedwrite(pmd_clear_savedwrite(pmd))));
WARN_ON(pmd_savedwrite(pmd_clear_savedwrite(pmd_mk_savedwrite(pmd))));
}
@@ -281,13 +291,14 @@ static void __init pmd_savedwrite_tests(
static void __init pud_basic_tests(struct mm_struct *mm, unsigned long pfn, int idx)
{
pgprot_t prot = protection_map[idx];
- pud_t pud = pfn_pud(pfn, prot);
unsigned long val = idx, *ptr = &val;
+ pud_t pud;
if (!has_transparent_hugepage())
return;
pr_debug("Validating PUD basic (%pGv)\n", ptr);
+ pud = pfn_pud(pfn, prot);
/*
* This test needs to be executed after the given page table entry
@@ -323,7 +334,7 @@ static void __init pud_advanced_tests(st
unsigned long pfn, unsigned long vaddr,
pgprot_t prot)
{
- pud_t pud = pfn_pud(pfn, prot);
+ pud_t pud;
if (!has_transparent_hugepage())
return;
@@ -332,6 +343,7 @@ static void __init pud_advanced_tests(st
/* Align the address wrt HPAGE_PUD_SIZE */
vaddr &= HPAGE_PUD_MASK;
+ pud = pfn_pud(pfn, prot);
set_pud_at(mm, vaddr, pudp, pud);
pudp_set_wrprotect(mm, vaddr, pudp);
pud = READ_ONCE(*pudp);
@@ -370,9 +382,13 @@ static void __init pud_advanced_tests(st
static void __init pud_leaf_tests(unsigned long pfn, pgprot_t prot)
{
- pud_t pud = pfn_pud(pfn, prot);
+ pud_t pud;
+
+ if (!has_transparent_hugepage())
+ return;
pr_debug("Validating PUD leaf\n");
+ pud = pfn_pud(pfn, prot);
/*
* PUD based THP is a leaf entry.
*/
@@ -654,12 +670,16 @@ static void __init pte_protnone_tests(un
#ifdef CONFIG_TRANSPARENT_HUGEPAGE
static void __init pmd_protnone_tests(unsigned long pfn, pgprot_t prot)
{
- pmd_t pmd = pmd_mkhuge(pfn_pmd(pfn, prot));
+ pmd_t pmd;
if (!IS_ENABLED(CONFIG_NUMA_BALANCING))
return;
+ if (!has_transparent_hugepage())
+ return;
+
pr_debug("Validating PMD protnone\n");
+ pmd = pmd_mkhuge(pfn_pmd(pfn, prot));
WARN_ON(!pmd_protnone(pmd));
WARN_ON(!pmd_present(pmd));
}
@@ -679,18 +699,26 @@ static void __init pte_devmap_tests(unsi
#ifdef CONFIG_TRANSPARENT_HUGEPAGE
static void __init pmd_devmap_tests(unsigned long pfn, pgprot_t prot)
{
- pmd_t pmd = pfn_pmd(pfn, prot);
+ pmd_t pmd;
+
+ if (!has_transparent_hugepage())
+ return;
pr_debug("Validating PMD devmap\n");
+ pmd = pfn_pmd(pfn, prot);
WARN_ON(!pmd_devmap(pmd_mkdevmap(pmd)));
}
#ifdef CONFIG_HAVE_ARCH_TRANSPARENT_HUGEPAGE_PUD
static void __init pud_devmap_tests(unsigned long pfn, pgprot_t prot)
{
- pud_t pud = pfn_pud(pfn, prot);
+ pud_t pud;
+
+ if (!has_transparent_hugepage())
+ return;
pr_debug("Validating PUD devmap\n");
+ pud = pfn_pud(pfn, prot);
WARN_ON(!pud_devmap(pud_mkdevmap(pud)));
}
#else /* !CONFIG_HAVE_ARCH_TRANSPARENT_HUGEPAGE_PUD */
@@ -733,25 +761,33 @@ static void __init pte_swap_soft_dirty_t
#ifdef CONFIG_TRANSPARENT_HUGEPAGE
static void __init pmd_soft_dirty_tests(unsigned long pfn, pgprot_t prot)
{
- pmd_t pmd = pfn_pmd(pfn, prot);
+ pmd_t pmd;
if (!IS_ENABLED(CONFIG_MEM_SOFT_DIRTY))
return;
+ if (!has_transparent_hugepage())
+ return;
+
pr_debug("Validating PMD soft dirty\n");
+ pmd = pfn_pmd(pfn, prot);
WARN_ON(!pmd_soft_dirty(pmd_mksoft_dirty(pmd)));
WARN_ON(pmd_soft_dirty(pmd_clear_soft_dirty(pmd)));
}
static void __init pmd_swap_soft_dirty_tests(unsigned long pfn, pgprot_t prot)
{
- pmd_t pmd = pfn_pmd(pfn, prot);
+ pmd_t pmd;
if (!IS_ENABLED(CONFIG_MEM_SOFT_DIRTY) ||
!IS_ENABLED(CONFIG_ARCH_ENABLE_THP_MIGRATION))
return;
+ if (!has_transparent_hugepage())
+ return;
+
pr_debug("Validating PMD swap soft dirty\n");
+ pmd = pfn_pmd(pfn, prot);
WARN_ON(!pmd_swp_soft_dirty(pmd_swp_mksoft_dirty(pmd)));
WARN_ON(pmd_swp_soft_dirty(pmd_swp_clear_soft_dirty(pmd)));
}
@@ -780,6 +816,9 @@ static void __init pmd_swap_tests(unsign
swp_entry_t swp;
pmd_t pmd;
+ if (!has_transparent_hugepage())
+ return;
+
pr_debug("Validating PMD swap\n");
pmd = pfn_pmd(pfn, prot);
swp = __pmd_to_swp_entry(pmd);
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 037/192] mm: mmap_lock: use local locks instead of disabling preemption
2021-06-29 2:32 incoming Andrew Morton
` (35 preceding siblings ...)
2021-06-29 2:35 ` [patch 036/192] mm/debug_vm_pgtable: ensure THP availability via has_transparent_hugepage() Andrew Morton
@ 2021-06-29 2:35 ` Andrew Morton
2021-06-29 2:35 ` [patch 038/192] mm/page_reporting: fix code style in __page_reporting_request() Andrew Morton
` (154 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:35 UTC (permalink / raw)
To: akpm, axelrasmussen, linux-mm, mm-commits, nsaenzju, rostedt,
torvalds, vbabka
From: Nicolas Saenz Julienne <nsaenzju@redhat.com>
Subject: mm: mmap_lock: use local locks instead of disabling preemption
mmap_lock will explicitly disable/enable preemption upon manipulating its
local CPU variables. This is to be expected, but in this case, it doesn't
play well with PREEMPT_RT. The preemption disabled code section also
takes a spin-lock. Spin-locks in RT systems will try to schedule, which
is exactly what we're trying to avoid.
To mitigate this, convert the explicit preemption handling to local_locks.
Which are RT aware, and will disable migration instead of preemption when
PREEMPT_RT=y.
The faulty call trace looks like the following:
__mmap_lock_do_trace_*()
preempt_disable()
get_mm_memcg_path()
cgroup_path()
kernfs_path_from_node()
spin_lock_irqsave() /* Scheduling while atomic! */
Link: https://lkml.kernel.org/r/20210604163506.2103900-1-nsaenzju@redhat.com
Fixes: 2b5067a8143e3 ("mm: mmap_lock: add tracepoints around lock acquisition ")
Signed-off-by: Nicolas Saenz Julienne <nsaenzju@redhat.com>
Tested-by: Axel Rasmussen <axelrasmussen@google.com>
Reviewed-by: Axel Rasmussen <axelrasmussen@google.com>
Cc: Vlastimil Babka <vbabka@suse.cz>
Cc: Steven Rostedt <rostedt@goodmis.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/mmap_lock.c | 33 ++++++++++++++++++++++-----------
1 file changed, 22 insertions(+), 11 deletions(-)
--- a/mm/mmap_lock.c~mm-mmap_lock-use-local-locks-instead-of-disabling-preemption
+++ a/mm/mmap_lock.c
@@ -11,6 +11,7 @@
#include <linux/rcupdate.h>
#include <linux/smp.h>
#include <linux/trace_events.h>
+#include <linux/local_lock.h>
EXPORT_TRACEPOINT_SYMBOL(mmap_lock_start_locking);
EXPORT_TRACEPOINT_SYMBOL(mmap_lock_acquire_returned);
@@ -39,21 +40,30 @@ static int reg_refcount; /* Protected by
*/
#define CONTEXT_COUNT 4
-static DEFINE_PER_CPU(char __rcu *, memcg_path_buf);
+struct memcg_path {
+ local_lock_t lock;
+ char __rcu *buf;
+ local_t buf_idx;
+};
+static DEFINE_PER_CPU(struct memcg_path, memcg_paths) = {
+ .lock = INIT_LOCAL_LOCK(lock),
+ .buf_idx = LOCAL_INIT(0),
+};
+
static char **tmp_bufs;
-static DEFINE_PER_CPU(int, memcg_path_buf_idx);
/* Called with reg_lock held. */
static void free_memcg_path_bufs(void)
{
+ struct memcg_path *memcg_path;
int cpu;
char **old = tmp_bufs;
for_each_possible_cpu(cpu) {
- *(old++) = rcu_dereference_protected(
- per_cpu(memcg_path_buf, cpu),
+ memcg_path = per_cpu_ptr(&memcg_paths, cpu);
+ *(old++) = rcu_dereference_protected(memcg_path->buf,
lockdep_is_held(®_lock));
- rcu_assign_pointer(per_cpu(memcg_path_buf, cpu), NULL);
+ rcu_assign_pointer(memcg_path->buf, NULL);
}
/* Wait for inflight memcg_path_buf users to finish. */
@@ -88,7 +98,7 @@ int trace_mmap_lock_reg(void)
new = kmalloc(MEMCG_PATH_BUF_SIZE * CONTEXT_COUNT, GFP_KERNEL);
if (new == NULL)
goto out_fail_free;
- rcu_assign_pointer(per_cpu(memcg_path_buf, cpu), new);
+ rcu_assign_pointer(per_cpu_ptr(&memcg_paths, cpu)->buf, new);
/* Don't need to wait for inflights, they'd have gotten NULL. */
}
@@ -122,23 +132,24 @@ out:
static inline char *get_memcg_path_buf(void)
{
+ struct memcg_path *memcg_path = this_cpu_ptr(&memcg_paths);
char *buf;
int idx;
rcu_read_lock();
- buf = rcu_dereference(*this_cpu_ptr(&memcg_path_buf));
+ buf = rcu_dereference(memcg_path->buf);
if (buf == NULL) {
rcu_read_unlock();
return NULL;
}
- idx = this_cpu_add_return(memcg_path_buf_idx, MEMCG_PATH_BUF_SIZE) -
+ idx = local_add_return(MEMCG_PATH_BUF_SIZE, &memcg_path->buf_idx) -
MEMCG_PATH_BUF_SIZE;
return &buf[idx];
}
static inline void put_memcg_path_buf(void)
{
- this_cpu_sub(memcg_path_buf_idx, MEMCG_PATH_BUF_SIZE);
+ local_sub(MEMCG_PATH_BUF_SIZE, &this_cpu_ptr(&memcg_paths)->buf_idx);
rcu_read_unlock();
}
@@ -179,14 +190,14 @@ out:
#define TRACE_MMAP_LOCK_EVENT(type, mm, ...) \
do { \
const char *memcg_path; \
- preempt_disable(); \
+ local_lock(&memcg_paths.lock); \
memcg_path = get_mm_memcg_path(mm); \
trace_mmap_lock_##type(mm, \
memcg_path != NULL ? memcg_path : "", \
##__VA_ARGS__); \
if (likely(memcg_path != NULL)) \
put_memcg_path_buf(); \
- preempt_enable(); \
+ local_unlock(&memcg_paths.lock); \
} while (0)
#else /* !CONFIG_MEMCG */
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 038/192] mm/page_reporting: fix code style in __page_reporting_request()
2021-06-29 2:32 incoming Andrew Morton
` (36 preceding siblings ...)
2021-06-29 2:35 ` [patch 037/192] mm: mmap_lock: use local locks instead of disabling preemption Andrew Morton
@ 2021-06-29 2:35 ` Andrew Morton
2021-06-29 2:35 ` [patch 039/192] mm/page_reporting: export reporting order as module parameter Andrew Morton
` (153 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:35 UTC (permalink / raw)
To: akpm, alexanderduyck, anshuman.khandual, catalin.marinas, david,
gshan, linux-mm, mm-commits, mst, torvalds, will
From: Gavin Shan <gshan@redhat.com>
Subject: mm/page_reporting: fix code style in __page_reporting_request()
Patch series "mm/page_reporting: Make page reporting work on arm64 with 64KB page size", v4.
The page reporting threshold is currently equal to @pageblock_order, which
is 13 and 512MB on arm64 with 64KB base page size selected. The page
reporting won't be triggered if the freeing page can't come up with a free
area like that huge. The condition is hard to be met, especially when the
system memory becomes fragmented.
This series intends to solve the issue by having page reporting threshold
as 5 (2MB) on arm64 with 64KB base page size. The patches are organized
as:
PATCH[1/4] Fix some coding style in __page_reporting_request().
PATCH[2/4] Represents page reporting order with variable so that it can
be exported as module parameter.
PATCH[3/4] Allows the device driver (e.g. virtio_balloon) to specify
the page reporting order when the device info is registered.
PATCH[4/4] Specifies the page reporting order to 5, corresponding to
2MB in size on ARM64 when 64KB base page size is used.
This patch (of 4):
The lines of comments would be starting with one, instead two space. This
corrects the style.
Link: https://lkml.kernel.org/r/20210625014710.42954-1-gshan@redhat.com
Link: https://lkml.kernel.org/r/20210625014710.42954-2-gshan@redhat.com
Signed-off-by: Gavin Shan <gshan@redhat.com>
Reviewed-by: Alexander Duyck <alexanderduyck@fb.com>
Cc: David Hildenbrand <david@redhat.com>
Cc: "Michael S. Tsirkin" <mst@redhat.com>
Cc: Anshuman Khandual <anshuman.khandual@arm.com>
Cc: Catalin Marinas <catalin.marinas@arm.com>
Cc: Will Deacon <will@kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/page_reporting.c | 4 ++--
1 file changed, 2 insertions(+), 2 deletions(-)
--- a/mm/page_reporting.c~mm-page_reporting-fix-code-style-in-__page_reporting_request
+++ a/mm/page_reporting.c
@@ -31,8 +31,8 @@ __page_reporting_request(struct page_rep
return;
/*
- * If reporting is already active there is nothing we need to do.
- * Test against 0 as that represents PAGE_REPORTING_IDLE.
+ * If reporting is already active there is nothing we need to do.
+ * Test against 0 as that represents PAGE_REPORTING_IDLE.
*/
state = atomic_xchg(&prdev->state, PAGE_REPORTING_REQUESTED);
if (state != PAGE_REPORTING_IDLE)
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 039/192] mm/page_reporting: export reporting order as module parameter
2021-06-29 2:32 incoming Andrew Morton
` (37 preceding siblings ...)
2021-06-29 2:35 ` [patch 038/192] mm/page_reporting: fix code style in __page_reporting_request() Andrew Morton
@ 2021-06-29 2:35 ` Andrew Morton
2021-06-29 2:35 ` [patch 040/192] mm/page_reporting: allow driver to specify reporting order Andrew Morton
` (152 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:35 UTC (permalink / raw)
To: akpm, alexanderduyck, anshuman.khandual, catalin.marinas, david,
gshan, linux-mm, mm-commits, mst, torvalds, will
From: Gavin Shan <gshan@redhat.com>
Subject: mm/page_reporting: export reporting order as module parameter
The macro PAGE_REPORTING_MIN_ORDER is defined as the page reporting
threshold. It can't be adjusted at runtime.
This introduces a variable (@page_reporting_order) to replace the marcro
(PAGE_REPORTING_MIN_ORDER). MAX_ORDER is assigned to it initially,
meaning the page reporting is disabled. It will be specified by driver if
valid one is provided. Otherwise, it will fall back to @pageblock_order.
It's also exported so that the page reporting order can be adjusted at
runtime.
Link: https://lkml.kernel.org/r/20210625014710.42954-3-gshan@redhat.com
Signed-off-by: Gavin Shan <gshan@redhat.com>
Suggested-by: David Hildenbrand <david@redhat.com>
Reviewed-by: Alexander Duyck <alexanderduyck@fb.com>
Cc: Anshuman Khandual <anshuman.khandual@arm.com>
Cc: Catalin Marinas <catalin.marinas@arm.com>
Cc: "Michael S. Tsirkin" <mst@redhat.com>
Cc: Will Deacon <will@kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
Documentation/admin-guide/kernel-parameters.txt | 6 ++++++
mm/page_reporting.c | 9 +++++++--
mm/page_reporting.h | 5 ++---
3 files changed, 15 insertions(+), 5 deletions(-)
--- a/Documentation/admin-guide/kernel-parameters.txt~mm-page_reporting-export-reporting-order-as-module-parameter
+++ a/Documentation/admin-guide/kernel-parameters.txt
@@ -3566,6 +3566,12 @@
off: turn off poisoning (default)
on: turn on poisoning
+ page_reporting.page_reporting_order=
+ [KNL] Minimal page reporting order
+ Format: <integer>
+ Adjust the minimal page reporting order. The page
+ reporting is disabled when it exceeds (MAX_ORDER-1).
+
panic= [KNL] Kernel behaviour on panic: delay <timeout>
timeout > 0: seconds before rebooting
timeout = 0: wait forever
--- a/mm/page_reporting.c~mm-page_reporting-export-reporting-order-as-module-parameter
+++ a/mm/page_reporting.c
@@ -4,12 +4,17 @@
#include <linux/page_reporting.h>
#include <linux/gfp.h>
#include <linux/export.h>
+#include <linux/module.h>
#include <linux/delay.h>
#include <linux/scatterlist.h>
#include "page_reporting.h"
#include "internal.h"
+unsigned int page_reporting_order = MAX_ORDER;
+module_param(page_reporting_order, uint, 0644);
+MODULE_PARM_DESC(page_reporting_order, "Set page reporting order");
+
#define PAGE_REPORTING_DELAY (2 * HZ)
static struct page_reporting_dev_info __rcu *pr_dev_info __read_mostly;
@@ -229,7 +234,7 @@ page_reporting_process_zone(struct page_
/* Generate minimum watermark to be able to guarantee progress */
watermark = low_wmark_pages(zone) +
- (PAGE_REPORTING_CAPACITY << PAGE_REPORTING_MIN_ORDER);
+ (PAGE_REPORTING_CAPACITY << page_reporting_order);
/*
* Cancel request if insufficient free memory or if we failed
@@ -239,7 +244,7 @@ page_reporting_process_zone(struct page_
return err;
/* Process each free list starting from lowest order/mt */
- for (order = PAGE_REPORTING_MIN_ORDER; order < MAX_ORDER; order++) {
+ for (order = page_reporting_order; order < MAX_ORDER; order++) {
for (mt = 0; mt < MIGRATE_TYPES; mt++) {
/* We do not pull pages from the isolate free list */
if (is_migrate_isolate(mt))
--- a/mm/page_reporting.h~mm-page_reporting-export-reporting-order-as-module-parameter
+++ a/mm/page_reporting.h
@@ -10,10 +10,9 @@
#include <linux/pgtable.h>
#include <linux/scatterlist.h>
-#define PAGE_REPORTING_MIN_ORDER pageblock_order
-
#ifdef CONFIG_PAGE_REPORTING
DECLARE_STATIC_KEY_FALSE(page_reporting_enabled);
+extern unsigned int page_reporting_order;
void __page_reporting_notify(void);
static inline bool page_reported(struct page *page)
@@ -38,7 +37,7 @@ static inline void page_reporting_notify
return;
/* Determine if we have crossed reporting threshold */
- if (order < PAGE_REPORTING_MIN_ORDER)
+ if (order < page_reporting_order)
return;
/* This will add a few cycles, but should be called infrequently */
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 040/192] mm/page_reporting: allow driver to specify reporting order
2021-06-29 2:32 incoming Andrew Morton
` (38 preceding siblings ...)
2021-06-29 2:35 ` [patch 039/192] mm/page_reporting: export reporting order as module parameter Andrew Morton
@ 2021-06-29 2:35 ` Andrew Morton
2021-06-29 2:35 ` [patch 041/192] virtio_balloon: specify page reporting order if needed Andrew Morton
` (151 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:35 UTC (permalink / raw)
To: akpm, alexanderduyck, anshuman.khandual, catalin.marinas, david,
gshan, linux-mm, mm-commits, mst, torvalds, will
From: Gavin Shan <gshan@redhat.com>
Subject: mm/page_reporting: allow driver to specify reporting order
The page reporting order (threshold) is sticky to @pageblock_order by
default. The page reporting can never be triggered because the freeing
page can't come up with a free area like that huge. The situation becomes
worse when the system memory becomes heavily fragmented.
For example, the following configurations are used on ARM64 when 64KB base
page size is enabled. In this specific case, the page reporting won't be
triggered until the freeing page comes up with a 512MB free area. That's
hard to be met, especially when the system memory becomes heavily
fragmented.
PAGE_SIZE: 64KB
HPAGE_SIZE: 512MB
pageblock_order: 13 (512MB)
MAX_ORDER: 14
This allows the drivers to specify the page reporting order when the page
reporting device is registered. It falls back to @pageblock_order if it's
not specified by the driver. The existing users (hv_balloon and
virtio_balloon) don't specify it and @pageblock_order is still taken as
their page reporting order. So this shouldn't introduce any functional
changes.
Link: https://lkml.kernel.org/r/20210625014710.42954-4-gshan@redhat.com
Signed-off-by: Gavin Shan <gshan@redhat.com>
Reviewed-by: Alexander Duyck <alexanderduyck@fb.com>
Cc: Anshuman Khandual <anshuman.khandual@arm.com>
Cc: Catalin Marinas <catalin.marinas@arm.com>
Cc: David Hildenbrand <david@redhat.com>
Cc: "Michael S. Tsirkin" <mst@redhat.com>
Cc: Will Deacon <will@kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/linux/page_reporting.h | 3 +++
mm/page_reporting.c | 6 ++++++
2 files changed, 9 insertions(+)
--- a/include/linux/page_reporting.h~mm-page_reporting-allow-driver-to-specify-reporting-order
+++ a/include/linux/page_reporting.h
@@ -18,6 +18,9 @@ struct page_reporting_dev_info {
/* Current state of page reporting */
atomic_t state;
+
+ /* Minimal order of page reporting */
+ unsigned int order;
};
/* Tear-down and bring-up for page reporting devices */
--- a/mm/page_reporting.c~mm-page_reporting-allow-driver-to-specify-reporting-order
+++ a/mm/page_reporting.c
@@ -329,6 +329,12 @@ int page_reporting_register(struct page_
goto err_out;
}
+ /*
+ * Update the page reporting order if it's specified by driver.
+ * Otherwise, it falls back to @pageblock_order.
+ */
+ page_reporting_order = prdev->order ? : pageblock_order;
+
/* initialize state and work structures */
atomic_set(&prdev->state, PAGE_REPORTING_IDLE);
INIT_DELAYED_WORK(&prdev->work, &page_reporting_process);
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 041/192] virtio_balloon: specify page reporting order if needed
2021-06-29 2:32 incoming Andrew Morton
` (39 preceding siblings ...)
2021-06-29 2:35 ` [patch 040/192] mm/page_reporting: allow driver to specify reporting order Andrew Morton
@ 2021-06-29 2:35 ` Andrew Morton
2021-06-29 2:35 ` [patch 042/192] mm: page-writeback: kill get_writeback_state() comments Andrew Morton
` (150 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:35 UTC (permalink / raw)
To: akpm, alexanderduyck, anshuman.khandual, catalin.marinas, david,
gshan, linux-mm, mm-commits, mst, torvalds, will
From: Gavin Shan <gshan@redhat.com>
Subject: virtio_balloon: specify page reporting order if needed
The page reporting won't be triggered if the freeing page can't come up
with a free area, whose size is equal or bigger than the threshold (page
reporting order). The default page reporting order, equal to
@pageblock_order, is too huge on some architectures to trigger page
reporting. One example is ARM64 when 64KB base page size is used.
PAGE_SIZE: 64KB
pageblock_order: 13 (512MB)
MAX_ORDER: 14
This specifies the page reporting order to 5 (2MB) for this specific case
so that page reporting can be triggered.
Link: https://lkml.kernel.org/r/20210625014710.42954-5-gshan@redhat.com
Signed-off-by: Gavin Shan <gshan@redhat.com>
Reviewed-by: Alexander Duyck <alexanderduyck@fb.com>
Cc: Michael S. Tsirkin <mst@redhat.com>
Cc: David Hildenbrand <david@redhat.com>
Cc: Anshuman Khandual <anshuman.khandual@arm.com>
Cc: Catalin Marinas <catalin.marinas@arm.com>
Cc: Will Deacon <will@kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
drivers/virtio/virtio_balloon.c | 17 +++++++++++++++++
1 file changed, 17 insertions(+)
--- a/drivers/virtio/virtio_balloon.c~virtio_balloon-specify-page-reporting-order-if-needed
+++ a/drivers/virtio/virtio_balloon.c
@@ -993,6 +993,23 @@ static int virtballoon_probe(struct virt
goto out_unregister_oom;
}
+ /*
+ * The default page reporting order is @pageblock_order, which
+ * corresponds to 512MB in size on ARM64 when 64KB base page
+ * size is used. The page reporting won't be triggered if the
+ * freeing page can't come up with a free area like that huge.
+ * So we specify the page reporting order to 5, corresponding
+ * to 2MB. It helps to avoid THP splitting if 4KB base page
+ * size is used by host.
+ *
+ * Ideally, the page reporting order is selected based on the
+ * host's base page size. However, it needs more work to report
+ * that value. The hard-coded order would be fine currently.
+ */
+#if defined(CONFIG_ARM64) && defined(CONFIG_ARM64_64K_PAGES)
+ vb->pr_dev_info.order = 5;
+#endif
+
err = page_reporting_register(&vb->pr_dev_info);
if (err)
goto out_unregister_oom;
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 042/192] mm: page-writeback: kill get_writeback_state() comments
2021-06-29 2:32 incoming Andrew Morton
` (40 preceding siblings ...)
2021-06-29 2:35 ` [patch 041/192] virtio_balloon: specify page reporting order if needed Andrew Morton
@ 2021-06-29 2:35 ` Andrew Morton
2021-06-29 2:35 ` [patch 043/192] mm/page-writeback: Fix performance when BDI's share of ratio is 0 Andrew Morton
` (149 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:35 UTC (permalink / raw)
To: akpm, linux-mm, mm-commits, torvalds, wangkefeng.wang
From: Kefeng Wang <wangkefeng.wang@huawei.com>
Subject: mm: page-writeback: kill get_writeback_state() comments
The get_writeback_state() has gone since 2006, kill related comments.
Link: https://lkml.kernel.org/r/20210508125026.56600-1-wangkefeng.wang@huawei.com
Signed-off-by: Kefeng Wang <wangkefeng.wang@huawei.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/page-writeback.c | 9 +++------
1 file changed, 3 insertions(+), 6 deletions(-)
--- a/mm/page-writeback.c~mm-page-writeback-kill-get_writeback_state-comments
+++ a/mm/page-writeback.c
@@ -1869,10 +1869,9 @@ DEFINE_PER_CPU(int, dirty_throttle_leaks
* which was newly dirtied. The function will periodically check the system's
* dirty state and will initiate writeback if needed.
*
- * On really big machines, get_writeback_state is expensive, so try to avoid
- * calling it too often (ratelimiting). But once we're over the dirty memory
- * limit we decrease the ratelimiting by a lot, to prevent individual processes
- * from overshooting the limit by (ratelimit_pages) each.
+ * Once we're over the dirty memory limit we decrease the ratelimiting
+ * by a lot, to prevent individual processes from overshooting the limit
+ * by (ratelimit_pages) each.
*/
void balance_dirty_pages_ratelimited(struct address_space *mapping)
{
@@ -2045,8 +2044,6 @@ void laptop_sync_completion(void)
/*
* If ratelimit_pages is too high then we can get into dirty-data overload
* if a large number of processes all perform writes at the same time.
- * If it is too low then SMP machines will call the (expensive)
- * get_writeback_state too often.
*
* Here we set ratelimit_pages to a level which ensures that when all CPUs are
* dirtying in parallel, we cannot go more than 3% (1/32) over the dirty memory
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 043/192] mm/page-writeback: Fix performance when BDI's share of ratio is 0.
2021-06-29 2:32 incoming Andrew Morton
` (41 preceding siblings ...)
2021-06-29 2:35 ` [patch 042/192] mm: page-writeback: kill get_writeback_state() comments Andrew Morton
@ 2021-06-29 2:35 ` Andrew Morton
2021-06-29 2:35 ` [patch 044/192] mm/page-writeback: update the comment of Dirty position control Andrew Morton
` (148 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:35 UTC (permalink / raw)
To: akpm, axboe, jack, linux-mm, mm-commits, mszeredi, sedat.dilek,
tj, torvalds, wuchi.zero
From: Chi Wu <wuchi.zero@gmail.com>
Subject: mm/page-writeback: Fix performance when BDI's share of ratio is 0.
Fix performance when BDI's share of ratio is 0.
The issue is similar to commit 74d369443325 ("writeback: Fix
performance regression in wb_over_bg_thresh()").
Balance_dirty_pages and the writeback worker will also disagree on
whether writeback when a BDI uses BDI_CAP_STRICTLIMIT and BDI's share
of the thresh ratio is zero.
For example, A thread on cpu0 writes 32 pages and then
balance_dirty_pages, it will wake up background writeback and pauses
because wb_dirty > wb->wb_thresh = 0 (share of thresh ratio is zero).
A thread may runs on cpu0 again because scheduler prefers pre_cpu.
Then writeback worker may runs on other cpus(1,2..) which causes the
value of wb_stat(wb, WB_RECLAIMABLE) in wb_over_bg_thresh is 0 and does
not writeback and returns.
Thus, balance_dirty_pages keeps looping, sleeping and then waking up the
worker who will do nothing. It remains stuck in this state until the
writeback worker hit the right dirty cpu or the dirty pages expire.
The fix that we should get the wb_stat_sum radically when thresh is low.
Link: https://lkml.kernel.org/r/20210428225046.16301-1-wuchi.zero@gmail.com
Signed-off-by: Chi Wu <wuchi.zero@gmail.com>
Reviewed-by: Jan Kara <jack@suse.cz>
Cc: Tejun Heo <tj@kernel.org>
Cc: Miklos Szeredi <mszeredi@redhat.com>
Cc: Sedat Dilek <sedat.dilek@gmail.com>
Cc: Jens Axboe <axboe@fb.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/page-writeback.c | 20 ++++++++++++++++----
1 file changed, 16 insertions(+), 4 deletions(-)
--- a/mm/page-writeback.c~mm-page-writeback-fix-performance-when-bdis-share-of-ratio-is-0
+++ a/mm/page-writeback.c
@@ -1944,6 +1944,8 @@ bool wb_over_bg_thresh(struct bdi_writeb
struct dirty_throttle_control * const gdtc = &gdtc_stor;
struct dirty_throttle_control * const mdtc = mdtc_valid(&mdtc_stor) ?
&mdtc_stor : NULL;
+ unsigned long reclaimable;
+ unsigned long thresh;
/*
* Similar to balance_dirty_pages() but ignores pages being written
@@ -1956,8 +1958,13 @@ bool wb_over_bg_thresh(struct bdi_writeb
if (gdtc->dirty > gdtc->bg_thresh)
return true;
- if (wb_stat(wb, WB_RECLAIMABLE) >
- wb_calc_thresh(gdtc->wb, gdtc->bg_thresh))
+ thresh = wb_calc_thresh(gdtc->wb, gdtc->bg_thresh);
+ if (thresh < 2 * wb_stat_error())
+ reclaimable = wb_stat_sum(wb, WB_RECLAIMABLE);
+ else
+ reclaimable = wb_stat(wb, WB_RECLAIMABLE);
+
+ if (reclaimable > thresh)
return true;
if (mdtc) {
@@ -1971,8 +1978,13 @@ bool wb_over_bg_thresh(struct bdi_writeb
if (mdtc->dirty > mdtc->bg_thresh)
return true;
- if (wb_stat(wb, WB_RECLAIMABLE) >
- wb_calc_thresh(mdtc->wb, mdtc->bg_thresh))
+ thresh = wb_calc_thresh(mdtc->wb, mdtc->bg_thresh);
+ if (thresh < 2 * wb_stat_error())
+ reclaimable = wb_stat_sum(wb, WB_RECLAIMABLE);
+ else
+ reclaimable = wb_stat(wb, WB_RECLAIMABLE);
+
+ if (reclaimable > thresh)
return true;
}
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 044/192] mm/page-writeback: update the comment of Dirty position control
2021-06-29 2:32 incoming Andrew Morton
` (42 preceding siblings ...)
2021-06-29 2:35 ` [patch 043/192] mm/page-writeback: Fix performance when BDI's share of ratio is 0 Andrew Morton
@ 2021-06-29 2:35 ` Andrew Morton
2021-06-29 2:35 ` [patch 045/192] mm/page-writeback: use __this_cpu_inc() in account_page_dirtied() Andrew Morton
` (147 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:35 UTC (permalink / raw)
To: akpm, axboe, hcochran, jack, linux-mm, mm-commits, mszeredi,
sedat.dilek, tj, torvalds, wuchi.zero
From: Chi Wu <wuchi.zero@gmail.com>
Subject: mm/page-writeback: update the comment of Dirty position control
As the value of pos_ratio_polynom() clamp between 0 and 2LL <<
RATELIMIT_CALC_SHIFT, the global control line should be consistent with
it.
Link: https://lkml.kernel.org/r/20210511103606.3732-1-wuchi.zero@gmail.com
Signed-off-by: Chi Wu <wuchi.zero@gmail.com>
Reviewed-by: Jan Kara <jack@suse.cz>
Cc: Jens Axboe <axboe@fb.com>
Cc: Howard Cochran <hcochran@kernelspring.com>
Cc: Miklos Szeredi <mszeredi@redhat.com>
Cc: Sedat Dilek <sedat.dilek@gmail.com>
Cc: Tejun Heo <tj@kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/page-writeback.c | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
--- a/mm/page-writeback.c~mm-page-writeback-update-the-comment-of-dirty-position-control
+++ a/mm/page-writeback.c
@@ -845,7 +845,7 @@ static long long pos_ratio_polynom(unsig
* ^ pos_ratio
* |
* | |<===== global dirty control scope ======>|
- * 2.0 .............*
+ * 2.0 * * * * * * *
* | .*
* | . *
* | . *
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 045/192] mm/page-writeback: use __this_cpu_inc() in account_page_dirtied()
2021-06-29 2:32 incoming Andrew Morton
` (43 preceding siblings ...)
2021-06-29 2:35 ` [patch 044/192] mm/page-writeback: update the comment of Dirty position control Andrew Morton
@ 2021-06-29 2:35 ` Andrew Morton
2021-06-29 2:35 ` [patch 046/192] writeback, cgroup: do not switch inodes with I_WILL_FREE flag Andrew Morton
` (146 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:35 UTC (permalink / raw)
To: akpm, hcochran, jack, linux-mm, mm-commits, mszeredi,
sedat.dilek, tj, torvalds, wuchi.zero
From: Chi Wu <wuchi.zero@gmail.com>
Subject: mm/page-writeback: use __this_cpu_inc() in account_page_dirtied()
As account_page_dirtied() was always protected by xa_lock_irqsave(), so
using __this_cpu_inc() is better.
Link: https://lkml.kernel.org/r/20210512144742.4764-1-wuchi.zero@gmail.com
Signed-off-by: Chi Wu <wuchi.zero@gmail.com>
Reviewed-by: Jan Kara <jack@suse.cz>
Cc: Howard Cochran <hcochran@kernelspring.com>
Cc: Miklos Szeredi <mszeredi@redhat.com>
Cc: Sedat Dilek <sedat.dilek@gmail.com>
Cc: Tejun Heo <tj@kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/page-writeback.c | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
--- a/mm/page-writeback.c~mm-page-writeback-use-__this_cpu_inc-in-account_page_dirtied
+++ a/mm/page-writeback.c
@@ -2445,7 +2445,7 @@ void account_page_dirtied(struct page *p
inc_wb_stat(wb, WB_DIRTIED);
task_io_account_write(PAGE_SIZE);
current->nr_dirtied++;
- this_cpu_inc(bdp_ratelimits);
+ __this_cpu_inc(bdp_ratelimits);
mem_cgroup_track_foreign_dirty(page, wb);
}
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 046/192] writeback, cgroup: do not switch inodes with I_WILL_FREE flag
2021-06-29 2:32 incoming Andrew Morton
` (44 preceding siblings ...)
2021-06-29 2:35 ` [patch 045/192] mm/page-writeback: use __this_cpu_inc() in account_page_dirtied() Andrew Morton
@ 2021-06-29 2:35 ` Andrew Morton
2021-06-29 2:35 ` [patch 047/192] writeback, cgroup: add smp_mb() to cgroup_writeback_umount() Andrew Morton
` (145 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:35 UTC (permalink / raw)
To: akpm, axboe, dchinner, dennis, guro, jack, jack, linux-mm,
mm-commits, tj, torvalds, viro
From: Roman Gushchin <guro@fb.com>
Subject: writeback, cgroup: do not switch inodes with I_WILL_FREE flag
Patch series "cgroup, blkcg: prevent dirty inodes to pin dying memory cgroups", v9.
When an inode is getting dirty for the first time it's associated with a
wb structure (see __inode_attach_wb()). It can later be switched to
another wb (if e.g. some other cgroup is writing a lot of data to the
same inode), but otherwise stays attached to the original wb until being
reclaimed.
The problem is that the wb structure holds a reference to the original
memory and blkcg cgroups. So if an inode has been dirty once and later is
actively used in read-only mode, it has a good chance to pin down the
original memory and blkcg cgroups forever. This is often the case with
services bringing data for other services, e.g. updating some rpm
packages.
In the real life it becomes a problem due to a large size of the memcg
structure, which can easily be 1000x larger than an inode. Also a really
large number of dying cgroups can raise different scalability issues, e.g.
making the memory reclaim costly and less effective.
To solve the problem inodes should be eventually detached from the
corresponding writeback structure. It's inefficient to do it after every
writeback completion. Instead it can be done whenever the original memory
cgroup is offlined and writeback structure is getting killed. Scanning
over a (potentially long) list of inodes and detach them from the
writeback structure can take quite some time. To avoid scanning all
inodes, attached inodes are kept on a new list (b_attached). To make it
less noticeable to a user, the scanning and switching is performed from a
work context.
Big thanks to Jan Kara, Dennis Zhou, Hillf Danton and Tejun Heo for their
ideas and contribution to this patchset.
This patch (of 8):
If an inode's state has I_WILL_FREE flag set, the inode will be freed
soon, so there is no point in trying to switch the inode to a different
cgwb.
I_WILL_FREE was ignored since the introduction of the inode switching, so
it looks like it doesn't lead to any noticeable issues for a user. This
is why the patch is not intended for a stable backport.
Link: https://lkml.kernel.org/r/20210608230225.2078447-1-guro@fb.com
Link: https://lkml.kernel.org/r/20210608230225.2078447-2-guro@fb.com
Signed-off-by: Roman Gushchin <guro@fb.com>
Suggested-by: Jan Kara <jack@suse.cz>
Acked-by: Tejun Heo <tj@kernel.org>
Reviewed-by: Jan Kara <jack@suse.cz>
Acked-by: Dennis Zhou <dennis@kernel.org>
Cc: Alexander Viro <viro@zeniv.linux.org.uk>
Cc: Dave Chinner <dchinner@redhat.com>
Cc: Jan Kara <jack@suse.com>
Cc: Jens Axboe <axboe@kernel.dk>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
fs/fs-writeback.c | 8 ++++----
1 file changed, 4 insertions(+), 4 deletions(-)
--- a/fs/fs-writeback.c~writeback-cgroup-do-not-switch-inodes-with-i_will_free-flag
+++ a/fs/fs-writeback.c
@@ -389,10 +389,10 @@ static void inode_switch_wbs_work_fn(str
xa_lock_irq(&mapping->i_pages);
/*
- * Once I_FREEING is visible under i_lock, the eviction path owns
- * the inode and we shouldn't modify ->i_io_list.
+ * Once I_FREEING or I_WILL_FREE are visible under i_lock, the eviction
+ * path owns the inode and we shouldn't modify ->i_io_list.
*/
- if (unlikely(inode->i_state & I_FREEING))
+ if (unlikely(inode->i_state & (I_FREEING | I_WILL_FREE)))
goto skip_switch;
trace_inode_switch_wbs(inode, old_wb, new_wb);
@@ -517,7 +517,7 @@ static void inode_switch_wbs(struct inod
/* while holding I_WB_SWITCH, no one else can update the association */
spin_lock(&inode->i_lock);
if (!(inode->i_sb->s_flags & SB_ACTIVE) ||
- inode->i_state & (I_WB_SWITCH | I_FREEING) ||
+ inode->i_state & (I_WB_SWITCH | I_FREEING | I_WILL_FREE) ||
inode_to_wb(inode) == isw->new_wb) {
spin_unlock(&inode->i_lock);
goto out_free;
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 047/192] writeback, cgroup: add smp_mb() to cgroup_writeback_umount()
2021-06-29 2:32 incoming Andrew Morton
` (45 preceding siblings ...)
2021-06-29 2:35 ` [patch 046/192] writeback, cgroup: do not switch inodes with I_WILL_FREE flag Andrew Morton
@ 2021-06-29 2:35 ` Andrew Morton
2021-06-29 2:35 ` [patch 048/192] writeback, cgroup: increment isw_nr_in_flight before grabbing an inode Andrew Morton
` (144 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:35 UTC (permalink / raw)
To: akpm, axboe, dchinner, dennis, guro, jack, jack, linux-mm,
mm-commits, tj, torvalds, viro
From: Roman Gushchin <guro@fb.com>
Subject: writeback, cgroup: add smp_mb() to cgroup_writeback_umount()
A full memory barrier is required between clearing SB_ACTIVE flag in
generic_shutdown_super() and checking isw_nr_in_flight in
cgroup_writeback_umount(), otherwise a new switch operation might be
scheduled after atomic_read(&isw_nr_in_flight) returned 0. This would
result in a non-flushed isw_wq, and a potential crash.
The problem hasn't yet been seen in the real life and was discovered by
Jan Kara by looking into the code.
Link: https://lkml.kernel.org/r/20210608230225.2078447-3-guro@fb.com
Signed-off-by: Roman Gushchin <guro@fb.com>
Suggested-by: Jan Kara <jack@suse.cz>
Reviewed-by: Jan Kara <jack@suse.cz>
Cc: Alexander Viro <viro@zeniv.linux.org.uk>
Cc: Dave Chinner <dchinner@redhat.com>
Cc: Dennis Zhou <dennis@kernel.org>
Cc: Jan Kara <jack@suse.com>
Cc: Tejun Heo <tj@kernel.org>
Cc: Jens Axboe <axboe@kernel.dk>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
fs/fs-writeback.c | 6 ++++++
1 file changed, 6 insertions(+)
--- a/fs/fs-writeback.c~writeback-cgroup-add-smp_mb-to-cgroup_writeback_umount
+++ a/fs/fs-writeback.c
@@ -1000,6 +1000,12 @@ out_bdi_put:
*/
void cgroup_writeback_umount(void)
{
+ /*
+ * SB_ACTIVE should be reliably cleared before checking
+ * isw_nr_in_flight, see generic_shutdown_super().
+ */
+ smp_mb();
+
if (atomic_read(&isw_nr_in_flight)) {
/*
* Use rcu_barrier() to wait for all pending callbacks to
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 048/192] writeback, cgroup: increment isw_nr_in_flight before grabbing an inode
2021-06-29 2:32 incoming Andrew Morton
` (46 preceding siblings ...)
2021-06-29 2:35 ` [patch 047/192] writeback, cgroup: add smp_mb() to cgroup_writeback_umount() Andrew Morton
@ 2021-06-29 2:35 ` Andrew Morton
2021-06-29 2:35 ` [patch 049/192] writeback, cgroup: switch to rcu_work API in inode_switch_wbs() Andrew Morton
` (143 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:35 UTC (permalink / raw)
To: akpm, axboe, dchinner, dennis, guro, jack, jack, linux-mm,
mm-commits, tj, torvalds, viro
From: Roman Gushchin <guro@fb.com>
Subject: writeback, cgroup: increment isw_nr_in_flight before grabbing an inode
isw_nr_in_flight is used to determine whether the inode switch queue
should be flushed from the umount path. Currently it's increased after
grabbing an inode and even scheduling the switch work. It means the
umount path can walk past cleanup_offline_cgwb() with active inode
references, which can result in a "Busy inodes after unmount." message and
use-after-free issues (with inode->i_sb which gets freed).
Fix it by incrementing isw_nr_in_flight before doing anything with the
inode and decrementing in the case when switching wasn't scheduled.
The problem hasn't yet been seen in the real life and was discovered by
Jan Kara by looking into the code.
Link: https://lkml.kernel.org/r/20210608230225.2078447-4-guro@fb.com
Signed-off-by: Roman Gushchin <guro@fb.com>
Suggested-by: Jan Kara <jack@suse.com>
Reviewed-by: Jan Kara <jack@suse.cz>
Cc: Alexander Viro <viro@zeniv.linux.org.uk>
Cc: Dave Chinner <dchinner@redhat.com>
Cc: Dennis Zhou <dennis@kernel.org>
Cc: Tejun Heo <tj@kernel.org>
Cc: Jens Axboe <axboe@kernel.dk>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
fs/fs-writeback.c | 5 +++--
1 file changed, 3 insertions(+), 2 deletions(-)
--- a/fs/fs-writeback.c~writeback-cgroup-increment-isw_nr_in_flight-before-grabbing-an-inode
+++ a/fs/fs-writeback.c
@@ -505,6 +505,8 @@ static void inode_switch_wbs(struct inod
if (!isw)
return;
+ atomic_inc(&isw_nr_in_flight);
+
/* find and pin the new wb */
rcu_read_lock();
memcg_css = css_from_id(new_wb_id, &memory_cgrp_subsys);
@@ -535,11 +537,10 @@ static void inode_switch_wbs(struct inod
* Let's continue after I_WB_SWITCH is guaranteed to be visible.
*/
call_rcu(&isw->rcu_head, inode_switch_wbs_rcu_fn);
-
- atomic_inc(&isw_nr_in_flight);
return;
out_free:
+ atomic_dec(&isw_nr_in_flight);
if (isw->new_wb)
wb_put(isw->new_wb);
kfree(isw);
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 049/192] writeback, cgroup: switch to rcu_work API in inode_switch_wbs()
2021-06-29 2:32 incoming Andrew Morton
` (47 preceding siblings ...)
2021-06-29 2:35 ` [patch 048/192] writeback, cgroup: increment isw_nr_in_flight before grabbing an inode Andrew Morton
@ 2021-06-29 2:35 ` Andrew Morton
2021-06-29 2:35 ` [patch 050/192] writeback, cgroup: keep list of inodes attached to bdi_writeback Andrew Morton
` (142 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:35 UTC (permalink / raw)
To: akpm, axboe, dchinner, dennis, guro, jack, jack, linux-mm,
mm-commits, tj, torvalds, viro
From: Roman Gushchin <guro@fb.com>
Subject: writeback, cgroup: switch to rcu_work API in inode_switch_wbs()
Inode's wb switching requires two steps divided by an RCU grace period.
It's currently implemented as an RCU callback inode_switch_wbs_rcu_fn(),
which schedules inode_switch_wbs_work_fn() as a work.
Switching to the rcu_work API allows to do the same in a cleaner and
slightly shorter form.
Link: https://lkml.kernel.org/r/20210608230225.2078447-5-guro@fb.com
Signed-off-by: Roman Gushchin <guro@fb.com>
Reviewed-by: Jan Kara <jack@suse.cz>
Acked-by: Tejun Heo <tj@kernel.org>
Acked-by: Dennis Zhou <dennis@kernel.org>
Cc: Alexander Viro <viro@zeniv.linux.org.uk>
Cc: Dave Chinner <dchinner@redhat.com>
Cc: Jan Kara <jack@suse.com>
Cc: Jens Axboe <axboe@kernel.dk>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
fs/fs-writeback.c | 18 ++++--------------
1 file changed, 4 insertions(+), 14 deletions(-)
--- a/fs/fs-writeback.c~writeback-cgroup-switch-to-rcu_work-api-in-inode_switch_wbs
+++ a/fs/fs-writeback.c
@@ -335,8 +335,7 @@ struct inode_switch_wbs_context {
struct inode *inode;
struct bdi_writeback *new_wb;
- struct rcu_head rcu_head;
- struct work_struct work;
+ struct rcu_work work;
};
static void bdi_down_write_wb_switch_rwsem(struct backing_dev_info *bdi)
@@ -352,7 +351,7 @@ static void bdi_up_write_wb_switch_rwsem
static void inode_switch_wbs_work_fn(struct work_struct *work)
{
struct inode_switch_wbs_context *isw =
- container_of(work, struct inode_switch_wbs_context, work);
+ container_of(to_rcu_work(work), struct inode_switch_wbs_context, work);
struct inode *inode = isw->inode;
struct backing_dev_info *bdi = inode_to_bdi(inode);
struct address_space *mapping = inode->i_mapping;
@@ -469,16 +468,6 @@ skip_switch:
atomic_dec(&isw_nr_in_flight);
}
-static void inode_switch_wbs_rcu_fn(struct rcu_head *rcu_head)
-{
- struct inode_switch_wbs_context *isw = container_of(rcu_head,
- struct inode_switch_wbs_context, rcu_head);
-
- /* needs to grab bh-unsafe locks, bounce to work item */
- INIT_WORK(&isw->work, inode_switch_wbs_work_fn);
- queue_work(isw_wq, &isw->work);
-}
-
/**
* inode_switch_wbs - change the wb association of an inode
* @inode: target inode
@@ -536,7 +525,8 @@ static void inode_switch_wbs(struct inod
* lock so that stat transfer can synchronize against them.
* Let's continue after I_WB_SWITCH is guaranteed to be visible.
*/
- call_rcu(&isw->rcu_head, inode_switch_wbs_rcu_fn);
+ INIT_RCU_WORK(&isw->work, inode_switch_wbs_work_fn);
+ queue_rcu_work(isw_wq, &isw->work);
return;
out_free:
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 050/192] writeback, cgroup: keep list of inodes attached to bdi_writeback
2021-06-29 2:32 incoming Andrew Morton
` (48 preceding siblings ...)
2021-06-29 2:35 ` [patch 049/192] writeback, cgroup: switch to rcu_work API in inode_switch_wbs() Andrew Morton
@ 2021-06-29 2:35 ` Andrew Morton
2021-06-29 2:35 ` [patch 051/192] writeback, cgroup: split out the functional part of inode_switch_wbs_work_fn() Andrew Morton
` (141 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:35 UTC (permalink / raw)
To: akpm, axboe, dchinner, dennis, guro, jack, jack, linux-mm,
mm-commits, tj, torvalds, viro
From: Roman Gushchin <guro@fb.com>
Subject: writeback, cgroup: keep list of inodes attached to bdi_writeback
Currently there is no way to iterate over inodes attached to a specific
cgwb structure. It limits the ability to efficiently reclaim the
writeback structure itself and associated memory and block cgroup
structures without scanning all inodes belonging to a sb, which can be
prohibitively expensive.
While dirty/in-active-writeback an inode belongs to one of the
bdi_writeback's io lists: b_dirty, b_io, b_more_io and b_dirty_time. Once
cleaned up, it's removed from all io lists. So the inode->i_io_list can
be reused to maintain the list of inodes, attached to a bdi_writeback
structure.
This patch introduces a new wb->b_attached list, which contains all inodes
which were dirty at least once and are attached to the given cgwb. Inodes
attached to the root bdi_writeback structures are never placed on such
list. The following patch will use this list to try to release cgwbs
structures more efficiently.
Link: https://lkml.kernel.org/r/20210608230225.2078447-6-guro@fb.com
Signed-off-by: Roman Gushchin <guro@fb.com>
Suggested-by: Jan Kara <jack@suse.cz>
Reviewed-by: Jan Kara <jack@suse.cz>
Acked-by: Tejun Heo <tj@kernel.org>
Acked-by: Dennis Zhou <dennis@kernel.org>
Cc: Alexander Viro <viro@zeniv.linux.org.uk>
Cc: Dave Chinner <dchinner@redhat.com>
Cc: Jan Kara <jack@suse.com>
Cc: Jens Axboe <axboe@kernel.dk>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
fs/fs-writeback.c | 93 ++++++++++++++++++-----------
include/linux/backing-dev-defs.h | 1
mm/backing-dev.c | 2
3 files changed, 62 insertions(+), 34 deletions(-)
--- a/fs/fs-writeback.c~writeback-cgroup-keep-list-of-inodes-attached-to-bdi_writeback
+++ a/fs/fs-writeback.c
@@ -131,25 +131,6 @@ static bool inode_io_list_move_locked(st
return false;
}
-/**
- * inode_io_list_del_locked - remove an inode from its bdi_writeback IO list
- * @inode: inode to be removed
- * @wb: bdi_writeback @inode is being removed from
- *
- * Remove @inode which may be on one of @wb->b_{dirty|io|more_io} lists and
- * clear %WB_has_dirty_io if all are empty afterwards.
- */
-static void inode_io_list_del_locked(struct inode *inode,
- struct bdi_writeback *wb)
-{
- assert_spin_locked(&wb->list_lock);
- assert_spin_locked(&inode->i_lock);
-
- inode->i_state &= ~I_SYNC_QUEUED;
- list_del_init(&inode->i_io_list);
- wb_io_lists_depopulated(wb);
-}
-
static void wb_wakeup(struct bdi_writeback *wb)
{
spin_lock_bh(&wb->work_lock);
@@ -279,6 +260,28 @@ void __inode_attach_wb(struct inode *ino
EXPORT_SYMBOL_GPL(__inode_attach_wb);
/**
+ * inode_cgwb_move_to_attached - put the inode onto wb->b_attached list
+ * @inode: inode of interest with i_lock held
+ * @wb: target bdi_writeback
+ *
+ * Remove the inode from wb's io lists and if necessarily put onto b_attached
+ * list. Only inodes attached to cgwb's are kept on this list.
+ */
+static void inode_cgwb_move_to_attached(struct inode *inode,
+ struct bdi_writeback *wb)
+{
+ assert_spin_locked(&wb->list_lock);
+ assert_spin_locked(&inode->i_lock);
+
+ inode->i_state &= ~I_SYNC_QUEUED;
+ if (wb != &wb->bdi->wb)
+ list_move(&inode->i_io_list, &wb->b_attached);
+ else
+ list_del_init(&inode->i_io_list);
+ wb_io_lists_depopulated(wb);
+}
+
+/**
* locked_inode_to_wb_and_lock_list - determine a locked inode's wb and lock it
* @inode: inode of interest with i_lock held
*
@@ -418,21 +421,28 @@ static void inode_switch_wbs_work_fn(str
wb_get(new_wb);
/*
- * Transfer to @new_wb's IO list if necessary. The specific list
- * @inode was on is ignored and the inode is put on ->b_dirty which
- * is always correct including from ->b_dirty_time. The transfer
- * preserves @inode->dirtied_when ordering.
+ * Transfer to @new_wb's IO list if necessary. If the @inode is dirty,
+ * the specific list @inode was on is ignored and the @inode is put on
+ * ->b_dirty which is always correct including from ->b_dirty_time.
+ * The transfer preserves @inode->dirtied_when ordering. If the @inode
+ * was clean, it means it was on the b_attached list, so move it onto
+ * the b_attached list of @new_wb.
*/
if (!list_empty(&inode->i_io_list)) {
- struct inode *pos;
-
- inode_io_list_del_locked(inode, old_wb);
inode->i_wb = new_wb;
- list_for_each_entry(pos, &new_wb->b_dirty, i_io_list)
- if (time_after_eq(inode->dirtied_when,
- pos->dirtied_when))
- break;
- inode_io_list_move_locked(inode, new_wb, pos->i_io_list.prev);
+
+ if (inode->i_state & I_DIRTY_ALL) {
+ struct inode *pos;
+
+ list_for_each_entry(pos, &new_wb->b_dirty, i_io_list)
+ if (time_after_eq(inode->dirtied_when,
+ pos->dirtied_when))
+ break;
+ inode_io_list_move_locked(inode, new_wb,
+ pos->i_io_list.prev);
+ } else {
+ inode_cgwb_move_to_attached(inode, new_wb);
+ }
} else {
inode->i_wb = new_wb;
}
@@ -1021,6 +1031,17 @@ fs_initcall(cgroup_writeback_init);
static void bdi_down_write_wb_switch_rwsem(struct backing_dev_info *bdi) { }
static void bdi_up_write_wb_switch_rwsem(struct backing_dev_info *bdi) { }
+static void inode_cgwb_move_to_attached(struct inode *inode,
+ struct bdi_writeback *wb)
+{
+ assert_spin_locked(&wb->list_lock);
+ assert_spin_locked(&inode->i_lock);
+
+ inode->i_state &= ~I_SYNC_QUEUED;
+ list_del_init(&inode->i_io_list);
+ wb_io_lists_depopulated(wb);
+}
+
static struct bdi_writeback *
locked_inode_to_wb_and_lock_list(struct inode *inode)
__releases(&inode->i_lock)
@@ -1121,7 +1142,11 @@ void inode_io_list_del(struct inode *ino
wb = inode_to_wb_and_lock_list(inode);
spin_lock(&inode->i_lock);
- inode_io_list_del_locked(inode, wb);
+
+ inode->i_state &= ~I_SYNC_QUEUED;
+ list_del_init(&inode->i_io_list);
+ wb_io_lists_depopulated(wb);
+
spin_unlock(&inode->i_lock);
spin_unlock(&wb->list_lock);
}
@@ -1434,7 +1459,7 @@ static void requeue_inode(struct inode *
inode->i_state &= ~I_SYNC_QUEUED;
} else {
/* The inode is clean. Remove from writeback lists. */
- inode_io_list_del_locked(inode, wb);
+ inode_cgwb_move_to_attached(inode, wb);
}
}
@@ -1586,7 +1611,7 @@ static int writeback_single_inode(struct
* responsible for the writeback lists.
*/
if (!(inode->i_state & I_DIRTY_ALL))
- inode_io_list_del_locked(inode, wb);
+ inode_cgwb_move_to_attached(inode, wb);
spin_unlock(&wb->list_lock);
inode_sync_complete(inode);
out:
--- a/include/linux/backing-dev-defs.h~writeback-cgroup-keep-list-of-inodes-attached-to-bdi_writeback
+++ a/include/linux/backing-dev-defs.h
@@ -154,6 +154,7 @@ struct bdi_writeback {
struct cgroup_subsys_state *blkcg_css; /* and blkcg */
struct list_head memcg_node; /* anchored at memcg->cgwb_list */
struct list_head blkcg_node; /* anchored at blkcg->cgwb_list */
+ struct list_head b_attached; /* attached inodes, protected by list_lock */
union {
struct work_struct release_work;
--- a/mm/backing-dev.c~writeback-cgroup-keep-list-of-inodes-attached-to-bdi_writeback
+++ a/mm/backing-dev.c
@@ -396,6 +396,7 @@ static void cgwb_release_workfn(struct w
fprop_local_destroy_percpu(&wb->memcg_completions);
percpu_ref_exit(&wb->refcnt);
wb_exit(wb);
+ WARN_ON_ONCE(!list_empty(&wb->b_attached));
kfree_rcu(wb, rcu);
}
@@ -472,6 +473,7 @@ static int cgwb_create(struct backing_de
wb->memcg_css = memcg_css;
wb->blkcg_css = blkcg_css;
+ INIT_LIST_HEAD(&wb->b_attached);
INIT_WORK(&wb->release_work, cgwb_release_workfn);
set_bit(WB_registered, &wb->state);
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 051/192] writeback, cgroup: split out the functional part of inode_switch_wbs_work_fn()
2021-06-29 2:32 incoming Andrew Morton
` (49 preceding siblings ...)
2021-06-29 2:35 ` [patch 050/192] writeback, cgroup: keep list of inodes attached to bdi_writeback Andrew Morton
@ 2021-06-29 2:35 ` Andrew Morton
2021-06-29 2:35 ` [patch 052/192] writeback, cgroup: support switching multiple inodes at once Andrew Morton
` (140 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:35 UTC (permalink / raw)
To: akpm, axboe, dchinner, dennis, guro, jack, jack, linux-mm,
mm-commits, tj, torvalds, viro
From: Roman Gushchin <guro@fb.com>
Subject: writeback, cgroup: split out the functional part of inode_switch_wbs_work_fn()
Split out the functional part of the inode_switch_wbs_work_fn() function
as inode_do switch_wbs() to reuse it later for switching inodes attached
to dying cgwbs.
This commit doesn't bring any functional changes.
Link: https://lkml.kernel.org/r/20210608230225.2078447-7-guro@fb.com
Signed-off-by: Roman Gushchin <guro@fb.com>
Reviewed-by: Jan Kara <jack@suse.cz>
Acked-by: Tejun Heo <tj@kernel.org>
Acked-by: Dennis Zhou <dennis@kernel.org>
Cc: Alexander Viro <viro@zeniv.linux.org.uk>
Cc: Dave Chinner <dchinner@redhat.com>
Cc: Jan Kara <jack@suse.com>
Cc: Jens Axboe <axboe@kernel.dk>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
fs/fs-writeback.c | 19 +++++++++++--------
1 file changed, 11 insertions(+), 8 deletions(-)
--- a/fs/fs-writeback.c~writeback-cgroup-split-out-the-functional-part-of-inode_switch_wbs_work_fn
+++ a/fs/fs-writeback.c
@@ -351,15 +351,12 @@ static void bdi_up_write_wb_switch_rwsem
up_write(&bdi->wb_switch_rwsem);
}
-static void inode_switch_wbs_work_fn(struct work_struct *work)
+static void inode_do_switch_wbs(struct inode *inode,
+ struct bdi_writeback *new_wb)
{
- struct inode_switch_wbs_context *isw =
- container_of(to_rcu_work(work), struct inode_switch_wbs_context, work);
- struct inode *inode = isw->inode;
struct backing_dev_info *bdi = inode_to_bdi(inode);
struct address_space *mapping = inode->i_mapping;
struct bdi_writeback *old_wb = inode->i_wb;
- struct bdi_writeback *new_wb = isw->new_wb;
XA_STATE(xas, &mapping->i_pages, 0);
struct page *page;
bool switched = false;
@@ -470,11 +467,17 @@ skip_switch:
wb_wakeup(new_wb);
wb_put(old_wb);
}
- wb_put(new_wb);
+}
- iput(inode);
- kfree(isw);
+static void inode_switch_wbs_work_fn(struct work_struct *work)
+{
+ struct inode_switch_wbs_context *isw =
+ container_of(to_rcu_work(work), struct inode_switch_wbs_context, work);
+ inode_do_switch_wbs(isw->inode, isw->new_wb);
+ wb_put(isw->new_wb);
+ iput(isw->inode);
+ kfree(isw);
atomic_dec(&isw_nr_in_flight);
}
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 052/192] writeback, cgroup: support switching multiple inodes at once
2021-06-29 2:32 incoming Andrew Morton
` (50 preceding siblings ...)
2021-06-29 2:35 ` [patch 051/192] writeback, cgroup: split out the functional part of inode_switch_wbs_work_fn() Andrew Morton
@ 2021-06-29 2:35 ` Andrew Morton
2021-06-29 2:36 ` [patch 053/192] writeback, cgroup: release dying cgwbs by switching attached inodes Andrew Morton
` (139 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:35 UTC (permalink / raw)
To: akpm, axboe, dchinner, dennis, guro, jack, jack, linux-mm,
mm-commits, tj, torvalds, viro
From: Roman Gushchin <guro@fb.com>
Subject: writeback, cgroup: support switching multiple inodes at once
Currently only a single inode can be switched to another writeback
structure at once. That means to switch an inode a separate
inode_switch_wbs_context structure must be allocated, and a separate rcu
callback and work must be scheduled.
It's fine for the existing ad-hoc switching, which is not happening that
often, but sub-optimal for massive switching required in order to release
a writeback structure. To prepare for it, let's add a support for
switching multiple inodes at once.
Instead of containing a single inode pointer, inode_switch_wbs_context
will contain a NULL-terminated array of inode pointers.
inode_do_switch_wbs() will be called for each inode.
To optimize the locking bdi->wb_switch_rwsem, old_wb's and new_wb's
list_locks will be acquired and released only once altogether for all
inodes. wb_wakeup() will be also be called only once. Instead of calling
wb_put(old_wb) after each successful switch, wb_put_many() is introduced
and used.
Link: https://lkml.kernel.org/r/20210608230225.2078447-8-guro@fb.com
Signed-off-by: Roman Gushchin <guro@fb.com>
Acked-by: Tejun Heo <tj@kernel.org>
Reviewed-by: Jan Kara <jack@suse.cz>
Acked-by: Dennis Zhou <dennis@kernel.org>
Cc: Alexander Viro <viro@zeniv.linux.org.uk>
Cc: Dave Chinner <dchinner@redhat.com>
Cc: Jan Kara <jack@suse.com>
Cc: Jens Axboe <axboe@kernel.dk>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
fs/fs-writeback.c | 106 +++++++++++++++++------------
include/linux/backing-dev-defs.h | 18 ++++
2 files changed, 80 insertions(+), 44 deletions(-)
--- a/fs/fs-writeback.c~writeback-cgroup-support-switching-multiple-inodes-at-once
+++ a/fs/fs-writeback.c
@@ -335,10 +335,18 @@ static struct bdi_writeback *inode_to_wb
}
struct inode_switch_wbs_context {
- struct inode *inode;
- struct bdi_writeback *new_wb;
-
struct rcu_work work;
+
+ /*
+ * Multiple inodes can be switched at once. The switching procedure
+ * consists of two parts, separated by a RCU grace period. To make
+ * sure that the second part is executed for each inode gone through
+ * the first part, all inode pointers are placed into a NULL-terminated
+ * array embedded into struct inode_switch_wbs_context. Otherwise
+ * an inode could be left in a non-consistent state.
+ */
+ struct bdi_writeback *new_wb;
+ struct inode *inodes[];
};
static void bdi_down_write_wb_switch_rwsem(struct backing_dev_info *bdi)
@@ -351,39 +359,15 @@ static void bdi_up_write_wb_switch_rwsem
up_write(&bdi->wb_switch_rwsem);
}
-static void inode_do_switch_wbs(struct inode *inode,
+static bool inode_do_switch_wbs(struct inode *inode,
+ struct bdi_writeback *old_wb,
struct bdi_writeback *new_wb)
{
- struct backing_dev_info *bdi = inode_to_bdi(inode);
struct address_space *mapping = inode->i_mapping;
- struct bdi_writeback *old_wb = inode->i_wb;
XA_STATE(xas, &mapping->i_pages, 0);
struct page *page;
bool switched = false;
- /*
- * If @inode switches cgwb membership while sync_inodes_sb() is
- * being issued, sync_inodes_sb() might miss it. Synchronize.
- */
- down_read(&bdi->wb_switch_rwsem);
-
- /*
- * By the time control reaches here, RCU grace period has passed
- * since I_WB_SWITCH assertion and all wb stat update transactions
- * between unlocked_inode_to_wb_begin/end() are guaranteed to be
- * synchronizing against the i_pages lock.
- *
- * Grabbing old_wb->list_lock, inode->i_lock and the i_pages lock
- * gives us exclusion against all wb related operations on @inode
- * including IO list manipulations and stat updates.
- */
- if (old_wb < new_wb) {
- spin_lock(&old_wb->list_lock);
- spin_lock_nested(&new_wb->list_lock, SINGLE_DEPTH_NESTING);
- } else {
- spin_lock(&new_wb->list_lock);
- spin_lock_nested(&old_wb->list_lock, SINGLE_DEPTH_NESTING);
- }
spin_lock(&inode->i_lock);
xa_lock_irq(&mapping->i_pages);
@@ -458,25 +442,63 @@ skip_switch:
xa_unlock_irq(&mapping->i_pages);
spin_unlock(&inode->i_lock);
- spin_unlock(&new_wb->list_lock);
- spin_unlock(&old_wb->list_lock);
-
- up_read(&bdi->wb_switch_rwsem);
- if (switched) {
- wb_wakeup(new_wb);
- wb_put(old_wb);
- }
+ return switched;
}
static void inode_switch_wbs_work_fn(struct work_struct *work)
{
struct inode_switch_wbs_context *isw =
container_of(to_rcu_work(work), struct inode_switch_wbs_context, work);
+ struct backing_dev_info *bdi = inode_to_bdi(isw->inodes[0]);
+ struct bdi_writeback *old_wb = isw->inodes[0]->i_wb;
+ struct bdi_writeback *new_wb = isw->new_wb;
+ unsigned long nr_switched = 0;
+ struct inode **inodep;
+
+ /*
+ * If @inode switches cgwb membership while sync_inodes_sb() is
+ * being issued, sync_inodes_sb() might miss it. Synchronize.
+ */
+ down_read(&bdi->wb_switch_rwsem);
+
+ /*
+ * By the time control reaches here, RCU grace period has passed
+ * since I_WB_SWITCH assertion and all wb stat update transactions
+ * between unlocked_inode_to_wb_begin/end() are guaranteed to be
+ * synchronizing against the i_pages lock.
+ *
+ * Grabbing old_wb->list_lock, inode->i_lock and the i_pages lock
+ * gives us exclusion against all wb related operations on @inode
+ * including IO list manipulations and stat updates.
+ */
+ if (old_wb < new_wb) {
+ spin_lock(&old_wb->list_lock);
+ spin_lock_nested(&new_wb->list_lock, SINGLE_DEPTH_NESTING);
+ } else {
+ spin_lock(&new_wb->list_lock);
+ spin_lock_nested(&old_wb->list_lock, SINGLE_DEPTH_NESTING);
+ }
+
+ for (inodep = isw->inodes; *inodep; inodep++) {
+ WARN_ON_ONCE((*inodep)->i_wb != old_wb);
+ if (inode_do_switch_wbs(*inodep, old_wb, new_wb))
+ nr_switched++;
+ }
+
+ spin_unlock(&new_wb->list_lock);
+ spin_unlock(&old_wb->list_lock);
+
+ up_read(&bdi->wb_switch_rwsem);
+
+ if (nr_switched) {
+ wb_wakeup(new_wb);
+ wb_put_many(old_wb, nr_switched);
+ }
- inode_do_switch_wbs(isw->inode, isw->new_wb);
- wb_put(isw->new_wb);
- iput(isw->inode);
+ for (inodep = isw->inodes; *inodep; inodep++)
+ iput(*inodep);
+ wb_put(new_wb);
kfree(isw);
atomic_dec(&isw_nr_in_flight);
}
@@ -503,7 +525,7 @@ static void inode_switch_wbs(struct inod
if (atomic_read(&isw_nr_in_flight) > WB_FRN_MAX_IN_FLIGHT)
return;
- isw = kzalloc(sizeof(*isw), GFP_ATOMIC);
+ isw = kzalloc(sizeof(*isw) + 2 * sizeof(struct inode *), GFP_ATOMIC);
if (!isw)
return;
@@ -530,7 +552,7 @@ static void inode_switch_wbs(struct inod
__iget(inode);
spin_unlock(&inode->i_lock);
- isw->inode = inode;
+ isw->inodes[0] = inode;
/*
* In addition to synchronizing among switchers, I_WB_SWITCH tells
--- a/include/linux/backing-dev-defs.h~writeback-cgroup-support-switching-multiple-inodes-at-once
+++ a/include/linux/backing-dev-defs.h
@@ -240,8 +240,9 @@ static inline void wb_get(struct bdi_wri
/**
* wb_put - decrement a wb's refcount
* @wb: bdi_writeback to put
+ * @nr: number of references to put
*/
-static inline void wb_put(struct bdi_writeback *wb)
+static inline void wb_put_many(struct bdi_writeback *wb, unsigned long nr)
{
if (WARN_ON_ONCE(!wb->bdi)) {
/*
@@ -252,7 +253,16 @@ static inline void wb_put(struct bdi_wri
}
if (wb != &wb->bdi->wb)
- percpu_ref_put(&wb->refcnt);
+ percpu_ref_put_many(&wb->refcnt, nr);
+}
+
+/**
+ * wb_put - decrement a wb's refcount
+ * @wb: bdi_writeback to put
+ */
+static inline void wb_put(struct bdi_writeback *wb)
+{
+ wb_put_many(wb, 1);
}
/**
@@ -281,6 +291,10 @@ static inline void wb_put(struct bdi_wri
{
}
+static inline void wb_put_many(struct bdi_writeback *wb, unsigned long nr)
+{
+}
+
static inline bool wb_dying(struct bdi_writeback *wb)
{
return false;
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 053/192] writeback, cgroup: release dying cgwbs by switching attached inodes
2021-06-29 2:32 incoming Andrew Morton
` (51 preceding siblings ...)
2021-06-29 2:35 ` [patch 052/192] writeback, cgroup: support switching multiple inodes at once Andrew Morton
@ 2021-06-29 2:36 ` Andrew Morton
2021-06-29 2:36 ` [patch 054/192] fs: unexport __set_page_dirty Andrew Morton
` (138 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:36 UTC (permalink / raw)
To: akpm, axboe, dchinner, dennis, guro, jack, jack, linux-mm,
mm-commits, tj, torvalds, viro, willy
From: Roman Gushchin <guro@fb.com>
Subject: writeback, cgroup: release dying cgwbs by switching attached inodes
Asynchronously try to release dying cgwbs by switching attached inodes to
the nearest living ancestor wb. It helps to get rid of per-cgroup
writeback structures themselves and of pinned memory and block cgroups,
which are significantly larger structures (mostly due to large per-cpu
statistics data). This prevents memory waste and helps to avoid different
scalability problems caused by large piles of dying cgroups.
Reuse the existing mechanism of inode switching used for foreign inode
detection. To speed things up batch up to 115 inode switching in a single
operation (the maximum number is selected so that the resulting struct
inode_switch_wbs_context can fit into 1024 bytes). Because every
switching consists of two steps divided by an RCU grace period, it would
be too slow without batching. Please note that the whole batch counts as
a single operation (when increasing/decreasing isw_nr_in_flight). This
allows to keep umounting working (flush the switching queue), however
prevents cleanups from consuming the whole switching quota and effectively
blocking the frn switching.
A cgwb cleanup operation can fail due to different reasons (e.g. not
enough memory, the cgwb has an in-flight/pending io, an attached inode in
a wrong state, etc). In this case the next scheduled cleanup will make a
new attempt. An attempt is made each time a new cgwb is offlined (in
other words a memcg and/or a blkcg is deleted by a user). In the future
an additional attempt scheduled by a timer can be implemented.
[guro@fb.com: replace open-coded "115" with arithmetic]
Link: https://lkml.kernel.org/r/YMEcSBcq/VXMiPPO@carbon.dhcp.thefacebook.com
[guro@fb.com: add smp_mb() to inode_prepare_wbs_switch()]
Link: https://lkml.kernel.org/r/YMFa+guFw7OFjf3X@carbon.dhcp.thefacebook.com
[willy@infradead.org: fix documentation]
Link: https://lkml.kernel.org/r/20210615200242.1716568-2-willy@infradead.org
Link: https://lkml.kernel.org/r/20210608230225.2078447-9-guro@fb.com
Signed-off-by: Roman Gushchin <guro@fb.com>
Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org>
Acked-by: Tejun Heo <tj@kernel.org>
Acked-by: Dennis Zhou <dennis@kernel.org>
Reviewed-by: Jan Kara <jack@suse.cz>
Cc: Alexander Viro <viro@zeniv.linux.org.uk>
Cc: Dave Chinner <dchinner@redhat.com>
Cc: Jan Kara <jack@suse.com>
Cc: Jens Axboe <axboe@kernel.dk>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
fs/fs-writeback.c | 111 ++++++++++++++++++++++++++---
include/linux/backing-dev-defs.h | 1
include/linux/writeback.h | 1
mm/backing-dev.c | 64 ++++++++++++++++
4 files changed, 165 insertions(+), 12 deletions(-)
--- a/fs/fs-writeback.c~writeback-cgroup-release-dying-cgwbs-by-switching-attached-inodes
+++ a/fs/fs-writeback.c
@@ -225,6 +225,13 @@ void wb_wait_for_completion(struct wb_co
/* one round can affect upto 5 slots */
#define WB_FRN_MAX_IN_FLIGHT 1024 /* don't queue too many concurrently */
+/*
+ * Maximum inodes per isw. A specific value has been chosen to make
+ * struct inode_switch_wbs_context fit into 1024 bytes kmalloc.
+ */
+#define WB_MAX_INODES_PER_ISW ((1024UL - sizeof(struct inode_switch_wbs_context)) \
+ / sizeof(struct inode *))
+
static atomic_t isw_nr_in_flight = ATOMIC_INIT(0);
static struct workqueue_struct *isw_wq;
@@ -503,6 +510,32 @@ static void inode_switch_wbs_work_fn(str
atomic_dec(&isw_nr_in_flight);
}
+static bool inode_prepare_wbs_switch(struct inode *inode,
+ struct bdi_writeback *new_wb)
+{
+ /*
+ * Paired with smp_mb() in cgroup_writeback_umount().
+ * isw_nr_in_flight must be increased before checking SB_ACTIVE and
+ * grabbing an inode, otherwise isw_nr_in_flight can be observed as 0
+ * in cgroup_writeback_umount() and the isw_wq will be not flushed.
+ */
+ smp_mb();
+
+ /* while holding I_WB_SWITCH, no one else can update the association */
+ spin_lock(&inode->i_lock);
+ if (!(inode->i_sb->s_flags & SB_ACTIVE) ||
+ inode->i_state & (I_WB_SWITCH | I_FREEING | I_WILL_FREE) ||
+ inode_to_wb(inode) == new_wb) {
+ spin_unlock(&inode->i_lock);
+ return false;
+ }
+ inode->i_state |= I_WB_SWITCH;
+ __iget(inode);
+ spin_unlock(&inode->i_lock);
+
+ return true;
+}
+
/**
* inode_switch_wbs - change the wb association of an inode
* @inode: target inode
@@ -540,17 +573,8 @@ static void inode_switch_wbs(struct inod
if (!isw->new_wb)
goto out_free;
- /* while holding I_WB_SWITCH, no one else can update the association */
- spin_lock(&inode->i_lock);
- if (!(inode->i_sb->s_flags & SB_ACTIVE) ||
- inode->i_state & (I_WB_SWITCH | I_FREEING | I_WILL_FREE) ||
- inode_to_wb(inode) == isw->new_wb) {
- spin_unlock(&inode->i_lock);
+ if (!inode_prepare_wbs_switch(inode, isw->new_wb))
goto out_free;
- }
- inode->i_state |= I_WB_SWITCH;
- __iget(inode);
- spin_unlock(&inode->i_lock);
isw->inodes[0] = inode;
@@ -572,6 +596,73 @@ out_free:
}
/**
+ * cleanup_offline_cgwb - detach associated inodes
+ * @wb: target wb
+ *
+ * Switch all inodes attached to @wb to a nearest living ancestor's wb in order
+ * to eventually release the dying @wb. Returns %true if not all inodes were
+ * switched and the function has to be restarted.
+ */
+bool cleanup_offline_cgwb(struct bdi_writeback *wb)
+{
+ struct cgroup_subsys_state *memcg_css;
+ struct inode_switch_wbs_context *isw;
+ struct inode *inode;
+ int nr;
+ bool restart = false;
+
+ isw = kzalloc(sizeof(*isw) + WB_MAX_INODES_PER_ISW *
+ sizeof(struct inode *), GFP_KERNEL);
+ if (!isw)
+ return restart;
+
+ atomic_inc(&isw_nr_in_flight);
+
+ for (memcg_css = wb->memcg_css->parent; memcg_css;
+ memcg_css = memcg_css->parent) {
+ isw->new_wb = wb_get_create(wb->bdi, memcg_css, GFP_KERNEL);
+ if (isw->new_wb)
+ break;
+ }
+ if (unlikely(!isw->new_wb))
+ isw->new_wb = &wb->bdi->wb; /* wb_get() is noop for bdi's wb */
+
+ nr = 0;
+ spin_lock(&wb->list_lock);
+ list_for_each_entry(inode, &wb->b_attached, i_io_list) {
+ if (!inode_prepare_wbs_switch(inode, isw->new_wb))
+ continue;
+
+ isw->inodes[nr++] = inode;
+
+ if (nr >= WB_MAX_INODES_PER_ISW - 1) {
+ restart = true;
+ break;
+ }
+ }
+ spin_unlock(&wb->list_lock);
+
+ /* no attached inodes? bail out */
+ if (nr == 0) {
+ atomic_dec(&isw_nr_in_flight);
+ wb_put(isw->new_wb);
+ kfree(isw);
+ return restart;
+ }
+
+ /*
+ * In addition to synchronizing among switchers, I_WB_SWITCH tells
+ * the RCU protected stat update paths to grab the i_page
+ * lock so that stat transfer can synchronize against them.
+ * Let's continue after I_WB_SWITCH is guaranteed to be visible.
+ */
+ INIT_RCU_WORK(&isw->work, inode_switch_wbs_work_fn);
+ queue_rcu_work(isw_wq, &isw->work);
+
+ return restart;
+}
+
+/**
* wbc_attach_and_unlock_inode - associate wbc with target inode and unlock it
* @wbc: writeback_control of interest
* @inode: target inode
--- a/include/linux/backing-dev-defs.h~writeback-cgroup-release-dying-cgwbs-by-switching-attached-inodes
+++ a/include/linux/backing-dev-defs.h
@@ -155,6 +155,7 @@ struct bdi_writeback {
struct list_head memcg_node; /* anchored at memcg->cgwb_list */
struct list_head blkcg_node; /* anchored at blkcg->cgwb_list */
struct list_head b_attached; /* attached inodes, protected by list_lock */
+ struct list_head offline_node; /* anchored at offline_cgwbs */
union {
struct work_struct release_work;
--- a/include/linux/writeback.h~writeback-cgroup-release-dying-cgwbs-by-switching-attached-inodes
+++ a/include/linux/writeback.h
@@ -221,6 +221,7 @@ void wbc_account_cgroup_owner(struct wri
int cgroup_writeback_by_id(u64 bdi_id, int memcg_id, unsigned long nr_pages,
enum wb_reason reason, struct wb_completion *done);
void cgroup_writeback_umount(void);
+bool cleanup_offline_cgwb(struct bdi_writeback *wb);
/**
* inode_attach_wb - associate an inode with its wb
--- a/mm/backing-dev.c~writeback-cgroup-release-dying-cgwbs-by-switching-attached-inodes
+++ a/mm/backing-dev.c
@@ -371,12 +371,16 @@ static void wb_exit(struct bdi_writeback
#include <linux/memcontrol.h>
/*
- * cgwb_lock protects bdi->cgwb_tree, blkcg->cgwb_list, and memcg->cgwb_list.
- * bdi->cgwb_tree is also RCU protected.
+ * cgwb_lock protects bdi->cgwb_tree, blkcg->cgwb_list, offline_cgwbs and
+ * memcg->cgwb_list. bdi->cgwb_tree is also RCU protected.
*/
static DEFINE_SPINLOCK(cgwb_lock);
static struct workqueue_struct *cgwb_release_wq;
+static LIST_HEAD(offline_cgwbs);
+static void cleanup_offline_cgwbs_workfn(struct work_struct *work);
+static DECLARE_WORK(cleanup_offline_cgwbs_work, cleanup_offline_cgwbs_workfn);
+
static void cgwb_release_workfn(struct work_struct *work)
{
struct bdi_writeback *wb = container_of(work, struct bdi_writeback,
@@ -395,6 +399,11 @@ static void cgwb_release_workfn(struct w
fprop_local_destroy_percpu(&wb->memcg_completions);
percpu_ref_exit(&wb->refcnt);
+
+ spin_lock_irq(&cgwb_lock);
+ list_del(&wb->offline_node);
+ spin_unlock_irq(&cgwb_lock);
+
wb_exit(wb);
WARN_ON_ONCE(!list_empty(&wb->b_attached));
kfree_rcu(wb, rcu);
@@ -414,6 +423,7 @@ static void cgwb_kill(struct bdi_writeba
WARN_ON(!radix_tree_delete(&wb->bdi->cgwb_tree, wb->memcg_css->id));
list_del(&wb->memcg_node);
list_del(&wb->blkcg_node);
+ list_add(&wb->offline_node, &offline_cgwbs);
percpu_ref_kill(&wb->refcnt);
}
@@ -635,6 +645,54 @@ static void cgwb_bdi_unregister(struct b
mutex_unlock(&bdi->cgwb_release_mutex);
}
+/*
+ * cleanup_offline_cgwbs_workfn - try to release dying cgwbs
+ *
+ * Try to release dying cgwbs by switching attached inodes to the nearest
+ * living ancestor's writeback. Processed wbs are placed at the end
+ * of the list to guarantee the forward progress.
+ */
+static void cleanup_offline_cgwbs_workfn(struct work_struct *work)
+{
+ struct bdi_writeback *wb;
+ LIST_HEAD(processed);
+
+ spin_lock_irq(&cgwb_lock);
+
+ while (!list_empty(&offline_cgwbs)) {
+ wb = list_first_entry(&offline_cgwbs, struct bdi_writeback,
+ offline_node);
+ list_move(&wb->offline_node, &processed);
+
+ /*
+ * If wb is dirty, cleaning up the writeback by switching
+ * attached inodes will result in an effective removal of any
+ * bandwidth restrictions, which isn't the goal. Instead,
+ * it can be postponed until the next time, when all io
+ * will be likely completed. If in the meantime some inodes
+ * will get re-dirtied, they should be eventually switched to
+ * a new cgwb.
+ */
+ if (wb_has_dirty_io(wb))
+ continue;
+
+ if (!wb_tryget(wb))
+ continue;
+
+ spin_unlock_irq(&cgwb_lock);
+ while (cleanup_offline_cgwb(wb))
+ cond_resched();
+ spin_lock_irq(&cgwb_lock);
+
+ wb_put(wb);
+ }
+
+ if (!list_empty(&processed))
+ list_splice_tail(&processed, &offline_cgwbs);
+
+ spin_unlock_irq(&cgwb_lock);
+}
+
/**
* wb_memcg_offline - kill all wb's associated with a memcg being offlined
* @memcg: memcg being offlined
@@ -651,6 +709,8 @@ void wb_memcg_offline(struct mem_cgroup
cgwb_kill(wb);
memcg_cgwb_list->next = NULL; /* prevent new wb's */
spin_unlock_irq(&cgwb_lock);
+
+ queue_work(system_unbound_wq, &cleanup_offline_cgwbs_work);
}
/**
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 054/192] fs: unexport __set_page_dirty
2021-06-29 2:32 incoming Andrew Morton
` (52 preceding siblings ...)
2021-06-29 2:36 ` [patch 053/192] writeback, cgroup: release dying cgwbs by switching attached inodes Andrew Morton
@ 2021-06-29 2:36 ` Andrew Morton
2021-06-29 2:36 ` [patch 055/192] fs: move ramfs_aops to libfs Andrew Morton
` (137 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:36 UTC (permalink / raw)
To: akpm, gregkh, hch, jack, linux-mm, mm-commits, torvalds, viro, willy
From: Christoph Hellwig <hch@lst.de>
Subject: fs: unexport __set_page_dirty
Patch series "remove the implicit .set_page_dirty default".
This series cleans up a few lose ends around ->set_page_dirty, most
importantly removes the default to the buffer head based on if no method
is wired up.
This patch (of 3):
__set_page_dirty is only used by built-in code.
Link: https://lkml.kernel.org/r/20210614061512.3966143-1-hch@lst.de
Link: https://lkml.kernel.org/r/20210614061512.3966143-2-hch@lst.de
Signed-off-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Reviewed-by: Jan Kara <jack@suse.cz>
Reviewed-by: Matthew Wilcox (Oracle) <willy@infradead.org>
Cc: Al Viro <viro@zeniv.linux.org.uk>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
fs/buffer.c | 1 -
1 file changed, 1 deletion(-)
--- a/fs/buffer.c~fs-unexport-__set_page_dirty
+++ a/fs/buffer.c
@@ -611,7 +611,6 @@ void __set_page_dirty(struct page *page,
}
xa_unlock_irqrestore(&mapping->i_pages, flags);
}
-EXPORT_SYMBOL_GPL(__set_page_dirty);
/*
* Add a page to the dirty page list.
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 055/192] fs: move ramfs_aops to libfs
2021-06-29 2:32 incoming Andrew Morton
` (53 preceding siblings ...)
2021-06-29 2:36 ` [patch 054/192] fs: unexport __set_page_dirty Andrew Morton
@ 2021-06-29 2:36 ` Andrew Morton
2021-06-29 2:36 ` [patch 056/192] mm: require ->set_page_dirty to be explicitly wired up Andrew Morton
` (136 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:36 UTC (permalink / raw)
To: akpm, gregkh, hch, jack, linux-mm, mm-commits, torvalds, viro, willy
From: Christoph Hellwig <hch@lst.de>
Subject: fs: move ramfs_aops to libfs
Move the ramfs aops to libfs and reuse them for kernfs and configfs.
Thosw two did not wire up ->set_page_dirty before and now get
__set_page_dirty_no_writeback, which is the right one for no-writeback
address_space usage.
Drop the now unused exports of the libfs helpers only used for ramfs-style
pagecache usage.
Link: https://lkml.kernel.org/r/20210614061512.3966143-3-hch@lst.de
Signed-off-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Reviewed-by: Jan Kara <jack@suse.cz>
Cc: Al Viro <viro@zeniv.linux.org.uk>
Cc: Matthew Wilcox (Oracle) <willy@infradead.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
fs/configfs/inode.c | 8 +-------
fs/kernfs/inode.c | 8 +-------
fs/libfs.c | 17 +++++++++++++----
fs/ramfs/inode.c | 9 +--------
include/linux/fs.h | 5 +----
5 files changed, 17 insertions(+), 30 deletions(-)
--- a/fs/configfs/inode.c~fs-move-ramfs_aops-to-libfs
+++ a/fs/configfs/inode.c
@@ -28,12 +28,6 @@
static struct lock_class_key default_group_class[MAX_LOCK_DEPTH];
#endif
-static const struct address_space_operations configfs_aops = {
- .readpage = simple_readpage,
- .write_begin = simple_write_begin,
- .write_end = simple_write_end,
-};
-
static const struct inode_operations configfs_inode_operations ={
.setattr = configfs_setattr,
};
@@ -114,7 +108,7 @@ struct inode *configfs_new_inode(umode_t
struct inode * inode = new_inode(s);
if (inode) {
inode->i_ino = get_next_ino();
- inode->i_mapping->a_ops = &configfs_aops;
+ inode->i_mapping->a_ops = &ram_aops;
inode->i_op = &configfs_inode_operations;
if (sd->s_iattr) {
--- a/fs/kernfs/inode.c~fs-move-ramfs_aops-to-libfs
+++ a/fs/kernfs/inode.c
@@ -17,12 +17,6 @@
#include "kernfs-internal.h"
-static const struct address_space_operations kernfs_aops = {
- .readpage = simple_readpage,
- .write_begin = simple_write_begin,
- .write_end = simple_write_end,
-};
-
static const struct inode_operations kernfs_iops = {
.permission = kernfs_iop_permission,
.setattr = kernfs_iop_setattr,
@@ -203,7 +197,7 @@ static void kernfs_init_inode(struct ker
{
kernfs_get(kn);
inode->i_private = kn;
- inode->i_mapping->a_ops = &kernfs_aops;
+ inode->i_mapping->a_ops = &ram_aops;
inode->i_op = &kernfs_iops;
inode->i_generation = kernfs_gen(kn);
--- a/fs/libfs.c~fs-move-ramfs_aops-to-libfs
+++ a/fs/libfs.c
@@ -512,7 +512,7 @@ int simple_setattr(struct user_namespace
}
EXPORT_SYMBOL(simple_setattr);
-int simple_readpage(struct file *file, struct page *page)
+static int simple_readpage(struct file *file, struct page *page)
{
clear_highpage(page);
flush_dcache_page(page);
@@ -520,7 +520,6 @@ int simple_readpage(struct file *file, s
unlock_page(page);
return 0;
}
-EXPORT_SYMBOL(simple_readpage);
int simple_write_begin(struct file *file, struct address_space *mapping,
loff_t pos, unsigned len, unsigned flags,
@@ -568,7 +567,7 @@ EXPORT_SYMBOL(simple_write_begin);
*
* Use *ONLY* with simple_readpage()
*/
-int simple_write_end(struct file *file, struct address_space *mapping,
+static int simple_write_end(struct file *file, struct address_space *mapping,
loff_t pos, unsigned len, unsigned copied,
struct page *page, void *fsdata)
{
@@ -597,7 +596,17 @@ int simple_write_end(struct file *file,
return copied;
}
-EXPORT_SYMBOL(simple_write_end);
+
+/*
+ * Provides ramfs-style behavior: data in the pagecache, but no writeback.
+ */
+const struct address_space_operations ram_aops = {
+ .readpage = simple_readpage,
+ .write_begin = simple_write_begin,
+ .write_end = simple_write_end,
+ .set_page_dirty = __set_page_dirty_no_writeback,
+};
+EXPORT_SYMBOL(ram_aops);
/*
* the inodes created here are not hashed. If you use iunique to generate
--- a/fs/ramfs/inode.c~fs-move-ramfs_aops-to-libfs
+++ a/fs/ramfs/inode.c
@@ -53,13 +53,6 @@ struct ramfs_fs_info {
static const struct super_operations ramfs_ops;
static const struct inode_operations ramfs_dir_inode_operations;
-static const struct address_space_operations ramfs_aops = {
- .readpage = simple_readpage,
- .write_begin = simple_write_begin,
- .write_end = simple_write_end,
- .set_page_dirty = __set_page_dirty_no_writeback,
-};
-
struct inode *ramfs_get_inode(struct super_block *sb,
const struct inode *dir, umode_t mode, dev_t dev)
{
@@ -68,7 +61,7 @@ struct inode *ramfs_get_inode(struct sup
if (inode) {
inode->i_ino = get_next_ino();
inode_init_owner(&init_user_ns, inode, dir, mode);
- inode->i_mapping->a_ops = &ramfs_aops;
+ inode->i_mapping->a_ops = &ram_aops;
mapping_set_gfp_mask(inode->i_mapping, GFP_HIGHUSER);
mapping_set_unevictable(inode->i_mapping);
inode->i_atime = inode->i_mtime = inode->i_ctime = current_time(inode);
--- a/include/linux/fs.h~fs-move-ramfs_aops-to-libfs
+++ a/include/linux/fs.h
@@ -3422,13 +3422,10 @@ extern void noop_invalidatepage(struct p
unsigned int length);
extern ssize_t noop_direct_IO(struct kiocb *iocb, struct iov_iter *iter);
extern int simple_empty(struct dentry *);
-extern int simple_readpage(struct file *file, struct page *page);
extern int simple_write_begin(struct file *file, struct address_space *mapping,
loff_t pos, unsigned len, unsigned flags,
struct page **pagep, void **fsdata);
-extern int simple_write_end(struct file *file, struct address_space *mapping,
- loff_t pos, unsigned len, unsigned copied,
- struct page *page, void *fsdata);
+extern const struct address_space_operations ram_aops;
extern int always_delete_dentry(const struct dentry *);
extern struct inode *alloc_anon_inode(struct super_block *);
extern int simple_nosetlease(struct file *, long, struct file_lock **, void **);
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 056/192] mm: require ->set_page_dirty to be explicitly wired up
2021-06-29 2:32 incoming Andrew Morton
` (54 preceding siblings ...)
2021-06-29 2:36 ` [patch 055/192] fs: move ramfs_aops to libfs Andrew Morton
@ 2021-06-29 2:36 ` Andrew Morton
2021-06-29 2:36 ` [patch 057/192] mm/writeback: move __set_page_dirty() to core mm Andrew Morton
` (135 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:36 UTC (permalink / raw)
To: akpm, code, gregkh, hch, jack, linux-mm, mm-commits, torvalds,
viro, willy
From: Christoph Hellwig <hch@lst.de>
Subject: mm: require ->set_page_dirty to be explicitly wired up
Remove the CONFIG_BLOCK default to __set_page_dirty_buffers and just wire
that method up for the missing instances.
[hch@lst.de: ecryptfs: add a ->set_page_dirty cludge]
Link: https://lkml.kernel.org/r/20210624125250.536369-1-hch@lst.de
Link: https://lkml.kernel.org/r/20210614061512.3966143-4-hch@lst.de
Signed-off-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Reviewed-by: Jan Kara <jack@suse.cz>
Cc: Al Viro <viro@zeniv.linux.org.uk>
Cc: Matthew Wilcox (Oracle) <willy@infradead.org>
Cc: Tyler Hicks <code@tyhicks.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
fs/adfs/inode.c | 1 +
fs/affs/file.c | 2 ++
fs/bfs/file.c | 1 +
fs/block_dev.c | 1 +
fs/ecryptfs/mmap.c | 13 +++++++++++++
fs/exfat/inode.c | 1 +
fs/ext2/inode.c | 2 ++
fs/fat/inode.c | 1 +
fs/gfs2/meta_io.c | 2 ++
fs/hfs/inode.c | 2 ++
fs/hfsplus/inode.c | 2 ++
fs/hpfs/file.c | 1 +
fs/jfs/inode.c | 1 +
fs/minix/inode.c | 1 +
fs/nilfs2/mdt.c | 1 +
fs/ocfs2/aops.c | 1 +
fs/omfs/file.c | 1 +
fs/sysv/itree.c | 1 +
fs/udf/file.c | 1 +
fs/udf/inode.c | 1 +
fs/ufs/inode.c | 1 +
mm/page-writeback.c | 18 ++++--------------
22 files changed, 42 insertions(+), 14 deletions(-)
--- a/fs/adfs/inode.c~mm-require-set_page_dirty-to-be-explicitly-wire-up
+++ a/fs/adfs/inode.c
@@ -73,6 +73,7 @@ static sector_t _adfs_bmap(struct addres
}
static const struct address_space_operations adfs_aops = {
+ .set_page_dirty = __set_page_dirty_buffers,
.readpage = adfs_readpage,
.writepage = adfs_writepage,
.write_begin = adfs_write_begin,
--- a/fs/affs/file.c~mm-require-set_page_dirty-to-be-explicitly-wire-up
+++ a/fs/affs/file.c
@@ -453,6 +453,7 @@ static sector_t _affs_bmap(struct addres
}
const struct address_space_operations affs_aops = {
+ .set_page_dirty = __set_page_dirty_buffers,
.readpage = affs_readpage,
.writepage = affs_writepage,
.write_begin = affs_write_begin,
@@ -833,6 +834,7 @@ err_bh:
}
const struct address_space_operations affs_aops_ofs = {
+ .set_page_dirty = __set_page_dirty_buffers,
.readpage = affs_readpage_ofs,
//.writepage = affs_writepage_ofs,
.write_begin = affs_write_begin_ofs,
--- a/fs/bfs/file.c~mm-require-set_page_dirty-to-be-explicitly-wire-up
+++ a/fs/bfs/file.c
@@ -188,6 +188,7 @@ static sector_t bfs_bmap(struct address_
}
const struct address_space_operations bfs_aops = {
+ .set_page_dirty = __set_page_dirty_buffers,
.readpage = bfs_readpage,
.writepage = bfs_writepage,
.write_begin = bfs_write_begin,
--- a/fs/block_dev.c~mm-require-set_page_dirty-to-be-explicitly-wire-up
+++ a/fs/block_dev.c
@@ -1754,6 +1754,7 @@ static int blkdev_writepages(struct addr
}
static const struct address_space_operations def_blk_aops = {
+ .set_page_dirty = __set_page_dirty_buffers,
.readpage = blkdev_readpage,
.readahead = blkdev_readahead,
.writepage = blkdev_writepage,
--- a/fs/ecryptfs/mmap.c~mm-require-set_page_dirty-to-be-explicitly-wire-up
+++ a/fs/ecryptfs/mmap.c
@@ -533,7 +533,20 @@ static sector_t ecryptfs_bmap(struct add
return block;
}
+#include <linux/buffer_head.h>
+
const struct address_space_operations ecryptfs_aops = {
+ /*
+ * XXX: This is pretty broken for multiple reasons: ecryptfs does not
+ * actually use buffer_heads, and ecryptfs will crash without
+ * CONFIG_BLOCK. But it matches the behavior before the default for
+ * address_space_operations without the ->set_page_dirty method was
+ * cleaned up, so this is the best we can do without maintainer
+ * feedback.
+ */
+#ifdef CONFIG_BLOCK
+ .set_page_dirty = __set_page_dirty_buffers,
+#endif
.writepage = ecryptfs_writepage,
.readpage = ecryptfs_readpage,
.write_begin = ecryptfs_write_begin,
--- a/fs/exfat/inode.c~mm-require-set_page_dirty-to-be-explicitly-wire-up
+++ a/fs/exfat/inode.c
@@ -491,6 +491,7 @@ int exfat_block_truncate_page(struct ino
}
static const struct address_space_operations exfat_aops = {
+ .set_page_dirty = __set_page_dirty_buffers,
.readpage = exfat_readpage,
.readahead = exfat_readahead,
.writepage = exfat_writepage,
--- a/fs/ext2/inode.c~mm-require-set_page_dirty-to-be-explicitly-wire-up
+++ a/fs/ext2/inode.c
@@ -961,6 +961,7 @@ ext2_dax_writepages(struct address_space
}
const struct address_space_operations ext2_aops = {
+ .set_page_dirty = __set_page_dirty_buffers,
.readpage = ext2_readpage,
.readahead = ext2_readahead,
.writepage = ext2_writepage,
@@ -975,6 +976,7 @@ const struct address_space_operations ex
};
const struct address_space_operations ext2_nobh_aops = {
+ .set_page_dirty = __set_page_dirty_buffers,
.readpage = ext2_readpage,
.readahead = ext2_readahead,
.writepage = ext2_nobh_writepage,
--- a/fs/fat/inode.c~mm-require-set_page_dirty-to-be-explicitly-wire-up
+++ a/fs/fat/inode.c
@@ -342,6 +342,7 @@ int fat_block_truncate_page(struct inode
}
static const struct address_space_operations fat_aops = {
+ .set_page_dirty = __set_page_dirty_buffers,
.readpage = fat_readpage,
.readahead = fat_readahead,
.writepage = fat_writepage,
--- a/fs/gfs2/meta_io.c~mm-require-set_page_dirty-to-be-explicitly-wire-up
+++ a/fs/gfs2/meta_io.c
@@ -89,11 +89,13 @@ static int gfs2_aspace_writepage(struct
}
const struct address_space_operations gfs2_meta_aops = {
+ .set_page_dirty = __set_page_dirty_buffers,
.writepage = gfs2_aspace_writepage,
.releasepage = gfs2_releasepage,
};
const struct address_space_operations gfs2_rgrp_aops = {
+ .set_page_dirty = __set_page_dirty_buffers,
.writepage = gfs2_aspace_writepage,
.releasepage = gfs2_releasepage,
};
--- a/fs/hfs/inode.c~mm-require-set_page_dirty-to-be-explicitly-wire-up
+++ a/fs/hfs/inode.c
@@ -159,6 +159,7 @@ static int hfs_writepages(struct address
}
const struct address_space_operations hfs_btree_aops = {
+ .set_page_dirty = __set_page_dirty_buffers,
.readpage = hfs_readpage,
.writepage = hfs_writepage,
.write_begin = hfs_write_begin,
@@ -168,6 +169,7 @@ const struct address_space_operations hf
};
const struct address_space_operations hfs_aops = {
+ .set_page_dirty = __set_page_dirty_buffers,
.readpage = hfs_readpage,
.writepage = hfs_writepage,
.write_begin = hfs_write_begin,
--- a/fs/hfsplus/inode.c~mm-require-set_page_dirty-to-be-explicitly-wire-up
+++ a/fs/hfsplus/inode.c
@@ -156,6 +156,7 @@ static int hfsplus_writepages(struct add
}
const struct address_space_operations hfsplus_btree_aops = {
+ .set_page_dirty = __set_page_dirty_buffers,
.readpage = hfsplus_readpage,
.writepage = hfsplus_writepage,
.write_begin = hfsplus_write_begin,
@@ -165,6 +166,7 @@ const struct address_space_operations hf
};
const struct address_space_operations hfsplus_aops = {
+ .set_page_dirty = __set_page_dirty_buffers,
.readpage = hfsplus_readpage,
.writepage = hfsplus_writepage,
.write_begin = hfsplus_write_begin,
--- a/fs/hpfs/file.c~mm-require-set_page_dirty-to-be-explicitly-wire-up
+++ a/fs/hpfs/file.c
@@ -196,6 +196,7 @@ static int hpfs_fiemap(struct inode *ino
}
const struct address_space_operations hpfs_aops = {
+ .set_page_dirty = __set_page_dirty_buffers,
.readpage = hpfs_readpage,
.writepage = hpfs_writepage,
.readahead = hpfs_readahead,
--- a/fs/jfs/inode.c~mm-require-set_page_dirty-to-be-explicitly-wire-up
+++ a/fs/jfs/inode.c
@@ -356,6 +356,7 @@ static ssize_t jfs_direct_IO(struct kioc
}
const struct address_space_operations jfs_aops = {
+ .set_page_dirty = __set_page_dirty_buffers,
.readpage = jfs_readpage,
.readahead = jfs_readahead,
.writepage = jfs_writepage,
--- a/fs/minix/inode.c~mm-require-set_page_dirty-to-be-explicitly-wire-up
+++ a/fs/minix/inode.c
@@ -442,6 +442,7 @@ static sector_t minix_bmap(struct addres
}
static const struct address_space_operations minix_aops = {
+ .set_page_dirty = __set_page_dirty_buffers,
.readpage = minix_readpage,
.writepage = minix_writepage,
.write_begin = minix_write_begin,
--- a/fs/nilfs2/mdt.c~mm-require-set_page_dirty-to-be-explicitly-wire-up
+++ a/fs/nilfs2/mdt.c
@@ -434,6 +434,7 @@ nilfs_mdt_write_page(struct page *page,
static const struct address_space_operations def_mdt_aops = {
+ .set_page_dirty = __set_page_dirty_buffers,
.writepage = nilfs_mdt_write_page,
};
--- a/fs/ocfs2/aops.c~mm-require-set_page_dirty-to-be-explicitly-wire-up
+++ a/fs/ocfs2/aops.c
@@ -2453,6 +2453,7 @@ static ssize_t ocfs2_direct_IO(struct ki
}
const struct address_space_operations ocfs2_aops = {
+ .set_page_dirty = __set_page_dirty_buffers,
.readpage = ocfs2_readpage,
.readahead = ocfs2_readahead,
.writepage = ocfs2_writepage,
--- a/fs/omfs/file.c~mm-require-set_page_dirty-to-be-explicitly-wire-up
+++ a/fs/omfs/file.c
@@ -372,6 +372,7 @@ const struct inode_operations omfs_file_
};
const struct address_space_operations omfs_aops = {
+ .set_page_dirty = __set_page_dirty_buffers,
.readpage = omfs_readpage,
.readahead = omfs_readahead,
.writepage = omfs_writepage,
--- a/fs/sysv/itree.c~mm-require-set_page_dirty-to-be-explicitly-wire-up
+++ a/fs/sysv/itree.c
@@ -495,6 +495,7 @@ static sector_t sysv_bmap(struct address
}
const struct address_space_operations sysv_aops = {
+ .set_page_dirty = __set_page_dirty_buffers,
.readpage = sysv_readpage,
.writepage = sysv_writepage,
.write_begin = sysv_write_begin,
--- a/fs/udf/file.c~mm-require-set_page_dirty-to-be-explicitly-wire-up
+++ a/fs/udf/file.c
@@ -125,6 +125,7 @@ static int udf_adinicb_write_end(struct
}
const struct address_space_operations udf_adinicb_aops = {
+ .set_page_dirty = __set_page_dirty_buffers,
.readpage = udf_adinicb_readpage,
.writepage = udf_adinicb_writepage,
.write_begin = udf_adinicb_write_begin,
--- a/fs/udf/inode.c~mm-require-set_page_dirty-to-be-explicitly-wire-up
+++ a/fs/udf/inode.c
@@ -235,6 +235,7 @@ static sector_t udf_bmap(struct address_
}
const struct address_space_operations udf_aops = {
+ .set_page_dirty = __set_page_dirty_buffers,
.readpage = udf_readpage,
.readahead = udf_readahead,
.writepage = udf_writepage,
--- a/fs/ufs/inode.c~mm-require-set_page_dirty-to-be-explicitly-wire-up
+++ a/fs/ufs/inode.c
@@ -526,6 +526,7 @@ static sector_t ufs_bmap(struct address_
}
const struct address_space_operations ufs_aops = {
+ .set_page_dirty = __set_page_dirty_buffers,
.readpage = ufs_readpage,
.writepage = ufs_writepage,
.write_begin = ufs_write_begin,
--- a/mm/page-writeback.c~mm-require-set_page_dirty-to-be-explicitly-wire-up
+++ a/mm/page-writeback.c
@@ -32,7 +32,6 @@
#include <linux/sysctl.h>
#include <linux/cpu.h>
#include <linux/syscalls.h>
-#include <linux/buffer_head.h> /* __set_page_dirty_buffers */
#include <linux/pagevec.h>
#include <linux/timer.h>
#include <linux/sched/rt.h>
@@ -2555,13 +2554,9 @@ EXPORT_SYMBOL(redirty_page_for_writepage
/*
* Dirty a page.
*
- * For pages with a mapping this should be done under the page lock
- * for the benefit of asynchronous memory errors who prefer a consistent
- * dirty state. This rule can be broken in some special cases,
- * but should be better not to.
- *
- * If the mapping doesn't provide a set_page_dirty a_op, then
- * just fall through and assume that it wants buffer_heads.
+ * For pages with a mapping this should be done under the page lock for the
+ * benefit of asynchronous memory errors who prefer a consistent dirty state.
+ * This rule can be broken in some special cases, but should be better not to.
*/
int set_page_dirty(struct page *page)
{
@@ -2569,7 +2564,6 @@ int set_page_dirty(struct page *page)
page = compound_head(page);
if (likely(mapping)) {
- int (*spd)(struct page *) = mapping->a_ops->set_page_dirty;
/*
* readahead/lru_deactivate_page could remain
* PG_readahead/PG_reclaim due to race with end_page_writeback
@@ -2582,11 +2576,7 @@ int set_page_dirty(struct page *page)
*/
if (PageReclaim(page))
ClearPageReclaim(page);
-#ifdef CONFIG_BLOCK
- if (!spd)
- spd = __set_page_dirty_buffers;
-#endif
- return (*spd)(page);
+ return mapping->a_ops->set_page_dirty(page);
}
if (!PageDirty(page)) {
if (!TestSetPageDirty(page))
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 057/192] mm/writeback: move __set_page_dirty() to core mm
2021-06-29 2:32 incoming Andrew Morton
` (55 preceding siblings ...)
2021-06-29 2:36 ` [patch 056/192] mm: require ->set_page_dirty to be explicitly wired up Andrew Morton
@ 2021-06-29 2:36 ` Andrew Morton
2021-06-29 2:36 ` [patch 058/192] mm/writeback: use __set_page_dirty in __set_page_dirty_nobuffers Andrew Morton
` (134 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:36 UTC (permalink / raw)
To: akpm, dan.j.williams, gregkh, hch, jack, linux-mm, mm-commits,
torvalds, viro, willy
From: "Matthew Wilcox (Oracle)" <willy@infradead.org>
Subject: mm/writeback: move __set_page_dirty() to core mm
Patch series "Further set_page_dirty cleanups".
Prompted by Christoph's recent patches, here are some more patches to
improve the state of set_page_dirty(). They're all from the folio tree,
so they've been tested to a certain extent.
This patch (of 6):
Nothing in __set_page_dirty() is specific to buffer_head, so move it to
mm/page-writeback.c. That removes the only caller of
account_page_dirtied() outside of page-writeback.c, so make it static.
Link: https://lkml.kernel.org/r/20210615162342.1669332-1-willy@infradead.org
Link: https://lkml.kernel.org/r/20210615162342.1669332-2-willy@infradead.org
Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Cc: Jan Kara <jack@suse.cz>
Cc: Al Viro <viro@zeniv.linux.org.uk>
Cc: Dan Williams <dan.j.williams@intel.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
fs/buffer.c | 24 ------------------------
include/linux/mm.h | 1 -
mm/page-writeback.c | 27 ++++++++++++++++++++++++++-
3 files changed, 26 insertions(+), 26 deletions(-)
--- a/fs/buffer.c~mm-writeback-move-__set_page_dirty-to-core-mm
+++ a/fs/buffer.c
@@ -589,30 +589,6 @@ void mark_buffer_dirty_inode(struct buff
EXPORT_SYMBOL(mark_buffer_dirty_inode);
/*
- * Mark the page dirty, and set it dirty in the page cache, and mark the inode
- * dirty.
- *
- * If warn is true, then emit a warning if the page is not uptodate and has
- * not been truncated.
- *
- * The caller must hold lock_page_memcg().
- */
-void __set_page_dirty(struct page *page, struct address_space *mapping,
- int warn)
-{
- unsigned long flags;
-
- xa_lock_irqsave(&mapping->i_pages, flags);
- if (page->mapping) { /* Race with truncate? */
- WARN_ON_ONCE(warn && !PageUptodate(page));
- account_page_dirtied(page, mapping);
- __xa_set_mark(&mapping->i_pages, page_index(page),
- PAGECACHE_TAG_DIRTY);
- }
- xa_unlock_irqrestore(&mapping->i_pages, flags);
-}
-
-/*
* Add a page to the dirty page list.
*
* It is a sad fact of life that this function is called from several places
--- a/include/linux/mm.h~mm-writeback-move-__set_page_dirty-to-core-mm
+++ a/include/linux/mm.h
@@ -1855,7 +1855,6 @@ int __set_page_dirty_nobuffers(struct pa
int __set_page_dirty_no_writeback(struct page *page);
int redirty_page_for_writepage(struct writeback_control *wbc,
struct page *page);
-void account_page_dirtied(struct page *page, struct address_space *mapping);
void account_page_cleaned(struct page *page, struct address_space *mapping,
struct bdi_writeback *wb);
int set_page_dirty(struct page *page);
--- a/mm/page-writeback.c~mm-writeback-move-__set_page_dirty-to-core-mm
+++ a/mm/page-writeback.c
@@ -2425,7 +2425,8 @@ int __set_page_dirty_no_writeback(struct
*
* NOTE: This relies on being atomic wrt interrupts.
*/
-void account_page_dirtied(struct page *page, struct address_space *mapping)
+static void account_page_dirtied(struct page *page,
+ struct address_space *mapping)
{
struct inode *inode = mapping->host;
@@ -2467,6 +2468,30 @@ void account_page_cleaned(struct page *p
}
/*
+ * Mark the page dirty, and set it dirty in the page cache, and mark the inode
+ * dirty.
+ *
+ * If warn is true, then emit a warning if the page is not uptodate and has
+ * not been truncated.
+ *
+ * The caller must hold lock_page_memcg().
+ */
+void __set_page_dirty(struct page *page, struct address_space *mapping,
+ int warn)
+{
+ unsigned long flags;
+
+ xa_lock_irqsave(&mapping->i_pages, flags);
+ if (page->mapping) { /* Race with truncate? */
+ WARN_ON_ONCE(warn && !PageUptodate(page));
+ account_page_dirtied(page, mapping);
+ __xa_set_mark(&mapping->i_pages, page_index(page),
+ PAGECACHE_TAG_DIRTY);
+ }
+ xa_unlock_irqrestore(&mapping->i_pages, flags);
+}
+
+/*
* For address_spaces which do not use buffers. Just tag the page as dirty in
* the xarray.
*
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 058/192] mm/writeback: use __set_page_dirty in __set_page_dirty_nobuffers
2021-06-29 2:32 incoming Andrew Morton
` (56 preceding siblings ...)
2021-06-29 2:36 ` [patch 057/192] mm/writeback: move __set_page_dirty() to core mm Andrew Morton
@ 2021-06-29 2:36 ` Andrew Morton
2021-06-29 2:36 ` [patch 059/192] iomap: use __set_page_dirty_nobuffers Andrew Morton
` (133 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:36 UTC (permalink / raw)
To: akpm, dan.j.williams, gregkh, hch, jack, linux-mm, mm-commits,
torvalds, viro, willy
From: "Matthew Wilcox (Oracle)" <willy@infradead.org>
Subject: mm/writeback: use __set_page_dirty in __set_page_dirty_nobuffers
This is fundamentally the same code, so just call it instead of
duplicating it.
Link: https://lkml.kernel.org/r/20210615162342.1669332-3-willy@infradead.org
Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Cc: Al Viro <viro@zeniv.linux.org.uk>
Cc: Dan Williams <dan.j.williams@intel.com>
Cc: Jan Kara <jack@suse.cz>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/page-writeback.c | 10 +---------
1 file changed, 1 insertion(+), 9 deletions(-)
--- a/mm/page-writeback.c~mm-writeback-use-__set_page_dirty-in-__set_page_dirty_nobuffers
+++ a/mm/page-writeback.c
@@ -2508,20 +2508,12 @@ int __set_page_dirty_nobuffers(struct pa
lock_page_memcg(page);
if (!TestSetPageDirty(page)) {
struct address_space *mapping = page_mapping(page);
- unsigned long flags;
if (!mapping) {
unlock_page_memcg(page);
return 1;
}
-
- xa_lock_irqsave(&mapping->i_pages, flags);
- BUG_ON(page_mapping(page) != mapping);
- WARN_ON_ONCE(!PagePrivate(page) && !PageUptodate(page));
- account_page_dirtied(page, mapping);
- __xa_set_mark(&mapping->i_pages, page_index(page),
- PAGECACHE_TAG_DIRTY);
- xa_unlock_irqrestore(&mapping->i_pages, flags);
+ __set_page_dirty(page, mapping, !PagePrivate(page));
unlock_page_memcg(page);
if (mapping->host) {
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 059/192] iomap: use __set_page_dirty_nobuffers
2021-06-29 2:32 incoming Andrew Morton
` (57 preceding siblings ...)
2021-06-29 2:36 ` [patch 058/192] mm/writeback: use __set_page_dirty in __set_page_dirty_nobuffers Andrew Morton
@ 2021-06-29 2:36 ` Andrew Morton
2021-06-29 2:36 ` [patch 060/192] fs: remove anon_set_page_dirty() Andrew Morton
` (132 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:36 UTC (permalink / raw)
To: akpm, dan.j.williams, gregkh, hch, jack, linux-mm, mm-commits,
torvalds, viro, willy
From: "Matthew Wilcox (Oracle)" <willy@infradead.org>
Subject: iomap: use __set_page_dirty_nobuffers
The only difference between iomap_set_page_dirty() and
__set_page_dirty_nobuffers() is that the latter includes a debugging check
that a !Uptodate page has private data.
Link: https://lkml.kernel.org/r/20210615162342.1669332-4-willy@infradead.org
Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Cc: Al Viro <viro@zeniv.linux.org.uk>
Cc: Dan Williams <dan.j.williams@intel.com>
Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Cc: Jan Kara <jack@suse.cz>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
fs/gfs2/aops.c | 2 +-
fs/iomap/buffered-io.c | 27 +--------------------------
fs/xfs/xfs_aops.c | 2 +-
fs/zonefs/super.c | 2 +-
include/linux/iomap.h | 1 -
5 files changed, 4 insertions(+), 30 deletions(-)
--- a/fs/gfs2/aops.c~iomap-use-__set_page_dirty_nobuffers
+++ a/fs/gfs2/aops.c
@@ -784,7 +784,7 @@ static const struct address_space_operat
.writepages = gfs2_writepages,
.readpage = gfs2_readpage,
.readahead = gfs2_readahead,
- .set_page_dirty = iomap_set_page_dirty,
+ .set_page_dirty = __set_page_dirty_nobuffers,
.releasepage = iomap_releasepage,
.invalidatepage = iomap_invalidatepage,
.bmap = gfs2_bmap,
--- a/fs/iomap/buffered-io.c~iomap-use-__set_page_dirty_nobuffers
+++ a/fs/iomap/buffered-io.c
@@ -640,31 +640,6 @@ out_no_page:
return status;
}
-int
-iomap_set_page_dirty(struct page *page)
-{
- struct address_space *mapping = page_mapping(page);
- int newly_dirty;
-
- if (unlikely(!mapping))
- return !TestSetPageDirty(page);
-
- /*
- * Lock out page's memcg migration to keep PageDirty
- * synchronized with per-memcg dirty page counters.
- */
- lock_page_memcg(page);
- newly_dirty = !TestSetPageDirty(page);
- if (newly_dirty)
- __set_page_dirty(page, mapping, 0);
- unlock_page_memcg(page);
-
- if (newly_dirty)
- __mark_inode_dirty(mapping->host, I_DIRTY_PAGES);
- return newly_dirty;
-}
-EXPORT_SYMBOL_GPL(iomap_set_page_dirty);
-
static size_t __iomap_write_end(struct inode *inode, loff_t pos, size_t len,
size_t copied, struct page *page)
{
@@ -684,7 +659,7 @@ static size_t __iomap_write_end(struct i
if (unlikely(copied < len && !PageUptodate(page)))
return 0;
iomap_set_range_uptodate(page, offset_in_page(pos), len);
- iomap_set_page_dirty(page);
+ __set_page_dirty_nobuffers(page);
return copied;
}
--- a/fs/xfs/xfs_aops.c~iomap-use-__set_page_dirty_nobuffers
+++ a/fs/xfs/xfs_aops.c
@@ -561,7 +561,7 @@ const struct address_space_operations xf
.readahead = xfs_vm_readahead,
.writepage = xfs_vm_writepage,
.writepages = xfs_vm_writepages,
- .set_page_dirty = iomap_set_page_dirty,
+ .set_page_dirty = __set_page_dirty_nobuffers,
.releasepage = iomap_releasepage,
.invalidatepage = iomap_invalidatepage,
.bmap = xfs_vm_bmap,
--- a/fs/zonefs/super.c~iomap-use-__set_page_dirty_nobuffers
+++ a/fs/zonefs/super.c
@@ -185,7 +185,7 @@ static const struct address_space_operat
.readahead = zonefs_readahead,
.writepage = zonefs_writepage,
.writepages = zonefs_writepages,
- .set_page_dirty = iomap_set_page_dirty,
+ .set_page_dirty = __set_page_dirty_nobuffers,
.releasepage = iomap_releasepage,
.invalidatepage = iomap_invalidatepage,
.migratepage = iomap_migrate_page,
--- a/include/linux/iomap.h~iomap-use-__set_page_dirty_nobuffers
+++ a/include/linux/iomap.h
@@ -159,7 +159,6 @@ ssize_t iomap_file_buffered_write(struct
const struct iomap_ops *ops);
int iomap_readpage(struct page *page, const struct iomap_ops *ops);
void iomap_readahead(struct readahead_control *, const struct iomap_ops *ops);
-int iomap_set_page_dirty(struct page *page);
int iomap_is_partially_uptodate(struct page *page, unsigned long from,
unsigned long count);
int iomap_releasepage(struct page *page, gfp_t gfp_mask);
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 060/192] fs: remove anon_set_page_dirty()
2021-06-29 2:32 incoming Andrew Morton
` (58 preceding siblings ...)
2021-06-29 2:36 ` [patch 059/192] iomap: use __set_page_dirty_nobuffers Andrew Morton
@ 2021-06-29 2:36 ` Andrew Morton
2021-06-29 2:36 ` [patch 061/192] fs: remove noop_set_page_dirty() Andrew Morton
` (131 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:36 UTC (permalink / raw)
To: akpm, dan.j.williams, gregkh, hch, jack, linux-mm, mm-commits,
torvalds, viro, willy
From: "Matthew Wilcox (Oracle)" <willy@infradead.org>
Subject: fs: remove anon_set_page_dirty()
Use __set_page_dirty_no_writeback() instead. This will set the dirty bit
on the page, which will be used to avoid calling set_page_dirty() in the
future. It will have no effect on actually writing the page back, as the
pages are not on any LRU lists.
Link: https://lkml.kernel.org/r/20210615162342.1669332-5-willy@infradead.org
Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Cc: Al Viro <viro@zeniv.linux.org.uk>
Cc: Dan Williams <dan.j.williams@intel.com>
Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Cc: Jan Kara <jack@suse.cz>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
fs/libfs.c | 11 +----------
1 file changed, 1 insertion(+), 10 deletions(-)
--- a/fs/libfs.c~fs-remove-anon_set_page_dirty
+++ a/fs/libfs.c
@@ -1217,19 +1217,10 @@ void kfree_link(void *p)
}
EXPORT_SYMBOL(kfree_link);
-/*
- * nop .set_page_dirty method so that people can use .page_mkwrite on
- * anon inodes.
- */
-static int anon_set_page_dirty(struct page *page)
-{
- return 0;
-};
-
struct inode *alloc_anon_inode(struct super_block *s)
{
static const struct address_space_operations anon_aops = {
- .set_page_dirty = anon_set_page_dirty,
+ .set_page_dirty = __set_page_dirty_no_writeback,
};
struct inode *inode = new_inode_pseudo(s);
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 061/192] fs: remove noop_set_page_dirty()
2021-06-29 2:32 incoming Andrew Morton
` (59 preceding siblings ...)
2021-06-29 2:36 ` [patch 060/192] fs: remove anon_set_page_dirty() Andrew Morton
@ 2021-06-29 2:36 ` Andrew Morton
2021-06-29 2:36 ` [patch 062/192] mm: move page dirtying prototypes from mm.h Andrew Morton
` (130 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:36 UTC (permalink / raw)
To: akpm, dan.j.williams, gregkh, hch, jack, linux-mm, mm-commits,
torvalds, viro, willy
From: "Matthew Wilcox (Oracle)" <willy@infradead.org>
Subject: fs: remove noop_set_page_dirty()
Use __set_page_dirty_no_writeback() instead. This will set the dirty bit
on the page, which will be used to avoid calling set_page_dirty() in the
future. It will have no effect on actually writing the page back, as the
pages are not on any LRU lists.
[akpm@linux-foundation.org: export __set_page_dirty_no_writeback() to modules]
Link: https://lkml.kernel.org/r/20210615162342.1669332-6-willy@infradead.org
Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org>
Cc: Al Viro <viro@zeniv.linux.org.uk>
Cc: Christoph Hellwig <hch@lst.de>
Cc: Dan Williams <dan.j.williams@intel.com>
Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Cc: Jan Kara <jack@suse.cz>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
drivers/dax/device.c | 2 +-
fs/ext2/inode.c | 2 +-
fs/ext4/inode.c | 2 +-
fs/fuse/dax.c | 2 +-
fs/libfs.c | 16 ----------------
fs/xfs/xfs_aops.c | 2 +-
include/linux/fs.h | 1 -
mm/page-writeback.c | 1 +
8 files changed, 6 insertions(+), 22 deletions(-)
--- a/drivers/dax/device.c~fs-remove-noop_set_page_dirty
+++ a/drivers/dax/device.c
@@ -337,7 +337,7 @@ static unsigned long dax_get_unmapped_ar
}
static const struct address_space_operations dev_dax_aops = {
- .set_page_dirty = noop_set_page_dirty,
+ .set_page_dirty = __set_page_dirty_no_writeback,
.invalidatepage = noop_invalidatepage,
};
--- a/fs/ext2/inode.c~fs-remove-noop_set_page_dirty
+++ a/fs/ext2/inode.c
@@ -992,7 +992,7 @@ const struct address_space_operations ex
static const struct address_space_operations ext2_dax_aops = {
.writepages = ext2_dax_writepages,
.direct_IO = noop_direct_IO,
- .set_page_dirty = noop_set_page_dirty,
+ .set_page_dirty = __set_page_dirty_no_writeback,
.invalidatepage = noop_invalidatepage,
};
--- a/fs/ext4/inode.c~fs-remove-noop_set_page_dirty
+++ a/fs/ext4/inode.c
@@ -3701,7 +3701,7 @@ static const struct address_space_operat
static const struct address_space_operations ext4_dax_aops = {
.writepages = ext4_dax_writepages,
.direct_IO = noop_direct_IO,
- .set_page_dirty = noop_set_page_dirty,
+ .set_page_dirty = __set_page_dirty_no_writeback,
.bmap = ext4_bmap,
.invalidatepage = noop_invalidatepage,
.swap_activate = ext4_iomap_swap_activate,
--- a/fs/fuse/dax.c~fs-remove-noop_set_page_dirty
+++ a/fs/fuse/dax.c
@@ -1329,7 +1329,7 @@ bool fuse_dax_inode_alloc(struct super_b
static const struct address_space_operations fuse_dax_file_aops = {
.writepages = fuse_dax_writepages,
.direct_IO = noop_direct_IO,
- .set_page_dirty = noop_set_page_dirty,
+ .set_page_dirty = __set_page_dirty_no_writeback,
.invalidatepage = noop_invalidatepage,
};
--- a/fs/libfs.c~fs-remove-noop_set_page_dirty
+++ a/fs/libfs.c
@@ -1171,22 +1171,6 @@ int noop_fsync(struct file *file, loff_t
}
EXPORT_SYMBOL(noop_fsync);
-int noop_set_page_dirty(struct page *page)
-{
- /*
- * Unlike __set_page_dirty_no_writeback that handles dirty page
- * tracking in the page object, dax does all dirty tracking in
- * the inode address_space in response to mkwrite faults. In the
- * dax case we only need to worry about potentially dirty CPU
- * caches, not dirty page cache pages to write back.
- *
- * This callback is defined to prevent fallback to
- * __set_page_dirty_buffers() in set_page_dirty().
- */
- return 0;
-}
-EXPORT_SYMBOL_GPL(noop_set_page_dirty);
-
void noop_invalidatepage(struct page *page, unsigned int offset,
unsigned int length)
{
--- a/fs/xfs/xfs_aops.c~fs-remove-noop_set_page_dirty
+++ a/fs/xfs/xfs_aops.c
@@ -575,7 +575,7 @@ const struct address_space_operations xf
const struct address_space_operations xfs_dax_aops = {
.writepages = xfs_dax_writepages,
.direct_IO = noop_direct_IO,
- .set_page_dirty = noop_set_page_dirty,
+ .set_page_dirty = __set_page_dirty_no_writeback,
.invalidatepage = noop_invalidatepage,
.swap_activate = xfs_iomap_swapfile_activate,
};
--- a/include/linux/fs.h~fs-remove-noop_set_page_dirty
+++ a/include/linux/fs.h
@@ -3417,7 +3417,6 @@ extern int simple_rename(struct user_nam
extern void simple_recursive_removal(struct dentry *,
void (*callback)(struct dentry *));
extern int noop_fsync(struct file *, loff_t, loff_t, int);
-extern int noop_set_page_dirty(struct page *page);
extern void noop_invalidatepage(struct page *page, unsigned int offset,
unsigned int length);
extern ssize_t noop_direct_IO(struct kiocb *iocb, struct iov_iter *iter);
--- a/mm/page-writeback.c~fs-remove-noop_set_page_dirty
+++ a/mm/page-writeback.c
@@ -2417,6 +2417,7 @@ int __set_page_dirty_no_writeback(struct
return !TestSetPageDirty(page);
return 0;
}
+EXPORT_SYMBOL(__set_page_dirty_no_writeback);
/*
* Helper function for set_page_dirty family.
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 062/192] mm: move page dirtying prototypes from mm.h
2021-06-29 2:32 incoming Andrew Morton
` (60 preceding siblings ...)
2021-06-29 2:36 ` [patch 061/192] fs: remove noop_set_page_dirty() Andrew Morton
@ 2021-06-29 2:36 ` Andrew Morton
2021-06-29 2:36 ` [patch 063/192] mm/gup_benchmark: support threading Andrew Morton
` (129 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:36 UTC (permalink / raw)
To: akpm, dan.j.williams, gregkh, hch, jack, linux-mm, mm-commits,
torvalds, viro, willy
From: "Matthew Wilcox (Oracle)" <willy@infradead.org>
Subject: mm: move page dirtying prototypes from mm.h
These functions implement the address_space ->set_page_dirty operation and
should live in pagemap.h, not mm.h so that the rest of the kernel doesn't
get funny ideas about calling them directly.
Link: https://lkml.kernel.org/r/20210615162342.1669332-7-willy@infradead.org
Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Cc: Al Viro <viro@zeniv.linux.org.uk>
Cc: Dan Williams <dan.j.williams@intel.com>
Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Cc: Jan Kara <jack@suse.cz>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
fs/fuse/dax.c | 1 +
fs/zonefs/super.c | 2 +-
include/linux/mm.h | 3 ---
include/linux/pagemap.h | 4 ++++
4 files changed, 6 insertions(+), 4 deletions(-)
--- a/fs/fuse/dax.c~mm-move-page-dirtying-prototypes-from-mmh
+++ a/fs/fuse/dax.c
@@ -9,6 +9,7 @@
#include <linux/delay.h>
#include <linux/dax.h>
#include <linux/uio.h>
+#include <linux/pagemap.h>
#include <linux/pfn_t.h>
#include <linux/iomap.h>
#include <linux/interval_tree.h>
--- a/fs/zonefs/super.c~mm-move-page-dirtying-prototypes-from-mmh
+++ a/fs/zonefs/super.c
@@ -5,7 +5,7 @@
* Copyright (C) 2019 Western Digital Corporation or its affiliates.
*/
#include <linux/module.h>
-#include <linux/fs.h>
+#include <linux/pagemap.h>
#include <linux/magic.h>
#include <linux/iomap.h>
#include <linux/init.h>
--- a/include/linux/mm.h~mm-move-page-dirtying-prototypes-from-mmh
+++ a/include/linux/mm.h
@@ -1850,9 +1850,6 @@ extern int try_to_release_page(struct pa
extern void do_invalidatepage(struct page *page, unsigned int offset,
unsigned int length);
-void __set_page_dirty(struct page *, struct address_space *, int warn);
-int __set_page_dirty_nobuffers(struct page *page);
-int __set_page_dirty_no_writeback(struct page *page);
int redirty_page_for_writepage(struct writeback_control *wbc,
struct page *page);
void account_page_cleaned(struct page *page, struct address_space *mapping,
--- a/include/linux/pagemap.h~mm-move-page-dirtying-prototypes-from-mmh
+++ a/include/linux/pagemap.h
@@ -702,6 +702,10 @@ int wait_on_page_writeback_killable(stru
extern void end_page_writeback(struct page *page);
void wait_for_stable_page(struct page *page);
+void __set_page_dirty(struct page *, struct address_space *, int warn);
+int __set_page_dirty_nobuffers(struct page *page);
+int __set_page_dirty_no_writeback(struct page *page);
+
void page_endio(struct page *page, bool is_write, int err);
/**
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 063/192] mm/gup_benchmark: support threading
2021-06-29 2:32 incoming Andrew Morton
` (61 preceding siblings ...)
2021-06-29 2:36 ` [patch 062/192] mm: move page dirtying prototypes from mm.h Andrew Morton
@ 2021-06-29 2:36 ` Andrew Morton
2021-06-29 2:36 ` [patch 064/192] mm: gup: allow FOLL_PIN to scale in SMP Andrew Morton
` (128 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:36 UTC (permalink / raw)
To: aarcange, akpm, hughd, jack, jannh, jgg, jhubbard, kirill,
ktkhai, linux-mm, mhocko, mm-commits, oleg, peterx, torvalds,
willy
From: Peter Xu <peterx@redhat.com>
Subject: mm/gup_benchmark: support threading
Patch series "mm/gup: Fix pin page write cache bouncing on has_pinned", v2.
This series contains 3 patches, the 1st one enables threading for
gup_benchmark in the kselftest. The latter two patches are collected from
Andrea's local branch which can fix write cache bouncing issue with
pinning fast-gup.
To be explicit on the latter two patches:
- the 2nd patch fixes the perf degrade when introducing has_pinned, then
- the last patch tries to remove the has_pinned with a bit in mm->flags
For patch 3: originally I think we had a plan to reuse has_pinned into a
counter very soon, however that's not happening at least until today, so
maybe it proves that we can remove it until we really want such a counter
for whatever reason. As the commit message stated, it saves 4 bytes for
each mm without observable regressions.
Regarding testing: we can reference to the commit message of patch 2 for
some detailed testing with will-is-scale. Meanwhile I did patch 1 just
because then we can even easily verify the patchset using the existing
kselftest facilities or even regress test it in the future with the repo
if we want.
Below numbers are extra verification tests that I did besides commit
message of patch 2 using the new gup_benchmark and 256 cpus. Below test
is done on 40 cpus host with Intel(R) Xeon(R) CPU E5-2630 v4 @ 2.20GHz,
and I can get similar result (of course the write cache bouncing get
severe with even more cores).
After patch 1 applied (only test patch, so using old kernel):
$ sudo chrt -f 1 ./gup_test -a -m 512 -j 40
PIN_FAST_BENCHMARK: Time: get:459632 put:5990 us
PIN_FAST_BENCHMARK: Time: get:461967 put:5840 us
PIN_FAST_BENCHMARK: Time: get:464521 put:6140 us
PIN_FAST_BENCHMARK: Time: get:465176 put:7100 us
PIN_FAST_BENCHMARK: Time: get:465960 put:6733 us
PIN_FAST_BENCHMARK: Time: get:465324 put:6781 us
PIN_FAST_BENCHMARK: Time: get:466018 put:7130 us
PIN_FAST_BENCHMARK: Time: get:466362 put:7118 us
PIN_FAST_BENCHMARK: Time: get:465118 put:6975 us
PIN_FAST_BENCHMARK: Time: get:466422 put:6602 us
PIN_FAST_BENCHMARK: Time: get:465791 put:6818 us
PIN_FAST_BENCHMARK: Time: get:467091 put:6298 us
PIN_FAST_BENCHMARK: Time: get:467694 put:5432 us
PIN_FAST_BENCHMARK: Time: get:469575 put:5581 us
PIN_FAST_BENCHMARK: Time: get:468124 put:6055 us
PIN_FAST_BENCHMARK: Time: get:468877 put:6720 us
PIN_FAST_BENCHMARK: Time: get:467212 put:4961 us
PIN_FAST_BENCHMARK: Time: get:467834 put:6697 us
PIN_FAST_BENCHMARK: Time: get:470778 put:6398 us
PIN_FAST_BENCHMARK: Time: get:469788 put:6310 us
PIN_FAST_BENCHMARK: Time: get:488277 put:7113 us
PIN_FAST_BENCHMARK: Time: get:486613 put:7085 us
PIN_FAST_BENCHMARK: Time: get:486940 put:7202 us
PIN_FAST_BENCHMARK: Time: get:488728 put:7101 us
PIN_FAST_BENCHMARK: Time: get:487570 put:7327 us
PIN_FAST_BENCHMARK: Time: get:489260 put:7027 us
PIN_FAST_BENCHMARK: Time: get:488846 put:6866 us
PIN_FAST_BENCHMARK: Time: get:488521 put:6745 us
PIN_FAST_BENCHMARK: Time: get:489950 put:6459 us
PIN_FAST_BENCHMARK: Time: get:489777 put:6617 us
PIN_FAST_BENCHMARK: Time: get:488224 put:6591 us
PIN_FAST_BENCHMARK: Time: get:488644 put:6477 us
PIN_FAST_BENCHMARK: Time: get:488754 put:6711 us
PIN_FAST_BENCHMARK: Time: get:488875 put:6743 us
PIN_FAST_BENCHMARK: Time: get:489290 put:6657 us
PIN_FAST_BENCHMARK: Time: get:490264 put:6684 us
PIN_FAST_BENCHMARK: Time: get:489631 put:6737 us
PIN_FAST_BENCHMARK: Time: get:488434 put:6655 us
PIN_FAST_BENCHMARK: Time: get:492213 put:6297 us
PIN_FAST_BENCHMARK: Time: get:491124 put:6173 us
After the whole series applied (new fixed kernel):
$ sudo chrt -f 1 ./gup_test -a -m 512 -j 40
PIN_FAST_BENCHMARK: Time: get:82038 put:7041 us
PIN_FAST_BENCHMARK: Time: get:82144 put:6817 us
PIN_FAST_BENCHMARK: Time: get:83417 put:6674 us
PIN_FAST_BENCHMARK: Time: get:82540 put:6594 us
PIN_FAST_BENCHMARK: Time: get:83214 put:6681 us
PIN_FAST_BENCHMARK: Time: get:83444 put:6889 us
PIN_FAST_BENCHMARK: Time: get:83194 put:7499 us
PIN_FAST_BENCHMARK: Time: get:84876 put:7369 us
PIN_FAST_BENCHMARK: Time: get:86092 put:10289 us
PIN_FAST_BENCHMARK: Time: get:86153 put:10415 us
PIN_FAST_BENCHMARK: Time: get:85026 put:7751 us
PIN_FAST_BENCHMARK: Time: get:85458 put:7944 us
PIN_FAST_BENCHMARK: Time: get:85735 put:8154 us
PIN_FAST_BENCHMARK: Time: get:85851 put:8299 us
PIN_FAST_BENCHMARK: Time: get:86323 put:9617 us
PIN_FAST_BENCHMARK: Time: get:86288 put:10496 us
PIN_FAST_BENCHMARK: Time: get:87697 put:9346 us
PIN_FAST_BENCHMARK: Time: get:87980 put:8382 us
PIN_FAST_BENCHMARK: Time: get:88719 put:8400 us
PIN_FAST_BENCHMARK: Time: get:87616 put:8588 us
PIN_FAST_BENCHMARK: Time: get:86730 put:9563 us
PIN_FAST_BENCHMARK: Time: get:88167 put:8673 us
PIN_FAST_BENCHMARK: Time: get:86844 put:9777 us
PIN_FAST_BENCHMARK: Time: get:88068 put:11774 us
PIN_FAST_BENCHMARK: Time: get:86170 put:15676 us
PIN_FAST_BENCHMARK: Time: get:87967 put:12827 us
PIN_FAST_BENCHMARK: Time: get:95773 put:7652 us
PIN_FAST_BENCHMARK: Time: get:87734 put:13650 us
PIN_FAST_BENCHMARK: Time: get:89833 put:14237 us
PIN_FAST_BENCHMARK: Time: get:96186 put:8029 us
PIN_FAST_BENCHMARK: Time: get:95532 put:8886 us
PIN_FAST_BENCHMARK: Time: get:95351 put:5826 us
PIN_FAST_BENCHMARK: Time: get:96401 put:8407 us
PIN_FAST_BENCHMARK: Time: get:96473 put:8287 us
PIN_FAST_BENCHMARK: Time: get:97177 put:8430 us
PIN_FAST_BENCHMARK: Time: get:98120 put:5263 us
PIN_FAST_BENCHMARK: Time: get:96271 put:7757 us
PIN_FAST_BENCHMARK: Time: get:99628 put:10467 us
PIN_FAST_BENCHMARK: Time: get:99344 put:10045 us
PIN_FAST_BENCHMARK: Time: get:94212 put:15485 us
Summary:
Old kernel: 477729.97 (+-3.79%)
New kernel: 89144.65 (+-11.76%)
This patch (of 3):
Add a new parameter "-j N" to support concurrent gup test.
Link: https://lkml.kernel.org/r/20210507150553.208763-1-peterx@redhat.com
Link: https://lkml.kernel.org/r/20210507150553.208763-2-peterx@redhat.com
Signed-off-by: Peter Xu <peterx@redhat.com>
Reviewed-by: John Hubbard <jhubbard@nvidia.com>
Cc: Jan Kara <jack@suse.cz>
Cc: Michal Hocko <mhocko@suse.com>
Cc: Kirill Tkhai <ktkhai@virtuozzo.com>
Cc: Kirill Shutemov <kirill@shutemov.name>
Cc: Oleg Nesterov <oleg@redhat.com>
Cc: Jann Horn <jannh@google.com>
Cc: Andrea Arcangeli <aarcange@redhat.com>
Cc: Jason Gunthorpe <jgg@nvidia.com>
Cc: Matthew Wilcox <willy@infradead.org>
Cc: Hugh Dickins <hughd@google.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
tools/testing/selftests/vm/gup_test.c | 96 ++++++++++++++++--------
1 file changed, 65 insertions(+), 31 deletions(-)
--- a/tools/testing/selftests/vm/gup_test.c~mm-gup_benchmark-support-threading
+++ a/tools/testing/selftests/vm/gup_test.c
@@ -6,6 +6,8 @@
#include <sys/mman.h>
#include <sys/stat.h>
#include <sys/types.h>
+#include <pthread.h>
+#include <assert.h>
#include "../../../../mm/gup_test.h"
#define MB (1UL << 20)
@@ -15,6 +17,12 @@
#define FOLL_WRITE 0x01 /* check pte is writable */
#define FOLL_TOUCH 0x02 /* mark page accessed */
+static unsigned long cmd = GUP_FAST_BENCHMARK;
+static int gup_fd, repeats = 1;
+static unsigned long size = 128 * MB;
+/* Serialize prints */
+static pthread_mutex_t print_mutex = PTHREAD_MUTEX_INITIALIZER;
+
static char *cmd_to_str(unsigned long cmd)
{
switch (cmd) {
@@ -34,17 +42,55 @@ static char *cmd_to_str(unsigned long cm
return "Unknown command";
}
+void *gup_thread(void *data)
+{
+ struct gup_test gup = *(struct gup_test *)data;
+ int i;
+
+ /* Only report timing information on the *_BENCHMARK commands: */
+ if ((cmd == PIN_FAST_BENCHMARK) || (cmd == GUP_FAST_BENCHMARK) ||
+ (cmd == PIN_LONGTERM_BENCHMARK)) {
+ for (i = 0; i < repeats; i++) {
+ gup.size = size;
+ if (ioctl(gup_fd, cmd, &gup))
+ perror("ioctl"), exit(1);
+
+ pthread_mutex_lock(&print_mutex);
+ printf("%s: Time: get:%lld put:%lld us",
+ cmd_to_str(cmd), gup.get_delta_usec,
+ gup.put_delta_usec);
+ if (gup.size != size)
+ printf(", truncated (size: %lld)", gup.size);
+ printf("\n");
+ pthread_mutex_unlock(&print_mutex);
+ }
+ } else {
+ gup.size = size;
+ if (ioctl(gup_fd, cmd, &gup)) {
+ perror("ioctl");
+ exit(1);
+ }
+
+ pthread_mutex_lock(&print_mutex);
+ printf("%s: done\n", cmd_to_str(cmd));
+ if (gup.size != size)
+ printf("Truncated (size: %lld)\n", gup.size);
+ pthread_mutex_unlock(&print_mutex);
+ }
+
+ return NULL;
+}
+
int main(int argc, char **argv)
{
struct gup_test gup = { 0 };
- unsigned long size = 128 * MB;
- int i, fd, filed, opt, nr_pages = 1, thp = -1, repeats = 1, write = 1;
- unsigned long cmd = GUP_FAST_BENCHMARK;
+ int filed, i, opt, nr_pages = 1, thp = -1, write = 1, nthreads = 1, ret;
int flags = MAP_PRIVATE, touch = 0;
char *file = "/dev/zero";
+ pthread_t *tid;
char *p;
- while ((opt = getopt(argc, argv, "m:r:n:F:f:abctTLUuwWSHpz")) != -1) {
+ while ((opt = getopt(argc, argv, "m:r:n:F:f:abcj:tTLUuwWSHpz")) != -1) {
switch (opt) {
case 'a':
cmd = PIN_FAST_BENCHMARK;
@@ -74,6 +120,9 @@ int main(int argc, char **argv)
/* strtol, so you can pass flags in hex form */
gup.gup_flags = strtol(optarg, 0, 0);
break;
+ case 'j':
+ nthreads = atoi(optarg);
+ break;
case 'm':
size = atoi(optarg) * MB;
break;
@@ -154,8 +203,8 @@ int main(int argc, char **argv)
if (write)
gup.gup_flags |= FOLL_WRITE;
- fd = open("/sys/kernel/debug/gup_test", O_RDWR);
- if (fd == -1) {
+ gup_fd = open("/sys/kernel/debug/gup_test", O_RDWR);
+ if (gup_fd == -1) {
perror("open");
exit(1);
}
@@ -185,32 +234,17 @@ int main(int argc, char **argv)
p[0] = 0;
}
- /* Only report timing information on the *_BENCHMARK commands: */
- if ((cmd == PIN_FAST_BENCHMARK) || (cmd == GUP_FAST_BENCHMARK) ||
- (cmd == PIN_LONGTERM_BENCHMARK)) {
- for (i = 0; i < repeats; i++) {
- gup.size = size;
- if (ioctl(fd, cmd, &gup))
- perror("ioctl"), exit(1);
-
- printf("%s: Time: get:%lld put:%lld us",
- cmd_to_str(cmd), gup.get_delta_usec,
- gup.put_delta_usec);
- if (gup.size != size)
- printf(", truncated (size: %lld)", gup.size);
- printf("\n");
- }
- } else {
- gup.size = size;
- if (ioctl(fd, cmd, &gup)) {
- perror("ioctl");
- exit(1);
- }
-
- printf("%s: done\n", cmd_to_str(cmd));
- if (gup.size != size)
- printf("Truncated (size: %lld)\n", gup.size);
+ tid = malloc(sizeof(pthread_t) * nthreads);
+ assert(tid);
+ for (i = 0; i < nthreads; i++) {
+ ret = pthread_create(&tid[i], NULL, gup_thread, &gup);
+ assert(ret == 0);
+ }
+ for (i = 0; i < nthreads; i++) {
+ ret = pthread_join(tid[i], NULL);
+ assert(ret == 0);
}
+ free(tid);
return 0;
}
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 064/192] mm: gup: allow FOLL_PIN to scale in SMP
2021-06-29 2:32 incoming Andrew Morton
` (62 preceding siblings ...)
2021-06-29 2:36 ` [patch 063/192] mm/gup_benchmark: support threading Andrew Morton
@ 2021-06-29 2:36 ` Andrew Morton
2021-06-29 2:36 ` [patch 065/192] mm: gup: pack has_pinned in MMF_HAS_PINNED Andrew Morton
` (127 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:36 UTC (permalink / raw)
To: aarcange, akpm, hughd, jack, jannh, jgg, jhubbard, kirill,
ktkhai, linux-mm, mhocko, mm-commits, oleg, peterx, torvalds,
willy
From: Andrea Arcangeli <aarcange@redhat.com>
Subject: mm: gup: allow FOLL_PIN to scale in SMP
has_pinned cannot be written by each pin-fast or it won't scale in SMP.
This isn't "false sharing" strictly speaking (it's more like "true
non-sharing"), but it creates the same SMP scalability bottleneck of
"false sharing".
To verify the improvement, below test is done on 40 cpus host with
Intel(R) Xeon(R) CPU E5-2630 v4 @ 2.20GHz (must be with
CONFIG_GUP_TEST=y):
$ sudo chrt -f 1 ./gup_test -a -m 512 -j 40
Where we can get (average value for 40 threads):
Old kernel: 477729.97 (+- 3.79%)
New kernel: 89144.65 (+-11.76%)
On a similar condition with 256 cpus, this commits increases the SMP
scalability of pin_user_pages_fast() executed by different threads of the
same process by more than 4000%.
[peterx@redhat.com: rewrite commit message, add parentheses against "(A & B)"]
Link: https://lkml.kernel.org/r/20210507150553.208763-3-peterx@redhat.com
Signed-off-by: Andrea Arcangeli <aarcange@redhat.com>
Signed-off-by: Peter Xu <peterx@redhat.com>
Reviewed-by: John Hubbard <jhubbard@nvidia.com>
Cc: Hugh Dickins <hughd@google.com>
Cc: Jan Kara <jack@suse.cz>
Cc: Jann Horn <jannh@google.com>
Cc: Jason Gunthorpe <jgg@nvidia.com>
Cc: Kirill Shutemov <kirill@shutemov.name>
Cc: Kirill Tkhai <ktkhai@virtuozzo.com>
Cc: Matthew Wilcox <willy@infradead.org>
Cc: Michal Hocko <mhocko@suse.com>
Cc: Oleg Nesterov <oleg@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/gup.c | 4 ++--
1 file changed, 2 insertions(+), 2 deletions(-)
--- a/mm/gup.c~mm-gup-allow-foll_pin-to-scale-in-smp
+++ a/mm/gup.c
@@ -1320,7 +1320,7 @@ static __always_inline long __get_user_p
BUG_ON(*locked != 1);
}
- if (flags & FOLL_PIN)
+ if ((flags & FOLL_PIN) && !atomic_read(&mm->has_pinned))
atomic_set(&mm->has_pinned, 1);
/*
@@ -2641,7 +2641,7 @@ static int internal_get_user_pages_fast(
FOLL_FAST_ONLY)))
return -EINVAL;
- if (gup_flags & FOLL_PIN)
+ if ((gup_flags & FOLL_PIN) && !atomic_read(¤t->mm->has_pinned))
atomic_set(¤t->mm->has_pinned, 1);
if (!(gup_flags & FOLL_FAST_ONLY))
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 065/192] mm: gup: pack has_pinned in MMF_HAS_PINNED
2021-06-29 2:32 incoming Andrew Morton
` (63 preceding siblings ...)
2021-06-29 2:36 ` [patch 064/192] mm: gup: allow FOLL_PIN to scale in SMP Andrew Morton
@ 2021-06-29 2:36 ` Andrew Morton
2021-06-29 2:36 ` [patch 066/192] mm: pagewalk: fix walk for hugepage tables Andrew Morton
` (126 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:36 UTC (permalink / raw)
To: aarcange, akpm, hughd, jack, jannh, jgg, jhubbard, kirill,
ktkhai, linux-mm, mhocko, mm-commits, oleg, peterx, torvalds,
willy
From: Andrea Arcangeli <aarcange@redhat.com>
Subject: mm: gup: pack has_pinned in MMF_HAS_PINNED
has_pinned 32bit can be packed in the MMF_HAS_PINNED bit as a noop
cleanup.
Any atomic_inc/dec to the mm cacheline shared by all threads in pin-fast
would reintroduce a loss of SMP scalability to pin-fast, so there's no
future potential usefulness to keep an atomic in the mm for this.
set_bit(MMF_HAS_PINNED) will be theoretically a bit slower than WRITE_ONCE
(atomic_set is equivalent to WRITE_ONCE), but the set_bit (just like
atomic_set after this commit) has to be still issued only once per "mm",
so the difference between the two will be lost in the noise.
will-it-scale "mmap2" shows no change in performance with enterprise
config as expected.
will-it-scale "pin_fast" retains the > 4000% SMP scalability performance
improvement against upstream as expected.
This is a noop as far as overall performance and SMP scalability are
concerned.
[peterx@redhat.com: pack has_pinned in MMF_HAS_PINNED]
Link: https://lkml.kernel.org/r/YJqWESqyxa8OZA+2@t490s
[akpm@linux-foundation.org: coding style fixes]
[peterx@redhat.com: fix build for task_mmu.c, introduce mm_set_has_pinned_flag, fix comments]
Link: https://lkml.kernel.org/r/20210507150553.208763-4-peterx@redhat.com
Signed-off-by: Andrea Arcangeli <aarcange@redhat.com>
Signed-off-by: Peter Xu <peterx@redhat.com>
Reviewed-by: John Hubbard <jhubbard@nvidia.com>
Cc: Hugh Dickins <hughd@google.com>
Cc: Jan Kara <jack@suse.cz>
Cc: Jann Horn <jannh@google.com>
Cc: Jason Gunthorpe <jgg@nvidia.com>
Cc: Kirill Shutemov <kirill@shutemov.name>
Cc: Kirill Tkhai <ktkhai@virtuozzo.com>
Cc: Matthew Wilcox <willy@infradead.org>
Cc: Michal Hocko <mhocko@suse.com>
Cc: Oleg Nesterov <oleg@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
fs/proc/task_mmu.c | 2 +-
include/linux/mm.h | 2 +-
include/linux/mm_types.h | 10 ----------
include/linux/sched/coredump.h | 8 ++++++++
kernel/fork.c | 1 -
mm/gup.c | 19 +++++++++++++++----
6 files changed, 25 insertions(+), 17 deletions(-)
--- a/fs/proc/task_mmu.c~mm-gup-pack-has_pinned-in-mmf_has_pinned
+++ a/fs/proc/task_mmu.c
@@ -1047,7 +1047,7 @@ static inline bool pte_is_pinned(struct
return false;
if (!is_cow_mapping(vma->vm_flags))
return false;
- if (likely(!atomic_read(&vma->vm_mm->has_pinned)))
+ if (likely(!test_bit(MMF_HAS_PINNED, &vma->vm_mm->flags)))
return false;
page = vm_normal_page(vma, addr, pte);
if (!page)
--- a/include/linux/mm.h~mm-gup-pack-has_pinned-in-mmf_has_pinned
+++ a/include/linux/mm.h
@@ -1341,7 +1341,7 @@ static inline bool page_needs_cow_for_dm
if (!is_cow_mapping(vma->vm_flags))
return false;
- if (!atomic_read(&vma->vm_mm->has_pinned))
+ if (!test_bit(MMF_HAS_PINNED, &vma->vm_mm->flags))
return false;
return page_maybe_dma_pinned(page);
--- a/include/linux/mm_types.h~mm-gup-pack-has_pinned-in-mmf_has_pinned
+++ a/include/linux/mm_types.h
@@ -435,16 +435,6 @@ struct mm_struct {
*/
atomic_t mm_count;
- /**
- * @has_pinned: Whether this mm has pinned any pages. This can
- * be either replaced in the future by @pinned_vm when it
- * becomes stable, or grow into a counter on its own. We're
- * aggresive on this bit now - even if the pinned pages were
- * unpinned later on, we'll still keep this bit set for the
- * lifecycle of this mm just for simplicity.
- */
- atomic_t has_pinned;
-
#ifdef CONFIG_MMU
atomic_long_t pgtables_bytes; /* PTE page table pages */
#endif
--- a/include/linux/sched/coredump.h~mm-gup-pack-has_pinned-in-mmf_has_pinned
+++ a/include/linux/sched/coredump.h
@@ -73,6 +73,14 @@ static inline int get_dumpable(struct mm
#define MMF_OOM_VICTIM 25 /* mm is the oom victim */
#define MMF_OOM_REAP_QUEUED 26 /* mm was queued for oom_reaper */
#define MMF_MULTIPROCESS 27 /* mm is shared between processes */
+/*
+ * MMF_HAS_PINNED: Whether this mm has pinned any pages. This can be either
+ * replaced in the future by mm.pinned_vm when it becomes stable, or grow into
+ * a counter on its own. We're aggresive on this bit for now: even if the
+ * pinned pages were unpinned later on, we'll still keep this bit set for the
+ * lifecycle of this mm, just for simplicity.
+ */
+#define MMF_HAS_PINNED 28 /* FOLL_PIN has run, never cleared */
#define MMF_DISABLE_THP_MASK (1 << MMF_DISABLE_THP)
#define MMF_INIT_MASK (MMF_DUMPABLE_MASK | MMF_DUMP_FILTER_MASK |\
--- a/kernel/fork.c~mm-gup-pack-has_pinned-in-mmf_has_pinned
+++ a/kernel/fork.c
@@ -1029,7 +1029,6 @@ static struct mm_struct *mm_init(struct
mm_pgtables_bytes_init(mm);
mm->map_count = 0;
mm->locked_vm = 0;
- atomic_set(&mm->has_pinned, 0);
atomic64_set(&mm->pinned_vm, 0);
memset(&mm->rss_stat, 0, sizeof(mm->rss_stat));
spin_lock_init(&mm->page_table_lock);
--- a/mm/gup.c~mm-gup-pack-has_pinned-in-mmf_has_pinned
+++ a/mm/gup.c
@@ -420,6 +420,17 @@ void unpin_user_pages(struct page **page
}
EXPORT_SYMBOL(unpin_user_pages);
+/*
+ * Set the MMF_HAS_PINNED if not set yet; after set it'll be there for the mm's
+ * lifecycle. Avoid setting the bit unless necessary, or it might cause write
+ * cache bouncing on large SMP machines for concurrent pinned gups.
+ */
+static inline void mm_set_has_pinned_flag(unsigned long *mm_flags)
+{
+ if (!test_bit(MMF_HAS_PINNED, mm_flags))
+ set_bit(MMF_HAS_PINNED, mm_flags);
+}
+
#ifdef CONFIG_MMU
static struct page *no_page_table(struct vm_area_struct *vma,
unsigned int flags)
@@ -1320,8 +1331,8 @@ static __always_inline long __get_user_p
BUG_ON(*locked != 1);
}
- if ((flags & FOLL_PIN) && !atomic_read(&mm->has_pinned))
- atomic_set(&mm->has_pinned, 1);
+ if (flags & FOLL_PIN)
+ mm_set_has_pinned_flag(&mm->flags);
/*
* FOLL_PIN and FOLL_GET are mutually exclusive. Traditional behavior
@@ -2641,8 +2652,8 @@ static int internal_get_user_pages_fast(
FOLL_FAST_ONLY)))
return -EINVAL;
- if ((gup_flags & FOLL_PIN) && !atomic_read(¤t->mm->has_pinned))
- atomic_set(¤t->mm->has_pinned, 1);
+ if (gup_flags & FOLL_PIN)
+ mm_set_has_pinned_flag(¤t->mm->flags);
if (!(gup_flags & FOLL_FAST_ONLY))
might_lock_read(¤t->mm->mmap_lock);
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 066/192] mm: pagewalk: fix walk for hugepage tables
2021-06-29 2:32 incoming Andrew Morton
` (64 preceding siblings ...)
2021-06-29 2:36 ` [patch 065/192] mm: gup: pack has_pinned in MMF_HAS_PINNED Andrew Morton
@ 2021-06-29 2:36 ` Andrew Morton
2021-06-29 2:36 ` [patch 067/192] mm/swapfile: use percpu_ref to serialize against concurrent swapoff Andrew Morton
` (125 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:36 UTC (permalink / raw)
To: akpm, benh, christophe.leroy, dja, linux-mm, mm-commits, mpe,
oohall, paulus, steven.price, torvalds
From: Christophe Leroy <christophe.leroy@csgroup.eu>
Subject: mm: pagewalk: fix walk for hugepage tables
Pagewalk ignores hugepd entries and walk down the tables as if it was
traditionnal entries, leading to crazy result.
Add walk_hugepd_range() and use it to walk hugepage tables.
Link: https://lkml.kernel.org/r/38d04410700c8d02f28ba37e020b62c55d6f3d2c.1624597695.git.christophe.leroy@csgroup.eu
Signed-off-by: Christophe Leroy <christophe.leroy@csgroup.eu>
Reviewed-by: Steven Price <steven.price@arm.com>
Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
Cc: Paul Mackerras <paulus@samba.org>
Cc: Michael Ellerman <mpe@ellerman.id.au>
Cc: Daniel Axtens <dja@axtens.net>
Cc: "Oliver O'Halloran" <oohall@gmail.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/pagewalk.c | 58 +++++++++++++++++++++++++++++++++++++++++++-----
1 file changed, 53 insertions(+), 5 deletions(-)
--- a/mm/pagewalk.c~mm-pagewalk-fix-walk-for-hugepage-tables
+++ a/mm/pagewalk.c
@@ -58,6 +58,45 @@ static int walk_pte_range(pmd_t *pmd, un
return err;
}
+#ifdef CONFIG_ARCH_HAS_HUGEPD
+static int walk_hugepd_range(hugepd_t *phpd, unsigned long addr,
+ unsigned long end, struct mm_walk *walk, int pdshift)
+{
+ int err = 0;
+ const struct mm_walk_ops *ops = walk->ops;
+ int shift = hugepd_shift(*phpd);
+ int page_size = 1 << shift;
+
+ if (!ops->pte_entry)
+ return 0;
+
+ if (addr & (page_size - 1))
+ return 0;
+
+ for (;;) {
+ pte_t *pte;
+
+ spin_lock(&walk->mm->page_table_lock);
+ pte = hugepte_offset(*phpd, addr, pdshift);
+ err = ops->pte_entry(pte, addr, addr + page_size, walk);
+ spin_unlock(&walk->mm->page_table_lock);
+
+ if (err)
+ break;
+ if (addr >= end - page_size)
+ break;
+ addr += page_size;
+ }
+ return err;
+}
+#else
+static int walk_hugepd_range(hugepd_t *phpd, unsigned long addr,
+ unsigned long end, struct mm_walk *walk, int pdshift)
+{
+ return 0;
+}
+#endif
+
static int walk_pmd_range(pud_t *pud, unsigned long addr, unsigned long end,
struct mm_walk *walk)
{
@@ -108,7 +147,10 @@ again:
goto again;
}
- err = walk_pte_range(pmd, addr, next, walk);
+ if (is_hugepd(__hugepd(pmd_val(*pmd))))
+ err = walk_hugepd_range((hugepd_t *)pmd, addr, next, walk, PMD_SHIFT);
+ else
+ err = walk_pte_range(pmd, addr, next, walk);
if (err)
break;
} while (pmd++, addr = next, addr != end);
@@ -157,7 +199,10 @@ static int walk_pud_range(p4d_t *p4d, un
if (pud_none(*pud))
goto again;
- err = walk_pmd_range(pud, addr, next, walk);
+ if (is_hugepd(__hugepd(pud_val(*pud))))
+ err = walk_hugepd_range((hugepd_t *)pud, addr, next, walk, PUD_SHIFT);
+ else
+ err = walk_pmd_range(pud, addr, next, walk);
if (err)
break;
} while (pud++, addr = next, addr != end);
@@ -189,7 +234,9 @@ static int walk_p4d_range(pgd_t *pgd, un
if (err)
break;
}
- if (ops->pud_entry || ops->pmd_entry || ops->pte_entry)
+ if (is_hugepd(__hugepd(p4d_val(*p4d))))
+ err = walk_hugepd_range((hugepd_t *)p4d, addr, next, walk, P4D_SHIFT);
+ else if (ops->pud_entry || ops->pmd_entry || ops->pte_entry)
err = walk_pud_range(p4d, addr, next, walk);
if (err)
break;
@@ -224,8 +271,9 @@ static int walk_pgd_range(unsigned long
if (err)
break;
}
- if (ops->p4d_entry || ops->pud_entry || ops->pmd_entry ||
- ops->pte_entry)
+ if (is_hugepd(__hugepd(pgd_val(*pgd))))
+ err = walk_hugepd_range((hugepd_t *)pgd, addr, next, walk, PGDIR_SHIFT);
+ else if (ops->p4d_entry || ops->pud_entry || ops->pmd_entry || ops->pte_entry)
err = walk_p4d_range(pgd, addr, next, walk);
if (err)
break;
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 067/192] mm/swapfile: use percpu_ref to serialize against concurrent swapoff
2021-06-29 2:32 incoming Andrew Morton
` (65 preceding siblings ...)
2021-06-29 2:36 ` [patch 066/192] mm: pagewalk: fix walk for hugepage tables Andrew Morton
@ 2021-06-29 2:36 ` Andrew Morton
2021-06-29 2:36 ` [patch 068/192] swap: fix do_swap_page() race with swapoff Andrew Morton
` (124 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:36 UTC (permalink / raw)
To: akpm, alexs, david, dennis, hannes, hughd, iamjoonsoo.kim,
linmiaohe, linux-mm, mhocko, minchan, mm-commits,
richard.weiyang, shy828301, tim.c.chen, torvalds, willy,
ying.huang, yuzhao
From: Miaohe Lin <linmiaohe@huawei.com>
Subject: mm/swapfile: use percpu_ref to serialize against concurrent swapoff
Patch series "close various race windows for swap", v6.
When I was investigating the swap code, I found some possible race
windows. This series aims to fix all these races. But using current
get/put_swap_device() to guard against concurrent swapoff for
swap_readpage() looks terrible because swap_readpage() may take really
long time. And to reduce the performance overhead on the hot-path as much
as possible, it appears we can use the percpu_ref to close this race
window(as suggested by Huang, Ying). The patch 1 adds percpu_ref support
for swap and most of the remaining patches try to use this to close
various race windows. More details can be found in the respective
changelogs.
This patch (of 4):
Using current get/put_swap_device() to guard against concurrent swapoff
for some swap ops, e.g. swap_readpage(), looks terrible because they
might take really long time. This patch adds the percpu_ref support to
serialize against concurrent swapoff(as suggested by Huang, Ying). Also
we remove the SWP_VALID flag because it's used together with RCU solution.
Link: https://lkml.kernel.org/r/20210426123316.806267-1-linmiaohe@huawei.com
Link: https://lkml.kernel.org/r/20210426123316.806267-2-linmiaohe@huawei.com
Signed-off-by: Miaohe Lin <linmiaohe@huawei.com>
Reviewed-by: "Huang, Ying" <ying.huang@intel.com>
Cc: Alex Shi <alexs@kernel.org>
Cc: David Hildenbrand <david@redhat.com>
Cc: Dennis Zhou <dennis@kernel.org>
Cc: Hugh Dickins <hughd@google.com>
Cc: Johannes Weiner <hannes@cmpxchg.org>
Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Cc: Matthew Wilcox <willy@infradead.org>
Cc: Michal Hocko <mhocko@suse.com>
Cc: Minchan Kim <minchan@kernel.org>
Cc: Tim Chen <tim.c.chen@linux.intel.com>
Cc: Wei Yang <richard.weiyang@gmail.com>
Cc: Yang Shi <shy828301@gmail.com>
Cc: Yu Zhao <yuzhao@google.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/linux/swap.h | 5 +-
mm/swapfile.c | 79 +++++++++++++++++++++++++----------------
2 files changed, 52 insertions(+), 32 deletions(-)
--- a/include/linux/swap.h~mm-swapfile-use-percpu_ref-to-serialize-against-concurrent-swapoff
+++ a/include/linux/swap.h
@@ -177,7 +177,6 @@ enum {
SWP_PAGE_DISCARD = (1 << 10), /* freed swap page-cluster discards */
SWP_STABLE_WRITES = (1 << 11), /* no overwrite PG_writeback pages */
SWP_SYNCHRONOUS_IO = (1 << 12), /* synchronous IO is efficient */
- SWP_VALID = (1 << 13), /* swap is valid to be operated on? */
/* add others here before... */
SWP_SCANNING = (1 << 14), /* refcount in scan_swap_map */
};
@@ -240,6 +239,7 @@ struct swap_cluster_list {
* The in-memory structure used to track swap areas.
*/
struct swap_info_struct {
+ struct percpu_ref users; /* indicate and keep swap device valid. */
unsigned long flags; /* SWP_USED etc: see above */
signed short prio; /* swap priority of this type */
struct plist_node list; /* entry in swap_active_head */
@@ -260,6 +260,7 @@ struct swap_info_struct {
struct block_device *bdev; /* swap device or bdev of swap file */
struct file *swap_file; /* seldom referenced */
unsigned int old_block_size; /* seldom referenced */
+ struct completion comp; /* seldom referenced */
#ifdef CONFIG_FRONTSWAP
unsigned long *frontswap_map; /* frontswap in-use, one bit per page */
atomic_t frontswap_pages; /* frontswap pages in-use counter */
@@ -511,7 +512,7 @@ sector_t swap_page_sector(struct page *p
static inline void put_swap_device(struct swap_info_struct *si)
{
- rcu_read_unlock();
+ percpu_ref_put(&si->users);
}
#else /* CONFIG_SWAP */
--- a/mm/swapfile.c~mm-swapfile-use-percpu_ref-to-serialize-against-concurrent-swapoff
+++ a/mm/swapfile.c
@@ -39,6 +39,7 @@
#include <linux/export.h>
#include <linux/swap_slots.h>
#include <linux/sort.h>
+#include <linux/completion.h>
#include <asm/tlbflush.h>
#include <linux/swapops.h>
@@ -511,6 +512,14 @@ static void swap_discard_work(struct wor
spin_unlock(&si->lock);
}
+static void swap_users_ref_free(struct percpu_ref *ref)
+{
+ struct swap_info_struct *si;
+
+ si = container_of(ref, struct swap_info_struct, users);
+ complete(&si->comp);
+}
+
static void alloc_cluster(struct swap_info_struct *si, unsigned long idx)
{
struct swap_cluster_info *ci = si->cluster_info;
@@ -1270,18 +1279,12 @@ static unsigned char __swap_entry_free_l
* via preventing the swap device from being swapoff, until
* put_swap_device() is called. Otherwise return NULL.
*
- * The entirety of the RCU read critical section must come before the
- * return from or after the call to synchronize_rcu() in
- * enable_swap_info() or swapoff(). So if "si->flags & SWP_VALID" is
- * true, the si->map, si->cluster_info, etc. must be valid in the
- * critical section.
- *
* Notice that swapoff or swapoff+swapon can still happen before the
- * rcu_read_lock() in get_swap_device() or after the rcu_read_unlock()
- * in put_swap_device() if there isn't any other way to prevent
- * swapoff, such as page lock, page table lock, etc. The caller must
- * be prepared for that. For example, the following situation is
- * possible.
+ * percpu_ref_tryget_live() in get_swap_device() or after the
+ * percpu_ref_put() in put_swap_device() if there isn't any other way
+ * to prevent swapoff, such as page lock, page table lock, etc. The
+ * caller must be prepared for that. For example, the following
+ * situation is possible.
*
* CPU1 CPU2
* do_swap_page()
@@ -1309,21 +1312,27 @@ struct swap_info_struct *get_swap_device
si = swp_swap_info(entry);
if (!si)
goto bad_nofile;
-
- rcu_read_lock();
- if (data_race(!(si->flags & SWP_VALID)))
- goto unlock_out;
+ if (!percpu_ref_tryget_live(&si->users))
+ goto out;
+ /*
+ * Guarantee the si->users are checked before accessing other
+ * fields of swap_info_struct.
+ *
+ * Paired with the spin_unlock() after setup_swap_info() in
+ * enable_swap_info().
+ */
+ smp_rmb();
offset = swp_offset(entry);
if (offset >= si->max)
- goto unlock_out;
+ goto put_out;
return si;
bad_nofile:
pr_err("%s: %s%08lx\n", __func__, Bad_file, entry.val);
out:
return NULL;
-unlock_out:
- rcu_read_unlock();
+put_out:
+ percpu_ref_put(&si->users);
return NULL;
}
@@ -2466,7 +2475,7 @@ static void setup_swap_info(struct swap_
static void _enable_swap_info(struct swap_info_struct *p)
{
- p->flags |= SWP_WRITEOK | SWP_VALID;
+ p->flags |= SWP_WRITEOK;
atomic_long_add(p->pages, &nr_swap_pages);
total_swap_pages += p->pages;
@@ -2497,10 +2506,9 @@ static void enable_swap_info(struct swap
spin_unlock(&p->lock);
spin_unlock(&swap_lock);
/*
- * Guarantee swap_map, cluster_info, etc. fields are valid
- * between get/put_swap_device() if SWP_VALID bit is set
+ * Finished initializing swap device, now it's safe to reference it.
*/
- synchronize_rcu();
+ percpu_ref_resurrect(&p->users);
spin_lock(&swap_lock);
spin_lock(&p->lock);
_enable_swap_info(p);
@@ -2616,16 +2624,16 @@ SYSCALL_DEFINE1(swapoff, const char __us
reenable_swap_slots_cache_unlock();
- spin_lock(&swap_lock);
- spin_lock(&p->lock);
- p->flags &= ~SWP_VALID; /* mark swap device as invalid */
- spin_unlock(&p->lock);
- spin_unlock(&swap_lock);
/*
- * wait for swap operations protected by get/put_swap_device()
- * to complete
+ * Wait for swap operations protected by get/put_swap_device()
+ * to complete.
+ *
+ * We need synchronize_rcu() here to protect the accessing to
+ * the swap cache data structure.
*/
+ percpu_ref_kill(&p->users);
synchronize_rcu();
+ wait_for_completion(&p->comp);
flush_work(&p->discard_work);
@@ -2857,6 +2865,12 @@ static struct swap_info_struct *alloc_sw
if (!p)
return ERR_PTR(-ENOMEM);
+ if (percpu_ref_init(&p->users, swap_users_ref_free,
+ PERCPU_REF_INIT_DEAD, GFP_KERNEL)) {
+ kvfree(p);
+ return ERR_PTR(-ENOMEM);
+ }
+
spin_lock(&swap_lock);
for (type = 0; type < nr_swapfiles; type++) {
if (!(swap_info[type]->flags & SWP_USED))
@@ -2864,6 +2878,7 @@ static struct swap_info_struct *alloc_sw
}
if (type >= MAX_SWAPFILES) {
spin_unlock(&swap_lock);
+ percpu_ref_exit(&p->users);
kvfree(p);
return ERR_PTR(-EPERM);
}
@@ -2891,9 +2906,13 @@ static struct swap_info_struct *alloc_sw
plist_node_init(&p->avail_lists[i], 0);
p->flags = SWP_USED;
spin_unlock(&swap_lock);
- kvfree(defer);
+ if (defer) {
+ percpu_ref_exit(&defer->users);
+ kvfree(defer);
+ }
spin_lock_init(&p->lock);
spin_lock_init(&p->cont_lock);
+ init_completion(&p->comp);
return p;
}
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 068/192] swap: fix do_swap_page() race with swapoff
2021-06-29 2:32 incoming Andrew Morton
` (66 preceding siblings ...)
2021-06-29 2:36 ` [patch 067/192] mm/swapfile: use percpu_ref to serialize against concurrent swapoff Andrew Morton
@ 2021-06-29 2:36 ` Andrew Morton
2021-06-29 2:36 ` [patch 069/192] mm/swap: remove confusing checking for non_swap_entry() in swap_ra_info() Andrew Morton
` (123 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:36 UTC (permalink / raw)
To: akpm, alexs, david, dennis, hannes, hughd, iamjoonsoo.kim,
linmiaohe, linux-mm, mhocko, minchan, mm-commits,
richard.weiyang, shy828301, tim.c.chen, torvalds, willy,
ying.huang, yuzhao
From: Miaohe Lin <linmiaohe@huawei.com>
Subject: swap: fix do_swap_page() race with swapoff
When I was investigating the swap code, I found the below possible race
window:
CPU 1 CPU 2
----- -----
do_swap_page
if (data_race(si->flags & SWP_SYNCHRONOUS_IO)
swap_readpage
if (data_race(sis->flags & SWP_FS_OPS)) {
swapoff
..
p->swap_file = NULL;
..
struct file *swap_file = sis->swap_file;
struct address_space *mapping = swap_file->f_mapping;[oops!]
Note that for the pages that are swapped in through swap cache, this isn't
an issue. Because the page is locked, and the swap entry will be marked
with SWAP_HAS_CACHE, so swapoff() can not proceed until the page has been
unlocked.
Fix this race by using get/put_swap_device() to guard against concurrent
swapoff.
Link: https://lkml.kernel.org/r/20210426123316.806267-3-linmiaohe@huawei.com
Fixes: 0bcac06f27d7 ("mm,swap: skip swapcache for swapin of synchronous device")
Signed-off-by: Miaohe Lin <linmiaohe@huawei.com>
Reviewed-by: "Huang, Ying" <ying.huang@intel.com>
Cc: Alex Shi <alexs@kernel.org>
Cc: David Hildenbrand <david@redhat.com>
Cc: Dennis Zhou <dennis@kernel.org>
Cc: Hugh Dickins <hughd@google.com>
Cc: Johannes Weiner <hannes@cmpxchg.org>
Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Cc: Matthew Wilcox <willy@infradead.org>
Cc: Michal Hocko <mhocko@suse.com>
Cc: Minchan Kim <minchan@kernel.org>
Cc: Tim Chen <tim.c.chen@linux.intel.com>
Cc: Wei Yang <richard.weiyang@gmail.com>
Cc: Yang Shi <shy828301@gmail.com>
Cc: Yu Zhao <yuzhao@google.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/linux/swap.h | 9 +++++++++
mm/memory.c | 11 +++++++++--
2 files changed, 18 insertions(+), 2 deletions(-)
--- a/include/linux/swap.h~swap-fix-do_swap_page-race-with-swapoff
+++ a/include/linux/swap.h
@@ -527,6 +527,15 @@ static inline struct swap_info_struct *s
return NULL;
}
+static inline struct swap_info_struct *get_swap_device(swp_entry_t entry)
+{
+ return NULL;
+}
+
+static inline void put_swap_device(struct swap_info_struct *si)
+{
+}
+
#define swap_address_space(entry) (NULL)
#define get_nr_swap_pages() 0L
#define total_swap_pages 0L
--- a/mm/memory.c~swap-fix-do_swap_page-race-with-swapoff
+++ a/mm/memory.c
@@ -3353,6 +3353,7 @@ vm_fault_t do_swap_page(struct vm_fault
{
struct vm_area_struct *vma = vmf->vma;
struct page *page = NULL, *swapcache;
+ struct swap_info_struct *si = NULL;
swp_entry_t entry;
pte_t pte;
int locked;
@@ -3380,14 +3381,16 @@ vm_fault_t do_swap_page(struct vm_fault
goto out;
}
+ /* Prevent swapoff from happening to us. */
+ si = get_swap_device(entry);
+ if (unlikely(!si))
+ goto out;
delayacct_set_flag(current, DELAYACCT_PF_SWAPIN);
page = lookup_swap_cache(entry, vma, vmf->address);
swapcache = page;
if (!page) {
- struct swap_info_struct *si = swp_swap_info(entry);
-
if (data_race(si->flags & SWP_SYNCHRONOUS_IO) &&
__swap_count(entry) == 1) {
/* skip swapcache */
@@ -3556,6 +3559,8 @@ vm_fault_t do_swap_page(struct vm_fault
unlock:
pte_unmap_unlock(vmf->pte, vmf->ptl);
out:
+ if (si)
+ put_swap_device(si);
return ret;
out_nomap:
pte_unmap_unlock(vmf->pte, vmf->ptl);
@@ -3567,6 +3572,8 @@ out_release:
unlock_page(swapcache);
put_page(swapcache);
}
+ if (si)
+ put_swap_device(si);
return ret;
}
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 069/192] mm/swap: remove confusing checking for non_swap_entry() in swap_ra_info()
2021-06-29 2:32 incoming Andrew Morton
` (67 preceding siblings ...)
2021-06-29 2:36 ` [patch 068/192] swap: fix do_swap_page() race with swapoff Andrew Morton
@ 2021-06-29 2:36 ` Andrew Morton
2021-06-29 2:36 ` [patch 070/192] mm/shmem: fix shmem_swapin() race with swapoff Andrew Morton
` (122 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:36 UTC (permalink / raw)
To: akpm, alexs, david, dennis, hannes, hughd, iamjoonsoo.kim,
linmiaohe, linux-mm, mhocko, minchan, mm-commits,
richard.weiyang, shy828301, tim.c.chen, torvalds, willy,
ying.huang, yuzhao
From: Miaohe Lin <linmiaohe@huawei.com>
Subject: mm/swap: remove confusing checking for non_swap_entry() in swap_ra_info()
The non_swap_entry() was used for working with VMA based swap readahead
via commit ec560175c0b6 ("mm, swap: VMA based swap readahead"). At that
time, the non_swap_entry() checking is necessary because the function is
called before checking that in do_swap_page(). Then it's moved to
swap_ra_info() since commit eaf649ebc3ac ("mm: swap: clean up swap
readahead"). After that, the non_swap_entry() checking is unnecessary,
because swap_ra_info() is called after non_swap_entry() has been checked
already. The resulting code is confusing as the non_swap_entry() check
looks racy now because while we released the pte lock, somebody else might
have faulted in this pte. So we should check whether it's swap pte first
to guard against such race or swap_type will be unexpected. But the race
isn't important because it will not cause problem. We would have enough
checking when we really operate the PTE entries later. So we remove the
non_swap_entry() check here to avoid confusion.
Link: https://lkml.kernel.org/r/20210426123316.806267-4-linmiaohe@huawei.com
Signed-off-by: Miaohe Lin <linmiaohe@huawei.com>
Reviewed-by: "Huang, Ying" <ying.huang@intel.com>
Cc: Alex Shi <alexs@kernel.org>
Cc: David Hildenbrand <david@redhat.com>
Cc: Dennis Zhou <dennis@kernel.org>
Cc: Hugh Dickins <hughd@google.com>
Cc: Johannes Weiner <hannes@cmpxchg.org>
Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Cc: Matthew Wilcox <willy@infradead.org>
Cc: Michal Hocko <mhocko@suse.com>
Cc: Minchan Kim <minchan@kernel.org>
Cc: Tim Chen <tim.c.chen@linux.intel.com>
Cc: Wei Yang <richard.weiyang@gmail.com>
Cc: Yang Shi <shy828301@gmail.com>
Cc: Yu Zhao <yuzhao@google.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/swap_state.c | 6 ------
1 file changed, 6 deletions(-)
--- a/mm/swap_state.c~mm-swap-remove-confusing-checking-for-non_swap_entry-in-swap_ra_info
+++ a/mm/swap_state.c
@@ -721,7 +721,6 @@ static void swap_ra_info(struct vm_fault
{
struct vm_area_struct *vma = vmf->vma;
unsigned long ra_val;
- swp_entry_t entry;
unsigned long faddr, pfn, fpfn;
unsigned long start, end;
pte_t *pte, *orig_pte;
@@ -739,11 +738,6 @@ static void swap_ra_info(struct vm_fault
faddr = vmf->address;
orig_pte = pte = pte_offset_map(vmf->pmd, faddr);
- entry = pte_to_swp_entry(*pte);
- if ((unlikely(non_swap_entry(entry)))) {
- pte_unmap(orig_pte);
- return;
- }
fpfn = PFN_DOWN(faddr);
ra_val = GET_SWAP_RA_VAL(vma);
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 070/192] mm/shmem: fix shmem_swapin() race with swapoff
2021-06-29 2:32 incoming Andrew Morton
` (68 preceding siblings ...)
2021-06-29 2:36 ` [patch 069/192] mm/swap: remove confusing checking for non_swap_entry() in swap_ra_info() Andrew Morton
@ 2021-06-29 2:36 ` Andrew Morton
2021-06-29 2:37 ` [patch 071/192] mm/swapfile: move get_swap_page_of_type() under CONFIG_HIBERNATION Andrew Morton
` (121 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:36 UTC (permalink / raw)
To: akpm, alexs, david, dennis, hannes, hughd, iamjoonsoo.kim,
linmiaohe, linux-mm, mhocko, minchan, mm-commits,
richard.weiyang, shy828301, tim.c.chen, torvalds, willy,
ying.huang, yuzhao
From: Miaohe Lin <linmiaohe@huawei.com>
Subject: mm/shmem: fix shmem_swapin() race with swapoff
When I was investigating the swap code, I found the below possible race
window:
CPU 1 CPU 2
----- -----
shmem_swapin
swap_cluster_readahead
if (likely(si->flags & (SWP_BLKDEV | SWP_FS_OPS))) {
swapoff
..
si->swap_file = NULL;
..
struct inode *inode = si->swap_file->f_mapping->host;[oops!]
Close this race window by using get/put_swap_device() to guard against
concurrent swapoff.
Link: https://lkml.kernel.org/r/20210426123316.806267-5-linmiaohe@huawei.com
Fixes: 8fd2e0b505d1 ("mm: swap: check if swap backing device is congested or not")
Signed-off-by: Miaohe Lin <linmiaohe@huawei.com>
Reviewed-by: "Huang, Ying" <ying.huang@intel.com>
Cc: Dennis Zhou <dennis@kernel.org>
Cc: Tim Chen <tim.c.chen@linux.intel.com>
Cc: Hugh Dickins <hughd@google.com>
Cc: Johannes Weiner <hannes@cmpxchg.org>
Cc: Michal Hocko <mhocko@suse.com>
Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Cc: Alex Shi <alexs@kernel.org>
Cc: Matthew Wilcox <willy@infradead.org>
Cc: Minchan Kim <minchan@kernel.org>
Cc: Wei Yang <richard.weiyang@gmail.com>
Cc: Yang Shi <shy828301@gmail.com>
Cc: David Hildenbrand <david@redhat.com>
Cc: Yu Zhao <yuzhao@google.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/shmem.c | 14 +++++++++++++-
1 file changed, 13 insertions(+), 1 deletion(-)
--- a/mm/shmem.c~mm-shmem-fix-shmem_swapin-race-with-swapoff
+++ a/mm/shmem.c
@@ -1696,7 +1696,8 @@ static int shmem_swapin_page(struct inod
struct address_space *mapping = inode->i_mapping;
struct shmem_inode_info *info = SHMEM_I(inode);
struct mm_struct *charge_mm = vma ? vma->vm_mm : current->mm;
- struct page *page;
+ struct swap_info_struct *si;
+ struct page *page = NULL;
swp_entry_t swap;
int error;
@@ -1704,6 +1705,12 @@ static int shmem_swapin_page(struct inod
swap = radix_to_swp_entry(*pagep);
*pagep = NULL;
+ /* Prevent swapoff from happening to us. */
+ si = get_swap_device(swap);
+ if (!si) {
+ error = EINVAL;
+ goto failed;
+ }
/* Look it up and read it in.. */
page = lookup_swap_cache(swap, NULL, 0);
if (!page) {
@@ -1765,6 +1772,8 @@ static int shmem_swapin_page(struct inod
swap_free(swap);
*pagep = page;
+ if (si)
+ put_swap_device(si);
return 0;
failed:
if (!shmem_confirm_swap(mapping, index, swap))
@@ -1775,6 +1784,9 @@ unlock:
put_page(page);
}
+ if (si)
+ put_swap_device(si);
+
return error;
}
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 071/192] mm/swapfile: move get_swap_page_of_type() under CONFIG_HIBERNATION
2021-06-29 2:32 incoming Andrew Morton
` (69 preceding siblings ...)
2021-06-29 2:36 ` [patch 070/192] mm/shmem: fix shmem_swapin() race with swapoff Andrew Morton
@ 2021-06-29 2:37 ` Andrew Morton
2021-06-29 2:37 ` [patch 072/192] mm/swap: remove unused local variable nr_shadows Andrew Morton
` (120 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:37 UTC (permalink / raw)
To: akpm, hughd, linmiaohe, linux-mm, mm-commits, torvalds, willy
From: Miaohe Lin <linmiaohe@huawei.com>
Subject: mm/swapfile: move get_swap_page_of_type() under CONFIG_HIBERNATION
Patch series "Cleanups for swap", v2.
This series contains just cleanups to remove some unused variables, delete
meaningless forward declarations and so on. More details can be found in
the respective changelogs.
This patch (of 4):
We should move get_swap_page_of_type() under CONFIG_HIBERNATION since the
only caller of this function is now suspend routine.
[linmiaohe@huawei.com: move scan_swap_map() under CONFIG_HIBERNATION]
Link: https://lkml.kernel.org/r/20210521070855.2015094-1-linmiaohe@huawei.com
[linmiaohe@huawei.com: fold scan_swap_map() into the only caller get_swap_page_of_type()]
Link: https://lkml.kernel.org/r/20210527120328.3935132-1-linmiaohe@huawei.com
Link: https://lkml.kernel.org/r/20210520134022.1370406-1-linmiaohe@huawei.com
Link: https://lkml.kernel.org/r/20210520134022.1370406-2-linmiaohe@huawei.com
Signed-off-by: Miaohe Lin <linmiaohe@huawei.com>
Cc: Hugh Dickins <hughd@google.com>
Cc: Matthew Wilcox (Oracle) <willy@infradead.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/swapfile.c | 83 +++++++++++++++++-------------------------------
1 file changed, 31 insertions(+), 52 deletions(-)
--- a/mm/swapfile.c~mm-swapfile-move-get_swap_page_of_type-under-config_hibernation
+++ a/mm/swapfile.c
@@ -453,10 +453,10 @@ static void swap_cluster_schedule_discar
unsigned int idx)
{
/*
- * If scan_swap_map() can't find a free cluster, it will check
+ * If scan_swap_map_slots() can't find a free cluster, it will check
* si->swap_map directly. To make sure the discarding cluster isn't
- * taken by scan_swap_map(), mark the swap entries bad (occupied). It
- * will be cleared after discard
+ * taken by scan_swap_map_slots(), mark the swap entries bad (occupied).
+ * It will be cleared after discard
*/
memset(si->swap_map + idx * SWAPFILE_CLUSTER,
SWAP_MAP_BAD, SWAPFILE_CLUSTER);
@@ -589,7 +589,7 @@ static void dec_cluster_info_page(struct
}
/*
- * It's possible scan_swap_map() uses a free cluster in the middle of free
+ * It's possible scan_swap_map_slots() uses a free cluster in the middle of free
* cluster list. Avoiding such abuse to avoid list corruption.
*/
static bool
@@ -1037,21 +1037,6 @@ static void swap_free_cluster(struct swa
swap_range_free(si, offset, SWAPFILE_CLUSTER);
}
-static unsigned long scan_swap_map(struct swap_info_struct *si,
- unsigned char usage)
-{
- swp_entry_t entry;
- int n_ret;
-
- n_ret = scan_swap_map_slots(si, usage, 1, &entry);
-
- if (n_ret)
- return swp_offset(entry);
- else
- return 0;
-
-}
-
int get_swap_pages(int n_goal, swp_entry_t swp_entries[], int entry_size)
{
unsigned long size = swap_entry_size(entry_size);
@@ -1114,14 +1099,14 @@ start_over:
nextsi:
/*
* if we got here, it's likely that si was almost full before,
- * and since scan_swap_map() can drop the si->lock, multiple
- * callers probably all tried to get a page from the same si
- * and it filled up before we could get one; or, the si filled
- * up between us dropping swap_avail_lock and taking si->lock.
- * Since we dropped the swap_avail_lock, the swap_avail_head
- * list may have been modified; so if next is still in the
- * swap_avail_head list then try it, otherwise start over
- * if we have not gotten any slots.
+ * and since scan_swap_map_slots() can drop the si->lock,
+ * multiple callers probably all tried to get a page from the
+ * same si and it filled up before we could get one; or, the si
+ * filled up between us dropping swap_avail_lock and taking
+ * si->lock. Since we dropped the swap_avail_lock, the
+ * swap_avail_head list may have been modified; so if next is
+ * still in the swap_avail_head list then try it, otherwise
+ * start over if we have not gotten any slots.
*/
if (plist_node_empty(&next->avail_lists[node]))
goto start_over;
@@ -1137,30 +1122,6 @@ noswap:
return n_ret;
}
-/* The only caller of this function is now suspend routine */
-swp_entry_t get_swap_page_of_type(int type)
-{
- struct swap_info_struct *si = swap_type_to_swap_info(type);
- pgoff_t offset;
-
- if (!si)
- goto fail;
-
- spin_lock(&si->lock);
- if (si->flags & SWP_WRITEOK) {
- /* This is called for allocating swap entry, not cache */
- offset = scan_swap_map(si, 1);
- if (offset) {
- atomic_long_dec(&nr_swap_pages);
- spin_unlock(&si->lock);
- return swp_entry(type, offset);
- }
- }
- spin_unlock(&si->lock);
-fail:
- return (swp_entry_t) {0};
-}
-
static struct swap_info_struct *__swap_info_get(swp_entry_t entry)
{
struct swap_info_struct *p;
@@ -1812,6 +1773,24 @@ int free_swap_and_cache(swp_entry_t entr
}
#ifdef CONFIG_HIBERNATION
+
+swp_entry_t get_swap_page_of_type(int type)
+{
+ struct swap_info_struct *si = swap_type_to_swap_info(type);
+ swp_entry_t entry = {0};
+
+ if (!si)
+ goto fail;
+
+ /* This is called for allocating swap entry, not cache */
+ spin_lock(&si->lock);
+ if ((si->flags & SWP_WRITEOK) && scan_swap_map_slots(si, 1, 1, &entry))
+ atomic_long_dec(&nr_swap_pages);
+ spin_unlock(&si->lock);
+fail:
+ return entry;
+}
+
/*
* Find the swap type that corresponds to given device (if any).
*
@@ -2649,7 +2628,7 @@ SYSCALL_DEFINE1(swapoff, const char __us
spin_lock(&p->lock);
drain_mmlist();
- /* wait for anyone still in scan_swap_map */
+ /* wait for anyone still in scan_swap_map_slots */
p->highest_bit = 0; /* cuts scans short */
while (p->flags >= SWP_SCANNING) {
spin_unlock(&p->lock);
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 072/192] mm/swap: remove unused local variable nr_shadows
2021-06-29 2:32 incoming Andrew Morton
` (70 preceding siblings ...)
2021-06-29 2:37 ` [patch 071/192] mm/swapfile: move get_swap_page_of_type() under CONFIG_HIBERNATION Andrew Morton
@ 2021-06-29 2:37 ` Andrew Morton
2021-06-29 2:37 ` [patch 073/192] mm/swap_slots.c: delete meaningless forward declarations Andrew Morton
` (119 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:37 UTC (permalink / raw)
To: akpm, hughd, linmiaohe, linux-mm, mm-commits, torvalds, willy
From: Miaohe Lin <linmiaohe@huawei.com>
Subject: mm/swap: remove unused local variable nr_shadows
Since commit 55c653b71e8c ("mm: stop accounting shadow entries"),
nr_shadows is not used anymore.
Link: https://lkml.kernel.org/r/20210520134022.1370406-3-linmiaohe@huawei.com
Signed-off-by: Miaohe Lin <linmiaohe@huawei.com>
Reviewed-by: Matthew Wilcox (Oracle) <willy@infradead.org>
Cc: Hugh Dickins <hughd@google.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/swap_state.c | 5 -----
1 file changed, 5 deletions(-)
--- a/mm/swap_state.c~mm-swap-remove-unused-local-variable-nr_shadows
+++ a/mm/swap_state.c
@@ -114,8 +114,6 @@ int add_to_swap_cache(struct page *page,
SetPageSwapCache(page);
do {
- unsigned long nr_shadows = 0;
-
xas_lock_irq(&xas);
xas_create_range(&xas);
if (xas_error(&xas))
@@ -124,7 +122,6 @@ int add_to_swap_cache(struct page *page,
VM_BUG_ON_PAGE(xas.xa_index != idx + i, page);
old = xas_load(&xas);
if (xa_is_value(old)) {
- nr_shadows++;
if (shadowp)
*shadowp = old;
}
@@ -260,7 +257,6 @@ void clear_shadow_from_swap_cache(int ty
void *old;
for (;;) {
- unsigned long nr_shadows = 0;
swp_entry_t entry = swp_entry(type, curr);
struct address_space *address_space = swap_address_space(entry);
XA_STATE(xas, &address_space->i_pages, curr);
@@ -270,7 +266,6 @@ void clear_shadow_from_swap_cache(int ty
if (!xa_is_value(old))
continue;
xas_store(&xas, NULL);
- nr_shadows++;
}
xa_unlock_irq(&address_space->i_pages);
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 073/192] mm/swap_slots.c: delete meaningless forward declarations
2021-06-29 2:32 incoming Andrew Morton
` (71 preceding siblings ...)
2021-06-29 2:37 ` [patch 072/192] mm/swap: remove unused local variable nr_shadows Andrew Morton
@ 2021-06-29 2:37 ` Andrew Morton
2021-06-29 2:37 ` [patch 074/192] mm, swap: remove unnecessary smp_rmb() in swap_type_to_swap_info() Andrew Morton
` (118 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:37 UTC (permalink / raw)
To: akpm, hughd, linmiaohe, linux-mm, mm-commits, torvalds, willy
From: Miaohe Lin <linmiaohe@huawei.com>
Subject: mm/swap_slots.c: delete meaningless forward declarations
deactivate_swap_slots_cache() and reactivate_swap_slots_cache() are only
called below their implementations. So these forward declarations are
meaningless and should be removed.
Link: https://lkml.kernel.org/r/20210520134022.1370406-4-linmiaohe@huawei.com
Signed-off-by: Miaohe Lin <linmiaohe@huawei.com>
Cc: Hugh Dickins <hughd@google.com>
Cc: Matthew Wilcox (Oracle) <willy@infradead.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/swap_slots.c | 2 --
1 file changed, 2 deletions(-)
--- a/mm/swap_slots.c~mm-swap_slotsc-delete-meaningless-forward-declarations
+++ a/mm/swap_slots.c
@@ -43,8 +43,6 @@ static DEFINE_MUTEX(swap_slots_cache_mut
static DEFINE_MUTEX(swap_slots_cache_enable_mutex);
static void __drain_swap_slots_cache(unsigned int type);
-static void deactivate_swap_slots_cache(void);
-static void reactivate_swap_slots_cache(void);
#define use_swap_slot_cache (swap_slot_cache_active && swap_slot_cache_enabled)
#define SLOTS_CACHE 0x1
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 074/192] mm, swap: remove unnecessary smp_rmb() in swap_type_to_swap_info()
2021-06-29 2:32 incoming Andrew Morton
` (72 preceding siblings ...)
2021-06-29 2:37 ` [patch 073/192] mm/swap_slots.c: delete meaningless forward declarations Andrew Morton
@ 2021-06-29 2:37 ` Andrew Morton
2021-06-29 2:37 ` [patch 075/192] mm: free idle swap cache page after COW Andrew Morton
` (117 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:37 UTC (permalink / raw)
To: ak, akpm, andrea.parri, dan.carpenter, daniel.m.jordan,
dave.hansen, hughd, linmiaohe, linux-mm, mm-commits, osandov,
paulmck, peterz, tj, torvalds, will.deacon, ying.huang
From: Huang Ying <ying.huang@intel.com>
Subject: mm, swap: remove unnecessary smp_rmb() in swap_type_to_swap_info()
Before commit c10d38cc8d3e ("mm, swap: bounds check swap_info array
accesses to avoid NULL derefs"), the typical code to reference the
swap_info[] is as follows,
type = swp_type(swp_entry);
if (type >= nr_swapfiles)
/* handle invalid swp_entry */;
p = swap_info[type];
/* access fields of *p. OOPS! p may be NULL! */
Because the ordering isn't guaranteed, it's possible that swap_info[type]
is read before "nr_swapfiles". And that may result in NULL pointer
dereference.
So after commit c10d38cc8d3e, the code becomes,
struct swap_info_struct *swap_type_to_swap_info(int type)
{
if (type >= READ_ONCE(nr_swapfiles))
return NULL;
smp_rmb();
return READ_ONCE(swap_info[type]);
}
/* users */
type = swp_type(swp_entry);
p = swap_type_to_swap_info(type);
if (!p)
/* handle invalid swp_entry */;
/* dereference p */
Where the value of swap_info[type] (that is, "p") is checked to be
non-zero before being dereferenced. So, the NULL deferencing becomes
impossible even if "nr_swapfiles" is read after swap_info[type].
Therefore, the "smp_rmb()" becomes unnecessary.
And, we don't even need to read "nr_swapfiles" here. Because the non-zero
checking for "p" is sufficient. We just need to make sure we will not
access out of the boundary of the array. With the change, nr_swapfiles
will only be accessed with swap_lock held, except in
swapcache_free_entries(). Where the absolute correctness of the value
isn't needed, as described in the comments.
We still need to guarantee swap_info[type] is read before being
dereferenced. That can be satisfied via the data dependency ordering
enforced by READ_ONCE(swap_info[type]). This needs to be paired with
proper write barriers. So smp_store_release() is used in
alloc_swap_info() to guarantee the fields of *swap_info[type] is
initialized before swap_info[type] itself being written. Note that the
fields of *swap_info[type] is initialized to be 0 via kvzalloc() firstly.
The assignment and deferencing of swap_info[type] is like
rcu_assign_pointer() and rcu_dereference().
Link: https://lkml.kernel.org/r/20210520073301.1676294-1-ying.huang@intel.com
Signed-off-by: "Huang, Ying" <ying.huang@intel.com>
Cc: Daniel Jordan <daniel.m.jordan@oracle.com>
Cc: Dan Carpenter <dan.carpenter@oracle.com>
Cc: Andrea Parri <andrea.parri@amarulasolutions.com>
Cc: Peter Zijlstra (Intel) <peterz@infradead.org>
Cc: Andi Kleen <ak@linux.intel.com>
Cc: Dave Hansen <dave.hansen@linux.intel.com>
Cc: Omar Sandoval <osandov@fb.com>
Cc: Paul McKenney <paulmck@kernel.org>
Cc: Tejun Heo <tj@kernel.org>
Cc: Will Deacon <will.deacon@arm.com>
Cc: Miaohe Lin <linmiaohe@huawei.com>
Cc: Hugh Dickins <hughd@google.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/swapfile.c | 15 ++++++---------
1 file changed, 6 insertions(+), 9 deletions(-)
--- a/mm/swapfile.c~mm-swap-remove-unnecessary-smp_rmb-in-swap_type_to_swap_info
+++ a/mm/swapfile.c
@@ -100,11 +100,10 @@ atomic_t nr_rotate_swap = ATOMIC_INIT(0)
static struct swap_info_struct *swap_type_to_swap_info(int type)
{
- if (type >= READ_ONCE(nr_swapfiles))
+ if (type >= MAX_SWAPFILES)
return NULL;
- smp_rmb(); /* Pairs with smp_wmb in alloc_swap_info. */
- return READ_ONCE(swap_info[type]);
+ return READ_ONCE(swap_info[type]); /* rcu_dereference() */
}
static inline unsigned char swap_count(unsigned char ent)
@@ -2863,14 +2862,12 @@ static struct swap_info_struct *alloc_sw
}
if (type >= nr_swapfiles) {
p->type = type;
- WRITE_ONCE(swap_info[type], p);
/*
- * Write swap_info[type] before nr_swapfiles, in case a
- * racing procfs swap_start() or swap_next() is reading them.
- * (We never shrink nr_swapfiles, we never free this entry.)
+ * Publish the swap_info_struct after initializing it.
+ * Note that kvzalloc() above zeroes all its fields.
*/
- smp_wmb();
- WRITE_ONCE(nr_swapfiles, nr_swapfiles + 1);
+ smp_store_release(&swap_info[type], p); /* rcu_assign_pointer() */
+ nr_swapfiles++;
} else {
defer = p;
p = swap_info[type];
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 075/192] mm: free idle swap cache page after COW
2021-06-29 2:32 incoming Andrew Morton
` (73 preceding siblings ...)
2021-06-29 2:37 ` [patch 074/192] mm, swap: remove unnecessary smp_rmb() in swap_type_to_swap_info() Andrew Morton
@ 2021-06-29 2:37 ` Andrew Morton
2021-06-29 2:37 ` [patch 076/192] swap: check mapping_empty() for swap cache before being freed Andrew Morton
` (116 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:37 UTC (permalink / raw)
To: aarcange, akpm, dave.hansen, hannes, hughd, linux-mm, mgorman,
mhocko, mm-commits, peterx, riel, tim.c.chen, torvalds, willy,
ying.huang
From: Huang Ying <ying.huang@intel.com>
Subject: mm: free idle swap cache page after COW
With commit 09854ba94c6a ("mm: do_wp_page() simplification"), after COW,
the idle swap cache page (neither the page nor the corresponding swap
entry is mapped by any process) will be left in the LRU list, even if it's
in the active list or the head of the inactive list. So, the page
reclaimer may take quite some overhead to reclaim these actually unused
pages.
To help the page reclaiming, in this patch, after COW, the idle swap cache
page will be tried to be freed. To avoid to introduce much overhead to
the hot COW code path,
a) there's almost zero overhead for non-swap case via checking
PageSwapCache() firstly.
b) the page lock is acquired via trylock only.
To test the patch, we used pmbench memory accessing benchmark with
working-set larger than available memory on a 2-socket Intel server with a
NVMe SSD as swap device. Test results shows that the pmbench score
increases up to 23.8% with the decreased size of swap cache and swapin
throughput.
Link: https://lkml.kernel.org/r/20210601053143.1380078-1-ying.huang@intel.com
Signed-off-by: "Huang, Ying" <ying.huang@intel.com>
Suggested-by: Johannes Weiner <hannes@cmpxchg.org> [use free_swap_cache()]
Acked-by: Johannes Weiner <hannes@cmpxchg.org>
Cc: Hugh Dickins <hughd@google.com>
Cc: Matthew Wilcox <willy@infradead.org>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Peter Xu <peterx@redhat.com>
Cc: Hugh Dickins <hughd@google.com>
Cc: Mel Gorman <mgorman@suse.de>
Cc: Rik van Riel <riel@surriel.com>
Cc: Andrea Arcangeli <aarcange@redhat.com>
Cc: Michal Hocko <mhocko@kernel.org>
Cc: Dave Hansen <dave.hansen@intel.com>
Cc: Tim Chen <tim.c.chen@intel.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/linux/swap.h | 5 +++++
mm/memory.c | 2 ++
mm/swap_state.c | 2 +-
3 files changed, 8 insertions(+), 1 deletion(-)
--- a/include/linux/swap.h~mm-free-idle-swap-cache-page-after-cow
+++ a/include/linux/swap.h
@@ -446,6 +446,7 @@ extern void __delete_from_swap_cache(str
extern void delete_from_swap_cache(struct page *);
extern void clear_shadow_from_swap_cache(int type, unsigned long begin,
unsigned long end);
+extern void free_swap_cache(struct page *);
extern void free_page_and_swap_cache(struct page *);
extern void free_pages_and_swap_cache(struct page **, int);
extern struct page *lookup_swap_cache(swp_entry_t entry,
@@ -551,6 +552,10 @@ static inline void put_swap_device(struc
#define free_pages_and_swap_cache(pages, nr) \
release_pages((pages), (nr));
+static inline void free_swap_cache(struct page *page)
+{
+}
+
static inline void show_swap_cache_info(void)
{
}
--- a/mm/memory.c~mm-free-idle-swap-cache-page-after-cow
+++ a/mm/memory.c
@@ -3023,6 +3023,8 @@ static vm_fault_t wp_page_copy(struct vm
munlock_vma_page(old_page);
unlock_page(old_page);
}
+ if (page_copied)
+ free_swap_cache(old_page);
put_page(old_page);
}
return page_copied ? VM_FAULT_WRITE : 0;
--- a/mm/swap_state.c~mm-free-idle-swap-cache-page-after-cow
+++ a/mm/swap_state.c
@@ -286,7 +286,7 @@ void clear_shadow_from_swap_cache(int ty
* try_to_free_swap() _with_ the lock.
* - Marcelo
*/
-static inline void free_swap_cache(struct page *page)
+void free_swap_cache(struct page *page)
{
if (PageSwapCache(page) && !page_mapped(page) && trylock_page(page)) {
try_to_free_swap(page);
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 076/192] swap: check mapping_empty() for swap cache before being freed
2021-06-29 2:32 incoming Andrew Morton
` (74 preceding siblings ...)
2021-06-29 2:37 ` [patch 075/192] mm: free idle swap cache page after COW Andrew Morton
@ 2021-06-29 2:37 ` Andrew Morton
2021-06-29 2:37 ` [patch 077/192] mm/memcg: move mod_objcg_state() to memcontrol.c Andrew Morton
` (115 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:37 UTC (permalink / raw)
To: akpm, dan.j.williams, hannes, hch, hughd, iamjoonsoo.kim,
idryomov, linmiaohe, linux-mm, mgorman, mhocko, minchan,
mm-commits, torvalds, vbabka, willy, ying.huang
From: Huang Ying <ying.huang@intel.com>
Subject: swap: check mapping_empty() for swap cache before being freed
To check whether all pages and shadow entries in swap cache has been
removed before swap cache is freed.
Link: https://lkml.kernel.org/r/20210608005121.511140-1-ying.huang@intel.com
Signed-off-by: "Huang, Ying" <ying.huang@intel.com>
Cc: Miaohe Lin <linmiaohe@huawei.com>
Cc: Matthew Wilcox <willy@infradead.org>
Cc: Minchan Kim <minchan@kernel.org>
Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Cc: Johannes Weiner <hannes@cmpxchg.org>
Cc: Vlastimil Babka <vbabka@suse.cz>
Cc: Hugh Dickins <hughd@google.com>
Cc: Mel Gorman <mgorman@techsingularity.net>
Cc: Michal Hocko <mhocko@kernel.org>
Cc: Dan Williams <dan.j.williams@intel.com>
Cc: Christoph Hellwig <hch@lst.de>
Cc: Ilya Dryomov <idryomov@gmail.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/swap_state.c | 7 ++++++-
1 file changed, 6 insertions(+), 1 deletion(-)
--- a/mm/swap_state.c~swap-check-mapping_empty-for-swap-cache-before-being-freed
+++ a/mm/swap_state.c
@@ -693,7 +693,12 @@ int init_swap_address_space(unsigned int
void exit_swap_address_space(unsigned int type)
{
- kvfree(swapper_spaces[type]);
+ int i;
+ struct address_space *spaces = swapper_spaces[type];
+
+ for (i = 0; i < nr_swapper_spaces[type]; i++)
+ VM_WARN_ON_ONCE(!mapping_empty(&spaces[i]));
+ kvfree(spaces);
nr_swapper_spaces[type] = 0;
swapper_spaces[type] = NULL;
}
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 077/192] mm/memcg: move mod_objcg_state() to memcontrol.c
2021-06-29 2:32 incoming Andrew Morton
` (75 preceding siblings ...)
2021-06-29 2:37 ` [patch 076/192] swap: check mapping_empty() for swap cache before being freed Andrew Morton
@ 2021-06-29 2:37 ` Andrew Morton
2021-06-29 2:37 ` [patch 078/192] mm/memcg: cache vmstat data in percpu memcg_stock_pcp Andrew Morton
` (114 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:37 UTC (permalink / raw)
To: akpm, alex.shi, chris, cl, guro, hannes, iamjoonsoo.kim,
laoar.shao, linux-mm, longman, mhocko, mm-commits, msys.mizuma,
penberg, richard.weiyang, rientjes, shakeelb, songmuchun, tj,
torvalds, vbabka, vdavydov.dev, willy, zhengjun.xing
From: Waiman Long <longman@redhat.com>
Subject: mm/memcg: move mod_objcg_state() to memcontrol.c
Patch series "mm/memcg: Reduce kmemcache memory accounting overhead", v6.
With the recent introduction of the new slab memory controller, we
eliminate the need for having separate kmemcaches for each memory cgroup
and reduce overall kernel memory usage. However, we also add additional
memory accounting overhead to each call of kmem_cache_alloc() and
kmem_cache_free().
For workloads that require a lot of kmemcache allocations and
de-allocations, they may experience performance regression as illustrated
in [1] and [2].
A simple kernel module that performs repeated loop of 100,000,000
kmem_cache_alloc() and kmem_cache_free() of either a small 32-byte object
or a big 4k object at module init time with a batch size of 4 (4 kmalloc's
followed by 4 kfree's) is used for benchmarking. The benchmarking tool
was run on a kernel based on linux-next-20210419. The test was run on a
CascadeLake server with turbo-boosting disable to reduce run-to-run
variation.
The small object test exercises mainly the object stock charging and
vmstat update code paths. The large object test also exercises the
refill_obj_stock() and __memcg_kmem_charge()/__memcg_kmem_uncharge() code
paths.
With memory accounting disabled, the run time was 3.130s with both small
object big object tests.
With memory accounting enabled, both cgroup v1 and v2 showed similar
results in the small object test. The performance results of the large
object test, however, differed between cgroup v1 and v2.
The execution times with the application of various patches in the
patchset were:
Applied patches Run time Accounting overhead %age 1 %age 2
--------------- -------- ------------------- ------ ------
Small 32-byte object:
None 11.634s 8.504s 100.0% 271.7%
1-2 9.425s 6.295s 74.0% 201.1%
1-3 9.708s 6.578s 77.4% 210.2%
1-4 8.062s 4.932s 58.0% 157.6%
Large 4k object (v2):
None 22.107s 18.977s 100.0% 606.3%
1-2 20.960s 17.830s 94.0% 569.6%
1-3 14.238s 11.108s 58.5% 354.9%
1-4 11.329s 8.199s 43.2% 261.9%
Large 4k object (v1):
None 36.807s 33.677s 100.0% 1075.9%
1-2 36.648s 33.518s 99.5% 1070.9%
1-3 22.345s 19.215s 57.1% 613.9%
1-4 18.662s 15.532s 46.1% 496.2%
N.B. %age 1 = overhead/unpatched overhead
%age 2 = overhead/accounting disabled time
Patch 2 (vmstat data stock caching) helps in both the small object test
and the large v2 object test. It doesn't help much in v1 big object test.
Patch 3 (refill_obj_stock improvement) does help the small object test
but offer significant performance improvement for the large object test
(both v1 and v2).
Patch 4 (eliminating irq disable/enable) helps in all test cases.
To test for the extreme case, a multi-threaded kmalloc/kfree
microbenchmark was run on the 2-socket 48-core 96-thread system with
96 testing threads in the same memcg doing kmalloc+kfree of a 4k object
with accounting enabled for 10s. The total number of kmalloc+kfree done
in kilo operations per second (kops/s) were as follows:
Applied patches v1 kops/s v1 change v2 kops/s v2 change
--------------- --------- --------- --------- ---------
None 3,520 1.00X 6,242 1.00X
1-2 4,304 1.22X 8,478 1.36X
1-3 4,731 1.34X 418,142 66.99X
1-4 4,587 1.30X 438,838 70.30X
With memory accounting disabled, the kmalloc/kfree rate was 1,481,291
kop/s. This test shows how significant the memory accouting overhead
can be in some extreme situations.
For this multithreaded test, the improvement from patch 2 mainly
comes from the conditional atomic xchg of objcg->nr_charged_bytes in
mod_objcg_state(). By using an unconditional xchg, the operation rates
were similar to the unpatched kernel.
Patch 3 elminates the single highly contended cacheline of
objcg->nr_charged_bytes for cgroup v2 leading to a huge performance
improvement. Cgroup v1, however, still has another highly contended
cacheline in the shared page counter &memcg->kmem. So the improvement
is only modest.
Patch 4 helps in cgroup v2, but performs worse in cgroup v1 as
eliminating the irq_disable/irq_enable overhead seems to aggravate the
cacheline contention.
[1] https://lore.kernel.org/linux-mm/20210408193948.vfktg3azh2wrt56t@gabell/T/#u
[2] https://lore.kernel.org/lkml/20210114025151.GA22932@xsang-OptiPlex-9020/
This patch (of 4):
mod_objcg_state() is moved from mm/slab.h to mm/memcontrol.c so that
further optimization can be done to it in later patches without exposing
unnecessary details to other mm components.
Link: https://lkml.kernel.org/r/20210506150007.16288-1-longman@redhat.com
Link: https://lkml.kernel.org/r/20210506150007.16288-2-longman@redhat.com
Signed-off-by: Waiman Long <longman@redhat.com>
Acked-by: Johannes Weiner <hannes@cmpxchg.org>
Reviewed-by: Shakeel Butt <shakeelb@google.com>
Acked-by: Roman Gushchin <guro@fb.com>
Cc: Alex Shi <alex.shi@linux.alibaba.com>
Cc: Chris Down <chris@chrisdown.name>
Cc: Christoph Lameter <cl@linux.com>
Cc: David Rientjes <rientjes@google.com>
Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Cc: Masayoshi Mizuma <msys.mizuma@gmail.com>
Cc: Matthew Wilcox <willy@infradead.org>
Cc: Michal Hocko <mhocko@kernel.org>
Cc: Muchun Song <songmuchun@bytedance.com>
Cc: Pekka Enberg <penberg@kernel.org>
Cc: Tejun Heo <tj@kernel.org>
Cc: Vladimir Davydov <vdavydov.dev@gmail.com>
Cc: Vlastimil Babka <vbabka@suse.cz>
Cc: Wei Yang <richard.weiyang@gmail.com>
Cc: Xing Zhengjun <zhengjun.xing@linux.intel.com>
Cc: Yafang Shao <laoar.shao@gmail.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/memcontrol.c | 13 +++++++++++++
mm/slab.h | 16 ++--------------
2 files changed, 15 insertions(+), 14 deletions(-)
--- a/mm/memcontrol.c~mm-memcg-move-mod_objcg_state-to-memcontrolc
+++ a/mm/memcontrol.c
@@ -782,6 +782,19 @@ void __mod_lruvec_kmem_state(void *p, en
rcu_read_unlock();
}
+void mod_objcg_state(struct obj_cgroup *objcg, struct pglist_data *pgdat,
+ enum node_stat_item idx, int nr)
+{
+ struct mem_cgroup *memcg;
+ struct lruvec *lruvec;
+
+ rcu_read_lock();
+ memcg = obj_cgroup_memcg(objcg);
+ lruvec = mem_cgroup_lruvec(memcg, pgdat);
+ mod_memcg_lruvec_state(lruvec, idx, nr);
+ rcu_read_unlock();
+}
+
/**
* __count_memcg_events - account VM events in a cgroup
* @memcg: the memory cgroup
--- a/mm/slab.h~mm-memcg-move-mod_objcg_state-to-memcontrolc
+++ a/mm/slab.h
@@ -240,6 +240,8 @@ static inline bool kmem_cache_debug_flag
#ifdef CONFIG_MEMCG_KMEM
int memcg_alloc_page_obj_cgroups(struct page *page, struct kmem_cache *s,
gfp_t gfp, bool new_page);
+void mod_objcg_state(struct obj_cgroup *objcg, struct pglist_data *pgdat,
+ enum node_stat_item idx, int nr);
static inline void memcg_free_page_obj_cgroups(struct page *page)
{
@@ -284,20 +286,6 @@ static inline bool memcg_slab_pre_alloc_
return true;
}
-static inline void mod_objcg_state(struct obj_cgroup *objcg,
- struct pglist_data *pgdat,
- enum node_stat_item idx, int nr)
-{
- struct mem_cgroup *memcg;
- struct lruvec *lruvec;
-
- rcu_read_lock();
- memcg = obj_cgroup_memcg(objcg);
- lruvec = mem_cgroup_lruvec(memcg, pgdat);
- mod_memcg_lruvec_state(lruvec, idx, nr);
- rcu_read_unlock();
-}
-
static inline void memcg_slab_post_alloc_hook(struct kmem_cache *s,
struct obj_cgroup *objcg,
gfp_t flags, size_t size,
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 078/192] mm/memcg: cache vmstat data in percpu memcg_stock_pcp
2021-06-29 2:32 incoming Andrew Morton
` (76 preceding siblings ...)
2021-06-29 2:37 ` [patch 077/192] mm/memcg: move mod_objcg_state() to memcontrol.c Andrew Morton
@ 2021-06-29 2:37 ` Andrew Morton
2021-06-29 2:37 ` [patch 079/192] mm/memcg: improve refill_obj_stock() performance Andrew Morton
` (113 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:37 UTC (permalink / raw)
To: akpm, alex.shi, chris, cl, guro, hannes, iamjoonsoo.kim,
laoar.shao, linux-mm, longman, mhocko, mm-commits, msys.mizuma,
penberg, richard.weiyang, rientjes, shakeelb, songmuchun, tj,
torvalds, vbabka, vdavydov.dev, willy, zhengjun.xing
From: Waiman Long <longman@redhat.com>
Subject: mm/memcg: cache vmstat data in percpu memcg_stock_pcp
Before the new slab memory controller with per object byte charging,
charging and vmstat data update happen only when new slab pages are
allocated or freed. Now they are done with every kmem_cache_alloc() and
kmem_cache_free(). This causes additional overhead for workloads that
generate a lot of alloc and free calls.
The memcg_stock_pcp is used to cache byte charge for a specific obj_cgroup
to reduce that overhead. To further reducing it, this patch makes the
vmstat data cached in the memcg_stock_pcp structure as well until it
accumulates a page size worth of update or when other cached data change.
Caching the vmstat data in the per-cpu stock eliminates two writes to
non-hot cachelines for memcg specific as well as memcg-lruvecs specific
vmstat data by a write to a hot local stock cacheline.
On a 2-socket Cascade Lake server with instrumentation enabled and this
patch applied, it was found that about 20% (634400 out of 3243830) of the
time when mod_objcg_state() is called leads to an actual call to
__mod_objcg_state() after initial boot. When doing parallel kernel build,
the figure was about 17% (24329265 out of 142512465). So caching the
vmstat data reduces the number of calls to __mod_objcg_state() by more
than 80%.
Link: https://lkml.kernel.org/r/20210506150007.16288-3-longman@redhat.com
Signed-off-by: Waiman Long <longman@redhat.com>
Reviewed-by: Shakeel Butt <shakeelb@google.com>
Cc: Alex Shi <alex.shi@linux.alibaba.com>
Cc: Chris Down <chris@chrisdown.name>
Cc: Christoph Lameter <cl@linux.com>
Cc: David Rientjes <rientjes@google.com>
Cc: Johannes Weiner <hannes@cmpxchg.org>
Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Cc: Masayoshi Mizuma <msys.mizuma@gmail.com>
Cc: Matthew Wilcox <willy@infradead.org>
Cc: Michal Hocko <mhocko@kernel.org>
Cc: Muchun Song <songmuchun@bytedance.com>
Cc: Pekka Enberg <penberg@kernel.org>
Cc: Roman Gushchin <guro@fb.com>
Cc: Tejun Heo <tj@kernel.org>
Cc: Vladimir Davydov <vdavydov.dev@gmail.com>
Cc: Vlastimil Babka <vbabka@suse.cz>
Cc: Wei Yang <richard.weiyang@gmail.com>
Cc: Xing Zhengjun <zhengjun.xing@linux.intel.com>
Cc: Yafang Shao <laoar.shao@gmail.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/memcontrol.c | 90 ++++++++++++++++++++++++++++++++++++++++++++--
1 file changed, 87 insertions(+), 3 deletions(-)
--- a/mm/memcontrol.c~mm-memcg-cache-vmstat-data-in-percpu-memcg_stock_pcp
+++ a/mm/memcontrol.c
@@ -782,8 +782,9 @@ void __mod_lruvec_kmem_state(void *p, en
rcu_read_unlock();
}
-void mod_objcg_state(struct obj_cgroup *objcg, struct pglist_data *pgdat,
- enum node_stat_item idx, int nr)
+static inline void mod_objcg_mlstate(struct obj_cgroup *objcg,
+ struct pglist_data *pgdat,
+ enum node_stat_item idx, int nr)
{
struct mem_cgroup *memcg;
struct lruvec *lruvec;
@@ -791,7 +792,7 @@ void mod_objcg_state(struct obj_cgroup *
rcu_read_lock();
memcg = obj_cgroup_memcg(objcg);
lruvec = mem_cgroup_lruvec(memcg, pgdat);
- mod_memcg_lruvec_state(lruvec, idx, nr);
+ __mod_memcg_lruvec_state(lruvec, idx, nr);
rcu_read_unlock();
}
@@ -2059,7 +2060,10 @@ struct memcg_stock_pcp {
#ifdef CONFIG_MEMCG_KMEM
struct obj_cgroup *cached_objcg;
+ struct pglist_data *cached_pgdat;
unsigned int nr_bytes;
+ int nr_slab_reclaimable_b;
+ int nr_slab_unreclaimable_b;
#endif
struct work_struct work;
@@ -3008,6 +3012,67 @@ void __memcg_kmem_uncharge_page(struct p
obj_cgroup_put(objcg);
}
+void mod_objcg_state(struct obj_cgroup *objcg, struct pglist_data *pgdat,
+ enum node_stat_item idx, int nr)
+{
+ struct memcg_stock_pcp *stock;
+ unsigned long flags;
+ int *bytes;
+
+ local_irq_save(flags);
+ stock = this_cpu_ptr(&memcg_stock);
+
+ /*
+ * Save vmstat data in stock and skip vmstat array update unless
+ * accumulating over a page of vmstat data or when pgdat or idx
+ * changes.
+ */
+ if (stock->cached_objcg != objcg) {
+ drain_obj_stock(stock);
+ obj_cgroup_get(objcg);
+ stock->nr_bytes = atomic_read(&objcg->nr_charged_bytes)
+ ? atomic_xchg(&objcg->nr_charged_bytes, 0) : 0;
+ stock->cached_objcg = objcg;
+ stock->cached_pgdat = pgdat;
+ } else if (stock->cached_pgdat != pgdat) {
+ /* Flush the existing cached vmstat data */
+ if (stock->nr_slab_reclaimable_b) {
+ mod_objcg_mlstate(objcg, pgdat, NR_SLAB_RECLAIMABLE_B,
+ stock->nr_slab_reclaimable_b);
+ stock->nr_slab_reclaimable_b = 0;
+ }
+ if (stock->nr_slab_unreclaimable_b) {
+ mod_objcg_mlstate(objcg, pgdat, NR_SLAB_UNRECLAIMABLE_B,
+ stock->nr_slab_unreclaimable_b);
+ stock->nr_slab_unreclaimable_b = 0;
+ }
+ stock->cached_pgdat = pgdat;
+ }
+
+ bytes = (idx == NR_SLAB_RECLAIMABLE_B) ? &stock->nr_slab_reclaimable_b
+ : &stock->nr_slab_unreclaimable_b;
+ /*
+ * Even for large object >= PAGE_SIZE, the vmstat data will still be
+ * cached locally at least once before pushing it out.
+ */
+ if (!*bytes) {
+ *bytes = nr;
+ nr = 0;
+ } else {
+ *bytes += nr;
+ if (abs(*bytes) > PAGE_SIZE) {
+ nr = *bytes;
+ *bytes = 0;
+ } else {
+ nr = 0;
+ }
+ }
+ if (nr)
+ mod_objcg_mlstate(objcg, pgdat, idx, nr);
+
+ local_irq_restore(flags);
+}
+
static bool consume_obj_stock(struct obj_cgroup *objcg, unsigned int nr_bytes)
{
struct memcg_stock_pcp *stock;
@@ -3055,6 +3120,25 @@ static void drain_obj_stock(struct memcg
stock->nr_bytes = 0;
}
+ /*
+ * Flush the vmstat data in current stock
+ */
+ if (stock->nr_slab_reclaimable_b || stock->nr_slab_unreclaimable_b) {
+ if (stock->nr_slab_reclaimable_b) {
+ mod_objcg_mlstate(old, stock->cached_pgdat,
+ NR_SLAB_RECLAIMABLE_B,
+ stock->nr_slab_reclaimable_b);
+ stock->nr_slab_reclaimable_b = 0;
+ }
+ if (stock->nr_slab_unreclaimable_b) {
+ mod_objcg_mlstate(old, stock->cached_pgdat,
+ NR_SLAB_UNRECLAIMABLE_B,
+ stock->nr_slab_unreclaimable_b);
+ stock->nr_slab_unreclaimable_b = 0;
+ }
+ stock->cached_pgdat = NULL;
+ }
+
obj_cgroup_put(old);
stock->cached_objcg = NULL;
}
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 079/192] mm/memcg: improve refill_obj_stock() performance
2021-06-29 2:32 incoming Andrew Morton
` (77 preceding siblings ...)
2021-06-29 2:37 ` [patch 078/192] mm/memcg: cache vmstat data in percpu memcg_stock_pcp Andrew Morton
@ 2021-06-29 2:37 ` Andrew Morton
2021-06-29 2:37 ` [patch 080/192] mm/memcg: optimize user context object stock access Andrew Morton
` (112 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:37 UTC (permalink / raw)
To: akpm, alex.shi, chris, cl, guro, hannes, iamjoonsoo.kim,
laoar.shao, linux-mm, longman, mhocko, mm-commits, msys.mizuma,
penberg, richard.weiyang, rientjes, shakeelb, songmuchun, tj,
torvalds, vbabka, vdavydov.dev, willy, zhengjun.xing
From: Waiman Long <longman@redhat.com>
Subject: mm/memcg: improve refill_obj_stock() performance
There are two issues with the current refill_obj_stock() code. First of
all, when nr_bytes reaches over PAGE_SIZE, it calls drain_obj_stock() to
atomically flush out remaining bytes to obj_cgroup, clear cached_objcg and
do a obj_cgroup_put(). It is likely that the same obj_cgroup will be used
again which leads to another call to drain_obj_stock() and
obj_cgroup_get() as well as atomically retrieve the available byte from
obj_cgroup. That is costly. Instead, we should just uncharge the excess
pages, reduce the stock bytes and be done with it. The drain_obj_stock()
function should only be called when obj_cgroup changes.
Secondly, when charging an object of size not less than a page in
obj_cgroup_charge(), it is possible that the remaining bytes to be
refilled to the stock will overflow a page and cause refill_obj_stock() to
uncharge 1 page. To avoid the additional uncharge in this case, a new
allow_uncharge flag is added to refill_obj_stock() which will be set to
false when called from obj_cgroup_charge() so that an uncharge_pages()
call won't be issued right after a charge_pages() call unless the objcg
changes.
A multithreaded kmalloc+kfree microbenchmark on a 2-socket 48-core
96-thread x86-64 system with 96 testing threads were run. Before this
patch, the total number of kilo kmalloc+kfree operations done for a 4k
large object by all the testing threads per second were 4,304 kops/s
(cgroup v1) and 8,478 kops/s (cgroup v2). After applying this patch, the
number were 4,731 (cgroup v1) and 418,142 (cgroup v2) respectively. This
represents a performance improvement of 1.10X (cgroup v1) and 49.3X
(cgroup v2).
Link: https://lkml.kernel.org/r/20210506150007.16288-4-longman@redhat.com
Signed-off-by: Waiman Long <longman@redhat.com>
Reviewed-by: Shakeel Butt <shakeelb@google.com>
Cc: Alex Shi <alex.shi@linux.alibaba.com>
Cc: Chris Down <chris@chrisdown.name>
Cc: Christoph Lameter <cl@linux.com>
Cc: David Rientjes <rientjes@google.com>
Cc: Johannes Weiner <hannes@cmpxchg.org>
Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Cc: Masayoshi Mizuma <msys.mizuma@gmail.com>
Cc: Matthew Wilcox <willy@infradead.org>
Cc: Michal Hocko <mhocko@kernel.org>
Cc: Muchun Song <songmuchun@bytedance.com>
Cc: Pekka Enberg <penberg@kernel.org>
Cc: Roman Gushchin <guro@fb.com>
Cc: Tejun Heo <tj@kernel.org>
Cc: Vladimir Davydov <vdavydov.dev@gmail.com>
Cc: Vlastimil Babka <vbabka@suse.cz>
Cc: Wei Yang <richard.weiyang@gmail.com>
Cc: Xing Zhengjun <zhengjun.xing@linux.intel.com>
Cc: Yafang Shao <laoar.shao@gmail.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/memcontrol.c | 48 +++++++++++++++++++++++++++++++++-------------
1 file changed, 35 insertions(+), 13 deletions(-)
--- a/mm/memcontrol.c~mm-memcg-improve-refill_obj_stock-performance
+++ a/mm/memcontrol.c
@@ -3157,10 +3157,12 @@ static bool obj_stock_flush_required(str
return false;
}
-static void refill_obj_stock(struct obj_cgroup *objcg, unsigned int nr_bytes)
+static void refill_obj_stock(struct obj_cgroup *objcg, unsigned int nr_bytes,
+ bool allow_uncharge)
{
struct memcg_stock_pcp *stock;
unsigned long flags;
+ unsigned int nr_pages = 0;
local_irq_save(flags);
@@ -3169,14 +3171,21 @@ static void refill_obj_stock(struct obj_
drain_obj_stock(stock);
obj_cgroup_get(objcg);
stock->cached_objcg = objcg;
- stock->nr_bytes = atomic_xchg(&objcg->nr_charged_bytes, 0);
+ stock->nr_bytes = atomic_read(&objcg->nr_charged_bytes)
+ ? atomic_xchg(&objcg->nr_charged_bytes, 0) : 0;
+ allow_uncharge = true; /* Allow uncharge when objcg changes */
}
stock->nr_bytes += nr_bytes;
- if (stock->nr_bytes > PAGE_SIZE)
- drain_obj_stock(stock);
+ if (allow_uncharge && (stock->nr_bytes > PAGE_SIZE)) {
+ nr_pages = stock->nr_bytes >> PAGE_SHIFT;
+ stock->nr_bytes &= (PAGE_SIZE - 1);
+ }
local_irq_restore(flags);
+
+ if (nr_pages)
+ obj_cgroup_uncharge_pages(objcg, nr_pages);
}
int obj_cgroup_charge(struct obj_cgroup *objcg, gfp_t gfp, size_t size)
@@ -3188,14 +3197,27 @@ int obj_cgroup_charge(struct obj_cgroup
return 0;
/*
- * In theory, memcg->nr_charged_bytes can have enough
+ * In theory, objcg->nr_charged_bytes can have enough
* pre-charged bytes to satisfy the allocation. However,
- * flushing memcg->nr_charged_bytes requires two atomic
- * operations, and memcg->nr_charged_bytes can't be big,
- * so it's better to ignore it and try grab some new pages.
- * memcg->nr_charged_bytes will be flushed in
- * refill_obj_stock(), called from this function or
- * independently later.
+ * flushing objcg->nr_charged_bytes requires two atomic
+ * operations, and objcg->nr_charged_bytes can't be big.
+ * The shared objcg->nr_charged_bytes can also become a
+ * performance bottleneck if all tasks of the same memcg are
+ * trying to update it. So it's better to ignore it and try
+ * grab some new pages. The stock's nr_bytes will be flushed to
+ * objcg->nr_charged_bytes later on when objcg changes.
+ *
+ * The stock's nr_bytes may contain enough pre-charged bytes
+ * to allow one less page from being charged, but we can't rely
+ * on the pre-charged bytes not being changed outside of
+ * consume_obj_stock() or refill_obj_stock(). So ignore those
+ * pre-charged bytes as well when charging pages. To avoid a
+ * page uncharge right after a page charge, we set the
+ * allow_uncharge flag to false when calling refill_obj_stock()
+ * to temporarily allow the pre-charged bytes to exceed the page
+ * size limit. The maximum reachable value of the pre-charged
+ * bytes is (sizeof(object) + PAGE_SIZE - 2) if there is no data
+ * race.
*/
nr_pages = size >> PAGE_SHIFT;
nr_bytes = size & (PAGE_SIZE - 1);
@@ -3205,14 +3227,14 @@ int obj_cgroup_charge(struct obj_cgroup
ret = obj_cgroup_charge_pages(objcg, gfp, nr_pages);
if (!ret && nr_bytes)
- refill_obj_stock(objcg, PAGE_SIZE - nr_bytes);
+ refill_obj_stock(objcg, PAGE_SIZE - nr_bytes, false);
return ret;
}
void obj_cgroup_uncharge(struct obj_cgroup *objcg, size_t size)
{
- refill_obj_stock(objcg, size);
+ refill_obj_stock(objcg, size, true);
}
#endif /* CONFIG_MEMCG_KMEM */
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 080/192] mm/memcg: optimize user context object stock access
2021-06-29 2:32 incoming Andrew Morton
` (78 preceding siblings ...)
2021-06-29 2:37 ` [patch 079/192] mm/memcg: improve refill_obj_stock() performance Andrew Morton
@ 2021-06-29 2:37 ` Andrew Morton
2021-06-29 2:37 ` [patch 081/192] mm: memcg/slab: properly set up gfp flags for objcg pointer array Andrew Morton
` (111 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:37 UTC (permalink / raw)
To: akpm, alex.shi, chris, cl, guro, hannes, iamjoonsoo.kim,
laoar.shao, linux-mm, longman, mhocko, mm-commits, msys.mizuma,
penberg, richard.weiyang, rientjes, shakeelb, songmuchun, tj,
torvalds, vbabka, vdavydov.dev, willy, zhengjun.xing
From: Waiman Long <longman@redhat.com>
Subject: mm/memcg: optimize user context object stock access
Most kmem_cache_alloc() calls are from user context. With instrumentation
enabled, the measured amount of kmem_cache_alloc() calls from non-task
context was about 0.01% of the total.
The irq disable/enable sequence used in this case to access content from
object stock is slow. To optimize for user context access, there are now
two sets of object stocks (in the new obj_stock structure) for task
context and interrupt context access respectively.
The task context object stock can be accessed after disabling preemption
which is cheap in non-preempt kernel. The interrupt context object stock
can only be accessed after disabling interrupt. User context code can
access interrupt object stock, but not vice versa.
The downside of this change is that there are more data stored in local
object stocks and not reflected in the charge counter and the vmstat
arrays. However, this is a small price to pay for better performance.
[longman@redhat.com: fix potential uninitialized variable warning]
Link: https://lkml.kernel.org/r/20210526193602.8742-1-longman@redhat.com
[akpm@linux-foundation.org: coding style fixes]
Link: https://lkml.kernel.org/r/20210506150007.16288-5-longman@redhat.com
Signed-off-by: Waiman Long <longman@redhat.com>
Acked-by: Roman Gushchin <guro@fb.com>
Reviewed-by: Shakeel Butt <shakeelb@google.com>
Cc: Johannes Weiner <hannes@cmpxchg.org>
Cc: Michal Hocko <mhocko@kernel.org>
Cc: Vladimir Davydov <vdavydov.dev@gmail.com>
Cc: Tejun Heo <tj@kernel.org>
Cc: Christoph Lameter <cl@linux.com>
Cc: Pekka Enberg <penberg@kernel.org>
Cc: David Rientjes <rientjes@google.com>
Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Cc: Vlastimil Babka <vbabka@suse.cz>
Cc: Roman Gushchin <guro@fb.com>
Cc: Muchun Song <songmuchun@bytedance.com>
Cc: Alex Shi <alex.shi@linux.alibaba.com>
Cc: Chris Down <chris@chrisdown.name>
Cc: Yafang Shao <laoar.shao@gmail.com>
Cc: Wei Yang <richard.weiyang@gmail.com>
Cc: Masayoshi Mizuma <msys.mizuma@gmail.com>
Cc: Xing Zhengjun <zhengjun.xing@linux.intel.com>
Cc: Matthew Wilcox <willy@infradead.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/memcontrol.c | 100 +++++++++++++++++++++++++++++++++-------------
1 file changed, 72 insertions(+), 28 deletions(-)
--- a/mm/memcontrol.c~mm-memcg-optimize-user-context-object-stock-access
+++ a/mm/memcontrol.c
@@ -782,6 +782,10 @@ void __mod_lruvec_kmem_state(void *p, en
rcu_read_unlock();
}
+/*
+ * mod_objcg_mlstate() may be called with irq enabled, so
+ * mod_memcg_lruvec_state() should be used.
+ */
static inline void mod_objcg_mlstate(struct obj_cgroup *objcg,
struct pglist_data *pgdat,
enum node_stat_item idx, int nr)
@@ -792,7 +796,7 @@ static inline void mod_objcg_mlstate(str
rcu_read_lock();
memcg = obj_cgroup_memcg(objcg);
lruvec = mem_cgroup_lruvec(memcg, pgdat);
- __mod_memcg_lruvec_state(lruvec, idx, nr);
+ mod_memcg_lruvec_state(lruvec, idx, nr);
rcu_read_unlock();
}
@@ -2054,17 +2058,23 @@ void unlock_page_memcg(struct page *page
}
EXPORT_SYMBOL(unlock_page_memcg);
-struct memcg_stock_pcp {
- struct mem_cgroup *cached; /* this never be root cgroup */
- unsigned int nr_pages;
-
+struct obj_stock {
#ifdef CONFIG_MEMCG_KMEM
struct obj_cgroup *cached_objcg;
struct pglist_data *cached_pgdat;
unsigned int nr_bytes;
int nr_slab_reclaimable_b;
int nr_slab_unreclaimable_b;
+#else
+ int dummy[0];
#endif
+};
+
+struct memcg_stock_pcp {
+ struct mem_cgroup *cached; /* this never be root cgroup */
+ unsigned int nr_pages;
+ struct obj_stock task_obj;
+ struct obj_stock irq_obj;
struct work_struct work;
unsigned long flags;
@@ -2074,12 +2084,12 @@ static DEFINE_PER_CPU(struct memcg_stock
static DEFINE_MUTEX(percpu_charge_mutex);
#ifdef CONFIG_MEMCG_KMEM
-static void drain_obj_stock(struct memcg_stock_pcp *stock);
+static void drain_obj_stock(struct obj_stock *stock);
static bool obj_stock_flush_required(struct memcg_stock_pcp *stock,
struct mem_cgroup *root_memcg);
#else
-static inline void drain_obj_stock(struct memcg_stock_pcp *stock)
+static inline void drain_obj_stock(struct obj_stock *stock)
{
}
static bool obj_stock_flush_required(struct memcg_stock_pcp *stock,
@@ -2089,6 +2099,41 @@ static bool obj_stock_flush_required(str
}
#endif
+/*
+ * Most kmem_cache_alloc() calls are from user context. The irq disable/enable
+ * sequence used in this case to access content from object stock is slow.
+ * To optimize for user context access, there are now two object stocks for
+ * task context and interrupt context access respectively.
+ *
+ * The task context object stock can be accessed by disabling preemption only
+ * which is cheap in non-preempt kernel. The interrupt context object stock
+ * can only be accessed after disabling interrupt. User context code can
+ * access interrupt object stock, but not vice versa.
+ */
+static inline struct obj_stock *get_obj_stock(unsigned long *pflags)
+{
+ struct memcg_stock_pcp *stock;
+
+ if (likely(in_task())) {
+ *pflags = 0UL;
+ preempt_disable();
+ stock = this_cpu_ptr(&memcg_stock);
+ return &stock->task_obj;
+ }
+
+ local_irq_save(*pflags);
+ stock = this_cpu_ptr(&memcg_stock);
+ return &stock->irq_obj;
+}
+
+static inline void put_obj_stock(unsigned long flags)
+{
+ if (likely(in_task()))
+ preempt_enable();
+ else
+ local_irq_restore(flags);
+}
+
/**
* consume_stock: Try to consume stocked charge on this cpu.
* @memcg: memcg to consume from.
@@ -2155,7 +2200,9 @@ static void drain_local_stock(struct wor
local_irq_save(flags);
stock = this_cpu_ptr(&memcg_stock);
- drain_obj_stock(stock);
+ drain_obj_stock(&stock->irq_obj);
+ if (in_task())
+ drain_obj_stock(&stock->task_obj);
drain_stock(stock);
clear_bit(FLUSHING_CACHED_CHARGE, &stock->flags);
@@ -3015,13 +3062,10 @@ void __memcg_kmem_uncharge_page(struct p
void mod_objcg_state(struct obj_cgroup *objcg, struct pglist_data *pgdat,
enum node_stat_item idx, int nr)
{
- struct memcg_stock_pcp *stock;
unsigned long flags;
+ struct obj_stock *stock = get_obj_stock(&flags);
int *bytes;
- local_irq_save(flags);
- stock = this_cpu_ptr(&memcg_stock);
-
/*
* Save vmstat data in stock and skip vmstat array update unless
* accumulating over a page of vmstat data or when pgdat or idx
@@ -3070,29 +3114,26 @@ void mod_objcg_state(struct obj_cgroup *
if (nr)
mod_objcg_mlstate(objcg, pgdat, idx, nr);
- local_irq_restore(flags);
+ put_obj_stock(flags);
}
static bool consume_obj_stock(struct obj_cgroup *objcg, unsigned int nr_bytes)
{
- struct memcg_stock_pcp *stock;
unsigned long flags;
+ struct obj_stock *stock = get_obj_stock(&flags);
bool ret = false;
- local_irq_save(flags);
-
- stock = this_cpu_ptr(&memcg_stock);
if (objcg == stock->cached_objcg && stock->nr_bytes >= nr_bytes) {
stock->nr_bytes -= nr_bytes;
ret = true;
}
- local_irq_restore(flags);
+ put_obj_stock(flags);
return ret;
}
-static void drain_obj_stock(struct memcg_stock_pcp *stock)
+static void drain_obj_stock(struct obj_stock *stock)
{
struct obj_cgroup *old = stock->cached_objcg;
@@ -3148,8 +3189,13 @@ static bool obj_stock_flush_required(str
{
struct mem_cgroup *memcg;
- if (stock->cached_objcg) {
- memcg = obj_cgroup_memcg(stock->cached_objcg);
+ if (in_task() && stock->task_obj.cached_objcg) {
+ memcg = obj_cgroup_memcg(stock->task_obj.cached_objcg);
+ if (memcg && mem_cgroup_is_descendant(memcg, root_memcg))
+ return true;
+ }
+ if (stock->irq_obj.cached_objcg) {
+ memcg = obj_cgroup_memcg(stock->irq_obj.cached_objcg);
if (memcg && mem_cgroup_is_descendant(memcg, root_memcg))
return true;
}
@@ -3160,13 +3206,10 @@ static bool obj_stock_flush_required(str
static void refill_obj_stock(struct obj_cgroup *objcg, unsigned int nr_bytes,
bool allow_uncharge)
{
- struct memcg_stock_pcp *stock;
unsigned long flags;
+ struct obj_stock *stock = get_obj_stock(&flags);
unsigned int nr_pages = 0;
- local_irq_save(flags);
-
- stock = this_cpu_ptr(&memcg_stock);
if (stock->cached_objcg != objcg) { /* reset if necessary */
drain_obj_stock(stock);
obj_cgroup_get(objcg);
@@ -3182,7 +3225,7 @@ static void refill_obj_stock(struct obj_
stock->nr_bytes &= (PAGE_SIZE - 1);
}
- local_irq_restore(flags);
+ put_obj_stock(flags);
if (nr_pages)
obj_cgroup_uncharge_pages(objcg, nr_pages);
@@ -6790,6 +6833,7 @@ static void uncharge_page(struct page *p
unsigned long nr_pages;
struct mem_cgroup *memcg;
struct obj_cgroup *objcg;
+ bool use_objcg = PageMemcgKmem(page);
VM_BUG_ON_PAGE(PageLRU(page), page);
@@ -6798,7 +6842,7 @@ static void uncharge_page(struct page *p
* page memcg or objcg at this point, we have fully
* exclusive access to the page.
*/
- if (PageMemcgKmem(page)) {
+ if (use_objcg) {
objcg = __page_objcg(page);
/*
* This get matches the put at the end of the function and
@@ -6826,7 +6870,7 @@ static void uncharge_page(struct page *p
nr_pages = compound_nr(page);
- if (PageMemcgKmem(page)) {
+ if (use_objcg) {
ug->nr_memory += nr_pages;
ug->nr_kmem += nr_pages;
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 081/192] mm: memcg/slab: properly set up gfp flags for objcg pointer array
2021-06-29 2:32 incoming Andrew Morton
` (79 preceding siblings ...)
2021-06-29 2:37 ` [patch 080/192] mm/memcg: optimize user context object stock access Andrew Morton
@ 2021-06-29 2:37 ` Andrew Morton
2021-06-29 2:37 ` [patch 082/192] mm: memcg/slab: create a new set of kmalloc-cg-<n> caches Andrew Morton
` (110 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:37 UTC (permalink / raw)
To: akpm, cl, guro, hannes, iamjoonsoo.kim, linux-mm, longman,
mhocko, mm-commits, penberg, rientjes, shakeelb, torvalds,
vbabka, vdavydov.dev
From: Waiman Long <longman@redhat.com>
Subject: mm: memcg/slab: properly set up gfp flags for objcg pointer array
Patch series "mm: memcg/slab: Fix objcg pointer array handling problem", v4.
Since the merging of the new slab memory controller in v5.9, the page
structure stores a pointer to objcg pointer array for slab pages. When
the slab has no used objects, it can be freed in free_slab() which will
call kfree() to free the objcg pointer array in
memcg_alloc_page_obj_cgroups(). If it happens that the objcg pointer
array is the last used object in its slab, that slab may then be freed
which may caused kfree() to be called again.
With the right workload, the slab cache may be set up in a way that allows
the recursive kfree() calling loop to nest deep enough to cause a kernel
stack overflow and panic the system. In fact, we have a reproducer that
can cause kernel stack overflow on a s390 system involving kmalloc-rcl-256
and kmalloc-rcl-128 slabs with the following kfree() loop recursively
called 74 times:
[ 285.520739] [<000000000ec432fc>] kfree+0x4bc/0x560 [ 285.520740]
[<000000000ec43466>] __free_slab+0xc6/0x228 [ 285.520741]
[<000000000ec41fc2>] __slab_free+0x3c2/0x3e0 [ 285.520742]
[<000000000ec432fc>] kfree+0x4bc/0x560 : While investigating this issue, I
also found an issue on the allocation side. If the objcg pointer array
happen to come from the same slab or a circular dependency linkage is
formed with multiple slabs, those affected slabs can never be freed again.
This patch series addresses these two issues by introducing a new set of
kmalloc-cg-<n> caches split from kmalloc-<n> caches. The new set will
only contain non-reclaimable and non-dma objects that are accounted in
memory cgroups whereas the old set are now for unaccounted objects only.
By making this split, all the objcg pointer arrays will come from the
kmalloc-<n> caches, but those caches will never hold any objcg pointer
array. As a result, deeply nested kfree() call and the unfreeable slab
problems are now gone.
This patch (of 4):
Since the merging of the new slab memory controller in v5.9, the page
structure may store a pointer to obj_cgroup pointer array for slab pages.
Currently, only the __GFP_ACCOUNT bit is masked off. However, the array
is not readily reclaimable and doesn't need to come from the DMA buffer.
So those GFP bits should be masked off as well.
Do the flag bit clearing at memcg_alloc_page_obj_cgroups() to make sure
that it is consistently applied no matter where it is called.
Link: https://lkml.kernel.org/r/20210505200610.13943-1-longman@redhat.com
Link: https://lkml.kernel.org/r/20210505200610.13943-2-longman@redhat.com
Fixes: 286e04b8ed7a ("mm: memcg/slab: allocate obj_cgroups for non-root slab pages")
Signed-off-by: Waiman Long <longman@redhat.com>
Reviewed-by: Shakeel Butt <shakeelb@google.com>
Acked-by: Roman Gushchin <guro@fb.com>
Reviewed-by: Vlastimil Babka <vbabka@suse.cz>
Cc: Johannes Weiner <hannes@cmpxchg.org>
Cc: Michal Hocko <mhocko@kernel.org>
Cc: Vladimir Davydov <vdavydov.dev@gmail.com>
Cc: Christoph Lameter <cl@linux.com>
Cc: Pekka Enberg <penberg@kernel.org>
Cc: David Rientjes <rientjes@google.com>
Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/memcontrol.c | 8 ++++++++
mm/slab.h | 1 -
2 files changed, 8 insertions(+), 1 deletion(-)
--- a/mm/memcontrol.c~mm-memcg-slab-properly-set-up-gfp-flags-for-objcg-pointer-array
+++ a/mm/memcontrol.c
@@ -2803,6 +2803,13 @@ retry:
}
#ifdef CONFIG_MEMCG_KMEM
+/*
+ * The allocated objcg pointers array is not accounted directly.
+ * Moreover, it should not come from DMA buffer and is not readily
+ * reclaimable. So those GFP bits should be masked off.
+ */
+#define OBJCGS_CLEAR_MASK (__GFP_DMA | __GFP_RECLAIMABLE | __GFP_ACCOUNT)
+
int memcg_alloc_page_obj_cgroups(struct page *page, struct kmem_cache *s,
gfp_t gfp, bool new_page)
{
@@ -2810,6 +2817,7 @@ int memcg_alloc_page_obj_cgroups(struct
unsigned long memcg_data;
void *vec;
+ gfp &= ~OBJCGS_CLEAR_MASK;
vec = kcalloc_node(objects, sizeof(struct obj_cgroup *), gfp,
page_to_nid(page));
if (!vec)
--- a/mm/slab.h~mm-memcg-slab-properly-set-up-gfp-flags-for-objcg-pointer-array
+++ a/mm/slab.h
@@ -298,7 +298,6 @@ static inline void memcg_slab_post_alloc
if (!memcg_kmem_enabled() || !objcg)
return;
- flags &= ~__GFP_ACCOUNT;
for (i = 0; i < size; i++) {
if (likely(p[i])) {
page = virt_to_head_page(p[i]);
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 082/192] mm: memcg/slab: create a new set of kmalloc-cg-<n> caches
2021-06-29 2:32 incoming Andrew Morton
` (80 preceding siblings ...)
2021-06-29 2:37 ` [patch 081/192] mm: memcg/slab: properly set up gfp flags for objcg pointer array Andrew Morton
@ 2021-06-29 2:37 ` Andrew Morton
2021-06-29 2:37 ` [patch 083/192] mm: memcg/slab: disable cache merging for KMALLOC_NORMAL caches Andrew Morton
` (109 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:37 UTC (permalink / raw)
To: akpm, cl, guro, hannes, iamjoonsoo.kim, linux-mm, longman,
mhocko, mm-commits, penberg, rientjes, shakeelb, torvalds,
vbabka, vdavydov.dev
From: Waiman Long <longman@redhat.com>
Subject: mm: memcg/slab: create a new set of kmalloc-cg-<n> caches
There are currently two problems in the way the objcg pointer array
(memcg_data) in the page structure is being allocated and freed.
On its allocation, it is possible that the allocated objcg pointer
array comes from the same slab that requires memory accounting. If this
happens, the slab will never become empty again as there is at least
one object left (the obj_cgroup array) in the slab.
When it is freed, the objcg pointer array object may be the last one
in its slab and hence causes kfree() to be called again. With the
right workload, the slab cache may be set up in a way that allows the
recursive kfree() calling loop to nest deep enough to cause a kernel
stack overflow and panic the system.
One way to solve this problem is to split the kmalloc-<n> caches
(KMALLOC_NORMAL) into two separate sets - a new set of kmalloc-<n>
(KMALLOC_NORMAL) caches for unaccounted objects only and a new set of
kmalloc-cg-<n> (KMALLOC_CGROUP) caches for accounted objects only. All
the other caches can still allow a mix of accounted and unaccounted
objects.
With this change, all the objcg pointer array objects will come from
KMALLOC_NORMAL caches which won't have their objcg pointer arrays. So
both the recursive kfree() problem and non-freeable slab problem are
gone.
Since both the KMALLOC_NORMAL and KMALLOC_CGROUP caches no longer have
mixed accounted and unaccounted objects, this will slightly reduce the
number of objcg pointer arrays that need to be allocated and save a bit
of memory. On the other hand, creating a new set of kmalloc caches does
have the effect of reducing cache utilization. So it is properly a wash.
The new KMALLOC_CGROUP is added between KMALLOC_NORMAL and
KMALLOC_RECLAIM so that the first for loop in create_kmalloc_caches()
will include the newly added caches without change.
[vbabka@suse.cz: don't create kmalloc-cg caches with cgroup.memory=nokmem]
Link: https://lkml.kernel.org/r/20210512145107.6208-1-longman@redhat.com
Link: https://lkml.kernel.org/r/20210505200610.13943-3-longman@redhat.com
Signed-off-by: Waiman Long <longman@redhat.com>
Signed-off-by: Vlastimil Babka <vbabka@suse.cz>
Suggested-by: Vlastimil Babka <vbabka@suse.cz>
Reviewed-by: Shakeel Butt <shakeelb@google.com>
Acked-by: Roman Gushchin <guro@fb.com>
Cc: Christoph Lameter <cl@linux.com>
Cc: David Rientjes <rientjes@google.com>
Cc: Johannes Weiner <hannes@cmpxchg.org>
Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Cc: Michal Hocko <mhocko@kernel.org>
Cc: Pekka Enberg <penberg@kernel.org>
Cc: Vladimir Davydov <vdavydov.dev@gmail.com>
[longman@redhat.com: fix for CONFIG_ZONE_DMA=n]
Link: https://lkml.kernel.org/r/20210512145107.6208-1-longman@redhat.com
[akpm@linux-foundation.org: un-fat-finger v5 delta creation]
[longman@redhat.com: disable cache merging for KMALLOC_NORMAL caches]
Link: https://lkml.kernel.org/r/20210505200610.13943-4-longman@redhat.com
Signed-off-by: Waiman Long <longman@redhat.com>
Suggested-by: Roman Gushchin <guro@fb.com>
Acked-by: Roman Gushchin <guro@fb.com>
Reviewed-by: Shakeel Butt <shakeelb@google.com>
Reviewed-by: Vlastimil Babka <vbabka@suse.cz>
Cc: Christoph Lameter <cl@linux.com>
Cc: David Rientjes <rientjes@google.com>
Cc: Johannes Weiner <hannes@cmpxchg.org>
Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Cc: Michal Hocko <mhocko@kernel.org>
Cc: Pekka Enberg <penberg@kernel.org>
Cc: Vladimir Davydov <vdavydov.dev@gmail.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/linux/slab.h | 42 ++++++++++++++++++++++++++++++++---------
mm/internal.h | 5 ++++
mm/memcontrol.c | 2 -
mm/slab_common.c | 32 ++++++++++++++++++++++---------
4 files changed, 62 insertions(+), 19 deletions(-)
--- a/include/linux/slab.h~mm-memcg-slab-create-a-new-set-of-kmalloc-cg-n-caches
+++ a/include/linux/slab.h
@@ -305,9 +305,21 @@ static inline void __check_heap_object(c
/*
* Whenever changing this, take care of that kmalloc_type() and
* create_kmalloc_caches() still work as intended.
+ *
+ * KMALLOC_NORMAL can contain only unaccounted objects whereas KMALLOC_CGROUP
+ * is for accounted but unreclaimable and non-dma objects. All the other
+ * kmem caches can have both accounted and unaccounted objects.
*/
enum kmalloc_cache_type {
KMALLOC_NORMAL = 0,
+#ifndef CONFIG_ZONE_DMA
+ KMALLOC_DMA = KMALLOC_NORMAL,
+#endif
+#ifndef CONFIG_MEMCG_KMEM
+ KMALLOC_CGROUP = KMALLOC_NORMAL,
+#else
+ KMALLOC_CGROUP,
+#endif
KMALLOC_RECLAIM,
#ifdef CONFIG_ZONE_DMA
KMALLOC_DMA,
@@ -319,24 +331,36 @@ enum kmalloc_cache_type {
extern struct kmem_cache *
kmalloc_caches[NR_KMALLOC_TYPES][KMALLOC_SHIFT_HIGH + 1];
+/*
+ * Define gfp bits that should not be set for KMALLOC_NORMAL.
+ */
+#define KMALLOC_NOT_NORMAL_BITS \
+ (__GFP_RECLAIMABLE | \
+ (IS_ENABLED(CONFIG_ZONE_DMA) ? __GFP_DMA : 0) | \
+ (IS_ENABLED(CONFIG_MEMCG_KMEM) ? __GFP_ACCOUNT : 0))
+
static __always_inline enum kmalloc_cache_type kmalloc_type(gfp_t flags)
{
-#ifdef CONFIG_ZONE_DMA
/*
* The most common case is KMALLOC_NORMAL, so test for it
- * with a single branch for both flags.
+ * with a single branch for all the relevant flags.
*/
- if (likely((flags & (__GFP_DMA | __GFP_RECLAIMABLE)) == 0))
+ if (likely((flags & KMALLOC_NOT_NORMAL_BITS) == 0))
return KMALLOC_NORMAL;
/*
- * At least one of the flags has to be set. If both are, __GFP_DMA
- * is more important.
+ * At least one of the flags has to be set. Their priorities in
+ * decreasing order are:
+ * 1) __GFP_DMA
+ * 2) __GFP_RECLAIMABLE
+ * 3) __GFP_ACCOUNT
*/
- return flags & __GFP_DMA ? KMALLOC_DMA : KMALLOC_RECLAIM;
-#else
- return flags & __GFP_RECLAIMABLE ? KMALLOC_RECLAIM : KMALLOC_NORMAL;
-#endif
+ if (IS_ENABLED(CONFIG_ZONE_DMA) && (flags & __GFP_DMA))
+ return KMALLOC_DMA;
+ if (!IS_ENABLED(CONFIG_MEMCG_KMEM) || (flags & __GFP_RECLAIMABLE))
+ return KMALLOC_RECLAIM;
+ else
+ return KMALLOC_CGROUP;
}
/*
--- a/mm/internal.h~mm-memcg-slab-create-a-new-set-of-kmalloc-cg-n-caches
+++ a/mm/internal.h
@@ -116,6 +116,11 @@ extern void putback_lru_page(struct page
extern pmd_t *mm_find_pmd(struct mm_struct *mm, unsigned long address);
/*
+ * in mm/memcontrol.c:
+ */
+extern bool cgroup_memory_nokmem;
+
+/*
* in mm/page_alloc.c
*/
--- a/mm/memcontrol.c~mm-memcg-slab-create-a-new-set-of-kmalloc-cg-n-caches
+++ a/mm/memcontrol.c
@@ -83,7 +83,7 @@ DEFINE_PER_CPU(struct mem_cgroup *, int_
static bool cgroup_memory_nosocket;
/* Kernel memory accounting disabled? */
-static bool cgroup_memory_nokmem;
+bool cgroup_memory_nokmem;
/* Whether the swap controller is active */
#ifdef CONFIG_MEMCG_SWAP
--- a/mm/slab_common.c~mm-memcg-slab-create-a-new-set-of-kmalloc-cg-n-caches
+++ a/mm/slab_common.c
@@ -738,21 +738,25 @@ struct kmem_cache *kmalloc_slab(size_t s
}
#ifdef CONFIG_ZONE_DMA
-#define INIT_KMALLOC_INFO(__size, __short_size) \
-{ \
- .name[KMALLOC_NORMAL] = "kmalloc-" #__short_size, \
- .name[KMALLOC_RECLAIM] = "kmalloc-rcl-" #__short_size, \
- .name[KMALLOC_DMA] = "dma-kmalloc-" #__short_size, \
- .size = __size, \
-}
+#define KMALLOC_DMA_NAME(sz) .name[KMALLOC_DMA] = "dma-kmalloc-" #sz,
#else
+#define KMALLOC_DMA_NAME(sz)
+#endif
+
+#ifdef CONFIG_MEMCG_KMEM
+#define KMALLOC_CGROUP_NAME(sz) .name[KMALLOC_CGROUP] = "kmalloc-cg-" #sz,
+#else
+#define KMALLOC_CGROUP_NAME(sz)
+#endif
+
#define INIT_KMALLOC_INFO(__size, __short_size) \
{ \
.name[KMALLOC_NORMAL] = "kmalloc-" #__short_size, \
.name[KMALLOC_RECLAIM] = "kmalloc-rcl-" #__short_size, \
+ KMALLOC_CGROUP_NAME(__short_size) \
+ KMALLOC_DMA_NAME(__short_size) \
.size = __size, \
}
-#endif
/*
* kmalloc_info[] is to make slub_debug=,kmalloc-xx option work at boot time.
@@ -838,8 +842,15 @@ void __init setup_kmalloc_cache_index_ta
static void __init
new_kmalloc_cache(int idx, enum kmalloc_cache_type type, slab_flags_t flags)
{
- if (type == KMALLOC_RECLAIM)
+ if (type == KMALLOC_RECLAIM) {
flags |= SLAB_RECLAIM_ACCOUNT;
+ } else if (IS_ENABLED(CONFIG_MEMCG_KMEM) && (type == KMALLOC_CGROUP)) {
+ if (cgroup_memory_nokmem) {
+ kmalloc_caches[type][idx] = kmalloc_caches[KMALLOC_NORMAL][idx];
+ return;
+ }
+ flags |= SLAB_ACCOUNT;
+ }
kmalloc_caches[type][idx] = create_kmalloc_cache(
kmalloc_info[idx].name[type],
@@ -857,6 +868,9 @@ void __init create_kmalloc_caches(slab_f
int i;
enum kmalloc_cache_type type;
+ /*
+ * Including KMALLOC_CGROUP if CONFIG_MEMCG_KMEM defined
+ */
for (type = KMALLOC_NORMAL; type <= KMALLOC_RECLAIM; type++) {
for (i = KMALLOC_SHIFT_LOW; i <= KMALLOC_SHIFT_HIGH; i++) {
if (!kmalloc_caches[type][i])
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 083/192] mm: memcg/slab: disable cache merging for KMALLOC_NORMAL caches
2021-06-29 2:32 incoming Andrew Morton
` (81 preceding siblings ...)
2021-06-29 2:37 ` [patch 082/192] mm: memcg/slab: create a new set of kmalloc-cg-<n> caches Andrew Morton
@ 2021-06-29 2:37 ` Andrew Morton
2021-06-29 2:37 ` [patch 084/192] mm: memcontrol: fix root_mem_cgroup charging Andrew Morton
` (108 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:37 UTC (permalink / raw)
To: akpm, cl, guro, hannes, iamjoonsoo.kim, linux-mm, longman,
mhocko, mm-commits, penberg, rientjes, shakeelb, torvalds,
vbabka, vdavydov.dev
From: Waiman Long <longman@redhat.com>
Subject: mm: memcg/slab: disable cache merging for KMALLOC_NORMAL caches
The KMALLOC_NORMAL (kmalloc-<n>) caches are for unaccounted objects only
when CONFIG_MEMCG_KMEM is enabled. To make sure that this condition
remains true, we will have to prevent KMALOC_NORMAL caches to merge with
other kmem caches. This is now done by setting its refcount to -1 right
after its creation.
Link: https://lkml.kernel.org/r/20210505200610.13943-4-longman@redhat.com
Signed-off-by: Waiman Long <longman@redhat.com>
Suggested-by: Roman Gushchin <guro@fb.com>
Acked-by: Roman Gushchin <guro@fb.com>
Reviewed-by: Shakeel Butt <shakeelb@google.com>
Reviewed-by: Vlastimil Babka <vbabka@suse.cz>
Cc: Christoph Lameter <cl@linux.com>
Cc: David Rientjes <rientjes@google.com>
Cc: Johannes Weiner <hannes@cmpxchg.org>
Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Cc: Michal Hocko <mhocko@kernel.org>
Cc: Pekka Enberg <penberg@kernel.org>
Cc: Vladimir Davydov <vdavydov.dev@gmail.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/slab_common.c | 7 +++++++
1 file changed, 7 insertions(+)
--- a/mm/slab_common.c~mm-memcg-slab-disable-cache-merging-for-kmalloc_normal-caches
+++ a/mm/slab_common.c
@@ -856,6 +856,13 @@ new_kmalloc_cache(int idx, enum kmalloc_
kmalloc_info[idx].name[type],
kmalloc_info[idx].size, flags, 0,
kmalloc_info[idx].size);
+
+ /*
+ * If CONFIG_MEMCG_KMEM is enabled, disable cache merging for
+ * KMALLOC_NORMAL caches.
+ */
+ if (IS_ENABLED(CONFIG_MEMCG_KMEM) && (type == KMALLOC_NORMAL))
+ kmalloc_caches[type][idx]->refcount = -1;
}
/*
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 084/192] mm: memcontrol: fix root_mem_cgroup charging
2021-06-29 2:32 incoming Andrew Morton
` (82 preceding siblings ...)
2021-06-29 2:37 ` [patch 083/192] mm: memcg/slab: disable cache merging for KMALLOC_NORMAL caches Andrew Morton
@ 2021-06-29 2:37 ` Andrew Morton
2021-06-29 2:37 ` [patch 085/192] mm: memcontrol: fix page charging in page replacement Andrew Morton
` (107 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:37 UTC (permalink / raw)
To: akpm, duanxiongchun, guro, hannes, linux-mm, mhocko, mm-commits,
shakeelb, songmuchun, torvalds, vdavydov.dev
From: Muchun Song <songmuchun@bytedance.com>
Subject: mm: memcontrol: fix root_mem_cgroup charging
The below scenario can cause the page counters of the root_mem_cgroup to
be out of balance.
CPU0: CPU1:
objcg = get_obj_cgroup_from_current()
obj_cgroup_charge_pages(objcg)
memcg_reparent_objcgs()
// reparent to root_mem_cgroup
WRITE_ONCE(iter->memcg, parent)
// memcg == root_mem_cgroup
memcg = get_mem_cgroup_from_objcg(objcg)
// do not charge to the root_mem_cgroup
try_charge(memcg)
obj_cgroup_uncharge_pages(objcg)
memcg = get_mem_cgroup_from_objcg(objcg)
// uncharge from the root_mem_cgroup
refill_stock(memcg)
drain_stock(memcg)
page_counter_uncharge(&memcg->memory)
get_obj_cgroup_from_current() never returns a root_mem_cgroup's objcg, so
we never explicitly charge the root_mem_cgroup. And it's not going to
change. It's all about a race when we got an obj_cgroup pointing at some
non-root memcg, but before we were able to charge it, the cgroup was gone,
objcg was reparented to the root and so we're skipping the charging. Then
we store the objcg pointer and later use to uncharge the root_mem_cgroup.
This can cause the page counter to be less than the actual value.
Although we do not display the value (mem_cgroup_usage) so there shouldn't
be any actual problem, but there is a WARN_ON_ONCE in the
page_counter_cancel(). Who knows if it will trigger? So it is better to
fix it.
Link: https://lkml.kernel.org/r/20210425075410.19255-1-songmuchun@bytedance.com
Signed-off-by: Muchun Song <songmuchun@bytedance.com>
Acked-by: Roman Gushchin <guro@fb.com>
Reviewed-by: Shakeel Butt <shakeelb@google.com>
Cc: Xiongchun Duan <duanxiongchun@bytedance.com>
Cc: Johannes Weiner <hannes@cmpxchg.org>
Cc: Michal Hocko <mhocko@kernel.org>
Cc: Vladimir Davydov <vdavydov.dev@gmail.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/memcontrol.c | 17 ++++++++++++-----
1 file changed, 12 insertions(+), 5 deletions(-)
--- a/mm/memcontrol.c~mm-memcontrol-fix-root_mem_cgroup-charging
+++ a/mm/memcontrol.c
@@ -2568,8 +2568,8 @@ out:
css_put(&memcg->css);
}
-static int try_charge(struct mem_cgroup *memcg, gfp_t gfp_mask,
- unsigned int nr_pages)
+static int try_charge_memcg(struct mem_cgroup *memcg, gfp_t gfp_mask,
+ unsigned int nr_pages)
{
unsigned int batch = max(MEMCG_CHARGE_BATCH, nr_pages);
int nr_retries = MAX_RECLAIM_RETRIES;
@@ -2581,8 +2581,6 @@ static int try_charge(struct mem_cgroup
bool drained = false;
unsigned long pflags;
- if (mem_cgroup_is_root(memcg))
- return 0;
retry:
if (consume_stock(memcg, nr_pages))
return 0;
@@ -2762,6 +2760,15 @@ done_restock:
return 0;
}
+static inline int try_charge(struct mem_cgroup *memcg, gfp_t gfp_mask,
+ unsigned int nr_pages)
+{
+ if (mem_cgroup_is_root(memcg))
+ return 0;
+
+ return try_charge_memcg(memcg, gfp_mask, nr_pages);
+}
+
#if defined(CONFIG_MEMCG_KMEM) || defined(CONFIG_MMU)
static void cancel_charge(struct mem_cgroup *memcg, unsigned int nr_pages)
{
@@ -2997,7 +3004,7 @@ static int obj_cgroup_charge_pages(struc
memcg = get_mem_cgroup_from_objcg(objcg);
- ret = try_charge(memcg, gfp, nr_pages);
+ ret = try_charge_memcg(memcg, gfp, nr_pages);
if (ret)
goto out;
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 085/192] mm: memcontrol: fix page charging in page replacement
2021-06-29 2:32 incoming Andrew Morton
` (83 preceding siblings ...)
2021-06-29 2:37 ` [patch 084/192] mm: memcontrol: fix root_mem_cgroup charging Andrew Morton
@ 2021-06-29 2:37 ` Andrew Morton
2021-06-29 2:37 ` [patch 086/192] mm: memcontrol: bail out early when !mm in get_mem_cgroup_from_mm Andrew Morton
` (106 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:37 UTC (permalink / raw)
To: akpm, duanxiongchun, guro, hannes, linux-mm, mhocko, mm-commits,
shakeelb, songmuchun, torvalds, vdavydov.dev
From: Muchun Song <songmuchun@bytedance.com>
Subject: mm: memcontrol: fix page charging in page replacement
Patch series "memcontrol code cleanup and simplification", v3.
This patch (of 8):
The pages aren't accounted at the root level, so do not charge the page to
the root memcg in page replacement. Although we do not display the value
(mem_cgroup_usage) so there shouldn't be any actual problem, but there is
a WARN_ON_ONCE in the page_counter_cancel(). Who knows if it will
trigger? So it is better to fix it.
Link: https://lkml.kernel.org/r/20210417043538.9793-1-songmuchun@bytedance.com
Link: https://lkml.kernel.org/r/20210417043538.9793-2-songmuchun@bytedance.com
Signed-off-by: Muchun Song <songmuchun@bytedance.com>
Acked-by: Johannes Weiner <hannes@cmpxchg.org>
Reviewed-by: Shakeel Butt <shakeelb@google.com>
Acked-by: Roman Gushchin <guro@fb.com>
Acked-by: Michal Hocko <mhocko@suse.com>
Cc: Vladimir Davydov <vdavydov.dev@gmail.com>
Cc: Xiongchun Duan <duanxiongchun@bytedance.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/memcontrol.c | 8 +++++---
1 file changed, 5 insertions(+), 3 deletions(-)
--- a/mm/memcontrol.c~mm-memcontrol-fix-page-charging-in-page-replacement
+++ a/mm/memcontrol.c
@@ -6984,9 +6984,11 @@ void mem_cgroup_migrate(struct page *old
/* Force-charge the new page. The old one will be freed soon */
nr_pages = thp_nr_pages(newpage);
- page_counter_charge(&memcg->memory, nr_pages);
- if (do_memsw_account())
- page_counter_charge(&memcg->memsw, nr_pages);
+ if (!mem_cgroup_is_root(memcg)) {
+ page_counter_charge(&memcg->memory, nr_pages);
+ if (do_memsw_account())
+ page_counter_charge(&memcg->memsw, nr_pages);
+ }
css_get(&memcg->css);
commit_charge(newpage, memcg);
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 086/192] mm: memcontrol: bail out early when !mm in get_mem_cgroup_from_mm
2021-06-29 2:32 incoming Andrew Morton
` (84 preceding siblings ...)
2021-06-29 2:37 ` [patch 085/192] mm: memcontrol: fix page charging in page replacement Andrew Morton
@ 2021-06-29 2:37 ` Andrew Morton
2021-06-29 2:37 ` [patch 087/192] mm: memcontrol: remove the pgdata parameter of mem_cgroup_page_lruvec Andrew Morton
` (105 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:37 UTC (permalink / raw)
To: akpm, duanxiongchun, guro, hannes, linux-mm, mhocko, mm-commits,
shakeelb, songmuchun, torvalds, vdavydov.dev
From: Muchun Song <songmuchun@bytedance.com>
Subject: mm: memcontrol: bail out early when !mm in get_mem_cgroup_from_mm
When mm is NULL, we do not need to hold rcu lock and call css_tryget for
the root memcg. And we also do not need to check !mm in every loop of
while. So bail out early when !mm.
Link: https://lkml.kernel.org/r/20210417043538.9793-3-songmuchun@bytedance.com
Signed-off-by: Muchun Song <songmuchun@bytedance.com>
Acked-by: Johannes Weiner <hannes@cmpxchg.org>
Reviewed-by: Shakeel Butt <shakeelb@google.com>
Acked-by: Roman Gushchin <guro@fb.com>
Acked-by: Michal Hocko <mhocko@suse.com>
Cc: Vladimir Davydov <vdavydov.dev@gmail.com>
Cc: Xiongchun Duan <duanxiongchun@bytedance.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/memcontrol.c | 25 ++++++++++++++-----------
1 file changed, 14 insertions(+), 11 deletions(-)
--- a/mm/memcontrol.c~mm-memcontrol-bail-out-early-when-mm-in-get_mem_cgroup_from_mm
+++ a/mm/memcontrol.c
@@ -919,20 +919,23 @@ struct mem_cgroup *get_mem_cgroup_from_m
if (mem_cgroup_disabled())
return NULL;
+ /*
+ * Page cache insertions can happen without an
+ * actual mm context, e.g. during disk probing
+ * on boot, loopback IO, acct() writes etc.
+ *
+ * No need to css_get on root memcg as the reference
+ * counting is disabled on the root level in the
+ * cgroup core. See CSS_NO_REF.
+ */
+ if (unlikely(!mm))
+ return root_mem_cgroup;
+
rcu_read_lock();
do {
- /*
- * Page cache insertions can happen without an
- * actual mm context, e.g. during disk probing
- * on boot, loopback IO, acct() writes etc.
- */
- if (unlikely(!mm))
+ memcg = mem_cgroup_from_task(rcu_dereference(mm->owner));
+ if (unlikely(!memcg))
memcg = root_mem_cgroup;
- else {
- memcg = mem_cgroup_from_task(rcu_dereference(mm->owner));
- if (unlikely(!memcg))
- memcg = root_mem_cgroup;
- }
} while (!css_tryget(&memcg->css));
rcu_read_unlock();
return memcg;
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 087/192] mm: memcontrol: remove the pgdata parameter of mem_cgroup_page_lruvec
2021-06-29 2:32 incoming Andrew Morton
` (85 preceding siblings ...)
2021-06-29 2:37 ` [patch 086/192] mm: memcontrol: bail out early when !mm in get_mem_cgroup_from_mm Andrew Morton
@ 2021-06-29 2:37 ` Andrew Morton
2021-06-29 2:37 ` [patch 088/192] mm: memcontrol: simplify lruvec_holds_page_lru_lock Andrew Morton
` (104 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:37 UTC (permalink / raw)
To: akpm, duanxiongchun, guro, hannes, linux-mm, mhocko, mm-commits,
shakeelb, songmuchun, torvalds, vdavydov.dev
From: Muchun Song <songmuchun@bytedance.com>
Subject: mm: memcontrol: remove the pgdata parameter of mem_cgroup_page_lruvec
All the callers of mem_cgroup_page_lruvec() just pass page_pgdat(page) as
the 2nd parameter to it (except isolate_migratepages_block()). But for
isolate_migratepages_block(), the page_pgdat(page) is also equal to the
local variable of @pgdat. So mem_cgroup_page_lruvec() do not need the
pgdat parameter. Just remove it to simplify the code.
Link: https://lkml.kernel.org/r/20210417043538.9793-4-songmuchun@bytedance.com
Signed-off-by: Muchun Song <songmuchun@bytedance.com>
Acked-by: Johannes Weiner <hannes@cmpxchg.org>
Reviewed-by: Shakeel Butt <shakeelb@google.com>
Acked-by: Roman Gushchin <guro@fb.com>
Acked-by: Michal Hocko <mhocko@suse.com>
Cc: Vladimir Davydov <vdavydov.dev@gmail.com>
Cc: Xiongchun Duan <duanxiongchun@bytedance.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/linux/memcontrol.h | 10 +++++-----
mm/compaction.c | 2 +-
mm/memcontrol.c | 9 +++------
mm/swap.c | 2 +-
mm/workingset.c | 2 +-
5 files changed, 11 insertions(+), 14 deletions(-)
--- a/include/linux/memcontrol.h~mm-memcontrol-remove-the-pgdata-parameter-of-mem_cgroup_page_lruvec
+++ a/include/linux/memcontrol.h
@@ -743,13 +743,12 @@ out:
/**
* mem_cgroup_page_lruvec - return lruvec for isolating/putting an LRU page
* @page: the page
- * @pgdat: pgdat of the page
*
* This function relies on page->mem_cgroup being stable.
*/
-static inline struct lruvec *mem_cgroup_page_lruvec(struct page *page,
- struct pglist_data *pgdat)
+static inline struct lruvec *mem_cgroup_page_lruvec(struct page *page)
{
+ pg_data_t *pgdat = page_pgdat(page);
struct mem_cgroup *memcg = page_memcg(page);
VM_WARN_ON_ONCE_PAGE(!memcg && !mem_cgroup_disabled(), page);
@@ -1221,9 +1220,10 @@ static inline struct lruvec *mem_cgroup_
return &pgdat->__lruvec;
}
-static inline struct lruvec *mem_cgroup_page_lruvec(struct page *page,
- struct pglist_data *pgdat)
+static inline struct lruvec *mem_cgroup_page_lruvec(struct page *page)
{
+ pg_data_t *pgdat = page_pgdat(page);
+
return &pgdat->__lruvec;
}
--- a/mm/compaction.c~mm-memcontrol-remove-the-pgdata-parameter-of-mem_cgroup_page_lruvec
+++ a/mm/compaction.c
@@ -1028,7 +1028,7 @@ isolate_migratepages_block(struct compac
if (!TestClearPageLRU(page))
goto isolate_fail_put;
- lruvec = mem_cgroup_page_lruvec(page, pgdat);
+ lruvec = mem_cgroup_page_lruvec(page);
/* If we already hold the lock, we can skip some rechecking */
if (lruvec != locked) {
--- a/mm/memcontrol.c~mm-memcontrol-remove-the-pgdata-parameter-of-mem_cgroup_page_lruvec
+++ a/mm/memcontrol.c
@@ -1199,9 +1199,8 @@ void lruvec_memcg_debug(struct lruvec *l
struct lruvec *lock_page_lruvec(struct page *page)
{
struct lruvec *lruvec;
- struct pglist_data *pgdat = page_pgdat(page);
- lruvec = mem_cgroup_page_lruvec(page, pgdat);
+ lruvec = mem_cgroup_page_lruvec(page);
spin_lock(&lruvec->lru_lock);
lruvec_memcg_debug(lruvec, page);
@@ -1212,9 +1211,8 @@ struct lruvec *lock_page_lruvec(struct p
struct lruvec *lock_page_lruvec_irq(struct page *page)
{
struct lruvec *lruvec;
- struct pglist_data *pgdat = page_pgdat(page);
- lruvec = mem_cgroup_page_lruvec(page, pgdat);
+ lruvec = mem_cgroup_page_lruvec(page);
spin_lock_irq(&lruvec->lru_lock);
lruvec_memcg_debug(lruvec, page);
@@ -1225,9 +1223,8 @@ struct lruvec *lock_page_lruvec_irq(stru
struct lruvec *lock_page_lruvec_irqsave(struct page *page, unsigned long *flags)
{
struct lruvec *lruvec;
- struct pglist_data *pgdat = page_pgdat(page);
- lruvec = mem_cgroup_page_lruvec(page, pgdat);
+ lruvec = mem_cgroup_page_lruvec(page);
spin_lock_irqsave(&lruvec->lru_lock, *flags);
lruvec_memcg_debug(lruvec, page);
--- a/mm/swap.c~mm-memcontrol-remove-the-pgdata-parameter-of-mem_cgroup_page_lruvec
+++ a/mm/swap.c
@@ -313,7 +313,7 @@ void lru_note_cost(struct lruvec *lruvec
void lru_note_cost_page(struct page *page)
{
- lru_note_cost(mem_cgroup_page_lruvec(page, page_pgdat(page)),
+ lru_note_cost(mem_cgroup_page_lruvec(page),
page_is_file_lru(page), thp_nr_pages(page));
}
--- a/mm/workingset.c~mm-memcontrol-remove-the-pgdata-parameter-of-mem_cgroup_page_lruvec
+++ a/mm/workingset.c
@@ -408,7 +408,7 @@ void workingset_activation(struct page *
memcg = page_memcg_rcu(page);
if (!mem_cgroup_disabled() && !memcg)
goto out;
- lruvec = mem_cgroup_page_lruvec(page, page_pgdat(page));
+ lruvec = mem_cgroup_page_lruvec(page);
workingset_age_nonresident(lruvec, thp_nr_pages(page));
out:
rcu_read_unlock();
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 088/192] mm: memcontrol: simplify lruvec_holds_page_lru_lock
2021-06-29 2:32 incoming Andrew Morton
` (86 preceding siblings ...)
2021-06-29 2:37 ` [patch 087/192] mm: memcontrol: remove the pgdata parameter of mem_cgroup_page_lruvec Andrew Morton
@ 2021-06-29 2:37 ` Andrew Morton
2021-06-29 2:37 ` [patch 089/192] mm: memcontrol: rename lruvec_holds_page_lru_lock to page_matches_lruvec Andrew Morton
` (103 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:37 UTC (permalink / raw)
To: akpm, duanxiongchun, guro, hannes, linux-mm, mhocko, mm-commits,
shakeelb, songmuchun, torvalds, vdavydov.dev
From: Muchun Song <songmuchun@bytedance.com>
Subject: mm: memcontrol: simplify lruvec_holds_page_lru_lock
We already have a helper lruvec_memcg() to get the memcg from lruvec, we
do not need to do it ourselves in the lruvec_holds_page_lru_lock(). So
use lruvec_memcg() instead. And if mem_cgroup_disabled() returns false,
the page_memcg(page) (the LRU pages) cannot be NULL. So remove the odd
logic of "memcg = page_memcg(page) ? : root_mem_cgroup". And use
lruvec_pgdat to simplify the code. We can have a single definition for
this function that works for !CONFIG_MEMCG, CONFIG_MEMCG +
mem_cgroup_disabled() and CONFIG_MEMCG.
Link: https://lkml.kernel.org/r/20210417043538.9793-5-songmuchun@bytedance.com
Signed-off-by: Muchun Song <songmuchun@bytedance.com>
Acked-by: Johannes Weiner <hannes@cmpxchg.org>
Reviewed-by: Shakeel Butt <shakeelb@google.com>
Acked-by: Roman Gushchin <guro@fb.com>
Acked-by: Michal Hocko <mhocko@suse.com>
Cc: Vladimir Davydov <vdavydov.dev@gmail.com>
Cc: Xiongchun Duan <duanxiongchun@bytedance.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/linux/memcontrol.h | 31 +++++++------------------------
1 file changed, 7 insertions(+), 24 deletions(-)
--- a/include/linux/memcontrol.h~mm-memcontrol-simplify-lruvec_holds_page_lru_lock
+++ a/include/linux/memcontrol.h
@@ -755,22 +755,6 @@ static inline struct lruvec *mem_cgroup_
return mem_cgroup_lruvec(memcg, pgdat);
}
-static inline bool lruvec_holds_page_lru_lock(struct page *page,
- struct lruvec *lruvec)
-{
- pg_data_t *pgdat = page_pgdat(page);
- const struct mem_cgroup *memcg;
- struct mem_cgroup_per_node *mz;
-
- if (mem_cgroup_disabled())
- return lruvec == &pgdat->__lruvec;
-
- mz = container_of(lruvec, struct mem_cgroup_per_node, lruvec);
- memcg = page_memcg(page) ? : root_mem_cgroup;
-
- return lruvec->pgdat == pgdat && mz->memcg == memcg;
-}
-
struct mem_cgroup *mem_cgroup_from_task(struct task_struct *p);
struct mem_cgroup *get_mem_cgroup_from_mm(struct mm_struct *mm);
@@ -1227,14 +1211,6 @@ static inline struct lruvec *mem_cgroup_
return &pgdat->__lruvec;
}
-static inline bool lruvec_holds_page_lru_lock(struct page *page,
- struct lruvec *lruvec)
-{
- pg_data_t *pgdat = page_pgdat(page);
-
- return lruvec == &pgdat->__lruvec;
-}
-
static inline void lruvec_memcg_debug(struct lruvec *lruvec, struct page *page)
{
}
@@ -1516,6 +1492,13 @@ static inline void unlock_page_lruvec_ir
spin_unlock_irqrestore(&lruvec->lru_lock, flags);
}
+static inline bool lruvec_holds_page_lru_lock(struct page *page,
+ struct lruvec *lruvec)
+{
+ return lruvec_pgdat(lruvec) == page_pgdat(page) &&
+ lruvec_memcg(lruvec) == page_memcg(page);
+}
+
/* Don't lock again iff page's lruvec locked */
static inline struct lruvec *relock_page_lruvec_irq(struct page *page,
struct lruvec *locked_lruvec)
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 089/192] mm: memcontrol: rename lruvec_holds_page_lru_lock to page_matches_lruvec
2021-06-29 2:32 incoming Andrew Morton
` (87 preceding siblings ...)
2021-06-29 2:37 ` [patch 088/192] mm: memcontrol: simplify lruvec_holds_page_lru_lock Andrew Morton
@ 2021-06-29 2:37 ` Andrew Morton
2021-06-29 2:38 ` [patch 090/192] mm: memcontrol: simplify the logic of objcg pinning memcg Andrew Morton
` (102 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:37 UTC (permalink / raw)
To: akpm, duanxiongchun, guro, hannes, linux-mm, mhocko, mm-commits,
shakeelb, songmuchun, torvalds, vdavydov.dev
From: Muchun Song <songmuchun@bytedance.com>
Subject: mm: memcontrol: rename lruvec_holds_page_lru_lock to page_matches_lruvec
lruvec_holds_page_lru_lock() doesn't check anything about locking and is
used to check whether the page belongs to the lruvec. So rename it to
page_matches_lruvec().
Link: https://lkml.kernel.org/r/20210417043538.9793-6-songmuchun@bytedance.com
Signed-off-by: Muchun Song <songmuchun@bytedance.com>
Acked-by: Michal Hocko <mhocko@suse.com>
Acked-by: Johannes Weiner <hannes@cmpxchg.org>
Reviewed-by: Shakeel Butt <shakeelb@google.com>
Cc: Roman Gushchin <guro@fb.com>
Cc: Vladimir Davydov <vdavydov.dev@gmail.com>
Cc: Xiongchun Duan <duanxiongchun@bytedance.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/linux/memcontrol.h | 8 ++++----
mm/vmscan.c | 2 +-
2 files changed, 5 insertions(+), 5 deletions(-)
--- a/include/linux/memcontrol.h~mm-memcontrol-rename-lruvec_holds_page_lru_lock-to-page_matches_lruvec
+++ a/include/linux/memcontrol.h
@@ -1492,8 +1492,8 @@ static inline void unlock_page_lruvec_ir
spin_unlock_irqrestore(&lruvec->lru_lock, flags);
}
-static inline bool lruvec_holds_page_lru_lock(struct page *page,
- struct lruvec *lruvec)
+/* Test requires a stable page->memcg binding, see page_memcg() */
+static inline bool page_matches_lruvec(struct page *page, struct lruvec *lruvec)
{
return lruvec_pgdat(lruvec) == page_pgdat(page) &&
lruvec_memcg(lruvec) == page_memcg(page);
@@ -1504,7 +1504,7 @@ static inline struct lruvec *relock_page
struct lruvec *locked_lruvec)
{
if (locked_lruvec) {
- if (lruvec_holds_page_lru_lock(page, locked_lruvec))
+ if (page_matches_lruvec(page, locked_lruvec))
return locked_lruvec;
unlock_page_lruvec_irq(locked_lruvec);
@@ -1518,7 +1518,7 @@ static inline struct lruvec *relock_page
struct lruvec *locked_lruvec, unsigned long *flags)
{
if (locked_lruvec) {
- if (lruvec_holds_page_lru_lock(page, locked_lruvec))
+ if (page_matches_lruvec(page, locked_lruvec))
return locked_lruvec;
unlock_page_lruvec_irqrestore(locked_lruvec, *flags);
--- a/mm/vmscan.c~mm-memcontrol-rename-lruvec_holds_page_lru_lock-to-page_matches_lruvec
+++ a/mm/vmscan.c
@@ -2063,7 +2063,7 @@ static unsigned noinline_for_stack move_
* All pages were isolated from the same lruvec (and isolation
* inhibits memcg migration).
*/
- VM_BUG_ON_PAGE(!lruvec_holds_page_lru_lock(page, lruvec), page);
+ VM_BUG_ON_PAGE(!page_matches_lruvec(page, lruvec), page);
add_page_to_lru_list(page, lruvec);
nr_pages = thp_nr_pages(page);
nr_moved += nr_pages;
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 090/192] mm: memcontrol: simplify the logic of objcg pinning memcg
2021-06-29 2:32 incoming Andrew Morton
` (88 preceding siblings ...)
2021-06-29 2:37 ` [patch 089/192] mm: memcontrol: rename lruvec_holds_page_lru_lock to page_matches_lruvec Andrew Morton
@ 2021-06-29 2:38 ` Andrew Morton
2021-06-29 2:38 ` [patch 091/192] mm: memcontrol: move obj_cgroup_uncharge_pages() out of css_set_lock Andrew Morton
` (101 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:38 UTC (permalink / raw)
To: akpm, duanxiongchun, guro, hannes, linux-mm, mhocko, mm-commits,
shakeelb, songmuchun, torvalds, vdavydov.dev
From: Muchun Song <songmuchun@bytedance.com>
Subject: mm: memcontrol: simplify the logic of objcg pinning memcg
The obj_cgroup_release() and memcg_reparent_objcgs() are serialized by the
css_set_lock. We do not need to care about objcg->memcg being released in
the process of obj_cgroup_release(). So there is no need to pin memcg
before releasing objcg. Remove those pinning logic to simplfy the code.
There are only two places that modifies the objcg->memcg. One is the
initialization to objcg->memcg in the memcg_online_kmem(), another is
objcgs reparenting in the memcg_reparent_objcgs(). It is also impossible
for the two to run in parallel. So xchg() is unnecessary and it is enough
to use WRITE_ONCE().
Link: https://lkml.kernel.org/r/20210417043538.9793-7-songmuchun@bytedance.com
Signed-off-by: Muchun Song <songmuchun@bytedance.com>
Acked-by: Johannes Weiner <hannes@cmpxchg.org>
Reviewed-by: Shakeel Butt <shakeelb@google.com>
Acked-by: Roman Gushchin <guro@fb.com>
Cc: Michal Hocko <mhocko@suse.com>
Cc: Vladimir Davydov <vdavydov.dev@gmail.com>
Cc: Xiongchun Duan <duanxiongchun@bytedance.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/memcontrol.c | 20 ++++++--------------
1 file changed, 6 insertions(+), 14 deletions(-)
--- a/mm/memcontrol.c~mm-memcontrol-simplify-the-logic-of-objcg-pinning-memcg
+++ a/mm/memcontrol.c
@@ -261,7 +261,6 @@ static void obj_cgroup_uncharge_pages(st
static void obj_cgroup_release(struct percpu_ref *ref)
{
struct obj_cgroup *objcg = container_of(ref, struct obj_cgroup, refcnt);
- struct mem_cgroup *memcg;
unsigned int nr_bytes;
unsigned int nr_pages;
unsigned long flags;
@@ -291,11 +290,9 @@ static void obj_cgroup_release(struct pe
nr_pages = nr_bytes >> PAGE_SHIFT;
spin_lock_irqsave(&css_set_lock, flags);
- memcg = obj_cgroup_memcg(objcg);
if (nr_pages)
obj_cgroup_uncharge_pages(objcg, nr_pages);
list_del(&objcg->list);
- mem_cgroup_put(memcg);
spin_unlock_irqrestore(&css_set_lock, flags);
percpu_ref_exit(ref);
@@ -330,17 +327,12 @@ static void memcg_reparent_objcgs(struct
spin_lock_irq(&css_set_lock);
- /* Move active objcg to the parent's list */
- xchg(&objcg->memcg, parent);
- css_get(&parent->css);
- list_add(&objcg->list, &parent->objcg_list);
-
- /* Move already reparented objcgs to the parent's list */
- list_for_each_entry(iter, &memcg->objcg_list, list) {
- css_get(&parent->css);
- xchg(&iter->memcg, parent);
- css_put(&memcg->css);
- }
+ /* 1) Ready to reparent active objcg. */
+ list_add(&objcg->list, &memcg->objcg_list);
+ /* 2) Reparent active objcg and already reparented objcgs to parent. */
+ list_for_each_entry(iter, &memcg->objcg_list, list)
+ WRITE_ONCE(iter->memcg, parent);
+ /* 3) Move already reparented objcgs to the parent's list */
list_splice(&memcg->objcg_list, &parent->objcg_list);
spin_unlock_irq(&css_set_lock);
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 091/192] mm: memcontrol: move obj_cgroup_uncharge_pages() out of css_set_lock
2021-06-29 2:32 incoming Andrew Morton
` (89 preceding siblings ...)
2021-06-29 2:38 ` [patch 090/192] mm: memcontrol: simplify the logic of objcg pinning memcg Andrew Morton
@ 2021-06-29 2:38 ` Andrew Morton
2021-06-29 2:38 ` [patch 092/192] mm: vmscan: remove noinline_for_stack Andrew Morton
` (100 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:38 UTC (permalink / raw)
To: akpm, duanxiongchun, guro, hannes, linux-mm, mhocko, mm-commits,
shakeelb, songmuchun, torvalds, vdavydov.dev
From: Muchun Song <songmuchun@bytedance.com>
Subject: mm: memcontrol: move obj_cgroup_uncharge_pages() out of css_set_lock
The css_set_lock is used to guard the list of inherited objcgs. So there
is no need to uncharge kernel memory under css_set_lock. Just move it out
of the lock.
Link: https://lkml.kernel.org/r/20210417043538.9793-8-songmuchun@bytedance.com
Signed-off-by: Muchun Song <songmuchun@bytedance.com>
Reviewed-by: Shakeel Butt <shakeelb@google.com>
Acked-by: Roman Gushchin <guro@fb.com>
Acked-by: Johannes Weiner <hannes@cmpxchg.org>
Cc: Michal Hocko <mhocko@suse.com>
Cc: Vladimir Davydov <vdavydov.dev@gmail.com>
Cc: Xiongchun Duan <duanxiongchun@bytedance.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/memcontrol.c | 3 ++-
1 file changed, 2 insertions(+), 1 deletion(-)
--- a/mm/memcontrol.c~mm-memcontrol-move-obj_cgroup_uncharge_pages-out-of-css_set_lock
+++ a/mm/memcontrol.c
@@ -289,9 +289,10 @@ static void obj_cgroup_release(struct pe
WARN_ON_ONCE(nr_bytes & (PAGE_SIZE - 1));
nr_pages = nr_bytes >> PAGE_SHIFT;
- spin_lock_irqsave(&css_set_lock, flags);
if (nr_pages)
obj_cgroup_uncharge_pages(objcg, nr_pages);
+
+ spin_lock_irqsave(&css_set_lock, flags);
list_del(&objcg->list);
spin_unlock_irqrestore(&css_set_lock, flags);
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 092/192] mm: vmscan: remove noinline_for_stack
2021-06-29 2:32 incoming Andrew Morton
` (90 preceding siblings ...)
2021-06-29 2:38 ` [patch 091/192] mm: memcontrol: move obj_cgroup_uncharge_pages() out of css_set_lock Andrew Morton
@ 2021-06-29 2:38 ` Andrew Morton
2021-06-29 2:38 ` [patch 093/192] memcontrol: use flexible-array member Andrew Morton
` (99 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:38 UTC (permalink / raw)
To: akpm, duanxiongchun, guro, hannes, linux-mm, mhocko, mm-commits,
shakeelb, songmuchun, torvalds, vdavydov.dev
From: Muchun Song <songmuchun@bytedance.com>
Subject: mm: vmscan: remove noinline_for_stack
The noinline_for_stack is introduced by commit 666356297ec4 ("vmscan: set
up pagevec as late as possible in shrink_inactive_list()"), its purpose is
to delay the allocation of pagevec as late as possible to save stack
memory. But the commit 2bcf88796381 ("mm: take pagevecs off reclaim
stack") replace pagevecs by lists of pages_to_free. So we do not need
noinline_for_stack, just remove it (let the compiler decide whether to
inline).
Link: https://lkml.kernel.org/r/20210417043538.9793-9-songmuchun@bytedance.com
Signed-off-by: Muchun Song <songmuchun@bytedance.com>
Acked-by: Johannes Weiner <hannes@cmpxchg.org>
Acked-by: Roman Gushchin <guro@fb.com>
Reviewed-by: Shakeel Butt <shakeelb@google.com>
Acked-by: Michal Hocko <mhocko@suse.com>
Cc: Vladimir Davydov <vdavydov.dev@gmail.com>
Cc: Xiongchun Duan <duanxiongchun@bytedance.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/vmscan.c | 6 +++---
1 file changed, 3 insertions(+), 3 deletions(-)
--- a/mm/vmscan.c~mm-vmscan-remove-noinline_for_stack
+++ a/mm/vmscan.c
@@ -2015,8 +2015,8 @@ static int too_many_isolated(struct pgli
*
* Returns the number of pages moved to the given lruvec.
*/
-static unsigned noinline_for_stack move_pages_to_lru(struct lruvec *lruvec,
- struct list_head *list)
+static unsigned int move_pages_to_lru(struct lruvec *lruvec,
+ struct list_head *list)
{
int nr_pages, nr_moved = 0;
LIST_HEAD(pages_to_free);
@@ -2096,7 +2096,7 @@ static int current_may_throttle(void)
* shrink_inactive_list() is a helper for shrink_node(). It returns the number
* of reclaimed pages
*/
-static noinline_for_stack unsigned long
+static unsigned long
shrink_inactive_list(unsigned long nr_to_scan, struct lruvec *lruvec,
struct scan_control *sc, enum lru_list lru)
{
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 093/192] memcontrol: use flexible-array member
2021-06-29 2:32 incoming Andrew Morton
` (91 preceding siblings ...)
2021-06-29 2:38 ` [patch 092/192] mm: vmscan: remove noinline_for_stack Andrew Morton
@ 2021-06-29 2:38 ` Andrew Morton
2021-06-29 2:38 ` [patch 094/192] loop: use worker per cgroup instead of kworker Andrew Morton
` (98 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:38 UTC (permalink / raw)
To: akpm, alexander.h.duyck, alexs, guro, hannes, linux-mm, mhocko,
mm-commits, richard.weiyang, shakeelb, shy828301, songmuchun,
torvalds, wenhui
From: wenhuizhang <wenhui@gwmail.gwu.edu>
Subject: memcontrol: use flexible-array member
Change deprecated zero-length-and-one-element-arrays into flexible array
member.Zero-length and one-element arrays detected by Lukas's CodeChecker.
Zero/one element arrays cause undefined behaviours if sizeof() used.
Link: https://lkml.kernel.org/r/20210518200910.29912-1-wenhui@gwmail.gwu.edu
Signed-off-by: wenhuizhang <wenhui@gwmail.gwu.edu>
Reviewed-by: Muchun Song <songmuchun@bytedance.com>
Acked-by: Michal Hocko <mhocko@suse.com>
Cc: Shakeel Butt <shakeelb@google.com>
Cc: Johannes Weiner <hannes@cmpxchg.org>
Cc: Roman Gushchin <guro@fb.com>
Cc: Muchun Song <songmuchun@bytedance.com>
Cc: Yang Shi <shy828301@gmail.com>
Cc: Alex Shi <alexs@kernel.org>
Cc: Alexander Duyck <alexander.h.duyck@linux.intel.com>
Cc: Wei Yang <richard.weiyang@gmail.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/linux/memcontrol.h | 3 +--
1 file changed, 1 insertion(+), 2 deletions(-)
--- a/include/linux/memcontrol.h~memcontrol-use-flexible-array-member
+++ a/include/linux/memcontrol.h
@@ -349,8 +349,7 @@ struct mem_cgroup {
struct deferred_split deferred_split_queue;
#endif
- struct mem_cgroup_per_node *nodeinfo[0];
- /* WARNING: nodeinfo must be the last member here */
+ struct mem_cgroup_per_node *nodeinfo[];
};
/*
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 094/192] loop: use worker per cgroup instead of kworker
2021-06-29 2:32 incoming Andrew Morton
` (92 preceding siblings ...)
2021-06-29 2:38 ` [patch 093/192] memcontrol: use flexible-array member Andrew Morton
@ 2021-06-29 2:38 ` Andrew Morton
2021-06-29 2:38 ` [patch 095/192] mm: charge active memcg when no mm is set Andrew Morton
` (97 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:38 UTC (permalink / raw)
To: akpm, axboe, chris, hannes, linux-mm, mhocko, ming.lei,
mm-commits, schatzberg.dan, shakeelb, tj, torvalds
From: Dan Schatzberg <schatzberg.dan@gmail.com>
Subject: loop: use worker per cgroup instead of kworker
Patch series "Charge loop device i/o to issuing cgroup", v14.
The loop device runs all i/o to the backing file on a separate kworker
thread which results in all i/o being charged to the root cgroup. This
allows a loop device to be used to trivially bypass resource limits and
other policy. This patch series fixes this gap in accounting.
A simple script to demonstrate this behavior on cgroupv2 machine:
'''
#!/bin/bash
set -e
CGROUP=/sys/fs/cgroup/test.slice
LOOP_DEV=/dev/loop0
if [[ ! -d $CGROUP ]]
then
sudo mkdir $CGROUP
fi
grep oom_kill $CGROUP/memory.events
# Set a memory limit, write more than that limit to tmpfs -> OOM kill
sudo unshare -m bash -c "
echo \$\$ > $CGROUP/cgroup.procs;
echo 0 > $CGROUP/memory.swap.max;
echo 64M > $CGROUP/memory.max;
mount -t tmpfs -o size=512m tmpfs /tmp;
dd if=/dev/zero of=/tmp/file bs=1M count=256" || true
grep oom_kill $CGROUP/memory.events
# Set a memory limit, write more than that limit through loopback
# device -> no OOM kill
sudo unshare -m bash -c "
echo \$\$ > $CGROUP/cgroup.procs;
echo 0 > $CGROUP/memory.swap.max;
echo 64M > $CGROUP/memory.max;
mount -t tmpfs -o size=512m tmpfs /tmp;
truncate -s 512m /tmp/backing_file
losetup $LOOP_DEV /tmp/backing_file
dd if=/dev/zero of=$LOOP_DEV bs=1M count=256;
losetup -D $LOOP_DEV" || true
grep oom_kill $CGROUP/memory.events
'''
Naively charging cgroups could result in priority inversions through the
single kworker thread in the case where multiple cgroups are
reading/writing to the same loop device. This patch series does some
minor modification to the loop driver so that each cgroup can make forward
progress independently to avoid this inversion.
With this patch series applied, the above script triggers OOM kills when
writing through the loop device as expected.
This patch (of 3):
Existing uses of loop device may have multiple cgroups reading/writing to
the same device. Simply charging resources for I/O to the backing file
could result in priority inversion where one cgroup gets synchronously
blocked, holding up all other I/O to the loop device.
In order to avoid this priority inversion, we use a single workqueue where
each work item is a "struct loop_worker" which contains a queue of struct
loop_cmds to issue. The loop device maintains a tree mapping blk css_id
-> loop_worker. This allows each cgroup to independently make forward
progress issuing I/O to the backing file.
There is also a single queue for I/O associated with the rootcg which can
be used in cases of extreme memory shortage where we cannot allocate a
loop_worker.
The locking for the tree and queues is fairly heavy handed - we acquire a
per-loop-device spinlock any time either is accessed. The existing
implementation serializes all I/O through a single thread anyways, so I
don't believe this is any worse.
[colin.king@canonical.com: fixes]
Link: https://lkml.kernel.org/r/20210610173944.1203706-1-schatzberg.dan@gmail.com
Link: https://lkml.kernel.org/r/20210610173944.1203706-2-schatzberg.dan@gmail.com
Signed-off-by: Dan Schatzberg <schatzberg.dan@gmail.com>
Reviewed-by: Ming Lei <ming.lei@redhat.com>
Acked-by: Jens Axboe <axboe@kernel.dk>
Cc: Johannes Weiner <hannes@cmpxchg.org>
Cc: Michal Hocko <mhocko@suse.com>
Cc: Chris Down <chris@chrisdown.name>
Cc: Shakeel Butt <shakeelb@google.com>
Cc: Tejun Heo <tj@kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
drivers/block/loop.c | 209 +++++++++++++++++++++++++++++++++++------
drivers/block/loop.h | 12 +-
2 files changed, 187 insertions(+), 34 deletions(-)
--- a/drivers/block/loop.c~loop-use-worker-per-cgroup-instead-of-kworker
+++ a/drivers/block/loop.c
@@ -71,7 +71,6 @@
#include <linux/writeback.h>
#include <linux/completion.h>
#include <linux/highmem.h>
-#include <linux/kthread.h>
#include <linux/splice.h>
#include <linux/sysfs.h>
#include <linux/miscdevice.h>
@@ -84,6 +83,8 @@
#include <linux/uaccess.h>
+#define LOOP_IDLE_WORKER_TIMEOUT (60 * HZ)
+
static DEFINE_IDR(loop_index_idr);
static DEFINE_MUTEX(loop_ctl_mutex);
@@ -921,27 +922,95 @@ static void loop_config_discard(struct l
q->limits.discard_alignment = 0;
}
-static void loop_unprepare_queue(struct loop_device *lo)
+struct loop_worker {
+ struct rb_node rb_node;
+ struct work_struct work;
+ struct list_head cmd_list;
+ struct list_head idle_list;
+ struct loop_device *lo;
+ struct cgroup_subsys_state *css;
+ unsigned long last_ran_at;
+};
+
+static void loop_workfn(struct work_struct *work);
+static void loop_rootcg_workfn(struct work_struct *work);
+static void loop_free_idle_workers(struct timer_list *timer);
+
+#ifdef CONFIG_BLK_CGROUP
+static inline int queue_on_root_worker(struct cgroup_subsys_state *css)
{
- kthread_flush_worker(&lo->worker);
- kthread_stop(lo->worker_task);
+ return !css || css == blkcg_root_css;
}
-
-static int loop_kthread_worker_fn(void *worker_ptr)
+#else
+static inline int queue_on_root_worker(struct cgroup_subsys_state *css)
{
- current->flags |= PF_LOCAL_THROTTLE | PF_MEMALLOC_NOIO;
- return kthread_worker_fn(worker_ptr);
+ return !css;
}
+#endif
-static int loop_prepare_queue(struct loop_device *lo)
+static void loop_queue_work(struct loop_device *lo, struct loop_cmd *cmd)
{
- kthread_init_worker(&lo->worker);
- lo->worker_task = kthread_run(loop_kthread_worker_fn,
- &lo->worker, "loop%d", lo->lo_number);
- if (IS_ERR(lo->worker_task))
- return -ENOMEM;
- set_user_nice(lo->worker_task, MIN_NICE);
- return 0;
+ struct rb_node **node = &(lo->worker_tree.rb_node), *parent = NULL;
+ struct loop_worker *cur_worker, *worker = NULL;
+ struct work_struct *work;
+ struct list_head *cmd_list;
+
+ spin_lock_irq(&lo->lo_work_lock);
+
+ if (queue_on_root_worker(cmd->css))
+ goto queue_work;
+
+ node = &lo->worker_tree.rb_node;
+
+ while (*node) {
+ parent = *node;
+ cur_worker = container_of(*node, struct loop_worker, rb_node);
+ if (cur_worker->css == cmd->css) {
+ worker = cur_worker;
+ break;
+ } else if ((long)cur_worker->css < (long)cmd->css) {
+ node = &(*node)->rb_left;
+ } else {
+ node = &(*node)->rb_right;
+ }
+ }
+ if (worker)
+ goto queue_work;
+
+ worker = kzalloc(sizeof(struct loop_worker), GFP_NOWAIT | __GFP_NOWARN);
+ /*
+ * In the event we cannot allocate a worker, just queue on the
+ * rootcg worker
+ */
+ if (!worker)
+ goto queue_work;
+
+ worker->css = cmd->css;
+ css_get(worker->css);
+ INIT_WORK(&worker->work, loop_workfn);
+ INIT_LIST_HEAD(&worker->cmd_list);
+ INIT_LIST_HEAD(&worker->idle_list);
+ worker->lo = lo;
+ rb_link_node(&worker->rb_node, parent, node);
+ rb_insert_color(&worker->rb_node, &lo->worker_tree);
+queue_work:
+ if (worker) {
+ /*
+ * We need to remove from the idle list here while
+ * holding the lock so that the idle timer doesn't
+ * free the worker
+ */
+ if (!list_empty(&worker->idle_list))
+ list_del_init(&worker->idle_list);
+ work = &worker->work;
+ cmd_list = &worker->cmd_list;
+ } else {
+ work = &lo->rootcg_work;
+ cmd_list = &lo->rootcg_cmd_list;
+ }
+ list_add_tail(&cmd->list_entry, cmd_list);
+ queue_work(lo->workqueue, work);
+ spin_unlock_irq(&lo->lo_work_lock);
}
static void loop_update_rotational(struct loop_device *lo)
@@ -1127,12 +1196,23 @@ static int loop_configure(struct loop_de
!file->f_op->write_iter)
lo->lo_flags |= LO_FLAGS_READ_ONLY;
- error = loop_prepare_queue(lo);
- if (error)
+ lo->workqueue = alloc_workqueue("loop%d",
+ WQ_UNBOUND | WQ_FREEZABLE,
+ 0,
+ lo->lo_number);
+ if (!lo->workqueue) {
+ error = -ENOMEM;
goto out_unlock;
+ }
set_disk_ro(lo->lo_disk, (lo->lo_flags & LO_FLAGS_READ_ONLY) != 0);
+ INIT_WORK(&lo->rootcg_work, loop_rootcg_workfn);
+ INIT_LIST_HEAD(&lo->rootcg_cmd_list);
+ INIT_LIST_HEAD(&lo->idle_worker_list);
+ lo->worker_tree = RB_ROOT;
+ timer_setup(&lo->timer, loop_free_idle_workers,
+ TIMER_DEFERRABLE);
lo->use_dio = lo->lo_flags & LO_FLAGS_DIRECT_IO;
lo->lo_device = bdev;
lo->lo_backing_file = file;
@@ -1200,6 +1280,7 @@ static int __loop_clr_fd(struct loop_dev
int err = 0;
bool partscan = false;
int lo_number;
+ struct loop_worker *pos, *worker;
mutex_lock(&lo->lo_mutex);
if (WARN_ON_ONCE(lo->lo_state != Lo_rundown)) {
@@ -1219,6 +1300,18 @@ static int __loop_clr_fd(struct loop_dev
/* freeze request queue during the transition */
blk_mq_freeze_queue(lo->lo_queue);
+ destroy_workqueue(lo->workqueue);
+ spin_lock_irq(&lo->lo_work_lock);
+ list_for_each_entry_safe(worker, pos, &lo->idle_worker_list,
+ idle_list) {
+ list_del(&worker->idle_list);
+ rb_erase(&worker->rb_node, &lo->worker_tree);
+ css_put(worker->css);
+ kfree(worker);
+ }
+ spin_unlock_irq(&lo->lo_work_lock);
+ del_timer_sync(&lo->timer);
+
spin_lock_irq(&lo->lo_lock);
lo->lo_backing_file = NULL;
spin_unlock_irq(&lo->lo_lock);
@@ -1255,7 +1348,6 @@ static int __loop_clr_fd(struct loop_dev
partscan = lo->lo_flags & LO_FLAGS_PARTSCAN && bdev;
lo_number = lo->lo_number;
- loop_unprepare_queue(lo);
out_unlock:
mutex_unlock(&lo->lo_mutex);
if (partscan) {
@@ -2015,7 +2107,7 @@ static blk_status_t loop_queue_rq(struct
} else
#endif
cmd->css = NULL;
- kthread_queue_work(&lo->worker, &cmd->work);
+ loop_queue_work(lo, cmd);
return BLK_STS_OK;
}
@@ -2045,26 +2137,82 @@ static void loop_handle_cmd(struct loop_
}
}
-static void loop_queue_work(struct kthread_work *work)
+static void loop_set_timer(struct loop_device *lo)
+{
+ timer_reduce(&lo->timer, jiffies + LOOP_IDLE_WORKER_TIMEOUT);
+}
+
+static void loop_process_work(struct loop_worker *worker,
+ struct list_head *cmd_list, struct loop_device *lo)
{
- struct loop_cmd *cmd =
- container_of(work, struct loop_cmd, work);
+ int orig_flags = current->flags;
+ struct loop_cmd *cmd;
+
+ current->flags |= PF_LOCAL_THROTTLE | PF_MEMALLOC_NOIO;
+ spin_lock_irq(&lo->lo_work_lock);
+ while (!list_empty(cmd_list)) {
+ cmd = container_of(
+ cmd_list->next, struct loop_cmd, list_entry);
+ list_del(cmd_list->next);
+ spin_unlock_irq(&lo->lo_work_lock);
+
+ loop_handle_cmd(cmd);
+ cond_resched();
+
+ spin_lock_irq(&lo->lo_work_lock);
+ }
- loop_handle_cmd(cmd);
+ /*
+ * We only add to the idle list if there are no pending cmds
+ * *and* the worker will not run again which ensures that it
+ * is safe to free any worker on the idle list
+ */
+ if (worker && !work_pending(&worker->work)) {
+ worker->last_ran_at = jiffies;
+ list_add_tail(&worker->idle_list, &lo->idle_worker_list);
+ loop_set_timer(lo);
+ }
+ spin_unlock_irq(&lo->lo_work_lock);
+ current->flags = orig_flags;
}
-static int loop_init_request(struct blk_mq_tag_set *set, struct request *rq,
- unsigned int hctx_idx, unsigned int numa_node)
+static void loop_workfn(struct work_struct *work)
{
- struct loop_cmd *cmd = blk_mq_rq_to_pdu(rq);
+ struct loop_worker *worker =
+ container_of(work, struct loop_worker, work);
+ loop_process_work(worker, &worker->cmd_list, worker->lo);
+}
- kthread_init_work(&cmd->work, loop_queue_work);
- return 0;
+static void loop_rootcg_workfn(struct work_struct *work)
+{
+ struct loop_device *lo =
+ container_of(work, struct loop_device, rootcg_work);
+ loop_process_work(NULL, &lo->rootcg_cmd_list, lo);
+}
+
+static void loop_free_idle_workers(struct timer_list *timer)
+{
+ struct loop_device *lo = container_of(timer, struct loop_device, timer);
+ struct loop_worker *pos, *worker;
+
+ spin_lock_irq(&lo->lo_work_lock);
+ list_for_each_entry_safe(worker, pos, &lo->idle_worker_list,
+ idle_list) {
+ if (time_is_after_jiffies(worker->last_ran_at +
+ LOOP_IDLE_WORKER_TIMEOUT))
+ break;
+ list_del(&worker->idle_list);
+ rb_erase(&worker->rb_node, &lo->worker_tree);
+ css_put(worker->css);
+ kfree(worker);
+ }
+ if (!list_empty(&lo->idle_worker_list))
+ loop_set_timer(lo);
+ spin_unlock_irq(&lo->lo_work_lock);
}
static const struct blk_mq_ops loop_mq_ops = {
.queue_rq = loop_queue_rq,
- .init_request = loop_init_request,
.complete = lo_complete_rq,
};
@@ -2153,6 +2301,7 @@ static int loop_add(struct loop_device *
mutex_init(&lo->lo_mutex);
lo->lo_number = i;
spin_lock_init(&lo->lo_lock);
+ spin_lock_init(&lo->lo_work_lock);
disk->major = LOOP_MAJOR;
disk->first_minor = i << part_shift;
disk->fops = &lo_fops;
--- a/drivers/block/loop.h~loop-use-worker-per-cgroup-instead-of-kworker
+++ a/drivers/block/loop.h
@@ -14,7 +14,6 @@
#include <linux/blk-mq.h>
#include <linux/spinlock.h>
#include <linux/mutex.h>
-#include <linux/kthread.h>
#include <uapi/linux/loop.h>
/* Possible states of device */
@@ -55,8 +54,13 @@ struct loop_device {
spinlock_t lo_lock;
int lo_state;
- struct kthread_worker worker;
- struct task_struct *worker_task;
+ spinlock_t lo_work_lock;
+ struct workqueue_struct *workqueue;
+ struct work_struct rootcg_work;
+ struct list_head rootcg_cmd_list;
+ struct list_head idle_worker_list;
+ struct rb_root worker_tree;
+ struct timer_list timer;
bool use_dio;
bool sysfs_inited;
@@ -67,7 +71,7 @@ struct loop_device {
};
struct loop_cmd {
- struct kthread_work work;
+ struct list_head list_entry;
bool use_aio; /* use AIO interface to handle I/O */
atomic_t ref; /* only for aio */
long ret;
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 095/192] mm: charge active memcg when no mm is set
2021-06-29 2:32 incoming Andrew Morton
` (93 preceding siblings ...)
2021-06-29 2:38 ` [patch 094/192] loop: use worker per cgroup instead of kworker Andrew Morton
@ 2021-06-29 2:38 ` Andrew Morton
2021-06-29 2:38 ` [patch 096/192] loop: charge i/o to mem and blk cg Andrew Morton
` (96 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:38 UTC (permalink / raw)
To: akpm, axboe, chris, hannes, linux-mm, mhocko, ming.lei, mkoutny,
mm-commits, schatzberg.dan, shakeelb, tj, torvalds
From: Dan Schatzberg <schatzberg.dan@gmail.com>
Subject: mm: charge active memcg when no mm is set
set_active_memcg() worked for kernel allocations but was silently ignored
for user pages.
This patch establishes a precedence order for who gets charged:
1. If there is a memcg associated with the page already, that memcg is
charged. This happens during swapin.
2. If an explicit mm is passed, mm->memcg is charged. This happens
during page faults, which can be triggered in remote VMs (eg gup).
3. Otherwise consult the current process context. If there is an
active_memcg, use that. Otherwise, current->mm->memcg.
Previously, if a NULL mm was passed to mem_cgroup_charge (case 3) it would
always charge the root cgroup. Now it looks up the active_memcg first
(falling back to charging the root cgroup if not set).
Link: https://lkml.kernel.org/r/20210610173944.1203706-3-schatzberg.dan@gmail.com
Signed-off-by: Dan Schatzberg <schatzberg.dan@gmail.com>
Acked-by: Johannes Weiner <hannes@cmpxchg.org>
Acked-by: Tejun Heo <tj@kernel.org>
Acked-by: Chris Down <chris@chrisdown.name>
Acked-by: Jens Axboe <axboe@kernel.dk>
Reviewed-by: Shakeel Butt <shakeelb@google.com>
Reviewed-by: Michal Koutný <mkoutny@suse.com>
Cc: Michal Hocko <mhocko@suse.com>
Cc: Ming Lei <ming.lei@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/filemap.c | 2 +-
mm/memcontrol.c | 41 +++++++++++++++++++++++++++--------------
mm/shmem.c | 4 ++--
3 files changed, 30 insertions(+), 17 deletions(-)
--- a/mm/filemap.c~mm-charge-active-memcg-when-no-mm-is-set
+++ a/mm/filemap.c
@@ -872,7 +872,7 @@ noinline int __add_to_page_cache_locked(
page->index = offset;
if (!huge) {
- error = mem_cgroup_charge(page, current->mm, gfp);
+ error = mem_cgroup_charge(page, NULL, gfp);
if (error)
goto error;
charged = true;
--- a/mm/memcontrol.c~mm-charge-active-memcg-when-no-mm-is-set
+++ a/mm/memcontrol.c
@@ -897,13 +897,24 @@ struct mem_cgroup *mem_cgroup_from_task(
}
EXPORT_SYMBOL(mem_cgroup_from_task);
+static __always_inline struct mem_cgroup *active_memcg(void)
+{
+ if (in_interrupt())
+ return this_cpu_read(int_active_memcg);
+ else
+ return current->active_memcg;
+}
+
/**
* get_mem_cgroup_from_mm: Obtain a reference on given mm_struct's memcg.
* @mm: mm from which memcg should be extracted. It can be NULL.
*
- * Obtain a reference on mm->memcg and returns it if successful. Otherwise
- * root_mem_cgroup is returned. However if mem_cgroup is disabled, NULL is
- * returned.
+ * Obtain a reference on mm->memcg and returns it if successful. If mm
+ * is NULL, then the memcg is chosen as follows:
+ * 1) The active memcg, if set.
+ * 2) current->mm->memcg, if available
+ * 3) root memcg
+ * If mem_cgroup is disabled, NULL is returned.
*/
struct mem_cgroup *get_mem_cgroup_from_mm(struct mm_struct *mm)
{
@@ -921,8 +932,17 @@ struct mem_cgroup *get_mem_cgroup_from_m
* counting is disabled on the root level in the
* cgroup core. See CSS_NO_REF.
*/
- if (unlikely(!mm))
- return root_mem_cgroup;
+ if (unlikely(!mm)) {
+ memcg = active_memcg();
+ if (unlikely(memcg)) {
+ /* remote memcg must hold a ref */
+ css_get(&memcg->css);
+ return memcg;
+ }
+ mm = current->mm;
+ if (unlikely(!mm))
+ return root_mem_cgroup;
+ }
rcu_read_lock();
do {
@@ -935,14 +955,6 @@ struct mem_cgroup *get_mem_cgroup_from_m
}
EXPORT_SYMBOL(get_mem_cgroup_from_mm);
-static __always_inline struct mem_cgroup *active_memcg(void)
-{
- if (in_interrupt())
- return this_cpu_read(int_active_memcg);
- else
- return current->active_memcg;
-}
-
static __always_inline bool memcg_kmem_bypass(void)
{
/* Allow remote memcg charging from any context. */
@@ -6711,7 +6723,8 @@ out:
* @gfp_mask: reclaim mode
*
* Try to charge @page to the memcg that @mm belongs to, reclaiming
- * pages according to @gfp_mask if necessary.
+ * pages according to @gfp_mask if necessary. if @mm is NULL, try to
+ * charge to the active memcg.
*
* Do not use this for pages allocated for swapin.
*
--- a/mm/shmem.c~mm-charge-active-memcg-when-no-mm-is-set
+++ a/mm/shmem.c
@@ -1695,7 +1695,7 @@ static int shmem_swapin_page(struct inod
{
struct address_space *mapping = inode->i_mapping;
struct shmem_inode_info *info = SHMEM_I(inode);
- struct mm_struct *charge_mm = vma ? vma->vm_mm : current->mm;
+ struct mm_struct *charge_mm = vma ? vma->vm_mm : NULL;
struct swap_info_struct *si;
struct page *page = NULL;
swp_entry_t swap;
@@ -1828,7 +1828,7 @@ repeat:
}
sbinfo = SHMEM_SB(inode->i_sb);
- charge_mm = vma ? vma->vm_mm : current->mm;
+ charge_mm = vma ? vma->vm_mm : NULL;
page = pagecache_get_page(mapping, index,
FGP_ENTRY | FGP_HEAD | FGP_LOCK, 0);
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 096/192] loop: charge i/o to mem and blk cg
2021-06-29 2:32 incoming Andrew Morton
` (94 preceding siblings ...)
2021-06-29 2:38 ` [patch 095/192] mm: charge active memcg when no mm is set Andrew Morton
@ 2021-06-29 2:38 ` Andrew Morton
2021-06-29 2:38 ` [patch 097/192] mm: memcontrol: remove trailing semicolon in macros Andrew Morton
` (95 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:38 UTC (permalink / raw)
To: akpm, axboe, chris, hannes, linux-mm, mhocko, ming.lei,
mm-commits, schatzberg.dan, shakeelb, tj, torvalds
From: Dan Schatzberg <schatzberg.dan@gmail.com>
Subject: loop: charge i/o to mem and blk cg
The current code only associates with the existing blkcg when aio is used
to access the backing file. This patch covers all types of i/o to the
backing file and also associates the memcg so if the backing file is on
tmpfs, memory is charged appropriately.
This patch also exports cgroup_get_e_css and int_active_memcg so it can be
used by the loop module.
Link: https://lkml.kernel.org/r/20210610173944.1203706-4-schatzberg.dan@gmail.com
Signed-off-by: Dan Schatzberg <schatzberg.dan@gmail.com>
Acked-by: Johannes Weiner <hannes@cmpxchg.org>
Acked-by: Jens Axboe <axboe@kernel.dk>
Cc: Chris Down <chris@chrisdown.name>
Cc: Michal Hocko <mhocko@suse.com>
Cc: Ming Lei <ming.lei@redhat.com>
Cc: Shakeel Butt <shakeelb@google.com>
Cc: Tejun Heo <tj@kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
drivers/block/loop.c | 61 +++++++++++++++++++++++------------
drivers/block/loop.h | 3 +
include/linux/memcontrol.h | 6 +++
kernel/cgroup/cgroup.c | 1
mm/memcontrol.c | 1
5 files changed, 51 insertions(+), 21 deletions(-)
--- a/drivers/block/loop.c~loop-charge-i-o-to-mem-and-blk-cg
+++ a/drivers/block/loop.c
@@ -78,6 +78,7 @@
#include <linux/uio.h>
#include <linux/ioprio.h>
#include <linux/blk-cgroup.h>
+#include <linux/sched/mm.h>
#include "loop.h"
@@ -516,8 +517,6 @@ static void lo_rw_aio_complete(struct ki
{
struct loop_cmd *cmd = container_of(iocb, struct loop_cmd, iocb);
- if (cmd->css)
- css_put(cmd->css);
cmd->ret = ret;
lo_rw_aio_do_completion(cmd);
}
@@ -578,8 +577,6 @@ static int lo_rw_aio(struct loop_device
cmd->iocb.ki_complete = lo_rw_aio_complete;
cmd->iocb.ki_flags = IOCB_DIRECT;
cmd->iocb.ki_ioprio = IOPRIO_PRIO_VALUE(IOPRIO_CLASS_NONE, 0);
- if (cmd->css)
- kthread_associate_blkcg(cmd->css);
if (rw == WRITE)
ret = call_write_iter(file, &cmd->iocb, &iter);
@@ -587,7 +584,6 @@ static int lo_rw_aio(struct loop_device
ret = call_read_iter(file, &cmd->iocb, &iter);
lo_rw_aio_do_completion(cmd);
- kthread_associate_blkcg(NULL);
if (ret != -EIOCBQUEUED)
cmd->iocb.ki_complete(&cmd->iocb, ret, 0);
@@ -928,7 +924,7 @@ struct loop_worker {
struct list_head cmd_list;
struct list_head idle_list;
struct loop_device *lo;
- struct cgroup_subsys_state *css;
+ struct cgroup_subsys_state *blkcg_css;
unsigned long last_ran_at;
};
@@ -957,7 +953,7 @@ static void loop_queue_work(struct loop_
spin_lock_irq(&lo->lo_work_lock);
- if (queue_on_root_worker(cmd->css))
+ if (queue_on_root_worker(cmd->blkcg_css))
goto queue_work;
node = &lo->worker_tree.rb_node;
@@ -965,10 +961,10 @@ static void loop_queue_work(struct loop_
while (*node) {
parent = *node;
cur_worker = container_of(*node, struct loop_worker, rb_node);
- if (cur_worker->css == cmd->css) {
+ if (cur_worker->blkcg_css == cmd->blkcg_css) {
worker = cur_worker;
break;
- } else if ((long)cur_worker->css < (long)cmd->css) {
+ } else if ((long)cur_worker->blkcg_css < (long)cmd->blkcg_css) {
node = &(*node)->rb_left;
} else {
node = &(*node)->rb_right;
@@ -980,13 +976,18 @@ static void loop_queue_work(struct loop_
worker = kzalloc(sizeof(struct loop_worker), GFP_NOWAIT | __GFP_NOWARN);
/*
* In the event we cannot allocate a worker, just queue on the
- * rootcg worker
+ * rootcg worker and issue the I/O as the rootcg
*/
- if (!worker)
+ if (!worker) {
+ cmd->blkcg_css = NULL;
+ if (cmd->memcg_css)
+ css_put(cmd->memcg_css);
+ cmd->memcg_css = NULL;
goto queue_work;
+ }
- worker->css = cmd->css;
- css_get(worker->css);
+ worker->blkcg_css = cmd->blkcg_css;
+ css_get(worker->blkcg_css);
INIT_WORK(&worker->work, loop_workfn);
INIT_LIST_HEAD(&worker->cmd_list);
INIT_LIST_HEAD(&worker->idle_list);
@@ -1306,7 +1307,7 @@ static int __loop_clr_fd(struct loop_dev
idle_list) {
list_del(&worker->idle_list);
rb_erase(&worker->rb_node, &lo->worker_tree);
- css_put(worker->css);
+ css_put(worker->blkcg_css);
kfree(worker);
}
spin_unlock_irq(&lo->lo_work_lock);
@@ -2100,13 +2101,18 @@ static blk_status_t loop_queue_rq(struct
}
/* always use the first bio's css */
+ cmd->blkcg_css = NULL;
+ cmd->memcg_css = NULL;
#ifdef CONFIG_BLK_CGROUP
- if (cmd->use_aio && rq->bio && rq->bio->bi_blkg) {
- cmd->css = &bio_blkcg(rq->bio)->css;
- css_get(cmd->css);
- } else
+ if (rq->bio && rq->bio->bi_blkg) {
+ cmd->blkcg_css = &bio_blkcg(rq->bio)->css;
+#ifdef CONFIG_MEMCG
+ cmd->memcg_css =
+ cgroup_get_e_css(cmd->blkcg_css->cgroup,
+ &memory_cgrp_subsys);
+#endif
+ }
#endif
- cmd->css = NULL;
loop_queue_work(lo, cmd);
return BLK_STS_OK;
@@ -2118,13 +2124,28 @@ static void loop_handle_cmd(struct loop_
const bool write = op_is_write(req_op(rq));
struct loop_device *lo = rq->q->queuedata;
int ret = 0;
+ struct mem_cgroup *old_memcg = NULL;
if (write && (lo->lo_flags & LO_FLAGS_READ_ONLY)) {
ret = -EIO;
goto failed;
}
+ if (cmd->blkcg_css)
+ kthread_associate_blkcg(cmd->blkcg_css);
+ if (cmd->memcg_css)
+ old_memcg = set_active_memcg(
+ mem_cgroup_from_css(cmd->memcg_css));
+
ret = do_req_filebacked(lo, rq);
+
+ if (cmd->blkcg_css)
+ kthread_associate_blkcg(NULL);
+
+ if (cmd->memcg_css) {
+ set_active_memcg(old_memcg);
+ css_put(cmd->memcg_css);
+ }
failed:
/* complete non-aio request */
if (!cmd->use_aio || ret) {
@@ -2203,7 +2224,7 @@ static void loop_free_idle_workers(struc
break;
list_del(&worker->idle_list);
rb_erase(&worker->rb_node, &lo->worker_tree);
- css_put(worker->css);
+ css_put(worker->blkcg_css);
kfree(worker);
}
if (!list_empty(&lo->idle_worker_list))
--- a/drivers/block/loop.h~loop-charge-i-o-to-mem-and-blk-cg
+++ a/drivers/block/loop.h
@@ -77,7 +77,8 @@ struct loop_cmd {
long ret;
struct kiocb iocb;
struct bio_vec *bvec;
- struct cgroup_subsys_state *css;
+ struct cgroup_subsys_state *blkcg_css;
+ struct cgroup_subsys_state *memcg_css;
};
/* Support for loadable transfer modules */
--- a/include/linux/memcontrol.h~loop-charge-i-o-to-mem-and-blk-cg
+++ a/include/linux/memcontrol.h
@@ -1230,6 +1230,12 @@ static inline struct mem_cgroup *get_mem
return NULL;
}
+static inline
+struct mem_cgroup *mem_cgroup_from_css(struct cgroup_subsys_state *css)
+{
+ return NULL;
+}
+
static inline void mem_cgroup_put(struct mem_cgroup *memcg)
{
}
--- a/kernel/cgroup/cgroup.c~loop-charge-i-o-to-mem-and-blk-cg
+++ a/kernel/cgroup/cgroup.c
@@ -577,6 +577,7 @@ out_unlock:
rcu_read_unlock();
return css;
}
+EXPORT_SYMBOL_GPL(cgroup_get_e_css);
static void cgroup_get_live(struct cgroup *cgrp)
{
--- a/mm/memcontrol.c~loop-charge-i-o-to-mem-and-blk-cg
+++ a/mm/memcontrol.c
@@ -78,6 +78,7 @@ struct mem_cgroup *root_mem_cgroup __rea
/* Active memory cgroup to use from an interrupt context */
DEFINE_PER_CPU(struct mem_cgroup *, int_active_memcg);
+EXPORT_PER_CPU_SYMBOL_GPL(int_active_memcg);
/* Socket memory accounting disabled? */
static bool cgroup_memory_nosocket;
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 097/192] mm: memcontrol: remove trailing semicolon in macros
2021-06-29 2:32 incoming Andrew Morton
` (95 preceding siblings ...)
2021-06-29 2:38 ` [patch 096/192] loop: charge i/o to mem and blk cg Andrew Morton
@ 2021-06-29 2:38 ` Andrew Morton
2021-06-29 2:38 ` [patch 098/192] perf: MAP_EXECUTABLE does not indicate VM_MAYEXEC Andrew Morton
` (94 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:38 UTC (permalink / raw)
To: akpm, denghuilong, guro, hannes, linux-mm, mhocko, mm-commits,
shakeelb, torvalds
From: Huilong Deng <denghuilong@cdjrlc.com>
Subject: mm: memcontrol: remove trailing semicolon in macros
Macros should not use a trailing semicolon.
Link: https://lkml.kernel.org/r/20210614091530.22117-1-denghuilong@cdjrlc.com
Signed-off-by: Huilong Deng <denghuilong@cdjrlc.com>
Reviewed-by: Andrew Morton <akpm@linux-foundation.org>
Reviewed-by: Shakeel Butt <shakeelb@google.com>
Cc: Roman Gushchin <guro@fb.com>
Cc: Michal Hocko <mhocko@suse.com>
Cc: Johannes Weiner <hannes@cmpxchg.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/linux/memcontrol.h | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
--- a/include/linux/memcontrol.h~mm-memcontrol-remove-trailing-semicolon-in-macros
+++ a/include/linux/memcontrol.h
@@ -192,7 +192,7 @@ enum memcg_kmem_state {
struct memcg_padding {
char x[0];
} ____cacheline_internodealigned_in_smp;
-#define MEMCG_PADDING(name) struct memcg_padding name;
+#define MEMCG_PADDING(name) struct memcg_padding name
#else
#define MEMCG_PADDING(name)
#endif
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 098/192] perf: MAP_EXECUTABLE does not indicate VM_MAYEXEC
2021-06-29 2:32 incoming Andrew Morton
` (96 preceding siblings ...)
2021-06-29 2:38 ` [patch 097/192] mm: memcontrol: remove trailing semicolon in macros Andrew Morton
@ 2021-06-29 2:38 ` Andrew Morton
2021-06-29 2:38 ` [patch 099/192] binfmt: remove in-tree usage of MAP_EXECUTABLE Andrew Morton
` (93 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:38 UTC (permalink / raw)
To: acme, akpm, alexander.shishkin, bp, catalin.marinas, david,
dzickus, ebiederm, feng.tang, gerg, hpa, jolsa, keescook,
Kevin.Brodsky, linux-mm, mark.rutland, mhocko, mingo, mm-commits,
namhyung, peterz, rppt, tglx, torvalds, viro
From: David Hildenbrand <david@redhat.com>
Subject: perf: MAP_EXECUTABLE does not indicate VM_MAYEXEC
Patch series "perf/binfmt/mm: remove in-tree usage of MAP_EXECUTABLE".
Stumbling over the history of MAP_EXECUTABLE, I noticed that we still have
some in-tree users that we can get rid of.
This patch (of 3):
Before commit e9714acf8c43 ("mm: kill vma flag VM_EXECUTABLE and
mm->num_exe_file_vmas"), VM_EXECUTABLE indicated MAP_EXECUTABLE.
MAP_EXECUTABLE is nowadays essentially ignored by the kernel and does not
relate to VM_MAYEXEC.
Link: https://lkml.kernel.org/r/20210421093453.6904-1-david@redhat.com
Link: https://lkml.kernel.org/r/20210421093453.6904-2-david@redhat.com
Fixes: f972eb63b100 ("perf: Pass protection and flags bits through mmap2 interface")
Signed-off-by: David Hildenbrand <david@redhat.com>
Cc: Peter Zijlstra <peterz@infradead.org>
Acked-by: "Eric W. Biederman" <ebiederm@xmission.com>
Reviewed-by: Kees Cook <keescook@chromium.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Cc: Ingo Molnar <mingo@redhat.com>
Cc: Borislav Petkov <bp@alien8.de>
Cc: "H. Peter Anvin" <hpa@zytor.com>
Cc: Alexander Viro <viro@zeniv.linux.org.uk>
Cc: Arnaldo Carvalho de Melo <acme@kernel.org>
Cc: Mark Rutland <mark.rutland@arm.com>
Cc: Alexander Shishkin <alexander.shishkin@linux.intel.com>
Cc: Jiri Olsa <jolsa@redhat.com>
Cc: Namhyung Kim <namhyung@kernel.org>
Cc: Greg Ungerer <gerg@linux-m68k.org>
Cc: Mike Rapoport <rppt@kernel.org>
Cc: Catalin Marinas <catalin.marinas@arm.com>
Cc: Kevin Brodsky <Kevin.Brodsky@arm.com>
Cc: Michal Hocko <mhocko@suse.com>
Cc: Feng Tang <feng.tang@intel.com>
Cc: Don Zickus <dzickus@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
kernel/events/core.c | 2 --
1 file changed, 2 deletions(-)
--- a/kernel/events/core.c~perf-map_executable-does-not-indicate-vm_mayexec
+++ a/kernel/events/core.c
@@ -8301,8 +8301,6 @@ static void perf_event_mmap_event(struct
if (vma->vm_flags & VM_DENYWRITE)
flags |= MAP_DENYWRITE;
- if (vma->vm_flags & VM_MAYEXEC)
- flags |= MAP_EXECUTABLE;
if (vma->vm_flags & VM_LOCKED)
flags |= MAP_LOCKED;
if (is_vm_hugetlb_page(vma))
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 099/192] binfmt: remove in-tree usage of MAP_EXECUTABLE
2021-06-29 2:32 incoming Andrew Morton
` (97 preceding siblings ...)
2021-06-29 2:38 ` [patch 098/192] perf: MAP_EXECUTABLE does not indicate VM_MAYEXEC Andrew Morton
@ 2021-06-29 2:38 ` Andrew Morton
2021-06-29 2:38 ` [patch 100/192] mm: ignore MAP_EXECUTABLE in ksys_mmap_pgoff() Andrew Morton
` (92 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:38 UTC (permalink / raw)
To: acme, akpm, alexander.shishkin, bp, catalin.marinas, david,
dzickus, ebiederm, feng.tang, gerg, hpa, jolsa, keescook,
Kevin.Brodsky, linux-mm, mark.rutland, mhocko, mingo, mm-commits,
namhyung, peterz, rppt, tglx, torvalds, viro
From: David Hildenbrand <david@redhat.com>
Subject: binfmt: remove in-tree usage of MAP_EXECUTABLE
Ever since commit e9714acf8c43 ("mm: kill vma flag VM_EXECUTABLE and
mm->num_exe_file_vmas"), VM_EXECUTABLE is gone and MAP_EXECUTABLE is
essentially completely ignored. Let's remove all usage of MAP_EXECUTABLE.
[akpm@linux-foundation.org: fix blooper in fs/binfmt_aout.c. per David]
Link: https://lkml.kernel.org/r/20210421093453.6904-3-david@redhat.com
Signed-off-by: David Hildenbrand <david@redhat.com>
Acked-by: "Eric W. Biederman" <ebiederm@xmission.com>
Reviewed-by: Kees Cook <keescook@chromium.org>
Cc: Alexander Shishkin <alexander.shishkin@linux.intel.com>
Cc: Alexander Viro <viro@zeniv.linux.org.uk>
Cc: Arnaldo Carvalho de Melo <acme@kernel.org>
Cc: Borislav Petkov <bp@alien8.de>
Cc: Catalin Marinas <catalin.marinas@arm.com>
Cc: Don Zickus <dzickus@redhat.com>
Cc: Feng Tang <feng.tang@intel.com>
Cc: Greg Ungerer <gerg@linux-m68k.org>
Cc: "H. Peter Anvin" <hpa@zytor.com>
Cc: Ingo Molnar <mingo@redhat.com>
Cc: Jiri Olsa <jolsa@redhat.com>
Cc: Kevin Brodsky <Kevin.Brodsky@arm.com>
Cc: Mark Rutland <mark.rutland@arm.com>
Cc: Michal Hocko <mhocko@suse.com>
Cc: Mike Rapoport <rppt@kernel.org>
Cc: Namhyung Kim <namhyung@kernel.org>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
arch/x86/ia32/ia32_aout.c | 4 ++--
fs/binfmt_aout.c | 4 ++--
fs/binfmt_elf.c | 2 +-
fs/binfmt_elf_fdpic.c | 11 ++---------
fs/binfmt_flat.c | 2 +-
5 files changed, 8 insertions(+), 15 deletions(-)
--- a/arch/x86/ia32/ia32_aout.c~binfmt-remove-in-tree-usage-of-map_executable
+++ a/arch/x86/ia32/ia32_aout.c
@@ -203,7 +203,7 @@ static int load_aout_binary(struct linux
error = vm_mmap(bprm->file, N_TXTADDR(ex), ex.a_text,
PROT_READ | PROT_EXEC,
MAP_FIXED | MAP_PRIVATE | MAP_DENYWRITE |
- MAP_EXECUTABLE | MAP_32BIT,
+ MAP_32BIT,
fd_offset);
if (error != N_TXTADDR(ex))
@@ -212,7 +212,7 @@ static int load_aout_binary(struct linux
error = vm_mmap(bprm->file, N_DATADDR(ex), ex.a_data,
PROT_READ | PROT_WRITE | PROT_EXEC,
MAP_FIXED | MAP_PRIVATE | MAP_DENYWRITE |
- MAP_EXECUTABLE | MAP_32BIT,
+ MAP_32BIT,
fd_offset + ex.a_text);
if (error != N_DATADDR(ex))
return error;
--- a/fs/binfmt_aout.c~binfmt-remove-in-tree-usage-of-map_executable
+++ a/fs/binfmt_aout.c
@@ -222,7 +222,7 @@ static int load_aout_binary(struct linux
error = vm_mmap(bprm->file, N_TXTADDR(ex), ex.a_text,
PROT_READ | PROT_EXEC,
- MAP_FIXED | MAP_PRIVATE | MAP_DENYWRITE | MAP_EXECUTABLE,
+ MAP_FIXED | MAP_PRIVATE | MAP_DENYWRITE,
fd_offset);
if (error != N_TXTADDR(ex))
@@ -230,7 +230,7 @@ static int load_aout_binary(struct linux
error = vm_mmap(bprm->file, N_DATADDR(ex), ex.a_data,
PROT_READ | PROT_WRITE | PROT_EXEC,
- MAP_FIXED | MAP_PRIVATE | MAP_DENYWRITE | MAP_EXECUTABLE,
+ MAP_FIXED | MAP_PRIVATE | MAP_DENYWRITE,
fd_offset + ex.a_text);
if (error != N_DATADDR(ex))
return error;
--- a/fs/binfmt_elf.c~binfmt-remove-in-tree-usage-of-map_executable
+++ a/fs/binfmt_elf.c
@@ -1070,7 +1070,7 @@ out_free_interp:
elf_prot = make_prot(elf_ppnt->p_flags, &arch_state,
!!interpreter, false);
- elf_flags = MAP_PRIVATE | MAP_DENYWRITE | MAP_EXECUTABLE;
+ elf_flags = MAP_PRIVATE | MAP_DENYWRITE;
vaddr = elf_ppnt->p_vaddr;
/*
--- a/fs/binfmt_elf_fdpic.c~binfmt-remove-in-tree-usage-of-map_executable
+++ a/fs/binfmt_elf_fdpic.c
@@ -928,7 +928,7 @@ static int elf_fdpic_map_file_constdisp_
{
struct elf32_fdpic_loadseg *seg;
struct elf32_phdr *phdr;
- unsigned long load_addr, base = ULONG_MAX, top = 0, maddr = 0, mflags;
+ unsigned long load_addr, base = ULONG_MAX, top = 0, maddr = 0;
int loop, ret;
load_addr = params->load_addr;
@@ -948,12 +948,8 @@ static int elf_fdpic_map_file_constdisp_
}
/* allocate one big anon block for everything */
- mflags = MAP_PRIVATE;
- if (params->flags & ELF_FDPIC_FLAG_EXECUTABLE)
- mflags |= MAP_EXECUTABLE;
-
maddr = vm_mmap(NULL, load_addr, top - base,
- PROT_READ | PROT_WRITE | PROT_EXEC, mflags, 0);
+ PROT_READ | PROT_WRITE | PROT_EXEC, MAP_PRIVATE, 0);
if (IS_ERR_VALUE(maddr))
return (int) maddr;
@@ -1046,9 +1042,6 @@ static int elf_fdpic_map_file_by_direct_
if (phdr->p_flags & PF_X) prot |= PROT_EXEC;
flags = MAP_PRIVATE | MAP_DENYWRITE;
- if (params->flags & ELF_FDPIC_FLAG_EXECUTABLE)
- flags |= MAP_EXECUTABLE;
-
maddr = 0;
switch (params->flags & ELF_FDPIC_FLAG_ARRANGEMENT) {
--- a/fs/binfmt_flat.c~binfmt-remove-in-tree-usage-of-map_executable
+++ a/fs/binfmt_flat.c
@@ -573,7 +573,7 @@ static int load_flat_file(struct linux_b
pr_debug("ROM mapping of file (we hope)\n");
textpos = vm_mmap(bprm->file, 0, text_len, PROT_READ|PROT_EXEC,
- MAP_PRIVATE|MAP_EXECUTABLE, 0);
+ MAP_PRIVATE, 0);
if (!textpos || IS_ERR_VALUE(textpos)) {
ret = textpos;
if (!textpos)
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 100/192] mm: ignore MAP_EXECUTABLE in ksys_mmap_pgoff()
2021-06-29 2:32 incoming Andrew Morton
` (98 preceding siblings ...)
2021-06-29 2:38 ` [patch 099/192] binfmt: remove in-tree usage of MAP_EXECUTABLE Andrew Morton
@ 2021-06-29 2:38 ` Andrew Morton
2021-06-29 2:38 ` [patch 101/192] mm/mmap.c: logic of find_vma_intersection repeated in __do_munmap Andrew Morton
` (91 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:38 UTC (permalink / raw)
To: acme, akpm, alexander.shishkin, bp, catalin.marinas, david,
dzickus, ebiederm, feng.tang, gerg, hpa, jolsa, keescook,
Kevin.Brodsky, linux-mm, mark.rutland, mhocko, mingo, mm-commits,
namhyung, peterz, rppt, tglx, torvalds, viro
From: David Hildenbrand <david@redhat.com>
Subject: mm: ignore MAP_EXECUTABLE in ksys_mmap_pgoff()
Let's also remove masking off MAP_EXECUTABLE from ksys_mmap_pgoff(): the
last in-tree occurrence of MAP_EXECUTABLE is now in LEGACY_MAP_MASK, which
accepts the flag e.g., for MAP_SHARED_VALIDATE; however, the flag is
ignored throughout the kernel now.
Add a comment to LEGACY_MAP_MASK stating that MAP_EXECUTABLE is ignored.
Link: https://lkml.kernel.org/r/20210421093453.6904-4-david@redhat.com
Signed-off-by: David Hildenbrand <david@redhat.com>
Acked-by: "Eric W. Biederman" <ebiederm@xmission.com>
Reviewed-by: Kees Cook <keescook@chromium.org>
Cc: Alexander Shishkin <alexander.shishkin@linux.intel.com>
Cc: Alexander Viro <viro@zeniv.linux.org.uk>
Cc: Arnaldo Carvalho de Melo <acme@kernel.org>
Cc: Borislav Petkov <bp@alien8.de>
Cc: Catalin Marinas <catalin.marinas@arm.com>
Cc: Don Zickus <dzickus@redhat.com>
Cc: Feng Tang <feng.tang@intel.com>
Cc: Greg Ungerer <gerg@linux-m68k.org>
Cc: "H. Peter Anvin" <hpa@zytor.com>
Cc: Ingo Molnar <mingo@redhat.com>
Cc: Jiri Olsa <jolsa@redhat.com>
Cc: Kevin Brodsky <Kevin.Brodsky@arm.com>
Cc: Mark Rutland <mark.rutland@arm.com>
Cc: Michal Hocko <mhocko@suse.com>
Cc: Mike Rapoport <rppt@kernel.org>
Cc: Namhyung Kim <namhyung@kernel.org>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/linux/mman.h | 2 ++
mm/mmap.c | 2 +-
mm/nommu.c | 2 +-
3 files changed, 4 insertions(+), 2 deletions(-)
--- a/include/linux/mman.h~mm-ignore-map_executable-in-ksys_mmap_pgoff
+++ a/include/linux/mman.h
@@ -31,6 +31,8 @@
/*
* The historical set of flags that all mmap implementations implicitly
* support when a ->mmap_validate() op is not provided in file_operations.
+ *
+ * MAP_EXECUTABLE is completely ignored throughout the kernel.
*/
#define LEGACY_MAP_MASK (MAP_SHARED \
| MAP_PRIVATE \
--- a/mm/mmap.c~mm-ignore-map_executable-in-ksys_mmap_pgoff
+++ a/mm/mmap.c
@@ -1633,7 +1633,7 @@ unsigned long ksys_mmap_pgoff(unsigned l
return PTR_ERR(file);
}
- flags &= ~(MAP_EXECUTABLE | MAP_DENYWRITE);
+ flags &= ~MAP_DENYWRITE;
retval = vm_mmap_pgoff(file, addr, len, prot, flags, pgoff);
out_fput:
--- a/mm/nommu.c~mm-ignore-map_executable-in-ksys_mmap_pgoff
+++ a/mm/nommu.c
@@ -1296,7 +1296,7 @@ unsigned long ksys_mmap_pgoff(unsigned l
goto out;
}
- flags &= ~(MAP_EXECUTABLE | MAP_DENYWRITE);
+ flags &= ~MAP_DENYWRITE;
retval = vm_mmap_pgoff(file, addr, len, prot, flags, pgoff);
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 101/192] mm/mmap.c: logic of find_vma_intersection repeated in __do_munmap
2021-06-29 2:32 incoming Andrew Morton
` (99 preceding siblings ...)
2021-06-29 2:38 ` [patch 100/192] mm: ignore MAP_EXECUTABLE in ksys_mmap_pgoff() Andrew Morton
@ 2021-06-29 2:38 ` Andrew Morton
2021-06-29 2:38 ` [patch 102/192] mm/mmap: introduce unlock_range() for code cleanup Andrew Morton
` (90 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:38 UTC (permalink / raw)
To: akpm, david, gmjuareztello, linux-mm, mm-commits, torvalds
From: Gonzalo Matias Juarez Tello <gmjuareztello@gmail.com>
Subject: mm/mmap.c: logic of find_vma_intersection repeated in __do_munmap
Logic of find_vma_intersection() is repeated in __do_munmap().
Also, prev is assigned a value before checking vma->vm_start >= end which
might end up on a return statement making that assignment useless.
Calling find_vma_intersection() checks that condition and returns NULL if
no vma is found, hence only the !vma check is needed in __do_munmap().
Link: https://lkml.kernel.org/r/20210409162129.18313-1-gmjuareztello@gmail.com
Signed-off-by: Gonzalo Matias Juarez Tello <gmjuareztello@gmail.com>
Reviewed-by: Andrew Morton <akpm@linux-foundation.org>
Reviewed-by: David Hildenbrand <david@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/mmap.c | 9 ++-------
1 file changed, 2 insertions(+), 7 deletions(-)
--- a/mm/mmap.c~mm-mmapc-logic-of-find_vma_intersection-repeated-in-__do_munmap
+++ a/mm/mmap.c
@@ -2828,16 +2828,11 @@ int __do_munmap(struct mm_struct *mm, un
*/
arch_unmap(mm, start, end);
- /* Find the first overlapping VMA */
- vma = find_vma(mm, start);
+ /* Find the first overlapping VMA where start < vma->vm_end */
+ vma = find_vma_intersection(mm, start, end);
if (!vma)
return 0;
prev = vma->vm_prev;
- /* we have start < vma->vm_end */
-
- /* if it doesn't overlap, we have nothing.. */
- if (vma->vm_start >= end)
- return 0;
/*
* If we need to split any vma, do it now to save pain later.
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 102/192] mm/mmap: introduce unlock_range() for code cleanup
2021-06-29 2:32 incoming Andrew Morton
` (100 preceding siblings ...)
2021-06-29 2:38 ` [patch 101/192] mm/mmap.c: logic of find_vma_intersection repeated in __do_munmap Andrew Morton
@ 2021-06-29 2:38 ` Andrew Morton
2021-06-29 2:38 ` [patch 103/192] mm/mmap: use find_vma_intersection() in do_mmap() for overlap Andrew Morton
` (89 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:38 UTC (permalink / raw)
To: akpm, dbueso, Liam.Howlett, linux-mm, mm-commits, torvalds, willy
From: Liam Howlett <liam.howlett@oracle.com>
Subject: mm/mmap: introduce unlock_range() for code cleanup
Both __do_munmap() and exit_mmap() unlock a range of VMAs using almost
identical code blocks. Replace both blocks by a static inline function.
[akpm@linux-foundation.org: tweak code layout]
Link: https://lkml.kernel.org/r/20210510211021.2797427-1-Liam.Howlett@Oracle.com
Signed-off-by: Liam R. Howlett <Liam.Howlett@Oracle.com>
Reviewed-by: Matthew Wilcox (Oracle) <willy@infradead.org>
Reviewed-by: Davidlohr Bueso <dbueso@suse.de>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/mmap.c | 39 ++++++++++++++++++++-------------------
1 file changed, 20 insertions(+), 19 deletions(-)
--- a/mm/mmap.c~mm-mmap-introduce-unlock_range-for-code-cleanup
+++ a/mm/mmap.c
@@ -2802,6 +2802,22 @@ int split_vma(struct mm_struct *mm, stru
return __split_vma(mm, vma, addr, new_below);
}
+static inline void
+unlock_range(struct vm_area_struct *start, unsigned long limit)
+{
+ struct mm_struct *mm = start->vm_mm;
+ struct vm_area_struct *tmp = start;
+
+ while (tmp && tmp->vm_start < limit) {
+ if (tmp->vm_flags & VM_LOCKED) {
+ mm->locked_vm -= vma_pages(tmp);
+ munlock_vma_pages_all(tmp);
+ }
+
+ tmp = tmp->vm_next;
+ }
+}
+
/* Munmap is split into 2 main parts -- this part which finds
* what needs doing, and the areas themselves, which do the
* work. This now handles partial unmappings.
@@ -2885,17 +2901,8 @@ int __do_munmap(struct mm_struct *mm, un
/*
* unlock any mlock()ed ranges before detaching vmas
*/
- if (mm->locked_vm) {
- struct vm_area_struct *tmp = vma;
- while (tmp && tmp->vm_start < end) {
- if (tmp->vm_flags & VM_LOCKED) {
- mm->locked_vm -= vma_pages(tmp);
- munlock_vma_pages_all(tmp);
- }
-
- tmp = tmp->vm_next;
- }
- }
+ if (mm->locked_vm)
+ unlock_range(vma, end);
/* Detach vmas from rbtree */
if (!detach_vmas_to_be_unmapped(mm, vma, prev, end))
@@ -3180,14 +3187,8 @@ void exit_mmap(struct mm_struct *mm)
mmap_write_unlock(mm);
}
- if (mm->locked_vm) {
- vma = mm->mmap;
- while (vma) {
- if (vma->vm_flags & VM_LOCKED)
- munlock_vma_pages_all(vma);
- vma = vma->vm_next;
- }
- }
+ if (mm->locked_vm)
+ unlock_range(mm->mmap, ULONG_MAX);
arch_exit_mmap(mm);
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 103/192] mm/mmap: use find_vma_intersection() in do_mmap() for overlap
2021-06-29 2:32 incoming Andrew Morton
` (101 preceding siblings ...)
2021-06-29 2:38 ` [patch 102/192] mm/mmap: introduce unlock_range() for code cleanup Andrew Morton
@ 2021-06-29 2:38 ` Andrew Morton
2021-06-29 2:38 ` [patch 104/192] mm/memory.c: fix comment of finish_mkwrite_fault() Andrew Morton
` (88 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:38 UTC (permalink / raw)
To: akpm, Liam.Howlett, linux-mm, mm-commits, torvalds
From: Liam Howlett <liam.howlett@oracle.com>
Subject: mm/mmap: use find_vma_intersection() in do_mmap() for overlap
Using find_vma_intersection() avoids the need for a temporary variable and
makes the code cleaner.
Link: https://lkml.kernel.org/r/20210511014328.2902782-1-Liam.Howlett@Oracle.com
Signed-off-by: Liam R. Howlett <Liam.Howlett@Oracle.com>
Reviewed-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/mmap.c | 4 +---
1 file changed, 1 insertion(+), 3 deletions(-)
--- a/mm/mmap.c~mm-mmap-use-find_vma_intersection-in-do_mmap-for-overlap
+++ a/mm/mmap.c
@@ -1457,9 +1457,7 @@ unsigned long do_mmap(struct file *file,
return addr;
if (flags & MAP_FIXED_NOREPLACE) {
- struct vm_area_struct *vma = find_vma(mm, addr);
-
- if (vma && vma->vm_start < addr + len)
+ if (find_vma_intersection(mm, addr, addr + len))
return -EEXIST;
}
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 104/192] mm/memory.c: fix comment of finish_mkwrite_fault()
2021-06-29 2:32 incoming Andrew Morton
` (102 preceding siblings ...)
2021-06-29 2:38 ` [patch 103/192] mm/mmap: use find_vma_intersection() in do_mmap() for overlap Andrew Morton
@ 2021-06-29 2:38 ` Andrew Morton
2021-06-29 2:38 ` [patch 105/192] mm: add vma_lookup(), update find_vma_intersection() comments Andrew Morton
` (87 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:38 UTC (permalink / raw)
To: akpm, linux-mm, liu.xiang, mm-commits, torvalds
From: Liu Xiang <liu.xiang@zlingsmart.com>
Subject: mm/memory.c: fix comment of finish_mkwrite_fault()
Fix the return value in comment of finish_mkwrite_fault().
Link: https://lkml.kernel.org/r/20210513093931.15234-1-liu.xiang@zlingsmart.com
Signed-off-by: Liu Xiang <liu.xiang@zlingsmart.com>
Reviewed-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/memory.c | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
--- a/mm/memory.c~mm-memoryc-fix-comment-of-finish_mkwrite_fault
+++ a/mm/memory.c
@@ -3049,7 +3049,7 @@ oom:
* The function expects the page to be locked or other protection against
* concurrent faults / writeback (such as DAX radix tree locks).
*
- * Return: %VM_FAULT_WRITE on success, %0 when PTE got changed before
+ * Return: %0 on success, %VM_FAULT_NOPAGE when PTE got changed before
* we acquired PTE lock.
*/
vm_fault_t finish_mkwrite_fault(struct vm_fault *vmf)
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 105/192] mm: add vma_lookup(), update find_vma_intersection() comments
2021-06-29 2:32 incoming Andrew Morton
` (103 preceding siblings ...)
2021-06-29 2:38 ` [patch 104/192] mm/memory.c: fix comment of finish_mkwrite_fault() Andrew Morton
@ 2021-06-29 2:38 ` Andrew Morton
2021-06-29 2:38 ` [patch 106/192] drm/i915/selftests: use vma_lookup() in __igt_mmap() Andrew Morton
` (86 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:38 UTC (permalink / raw)
To: akpm, davem, david, dbueso, geert, ldufour, Liam.Howlett,
linux-mm, mm-commits, torvalds
From: Liam Howlett <liam.howlett@oracle.com>
Subject: mm: add vma_lookup(), update find_vma_intersection() comments
Patch series "mm: Add vma_lookup()", v2.
Many places in the kernel use find_vma() to get a vma and then check the
start address of the vma to ensure the next vma was not returned.
Other places use the find_vma_intersection() call with add, addr + 1 as
the range; looking for just the vma at a specific address.
The third use of find_vma() is by developers who do not know that the
function starts searching at the provided address upwards for the next
vma. This results in a bug that is often overlooked for a long time.
Adding the new vma_lookup() function will allow for cleaner code by
removing the find_vma() calls which check limits, making
find_vma_intersection() calls of a single address to be shorter, and
potentially reduce the incorrect uses of find_vma().
This patch (of 22):
Many places in the kernel use find_vma() to get a vma and then check the
start address of the vma to ensure the next vma was not returned.
Other places use the find_vma_intersection() call with add, addr + 1 as
the range; looking for just the vma at a specific address.
The third use of find_vma() is by developers who do not know that the
function starts searching at the provided address upwards for the next
vma. This results in a bug that is often overlooked for a long time.
Adding the new vma_lookup() function will allow for cleaner code by
removing the find_vma() calls which check limits, making
find_vma_intersection() calls of a single address to be shorter, and
potentially reduce the incorrect uses of find_vma().
Also change find_vma_intersection() comments and declaration to be of the
correct length and add kernel documentation style comment.
Link: https://lkml.kernel.org/r/20210521174745.2219620-1-Liam.Howlett@Oracle.com
Link: https://lkml.kernel.org/r/20210521174745.2219620-2-Liam.Howlett@Oracle.com
Signed-off-by: Liam R. Howlett <Liam.Howlett@Oracle.com>
Reviewed-by: Laurent Dufour <ldufour@linux.ibm.com>
Acked-by: David Hildenbrand <david@redhat.com>
Acked-by: Davidlohr Bueso <dbueso@suse.de>
Cc: David Miller <davem@davemloft.net>
Cc: Geert Uytterhoeven <geert@linux-m68k.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/linux/mm.h | 36 ++++++++++++++++++++++++++++++++----
1 file changed, 32 insertions(+), 4 deletions(-)
--- a/include/linux/mm.h~mm-add-vma_lookup-update-find_vma_intersection-comments
+++ a/include/linux/mm.h
@@ -2676,17 +2676,45 @@ extern struct vm_area_struct * find_vma(
extern struct vm_area_struct * find_vma_prev(struct mm_struct * mm, unsigned long addr,
struct vm_area_struct **pprev);
-/* Look up the first VMA which intersects the interval start_addr..end_addr-1,
- NULL if none. Assume start_addr < end_addr. */
-static inline struct vm_area_struct * find_vma_intersection(struct mm_struct * mm, unsigned long start_addr, unsigned long end_addr)
+/**
+ * find_vma_intersection() - Look up the first VMA which intersects the interval
+ * @mm: The process address space.
+ * @start_addr: The inclusive start user address.
+ * @end_addr: The exclusive end user address.
+ *
+ * Returns: The first VMA within the provided range, %NULL otherwise. Assumes
+ * start_addr < end_addr.
+ */
+static inline
+struct vm_area_struct *find_vma_intersection(struct mm_struct *mm,
+ unsigned long start_addr,
+ unsigned long end_addr)
{
- struct vm_area_struct * vma = find_vma(mm,start_addr);
+ struct vm_area_struct *vma = find_vma(mm, start_addr);
if (vma && end_addr <= vma->vm_start)
vma = NULL;
return vma;
}
+/**
+ * vma_lookup() - Find a VMA at a specific address
+ * @mm: The process address space.
+ * @addr: The user address.
+ *
+ * Return: The vm_area_struct at the given address, %NULL otherwise.
+ */
+static inline
+struct vm_area_struct *vma_lookup(struct mm_struct *mm, unsigned long addr)
+{
+ struct vm_area_struct *vma = find_vma(mm, addr);
+
+ if (vma && addr < vma->vm_start)
+ vma = NULL;
+
+ return vma;
+}
+
static inline unsigned long vm_start_gap(struct vm_area_struct *vma)
{
unsigned long vm_start = vma->vm_start;
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 106/192] drm/i915/selftests: use vma_lookup() in __igt_mmap()
2021-06-29 2:32 incoming Andrew Morton
` (104 preceding siblings ...)
2021-06-29 2:38 ` [patch 105/192] mm: add vma_lookup(), update find_vma_intersection() comments Andrew Morton
@ 2021-06-29 2:38 ` Andrew Morton
2021-06-29 2:38 ` [patch 107/192] arch/arc/kernel/troubleshoot: use vma_lookup() instead of find_vma() Andrew Morton
` (85 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:38 UTC (permalink / raw)
To: akpm, david, dbueso, geert, ldufour, Liam.Howlett, linux-mm,
mm-commits, torvalds
From: Liam Howlett <liam.howlett@oracle.com>
Subject: drm/i915/selftests: use vma_lookup() in __igt_mmap()
vma_lookup() will look up the vma at a specific address. find_vma() will
start the search for a specific address and continue upwards. This fixes
an issue with the selftest as the returned vma may not be the newly
created vma, but simply the vma at a higher address.
Link: https://lkml.kernel.org/r/20210521174745.2219620-3-Liam.Howlett@Oracle.com
Fixes: 6fedafacae1b (drm/i915/selftests: Wrap vm_mmap() around GEM
objects
Signed-off-by: Liam R. Howlett <Liam.Howlett@Oracle.com>
Reviewed-by: Laurent Dufour <ldufour@linux.ibm.com>
Acked-by: David Hildenbrand <david@redhat.com>
Acked-by: Davidlohr Bueso <dbueso@suse.de>
Cc: Geert Uytterhoeven <geert@linux-m68k.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
--- a/drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c~drm-i915-selftests-use-vma_lookup-in-__igt_mmap
+++ a/drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
@@ -871,7 +871,7 @@ static int __igt_mmap(struct drm_i915_pr
pr_debug("igt_mmap(%s, %d) @ %lx\n", obj->mm.region->name, type, addr);
- area = find_vma(current->mm, addr);
+ area = vma_lookup(current->mm, addr);
if (!area) {
pr_err("%s: Did not create a vm_area_struct for the mmap\n",
obj->mm.region->name);
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 107/192] arch/arc/kernel/troubleshoot: use vma_lookup() instead of find_vma()
2021-06-29 2:32 incoming Andrew Morton
` (105 preceding siblings ...)
2021-06-29 2:38 ` [patch 106/192] drm/i915/selftests: use vma_lookup() in __igt_mmap() Andrew Morton
@ 2021-06-29 2:38 ` Andrew Morton
2021-06-29 2:38 ` [patch 108/192] arch/arm64/kvm: use vma_lookup() instead of find_vma_intersection() Andrew Morton
` (84 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:38 UTC (permalink / raw)
To: akpm, david, dbueso, geert, ldufour, Liam.Howlett, linux-mm,
mm-commits, torvalds
From: Liam Howlett <liam.howlett@oracle.com>
Subject: arch/arc/kernel/troubleshoot: use vma_lookup() instead of find_vma()
Use vma_lookup() to find the VMA at a specific address. As vma_lookup()
will return NULL if the address is not within any VMA, the start address
no longer needs to be validated.
Link: https://lkml.kernel.org/r/20210521174745.2219620-4-Liam.Howlett@Oracle.com
Signed-off-by: Liam R. Howlett <Liam.Howlett@Oracle.com>
Reviewed-by: Laurent Dufour <ldufour@linux.ibm.com>
Acked-by: David Hildenbrand <david@redhat.com>
Acked-by: Davidlohr Bueso <dbueso@suse.de>
Cc: Geert Uytterhoeven <geert@linux-m68k.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
arch/arc/kernel/troubleshoot.c | 8 ++++----
1 file changed, 4 insertions(+), 4 deletions(-)
--- a/arch/arc/kernel/troubleshoot.c~arch-arc-kernel-troubleshoot-use-vma_lookup-instead-of-find_vma
+++ a/arch/arc/kernel/troubleshoot.c
@@ -83,12 +83,12 @@ static void show_faulting_vma(unsigned l
* non-inclusive vma
*/
mmap_read_lock(active_mm);
- vma = find_vma(active_mm, address);
+ vma = vma_lookup(active_mm, address);
- /* check against the find_vma( ) behaviour which returns the next VMA
- * if the container VMA is not found
+ /* Lookup the vma at the address and report if the container VMA is not
+ * found
*/
- if (vma && (vma->vm_start <= address)) {
+ if (vma) {
char buf[ARC_PATH_MAX];
char *nm = "?";
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 108/192] arch/arm64/kvm: use vma_lookup() instead of find_vma_intersection()
2021-06-29 2:32 incoming Andrew Morton
` (106 preceding siblings ...)
2021-06-29 2:38 ` [patch 107/192] arch/arc/kernel/troubleshoot: use vma_lookup() instead of find_vma() Andrew Morton
@ 2021-06-29 2:38 ` Andrew Morton
2021-06-29 2:39 ` [patch 109/192] arch/powerpc/kvm/book3s_hv_uvmem: " Andrew Morton
` (83 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:38 UTC (permalink / raw)
To: akpm, david, dbueso, geert, ldufour, Liam.Howlett, linux-mm,
mm-commits, torvalds
From: Liam Howlett <liam.howlett@oracle.com>
Subject: arch/arm64/kvm: use vma_lookup() instead of find_vma_intersection()
vma_lookup() finds the vma of a specific address with a cleaner interface
and is more readable.
Link: https://lkml.kernel.org/r/20210521174745.2219620-5-Liam.Howlett@Oracle.com
Signed-off-by: Liam R. Howlett <Liam.Howlett@Oracle.com>
Reviewed-by: Laurent Dufour <ldufour@linux.ibm.com>
Acked-by: David Hildenbrand <david@redhat.com>
Acked-by: Davidlohr Bueso <dbueso@suse.de>
Cc: Geert Uytterhoeven <geert@linux-m68k.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
arch/arm64/kvm/mmu.c | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
--- a/arch/arm64/kvm/mmu.c~arch-arm64-kvm-use-vma_lookup-instead-of-find_vma_intersection
+++ a/arch/arm64/kvm/mmu.c
@@ -855,7 +855,7 @@ static int user_mem_abort(struct kvm_vcp
/* Let's check if we will get back a huge page backed by hugetlbfs */
mmap_read_lock(current->mm);
- vma = find_vma_intersection(current->mm, hva, hva + 1);
+ vma = vma_lookup(current->mm, hva);
if (unlikely(!vma)) {
kvm_err("Failed to find VMA for hva 0x%lx\n", hva);
mmap_read_unlock(current->mm);
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 109/192] arch/powerpc/kvm/book3s_hv_uvmem: use vma_lookup() instead of find_vma_intersection()
2021-06-29 2:32 incoming Andrew Morton
` (107 preceding siblings ...)
2021-06-29 2:38 ` [patch 108/192] arch/arm64/kvm: use vma_lookup() instead of find_vma_intersection() Andrew Morton
@ 2021-06-29 2:39 ` Andrew Morton
2021-06-29 2:39 ` [patch 110/192] arch/powerpc/kvm/book3s: use vma_lookup() in kvmppc_hv_setup_htab_rma() Andrew Morton
` (82 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:39 UTC (permalink / raw)
To: akpm, david, dbueso, geert, ldufour, Liam.Howlett, linux-mm,
mm-commits, torvalds
From: Liam Howlett <liam.howlett@oracle.com>
Subject: arch/powerpc/kvm/book3s_hv_uvmem: use vma_lookup() instead of find_vma_intersection()
vma_lookup() finds the vma of a specific address with a cleaner interface
and is more readable.
Link: https://lkml.kernel.org/r/20210521174745.2219620-6-Liam.Howlett@Oracle.com
Signed-off-by: Liam R. Howlett <Liam.Howlett@Oracle.com>
Reviewed-by: Laurent Dufour <ldufour@linux.ibm.com>
Acked-by: David Hildenbrand <david@redhat.com>
Acked-by: Davidlohr Bueso <dbueso@suse.de>
Cc: Geert Uytterhoeven <geert@linux-m68k.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
arch/powerpc/kvm/book3s_hv_uvmem.c | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
--- a/arch/powerpc/kvm/book3s_hv_uvmem.c~arch-powerpc-kvm-book3s_hv_uvmem-use-vma_lookup-instead-of-find_vma_intersection
+++ a/arch/powerpc/kvm/book3s_hv_uvmem.c
@@ -614,7 +614,7 @@ void kvmppc_uvmem_drop_pages(const struc
/* Fetch the VMA if addr is not in the latest fetched one */
if (!vma || addr >= vma->vm_end) {
- vma = find_vma_intersection(kvm->mm, addr, addr+1);
+ vma = vma_lookup(kvm->mm, addr);
if (!vma) {
pr_err("Can't find VMA for gfn:0x%lx\n", gfn);
break;
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 110/192] arch/powerpc/kvm/book3s: use vma_lookup() in kvmppc_hv_setup_htab_rma()
2021-06-29 2:32 incoming Andrew Morton
` (108 preceding siblings ...)
2021-06-29 2:39 ` [patch 109/192] arch/powerpc/kvm/book3s_hv_uvmem: " Andrew Morton
@ 2021-06-29 2:39 ` Andrew Morton
2021-06-29 2:39 ` [patch 111/192] arch/mips/kernel/traps: use vma_lookup() instead of find_vma() Andrew Morton
` (81 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:39 UTC (permalink / raw)
To: akpm, david, dbueso, geert, ldufour, Liam.Howlett, linux-mm,
mm-commits, torvalds
From: Liam Howlett <liam.howlett@oracle.com>
Subject: arch/powerpc/kvm/book3s: use vma_lookup() in kvmppc_hv_setup_htab_rma()
Using vma_lookup() removes the requirement to check if the address is
within the returned vma. The code is easier to understand and more
compact.
Link: https://lkml.kernel.org/r/20210521174745.2219620-7-Liam.Howlett@Oracle.com
Signed-off-by: Liam R. Howlett <Liam.Howlett@Oracle.com>
Reviewed-by: Laurent Dufour <ldufour@linux.ibm.com>
Acked-by: David Hildenbrand <david@redhat.com>
Acked-by: Davidlohr Bueso <dbueso@suse.de>
Cc: Geert Uytterhoeven <geert@linux-m68k.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
arch/powerpc/kvm/book3s_hv.c | 4 ++--
1 file changed, 2 insertions(+), 2 deletions(-)
--- a/arch/powerpc/kvm/book3s_hv.c~arch-powerpc-kvm-book3s-use-vma_lookup-in-kvmppc_hv_setup_htab_rma
+++ a/arch/powerpc/kvm/book3s_hv.c
@@ -4758,8 +4758,8 @@ static int kvmppc_hv_setup_htab_rma(stru
/* Look up the VMA for the start of this memory slot */
hva = memslot->userspace_addr;
mmap_read_lock(kvm->mm);
- vma = find_vma(kvm->mm, hva);
- if (!vma || vma->vm_start > hva || (vma->vm_flags & VM_IO))
+ vma = vma_lookup(kvm->mm, hva);
+ if (!vma || (vma->vm_flags & VM_IO))
goto up_out;
psize = vma_kernel_pagesize(vma);
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 111/192] arch/mips/kernel/traps: use vma_lookup() instead of find_vma()
2021-06-29 2:32 incoming Andrew Morton
` (109 preceding siblings ...)
2021-06-29 2:39 ` [patch 110/192] arch/powerpc/kvm/book3s: use vma_lookup() in kvmppc_hv_setup_htab_rma() Andrew Morton
@ 2021-06-29 2:39 ` Andrew Morton
2021-06-29 2:39 ` [patch 112/192] arch/m68k/kernel/sys_m68k: use vma_lookup() in sys_cacheflush() Andrew Morton
` (80 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:39 UTC (permalink / raw)
To: akpm, david, dbueso, geert, ldufour, Liam.Howlett, linux-mm,
mm-commits, torvalds
From: Liam Howlett <liam.howlett@oracle.com>
Subject: arch/mips/kernel/traps: use vma_lookup() instead of find_vma()
Use vma_lookup() to find the VMA at a specific address. As vma_lookup()
will return NULL if the address is not within any VMA, the start address
no longer needs to be validated.
Link: https://lkml.kernel.org/r/20210521174745.2219620-8-Liam.Howlett@Oracle.com
Signed-off-by: Liam R. Howlett <Liam.Howlett@Oracle.com>
Reviewed-by: Laurent Dufour <ldufour@linux.ibm.com>
Acked-by: David Hildenbrand <david@redhat.com>
Acked-by: Davidlohr Bueso <dbueso@suse.de>
Cc: Geert Uytterhoeven <geert@linux-m68k.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
arch/mips/kernel/traps.c | 4 +---
1 file changed, 1 insertion(+), 3 deletions(-)
--- a/arch/mips/kernel/traps.c~arch-mips-kernel-traps-use-vma_lookup-instead-of-find_vma
+++ a/arch/mips/kernel/traps.c
@@ -784,7 +784,6 @@ void force_fcr31_sig(unsigned long fcr31
int process_fpemu_return(int sig, void __user *fault_addr, unsigned long fcr31)
{
int si_code;
- struct vm_area_struct *vma;
switch (sig) {
case 0:
@@ -800,8 +799,7 @@ int process_fpemu_return(int sig, void _
case SIGSEGV:
mmap_read_lock(current->mm);
- vma = find_vma(current->mm, (unsigned long)fault_addr);
- if (vma && (vma->vm_start <= (unsigned long)fault_addr))
+ if (vma_lookup(current->mm, (unsigned long)fault_addr))
si_code = SEGV_ACCERR;
else
si_code = SEGV_MAPERR;
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 112/192] arch/m68k/kernel/sys_m68k: use vma_lookup() in sys_cacheflush()
2021-06-29 2:32 incoming Andrew Morton
` (110 preceding siblings ...)
2021-06-29 2:39 ` [patch 111/192] arch/mips/kernel/traps: use vma_lookup() instead of find_vma() Andrew Morton
@ 2021-06-29 2:39 ` Andrew Morton
2021-06-29 2:39 ` [patch 113/192] x86/sgx: use vma_lookup() in sgx_encl_find() Andrew Morton
` (79 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:39 UTC (permalink / raw)
To: akpm, david, dbueso, geert, ldufour, Liam.Howlett, linux-mm,
mm-commits, torvalds
From: Liam Howlett <liam.howlett@oracle.com>
Subject: arch/m68k/kernel/sys_m68k: use vma_lookup() in sys_cacheflush()
Use vma_lookup() to find the VMA at a specific address. As vma_lookup()
will return NULL if the address is not within any VMA, the start address
no longer needs to be validated.
Link: https://lkml.kernel.org/r/20210521174745.2219620-9-Liam.Howlett@Oracle.com
Signed-off-by: Liam R. Howlett <Liam.Howlett@Oracle.com>
Reviewed-by: Laurent Dufour <ldufour@linux.ibm.com>
Acked-by: David Hildenbrand <david@redhat.com>
Acked-by: Davidlohr Bueso <dbueso@suse.de>
Acked-by: Geert Uytterhoeven <geert@linux-m68k.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
arch/m68k/kernel/sys_m68k.c | 4 ++--
1 file changed, 2 insertions(+), 2 deletions(-)
--- a/arch/m68k/kernel/sys_m68k.c~arch-m68k-kernel-sys_m68k-use-vma_lookup-in-sys_cacheflush
+++ a/arch/m68k/kernel/sys_m68k.c
@@ -402,8 +402,8 @@ sys_cacheflush (unsigned long addr, int
* to this process.
*/
mmap_read_lock(current->mm);
- vma = find_vma(current->mm, addr);
- if (!vma || addr < vma->vm_start || addr + len > vma->vm_end)
+ vma = vma_lookup(current->mm, addr);
+ if (!vma || addr + len > vma->vm_end)
goto out_unlock;
}
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 113/192] x86/sgx: use vma_lookup() in sgx_encl_find()
2021-06-29 2:32 incoming Andrew Morton
` (111 preceding siblings ...)
2021-06-29 2:39 ` [patch 112/192] arch/m68k/kernel/sys_m68k: use vma_lookup() in sys_cacheflush() Andrew Morton
@ 2021-06-29 2:39 ` Andrew Morton
2021-06-29 2:39 ` [patch 114/192] virt/kvm: use vma_lookup() instead of find_vma_intersection() Andrew Morton
` (78 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:39 UTC (permalink / raw)
To: akpm, david, dbueso, geert, ldufour, Liam.Howlett, linux-mm,
mm-commits, torvalds
From: Liam Howlett <liam.howlett@oracle.com>
Subject: x86/sgx: use vma_lookup() in sgx_encl_find()
Use vma_lookup() to find the VMA at a specific address. As vma_lookup()
will return NULL if the address is not within any VMA, the start address
no longer needs to be validated.
Link: https://lkml.kernel.org/r/20210521174745.2219620-10-Liam.Howlett@Oracle.com
Signed-off-by: Liam R. Howlett <Liam.Howlett@Oracle.com>
Reviewed-by: Laurent Dufour <ldufour@linux.ibm.com>
Acked-by: David Hildenbrand <david@redhat.com>
Acked-by: Davidlohr Bueso <dbueso@suse.de>
Cc: Geert Uytterhoeven <geert@linux-m68k.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
arch/x86/kernel/cpu/sgx/encl.h | 4 ++--
1 file changed, 2 insertions(+), 2 deletions(-)
--- a/arch/x86/kernel/cpu/sgx/encl.h~x86-sgx-use-vma_lookup-in-sgx_encl_find
+++ a/arch/x86/kernel/cpu/sgx/encl.h
@@ -91,8 +91,8 @@ static inline int sgx_encl_find(struct m
{
struct vm_area_struct *result;
- result = find_vma(mm, addr);
- if (!result || result->vm_ops != &sgx_vm_ops || addr < result->vm_start)
+ result = vma_lookup(mm, addr);
+ if (!result || result->vm_ops != &sgx_vm_ops)
return -EINVAL;
*vma = result;
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 114/192] virt/kvm: use vma_lookup() instead of find_vma_intersection()
2021-06-29 2:32 incoming Andrew Morton
` (112 preceding siblings ...)
2021-06-29 2:39 ` [patch 113/192] x86/sgx: use vma_lookup() in sgx_encl_find() Andrew Morton
@ 2021-06-29 2:39 ` Andrew Morton
2021-06-29 2:39 ` [patch 115/192] vfio: " Andrew Morton
` (77 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:39 UTC (permalink / raw)
To: akpm, david, dbueso, geert, ldufour, Liam.Howlett, linux-mm,
mm-commits, torvalds
From: Liam Howlett <liam.howlett@oracle.com>
Subject: virt/kvm: use vma_lookup() instead of find_vma_intersection()
vma_lookup() finds the vma of a specific address with a cleaner interface
and is more readable.
Link: https://lkml.kernel.org/r/20210521174745.2219620-11-Liam.Howlett@Oracle.com
Signed-off-by: Liam R. Howlett <Liam.Howlett@Oracle.com>
Reviewed-by: Laurent Dufour <ldufour@linux.ibm.com>
Acked-by: David Hildenbrand <david@redhat.com>
Acked-by: Davidlohr Bueso <dbueso@suse.de>
Cc: Geert Uytterhoeven <geert@linux-m68k.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
virt/kvm/kvm_main.c | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
--- a/virt/kvm/kvm_main.c~virt-kvm-use-vma_lookup-instead-of-find_vma_intersection
+++ a/virt/kvm/kvm_main.c
@@ -2170,7 +2170,7 @@ static kvm_pfn_t hva_to_pfn(unsigned lon
}
retry:
- vma = find_vma_intersection(current->mm, addr, addr + 1);
+ vma = vma_lookup(current->mm, addr);
if (vma == NULL)
pfn = KVM_PFN_ERR_FAULT;
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 115/192] vfio: use vma_lookup() instead of find_vma_intersection()
2021-06-29 2:32 incoming Andrew Morton
` (113 preceding siblings ...)
2021-06-29 2:39 ` [patch 114/192] virt/kvm: use vma_lookup() instead of find_vma_intersection() Andrew Morton
@ 2021-06-29 2:39 ` Andrew Morton
2021-06-29 2:39 ` [patch 116/192] net/ipv5/tcp: use vma_lookup() in tcp_zerocopy_receive() Andrew Morton
` (76 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:39 UTC (permalink / raw)
To: akpm, david, dbueso, geert, ldufour, Liam.Howlett, linux-mm,
mm-commits, torvalds
From: Liam Howlett <liam.howlett@oracle.com>
Subject: vfio: use vma_lookup() instead of find_vma_intersection()
vma_lookup() finds the vma of a specific address with a cleaner interface
and is more readable.
Link: https://lkml.kernel.org/r/20210521174745.2219620-12-Liam.Howlett@Oracle.com
Signed-off-by: Liam R. Howlett <Liam.Howlett@Oracle.com>
Reviewed-by: Laurent Dufour <ldufour@linux.ibm.com>
Acked-by: David Hildenbrand <david@redhat.com>
Acked-by: Davidlohr Bueso <dbueso@suse.de>
Cc: Geert Uytterhoeven <geert@linux-m68k.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
drivers/vfio/vfio_iommu_type1.c | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
--- a/drivers/vfio/vfio_iommu_type1.c~vfio-use-vma_lookup-instead-of-find_vma_intersection
+++ a/drivers/vfio/vfio_iommu_type1.c
@@ -567,7 +567,7 @@ static int vaddr_get_pfns(struct mm_stru
vaddr = untagged_addr(vaddr);
retry:
- vma = find_vma_intersection(mm, vaddr, vaddr + 1);
+ vma = vma_lookup(mm, vaddr);
if (vma && vma->vm_flags & VM_PFNMAP) {
ret = follow_fault_pfn(vma, mm, vaddr, pfn, prot & IOMMU_WRITE);
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 116/192] net/ipv5/tcp: use vma_lookup() in tcp_zerocopy_receive()
2021-06-29 2:32 incoming Andrew Morton
` (114 preceding siblings ...)
2021-06-29 2:39 ` [patch 115/192] vfio: " Andrew Morton
@ 2021-06-29 2:39 ` Andrew Morton
2021-06-29 2:39 ` [patch 117/192] drm/amdgpu: use vma_lookup() in amdgpu_ttm_tt_get_user_pages() Andrew Morton
` (75 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:39 UTC (permalink / raw)
To: akpm, davem, david, dbueso, geert, ldufour, Liam.Howlett,
linux-mm, mm-commits, torvalds
From: Liam Howlett <liam.howlett@oracle.com>
Subject: net/ipv5/tcp: use vma_lookup() in tcp_zerocopy_receive()
Use vma_lookup() to find the VMA at a specific address. As vma_lookup()
will return NULL if the address is not within any VMA, the start address
no longer needs to be validated.
Link: https://lkml.kernel.org/r/20210521174745.2219620-13-Liam.Howlett@Oracle.com
Signed-off-by: Liam R. Howlett <Liam.Howlett@Oracle.com>
Reviewed-by: Laurent Dufour <ldufour@linux.ibm.com>
Acked-by: David Hildenbrand <david@redhat.com>
Acked-by: Davidlohr Bueso <dbueso@suse.de>
Cc: David Miller <davem@davemloft.net>
Cc: Geert Uytterhoeven <geert@linux-m68k.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
net/ipv4/tcp.c | 4 ++--
1 file changed, 2 insertions(+), 2 deletions(-)
--- a/net/ipv4/tcp.c~net-ipv5-tcp-use-vma_lookup-in-tcp_zerocopy_receive
+++ a/net/ipv4/tcp.c
@@ -2095,8 +2095,8 @@ static int tcp_zerocopy_receive(struct s
mmap_read_lock(current->mm);
- vma = find_vma(current->mm, address);
- if (!vma || vma->vm_start > address || vma->vm_ops != &tcp_vm_ops) {
+ vma = vma_lookup(current->mm, address);
+ if (!vma || vma->vm_ops != &tcp_vm_ops) {
mmap_read_unlock(current->mm);
return -EINVAL;
}
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 117/192] drm/amdgpu: use vma_lookup() in amdgpu_ttm_tt_get_user_pages()
2021-06-29 2:32 incoming Andrew Morton
` (115 preceding siblings ...)
2021-06-29 2:39 ` [patch 116/192] net/ipv5/tcp: use vma_lookup() in tcp_zerocopy_receive() Andrew Morton
@ 2021-06-29 2:39 ` Andrew Morton
2021-06-29 2:39 ` [patch 118/192] media: videobuf2: use vma_lookup() in get_vaddr_frames() Andrew Morton
` (74 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:39 UTC (permalink / raw)
To: akpm, alexander.deucher, david, dbueso, geert, ldufour,
Liam.Howlett, linux-mm, mm-commits, torvalds
From: Liam Howlett <liam.howlett@oracle.com>
Subject: drm/amdgpu: use vma_lookup() in amdgpu_ttm_tt_get_user_pages()
Use vma_lookup() to find the VMA at a specific address. As vma_lookup()
will return NULL if the address is not within any VMA, the start address
no longer needs to be validated.
Link: https://lkml.kernel.org/r/20210521174745.2219620-14-Liam.Howlett@Oracle.com
Signed-off-by: Liam R. Howlett <Liam.Howlett@Oracle.com>
Reviewed-by: Laurent Dufour <ldufour@linux.ibm.com>
Acked-by: David Hildenbrand <david@redhat.com>
Acked-by: Davidlohr Bueso <dbueso@suse.de>
Acked-by: Alex Deucher <alexander.deucher@amd.com>
Cc: Geert Uytterhoeven <geert@linux-m68k.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c | 4 ++--
1 file changed, 2 insertions(+), 2 deletions(-)
--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c~drm-amdgpu-use-vma_lookup-in-amdgpu_ttm_tt_get_user_pages
+++ a/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
@@ -709,8 +709,8 @@ int amdgpu_ttm_tt_get_user_pages(struct
}
mmap_read_lock(mm);
- vma = find_vma(mm, start);
- if (unlikely(!vma || start < vma->vm_start)) {
+ vma = vma_lookup(mm, start);
+ if (unlikely(!vma)) {
r = -EFAULT;
goto out_unlock;
}
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 118/192] media: videobuf2: use vma_lookup() in get_vaddr_frames()
2021-06-29 2:32 incoming Andrew Morton
` (116 preceding siblings ...)
2021-06-29 2:39 ` [patch 117/192] drm/amdgpu: use vma_lookup() in amdgpu_ttm_tt_get_user_pages() Andrew Morton
@ 2021-06-29 2:39 ` Andrew Morton
2021-06-29 2:39 ` [patch 119/192] misc/sgi-gru/grufault: use vma_lookup() in gru_find_vma() Andrew Morton
` (73 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:39 UTC (permalink / raw)
To: akpm, david, dbueso, geert, ldufour, Liam.Howlett, linux-mm,
mm-commits, torvalds
From: Liam Howlett <liam.howlett@oracle.com>
Subject: media: videobuf2: use vma_lookup() in get_vaddr_frames()
vma_lookup() finds the vma of a specific address with a cleaner interface
and is more readable.
Link: https://lkml.kernel.org/r/20210521174745.2219620-15-Liam.Howlett@Oracle.com
Signed-off-by: Liam R. Howlett <Liam.Howlett@Oracle.com>
Reviewed-by: Laurent Dufour <ldufour@linux.ibm.com>
Acked-by: David Hildenbrand <david@redhat.com>
Acked-by: Davidlohr Bueso <dbueso@suse.de>
Cc: Geert Uytterhoeven <geert@linux-m68k.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
drivers/media/common/videobuf2/frame_vector.c | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
--- a/drivers/media/common/videobuf2/frame_vector.c~media-videobuf2-use-vma_lookup-in-get_vaddr_frames
+++ a/drivers/media/common/videobuf2/frame_vector.c
@@ -64,7 +64,7 @@ int get_vaddr_frames(unsigned long start
do {
unsigned long *nums = frame_vector_pfns(vec);
- vma = find_vma_intersection(mm, start, start + 1);
+ vma = vma_lookup(mm, start);
if (!vma)
break;
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 119/192] misc/sgi-gru/grufault: use vma_lookup() in gru_find_vma()
2021-06-29 2:32 incoming Andrew Morton
` (117 preceding siblings ...)
2021-06-29 2:39 ` [patch 118/192] media: videobuf2: use vma_lookup() in get_vaddr_frames() Andrew Morton
@ 2021-06-29 2:39 ` Andrew Morton
2021-06-29 2:39 ` [patch 120/192] kernel/events/uprobes: use vma_lookup() in find_active_uprobe() Andrew Morton
` (72 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:39 UTC (permalink / raw)
To: akpm, david, dbueso, geert, ldufour, Liam.Howlett, linux-mm,
mm-commits, torvalds
From: Liam Howlett <liam.howlett@oracle.com>
Subject: misc/sgi-gru/grufault: use vma_lookup() in gru_find_vma()
Use vma_lookup() to find the VMA at a specific address. As vma_lookup()
will return NULL if the address is not within any VMA, the start address
no longer needs to be validated.
Link: https://lkml.kernel.org/r/20210521174745.2219620-16-Liam.Howlett@Oracle.com
Signed-off-by: Liam R. Howlett <Liam.Howlett@Oracle.com>
Reviewed-by: Laurent Dufour <ldufour@linux.ibm.com>
Acked-by: David Hildenbrand <david@redhat.com>
Acked-by: Davidlohr Bueso <dbueso@suse.de>
Cc: Geert Uytterhoeven <geert@linux-m68k.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
drivers/misc/sgi-gru/grufault.c | 4 ++--
1 file changed, 2 insertions(+), 2 deletions(-)
--- a/drivers/misc/sgi-gru/grufault.c~misc-sgi-gru-grufault-use-vma_lookup-in-gru_find_vma
+++ a/drivers/misc/sgi-gru/grufault.c
@@ -49,8 +49,8 @@ struct vm_area_struct *gru_find_vma(unsi
{
struct vm_area_struct *vma;
- vma = find_vma(current->mm, vaddr);
- if (vma && vma->vm_start <= vaddr && vma->vm_ops == &gru_vm_ops)
+ vma = vma_lookup(current->mm, vaddr);
+ if (vma && vma->vm_ops == &gru_vm_ops)
return vma;
return NULL;
}
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 120/192] kernel/events/uprobes: use vma_lookup() in find_active_uprobe()
2021-06-29 2:32 incoming Andrew Morton
` (118 preceding siblings ...)
2021-06-29 2:39 ` [patch 119/192] misc/sgi-gru/grufault: use vma_lookup() in gru_find_vma() Andrew Morton
@ 2021-06-29 2:39 ` Andrew Morton
2021-06-29 2:39 ` [patch 121/192] lib/test_hmm: use vma_lookup() in dmirror_migrate() Andrew Morton
` (71 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:39 UTC (permalink / raw)
To: akpm, david, dbueso, geert, ldufour, Liam.Howlett, linux-mm,
mm-commits, torvalds
From: Liam Howlett <liam.howlett@oracle.com>
Subject: kernel/events/uprobes: use vma_lookup() in find_active_uprobe()
Use vma_lookup() to find the VMA at a specific address. As vma_lookup()
will return NULL if the address is not within any VMA, the start address
no longer needs to be validated.
Link: https://lkml.kernel.org/r/20210521174745.2219620-17-Liam.Howlett@Oracle.com
Signed-off-by: Liam R. Howlett <Liam.Howlett@Oracle.com>
Reviewed-by: Laurent Dufour <ldufour@linux.ibm.com>
Acked-by: David Hildenbrand <david@redhat.com>
Acked-by: Davidlohr Bueso <dbueso@suse.de>
Cc: Geert Uytterhoeven <geert@linux-m68k.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
kernel/events/uprobes.c | 4 ++--
1 file changed, 2 insertions(+), 2 deletions(-)
--- a/kernel/events/uprobes.c~kernel-events-uprobes-use-vma_lookup-in-find_active_uprobe
+++ a/kernel/events/uprobes.c
@@ -2046,8 +2046,8 @@ static struct uprobe *find_active_uprobe
struct vm_area_struct *vma;
mmap_read_lock(mm);
- vma = find_vma(mm, bp_vaddr);
- if (vma && vma->vm_start <= bp_vaddr) {
+ vma = vma_lookup(mm, bp_vaddr);
+ if (vma) {
if (valid_vma(vma, false)) {
struct inode *inode = file_inode(vma->vm_file);
loff_t offset = vaddr_to_offset(vma, bp_vaddr);
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 121/192] lib/test_hmm: use vma_lookup() in dmirror_migrate()
2021-06-29 2:32 incoming Andrew Morton
` (119 preceding siblings ...)
2021-06-29 2:39 ` [patch 120/192] kernel/events/uprobes: use vma_lookup() in find_active_uprobe() Andrew Morton
@ 2021-06-29 2:39 ` Andrew Morton
2021-06-29 2:39 ` [patch 122/192] mm/ksm: use vma_lookup() in find_mergeable_vma() Andrew Morton
` (70 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:39 UTC (permalink / raw)
To: akpm, david, dbueso, geert, ldufour, Liam.Howlett, linux-mm,
mm-commits, torvalds
From: Liam Howlett <liam.howlett@oracle.com>
Subject: lib/test_hmm: use vma_lookup() in dmirror_migrate()
Use vma_lookup() to find the VMA at a specific address. As vma_lookup()
will return NULL if the address is not within any VMA, the start address
no longer needs to be validated.
Link: https://lkml.kernel.org/r/20210521174745.2219620-18-Liam.Howlett@Oracle.com
Signed-off-by: Liam R. Howlett <Liam.Howlett@Oracle.com>
Reviewed-by: Laurent Dufour <ldufour@linux.ibm.com>
Acked-by: David Hildenbrand <david@redhat.com>
Acked-by: Davidlohr Bueso <dbueso@suse.de>
Cc: Geert Uytterhoeven <geert@linux-m68k.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
lib/test_hmm.c | 5 ++---
1 file changed, 2 insertions(+), 3 deletions(-)
--- a/lib/test_hmm.c~lib-test_hmm-use-vma_lookup-in-dmirror_migrate
+++ a/lib/test_hmm.c
@@ -686,9 +686,8 @@ static int dmirror_migrate(struct dmirro
mmap_read_lock(mm);
for (addr = start; addr < end; addr = next) {
- vma = find_vma(mm, addr);
- if (!vma || addr < vma->vm_start ||
- !(vma->vm_flags & VM_READ)) {
+ vma = vma_lookup(mm, addr);
+ if (!vma || !(vma->vm_flags & VM_READ)) {
ret = -EINVAL;
goto out;
}
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 122/192] mm/ksm: use vma_lookup() in find_mergeable_vma()
2021-06-29 2:32 incoming Andrew Morton
` (120 preceding siblings ...)
2021-06-29 2:39 ` [patch 121/192] lib/test_hmm: use vma_lookup() in dmirror_migrate() Andrew Morton
@ 2021-06-29 2:39 ` Andrew Morton
2021-06-29 2:39 ` [patch 123/192] mm/migrate: use vma_lookup() in do_pages_stat_array() Andrew Morton
` (69 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:39 UTC (permalink / raw)
To: akpm, david, dbueso, geert, ldufour, Liam.Howlett, linux-mm,
mm-commits, torvalds
From: Liam Howlett <liam.howlett@oracle.com>
Subject: mm/ksm: use vma_lookup() in find_mergeable_vma()
Use vma_lookup() to find the VMA at a specific address. As vma_lookup()
will return NULL if the address is not within any VMA, the start address
no longer needs to be validated.
Link: https://lkml.kernel.org/r/20210521174745.2219620-19-Liam.Howlett@Oracle.com
Signed-off-by: Liam R. Howlett <Liam.Howlett@Oracle.com>
Reviewed-by: Laurent Dufour <ldufour@linux.ibm.com>
Acked-by: David Hildenbrand <david@redhat.com>
Acked-by: Davidlohr Bueso <dbueso@suse.de>
Cc: Geert Uytterhoeven <geert@linux-m68k.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/ksm.c | 6 ++----
1 file changed, 2 insertions(+), 4 deletions(-)
--- a/mm/ksm.c~mm-ksm-use-vma_lookup-in-find_mergeable_vma
+++ a/mm/ksm.c
@@ -521,10 +521,8 @@ static struct vm_area_struct *find_merge
struct vm_area_struct *vma;
if (ksm_test_exit(mm))
return NULL;
- vma = find_vma(mm, addr);
- if (!vma || vma->vm_start > addr)
- return NULL;
- if (!(vma->vm_flags & VM_MERGEABLE) || !vma->anon_vma)
+ vma = vma_lookup(mm, addr);
+ if (!vma || !(vma->vm_flags & VM_MERGEABLE) || !vma->anon_vma)
return NULL;
return vma;
}
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 123/192] mm/migrate: use vma_lookup() in do_pages_stat_array()
2021-06-29 2:32 incoming Andrew Morton
` (121 preceding siblings ...)
2021-06-29 2:39 ` [patch 122/192] mm/ksm: use vma_lookup() in find_mergeable_vma() Andrew Morton
@ 2021-06-29 2:39 ` Andrew Morton
2021-06-29 2:39 ` [patch 124/192] mm/mremap: use vma_lookup() in vma_to_resize() Andrew Morton
` (68 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:39 UTC (permalink / raw)
To: akpm, david, dbueso, geert, ldufour, Liam.Howlett, linux-mm,
mm-commits, torvalds
From: Liam Howlett <liam.howlett@oracle.com>
Subject: mm/migrate: use vma_lookup() in do_pages_stat_array()
Use vma_lookup() to find the VMA at a specific address. As vma_lookup()
will return NULL if the address is not within any VMA, the start address
no longer needs to be validated.
Link: https://lkml.kernel.org/r/20210521174745.2219620-20-Liam.Howlett@Oracle.com
Signed-off-by: Liam R. Howlett <Liam.Howlett@Oracle.com>
Reviewed-by: Laurent Dufour <ldufour@linux.ibm.com>
Acked-by: David Hildenbrand <david@redhat.com>
Acked-by: Davidlohr Bueso <dbueso@suse.de>
Cc: Geert Uytterhoeven <geert@linux-m68k.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/migrate.c | 4 ++--
1 file changed, 2 insertions(+), 2 deletions(-)
--- a/mm/migrate.c~mm-migrate-use-vma_lookup-in-do_pages_stat_array
+++ a/mm/migrate.c
@@ -1834,8 +1834,8 @@ static void do_pages_stat_array(struct m
struct page *page;
int err = -EFAULT;
- vma = find_vma(mm, addr);
- if (!vma || addr < vma->vm_start)
+ vma = vma_lookup(mm, addr);
+ if (!vma)
goto set_status;
/* FOLL_DUMP to ignore special (like zero) pages */
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 124/192] mm/mremap: use vma_lookup() in vma_to_resize()
2021-06-29 2:32 incoming Andrew Morton
` (122 preceding siblings ...)
2021-06-29 2:39 ` [patch 123/192] mm/migrate: use vma_lookup() in do_pages_stat_array() Andrew Morton
@ 2021-06-29 2:39 ` Andrew Morton
2021-06-29 2:39 ` [patch 125/192] mm/memory.c: use vma_lookup() in __access_remote_vm() Andrew Morton
` (67 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:39 UTC (permalink / raw)
To: akpm, david, dbueso, geert, ldufour, Liam.Howlett, linux-mm,
mm-commits, torvalds
From: Liam Howlett <liam.howlett@oracle.com>
Subject: mm/mremap: use vma_lookup() in vma_to_resize()
Use vma_lookup() to find the VMA at a specific address. As vma_lookup()
will return NULL if the address is not within any VMA, the start address
no longer needs to be validated.
Link: https://lkml.kernel.org/r/20210521174745.2219620-21-Liam.Howlett@Oracle.com
Signed-off-by: Liam R. Howlett <Liam.Howlett@Oracle.com>
Reviewed-by: Laurent Dufour <ldufour@linux.ibm.com>
Acked-by: David Hildenbrand <david@redhat.com>
Acked-by: Davidlohr Bueso <dbueso@suse.de>
Cc: Geert Uytterhoeven <geert@linux-m68k.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/mremap.c | 5 +++--
1 file changed, 3 insertions(+), 2 deletions(-)
--- a/mm/mremap.c~mm-mremap-use-vma_lookup-in-vma_to_resize
+++ a/mm/mremap.c
@@ -634,10 +634,11 @@ static struct vm_area_struct *vma_to_res
unsigned long *p)
{
struct mm_struct *mm = current->mm;
- struct vm_area_struct *vma = find_vma(mm, addr);
+ struct vm_area_struct *vma;
unsigned long pgoff;
- if (!vma || vma->vm_start > addr)
+ vma = vma_lookup(mm, addr);
+ if (!vma)
return ERR_PTR(-EFAULT);
/*
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 125/192] mm/memory.c: use vma_lookup() in __access_remote_vm()
2021-06-29 2:32 incoming Andrew Morton
` (123 preceding siblings ...)
2021-06-29 2:39 ` [patch 124/192] mm/mremap: use vma_lookup() in vma_to_resize() Andrew Morton
@ 2021-06-29 2:39 ` Andrew Morton
2021-06-29 2:39 ` [patch 126/192] mm/mempolicy: " Andrew Morton
` (66 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:39 UTC (permalink / raw)
To: akpm, david, dbueso, geert, ldufour, Liam.Howlett, linux-mm,
mm-commits, torvalds
From: Liam Howlett <liam.howlett@oracle.com>
Subject: mm/memory.c: use vma_lookup() in __access_remote_vm()
Use vma_lookup() to find the VMA at a specific address. As vma_lookup()
will return NULL if the address is not within any VMA, the start address
no longer needs to be validated.
Link: https://lkml.kernel.org/r/20210521174745.2219620-22-Liam.Howlett@Oracle.com
Signed-off-by: Liam R. Howlett <Liam.Howlett@Oracle.com>
Reviewed-by: Laurent Dufour <ldufour@linux.ibm.com>
Acked-by: David Hildenbrand <david@redhat.com>
Acked-by: Davidlohr Bueso <dbueso@suse.de>
Cc: Geert Uytterhoeven <geert@linux-m68k.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/memory.c | 4 ++--
1 file changed, 2 insertions(+), 2 deletions(-)
--- a/mm/memory.c~mm-memoryc-use-vma_lookup-in-__access_remote_vm
+++ a/mm/memory.c
@@ -4994,8 +4994,8 @@ int __access_remote_vm(struct mm_struct
* Check if this is a VM_IO | VM_PFNMAP VMA, which
* we can access using slightly different code.
*/
- vma = find_vma(mm, addr);
- if (!vma || vma->vm_start > addr)
+ vma = vma_lookup(mm, addr);
+ if (!vma)
break;
if (vma->vm_ops && vma->vm_ops->access)
ret = vma->vm_ops->access(vma, addr, buf,
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 126/192] mm/mempolicy: use vma_lookup() in __access_remote_vm()
2021-06-29 2:32 incoming Andrew Morton
` (124 preceding siblings ...)
2021-06-29 2:39 ` [patch 125/192] mm/memory.c: use vma_lookup() in __access_remote_vm() Andrew Morton
@ 2021-06-29 2:39 ` Andrew Morton
2021-06-29 2:39 ` [patch 127/192] mm: update legacy flush_tlb_* to use vma Andrew Morton
` (65 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:39 UTC (permalink / raw)
To: akpm, david, dbueso, geert, ldufour, Liam.Howlett, linux-mm,
mm-commits, torvalds
From: Liam Howlett <liam.howlett@oracle.com>
Subject: mm/mempolicy: use vma_lookup() in __access_remote_vm()
vma_lookup() finds the vma of a specific address with a cleaner interface
and is more readable.
Link: https://lkml.kernel.org/r/20210521174745.2219620-23-Liam.Howlett@Oracle.com
Signed-off-by: Liam R. Howlett <Liam.Howlett@Oracle.com>
Reviewed-by: Laurent Dufour <ldufour@linux.ibm.com>
Acked-by: David Hildenbrand <david@redhat.com>
Acked-by: Davidlohr Bueso <dbueso@suse.de>
Cc: Geert Uytterhoeven <geert@linux-m68k.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/mempolicy.c | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
--- a/mm/mempolicy.c~mm-mempolicy-use-vma_lookup-in-__access_remote_vm
+++ a/mm/mempolicy.c
@@ -975,7 +975,7 @@ static long do_get_mempolicy(int *policy
* want to return MPOL_DEFAULT in this case.
*/
mmap_read_lock(mm);
- vma = find_vma_intersection(mm, addr, addr+1);
+ vma = vma_lookup(mm, addr);
if (!vma) {
mmap_read_unlock(mm);
return -EFAULT;
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 127/192] mm: update legacy flush_tlb_* to use vma
2021-06-29 2:32 incoming Andrew Morton
` (125 preceding siblings ...)
2021-06-29 2:39 ` [patch 126/192] mm/mempolicy: " Andrew Morton
@ 2021-06-29 2:39 ` Andrew Morton
2021-06-29 2:39 ` [patch 128/192] mm: improve mprotect(R|W) efficiency on pages referenced once Andrew Morton
` (64 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:39 UTC (permalink / raw)
To: akpm, chenli, chris, geert, jonas, linux-mm, linux, mm-commits, torvalds
From: Chen Li <chenli@uniontech.com>
Subject: mm: update legacy flush_tlb_* to use vma
1. These tlb flush functions have been using vma instead mm long time
ago, but there is still some coments use mm as parameter.
2. the actual struct we use is vm_area_struct instead of vma_struct.
3. remove unused flush_kern_tlb_page.
Link: https://lkml.kernel.org/r/87k0oaq311.wl-chenli@uniontech.com
Signed-off-by: Chen Li <chenli@uniontech.com>
Acked-by: Geert Uytterhoeven <geert@linux-m68k.org>
Cc: Russell King <linux@armlinux.org.uk>
Cc: Jonas Bonn <jonas@southpole.se>
Cc: Chris Zankel <chris@zankel.net>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
arch/arm/include/asm/tlbflush.h | 13 +++----------
arch/arm/mm/tlb-v6.S | 2 +-
arch/arm/mm/tlb-v7.S | 2 +-
arch/ia64/kernel/efi_stub.S | 2 +-
arch/m68k/include/asm/tlbflush.h | 2 +-
arch/openrisc/include/asm/tlbflush.h | 2 +-
arch/xtensa/include/asm/tlbflush.h | 4 ++--
7 files changed, 10 insertions(+), 17 deletions(-)
--- a/arch/arm/include/asm/tlbflush.h~mm-update-legacy-flush_tlb_-to-use-vma
+++ a/arch/arm/include/asm/tlbflush.h
@@ -253,7 +253,7 @@ extern struct cpu_tlb_fns cpu_tlb;
* space.
* - mm - mm_struct describing address space
*
- * flush_tlb_range(mm,start,end)
+ * flush_tlb_range(vma,start,end)
*
* Invalidate a range of TLB entries in the specified
* address space.
@@ -261,18 +261,11 @@ extern struct cpu_tlb_fns cpu_tlb;
* - start - start address (may not be aligned)
* - end - end address (exclusive, may not be aligned)
*
- * flush_tlb_page(vaddr,vma)
+ * flush_tlb_page(vma, uaddr)
*
* Invalidate the specified page in the specified address range.
+ * - vma - vm_area_struct describing address range
* - vaddr - virtual address (may not be aligned)
- * - vma - vma_struct describing address range
- *
- * flush_kern_tlb_page(kaddr)
- *
- * Invalidate the TLB entry for the specified page. The address
- * will be in the kernels virtual memory space. Current uses
- * only require the D-TLB to be invalidated.
- * - kaddr - Kernel virtual memory address
*/
/*
--- a/arch/arm/mm/tlb-v6.S~mm-update-legacy-flush_tlb_-to-use-vma
+++ a/arch/arm/mm/tlb-v6.S
@@ -24,7 +24,7 @@
*
* - start - start address (may not be aligned)
* - end - end address (exclusive, may not be aligned)
- * - vma - vma_struct describing address range
+ * - vma - vm_area_struct describing address range
*
* It is assumed that:
* - the "Invalidate single entry" instruction will invalidate
--- a/arch/arm/mm/tlb-v7.S~mm-update-legacy-flush_tlb_-to-use-vma
+++ a/arch/arm/mm/tlb-v7.S
@@ -23,7 +23,7 @@
*
* - start - start address (may not be aligned)
* - end - end address (exclusive, may not be aligned)
- * - vma - vma_struct describing address range
+ * - vma - vm_area_struct describing address range
*
* It is assumed that:
* - the "Invalidate single entry" instruction will invalidate
--- a/arch/ia64/kernel/efi_stub.S~mm-update-legacy-flush_tlb_-to-use-vma
+++ a/arch/ia64/kernel/efi_stub.S
@@ -7,7 +7,7 @@
*
* This stub allows us to make EFI calls in physical mode with interrupts
* turned off. We need this because we can't call SetVirtualMap() until
- * the kernel has booted far enough to allow allocation of struct vma_struct
+ * the kernel has booted far enough to allow allocation of struct vm_area_struct
* entries (which we would need to map stuff with memory attributes other
* than uncached or writeback...). Since the GetTime() service gets called
* earlier than that, we need to be able to make physical mode EFI calls from
--- a/arch/m68k/include/asm/tlbflush.h~mm-update-legacy-flush_tlb_-to-use-vma
+++ a/arch/m68k/include/asm/tlbflush.h
@@ -263,7 +263,7 @@ static inline void flush_tlb_page(struct
BUG();
}
-static inline void flush_tlb_range(struct mm_struct *mm,
+static inline void flush_tlb_range(struct vm_area_struct *vma,
unsigned long start, unsigned long end)
{
BUG();
--- a/arch/openrisc/include/asm/tlbflush.h~mm-update-legacy-flush_tlb_-to-use-vma
+++ a/arch/openrisc/include/asm/tlbflush.h
@@ -25,7 +25,7 @@
* - flush_tlb_all() flushes all processes TLBs
* - flush_tlb_mm(mm) flushes the specified mm context TLB's
* - flush_tlb_page(vma, vmaddr) flushes one page
- * - flush_tlb_range(mm, start, end) flushes a range of pages
+ * - flush_tlb_range(vma, start, end) flushes a range of pages
*/
extern void local_flush_tlb_all(void);
extern void local_flush_tlb_mm(struct mm_struct *mm);
--- a/arch/xtensa/include/asm/tlbflush.h~mm-update-legacy-flush_tlb_-to-use-vma
+++ a/arch/xtensa/include/asm/tlbflush.h
@@ -26,8 +26,8 @@
*
* - flush_tlb_all() flushes all processes TLB entries
* - flush_tlb_mm(mm) flushes the specified mm context TLB entries
- * - flush_tlb_page(mm, vmaddr) flushes a single page
- * - flush_tlb_range(mm, start, end) flushes a range of pages
+ * - flush_tlb_page(vma, page) flushes a single page
+ * - flush_tlb_range(vma, vmaddr, end) flushes a range of pages
*/
void local_flush_tlb_all(void);
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 128/192] mm: improve mprotect(R|W) efficiency on pages referenced once
2021-06-29 2:32 incoming Andrew Morton
` (126 preceding siblings ...)
2021-06-29 2:39 ` [patch 127/192] mm: update legacy flush_tlb_* to use vma Andrew Morton
@ 2021-06-29 2:39 ` Andrew Morton
2021-06-29 17:50 ` Linus Torvalds
2021-06-29 2:40 ` [patch 129/192] h8300: remove unused variable Andrew Morton
` (63 subsequent siblings)
191 siblings, 1 reply; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:39 UTC (permalink / raw)
To: aarcange, akpm, eugenis, kostyak, linux-mm, mm-commits, pcc,
peterx, torvalds
From: Peter Collingbourne <pcc@google.com>
Subject: mm: improve mprotect(R|W) efficiency on pages referenced once
In the Scudo memory allocator [1] we would like to be able to detect
use-after-free vulnerabilities involving large allocations by issuing
mprotect(PROT_NONE) on the memory region used for the allocation when it
is deallocated. Later on, after the memory region has been "quarantined"
for a sufficient period of time we would like to be able to use it for
another allocation by issuing mprotect(PROT_READ|PROT_WRITE).
Before this patch, after removing the write protection, any writes to the
memory region would result in page faults and entering the copy-on-write
code path, even in the usual case where the pages are only referenced by a
single PTE, harming performance unnecessarily. Make it so that any pages
in anonymous mappings that are only referenced by a single PTE are
immediately made writable during the mprotect so that we can avoid the
page faults.
This program shows the critical syscall sequence that we intend to use in
the allocator:
#include <string.h>
#include <sys/mman.h>
enum { kSize = 131072 };
int main(int argc, char **argv) {
char *addr = (char *)mmap(0, kSize, PROT_READ | PROT_WRITE,
MAP_ANONYMOUS | MAP_PRIVATE, -1, 0);
for (int i = 0; i != 100000; ++i) {
memset(addr, i, kSize);
mprotect((void *)addr, kSize, PROT_NONE);
mprotect((void *)addr, kSize, PROT_READ | PROT_WRITE);
}
}
The effect of this patch on the above program was measured on a
DragonBoard 845c by taking the median real time execution time of 10 runs.
Before: 2.94s
After: 0.66s
The effect was also measured using one of the microbenchmarks that we
normally use to benchmark the allocator [2], after modifying it to make
the appropriate mprotect calls [3]. With an allocation size of 131072
bytes to trigger the allocator's "large allocation" code path the
per-iteration time was measured as follows:
Before: 27450ns
After: 6010ns
This patch means that we do more work during the mprotect call itself in
exchange for less work when the pages are accessed. In the worst case,
the pages are not accessed at all. The effect of this patch in such cases
was measured using the following program:
#include <string.h>
#include <sys/mman.h>
enum { kSize = 131072 };
int main(int argc, char **argv) {
char *addr = (char *)mmap(0, kSize, PROT_READ | PROT_WRITE,
MAP_ANONYMOUS | MAP_PRIVATE, -1, 0);
memset(addr, 1, kSize);
for (int i = 0; i != 100000; ++i) {
#ifdef PAGE_FAULT
memset(addr + (i * 4096) % kSize, i, 4096);
#endif
mprotect((void *)addr, kSize, PROT_NONE);
mprotect((void *)addr, kSize, PROT_READ | PROT_WRITE);
}
}
With PAGE_FAULT undefined (0 pages touched after removing write
protection) the median real time execution time of 100 runs was measured
as follows:
Before: 0.330260s
After: 0.338836s
With PAGE_FAULT defined (1 page touched) the measurements were
as follows:
Before: 0.438048s
After: 0.355661s
So it seems that even with a single page fault the new approach is faster.
I saw similar results if I adjusted the programs to use a larger mapping
size. With kSize = 1048576 I get these numbers with PAGE_FAULT undefined:
Before: 1.428988s
After: 1.512016s
i.e. around 5.5%.
And these with PAGE_FAULT defined:
Before: 1.518559s
After: 1.524417s
i.e. about the same.
What I think we may conclude from these results is that for smaller
mappings the advantage of the previous approach, although measurable, is
wiped out by a single page fault. I think we may expect that there should
be at least one access resulting in a page fault (under the previous
approach) after making the pages writable, since the program presumably
made the pages writable for a reason.
For larger mappings we may guesstimate that the new approach wins if the
density of future page faults is > 0.4%. But for the mappings that are
large enough for density to matter (not just the absolute number of page
faults) it doesn't seem like the increase in mprotect latency would be
very large relative to the total mprotect execution time.
[pcc@google.com: add comments, prohibit optimization for NUMA pages]
Link: https://lkml.kernel.org/r/20210601185926.2623183-1-pcc@google.com
Link: https://lkml.kernel.org/r/20210527190453.1259020-1-pcc@google.com
Link: https://linux-review.googlesource.com/id/I98d75ef90e20330c578871c87494d64b1df3f1b8
Link: [1] https://source.android.com/devices/tech/debug/scudo
Link: [2] https://cs.android.com/android/platform/superproject/+/master:bionic/benchmarks/stdlib_benchmark.cpp;l=53;drc=e8693e78711e8f45ccd2b610e4dbe0b94d551cc9
Link: [3] https://github.com/pcc/llvm-project/commit/scudo-mprotect-secondary2
Signed-off-by: Peter Collingbourne <pcc@google.com>
Reviewed-by: Peter Xu <peterx@redhat.com>
Cc: Kostya Kortchinsky <kostyak@google.com>
Cc: Evgenii Stepanov <eugenis@google.com>
Cc: Andrea Arcangeli <aarcange@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/mprotect.c | 52 ++++++++++++++++++++++++++++++++++++++++++------
1 file changed, 46 insertions(+), 6 deletions(-)
--- a/mm/mprotect.c~mm-improve-mprotectrw-efficiency-on-pages-referenced-once
+++ a/mm/mprotect.c
@@ -35,6 +35,51 @@
#include "internal.h"
+/* Determine whether we can avoid taking write faults for known dirty pages. */
+static bool may_avoid_write_fault(pte_t pte, struct vm_area_struct *vma,
+ unsigned long cp_flags)
+{
+ /*
+ * The dirty accountable bit indicates that we can always make the page
+ * writable regardless of the number of references.
+ */
+ if (!(cp_flags & MM_CP_DIRTY_ACCT)) {
+ /* Otherwise, we must have exclusive access to the page. */
+ if (!(vma_is_anonymous(vma) && (vma->vm_flags & VM_WRITE)))
+ return false;
+
+ if (page_count(pte_page(pte)) != 1)
+ return false;
+ }
+
+ /*
+ * Don't do this optimization for clean pages as we need to be notified
+ * of the transition from clean to dirty.
+ */
+ if (!pte_dirty(pte))
+ return false;
+
+ /* Same for softdirty. */
+ if (!pte_soft_dirty(pte) && (vma->vm_flags & VM_SOFTDIRTY))
+ return false;
+
+ /*
+ * For userfaultfd the user program needs to monitor write faults so we
+ * can't do this optimization.
+ */
+ if (pte_uffd_wp(pte))
+ return false;
+
+ /*
+ * It is unclear whether this optimization can be done safely for NUMA
+ * pages.
+ */
+ if (cp_flags & MM_CP_PROT_NUMA)
+ return false;
+
+ return true;
+}
+
static unsigned long change_pte_range(struct vm_area_struct *vma, pmd_t *pmd,
unsigned long addr, unsigned long end, pgprot_t newprot,
unsigned long cp_flags)
@@ -43,7 +88,6 @@ static unsigned long change_pte_range(st
spinlock_t *ptl;
unsigned long pages = 0;
int target_node = NUMA_NO_NODE;
- bool dirty_accountable = cp_flags & MM_CP_DIRTY_ACCT;
bool prot_numa = cp_flags & MM_CP_PROT_NUMA;
bool uffd_wp = cp_flags & MM_CP_UFFD_WP;
bool uffd_wp_resolve = cp_flags & MM_CP_UFFD_WP_RESOLVE;
@@ -131,12 +175,8 @@ static unsigned long change_pte_range(st
ptent = pte_clear_uffd_wp(ptent);
}
- /* Avoid taking write faults for known dirty pages */
- if (dirty_accountable && pte_dirty(ptent) &&
- (pte_soft_dirty(ptent) ||
- !(vma->vm_flags & VM_SOFTDIRTY))) {
+ if (may_avoid_write_fault(ptent, vma, cp_flags))
ptent = pte_mkwrite(ptent);
- }
ptep_modify_prot_commit(vma, addr, pte, oldpte, ptent);
pages++;
} else if (is_swap_pte(oldpte)) {
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 129/192] h8300: remove unused variable
2021-06-29 2:32 incoming Andrew Morton
` (127 preceding siblings ...)
2021-06-29 2:39 ` [patch 128/192] mm: improve mprotect(R|W) efficiency on pages referenced once Andrew Morton
@ 2021-06-29 2:40 ` Andrew Morton
2021-06-29 2:40 ` [patch 130/192] mm/dmapool: use DEVICE_ATTR_RO macro Andrew Morton
` (62 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:40 UTC (permalink / raw)
To: akpm, jrdr.linux, linux-mm, lkp, mm-commits, rppt, torvalds, ysato
From: Souptick Joarder <jrdr.linux@gmail.com>
Subject: h8300: remove unused variable
Kernel test robot throws below warning ->
>> arch/h8300/kernel/setup.c:72:26:
warning: Unused variable: region [unusedVariable]
struct memblock_region *region;
Fixed it by removing unused variable.
Link: https://lkml.kernel.org/r/20210602185431.11416-1-jrdr.linux@gmail.com
Signed-off-by: Souptick Joarder <jrdr.linux@gmail.com>
Reported-by: kernel test robot <lkp@intel.com>
Acked-by: Mike Rapoport <rppt@linux.ibm.com>
Acked-by: Yoshinori Sato <ysato@users.sourceforge.jp>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
arch/h8300/kernel/setup.c | 2 --
1 file changed, 2 deletions(-)
--- a/arch/h8300/kernel/setup.c~h8300-remove-unused-variable
+++ a/arch/h8300/kernel/setup.c
@@ -69,8 +69,6 @@ void __init h8300_fdt_init(void *fdt, ch
static void __init bootmem_init(void)
{
- struct memblock_region *region;
-
memory_end = memory_start = 0;
/* Find main memory where is the kernel */
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 130/192] mm/dmapool: use DEVICE_ATTR_RO macro
2021-06-29 2:32 incoming Andrew Morton
` (128 preceding siblings ...)
2021-06-29 2:40 ` [patch 129/192] h8300: remove unused variable Andrew Morton
@ 2021-06-29 2:40 ` Andrew Morton
2021-06-29 2:40 ` [patch 131/192] mm, tracing: unify PFN format strings Andrew Morton
` (61 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:40 UTC (permalink / raw)
To: akpm, andriy.shevchenko, linux-mm, mm-commits, torvalds, yuehaibing
From: YueHaibing <yuehaibing@huawei.com>
Subject: mm/dmapool: use DEVICE_ATTR_RO macro
Use DEVICE_ATTR_RO() helper instead of plain DEVICE_ATTR(), which makes
the code a bit shorter and easier to read.
Link: https://lkml.kernel.org/r/20210524112852.34716-1-yuehaibing@huawei.com
Signed-off-by: YueHaibing <yuehaibing@huawei.com>
Reviewed-by: Andy Shevchenko <andriy.shevchenko@linux.intel.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/dmapool.c | 5 ++---
1 file changed, 2 insertions(+), 3 deletions(-)
--- a/mm/dmapool.c~mm-dmapool-use-device_attr_ro-macro
+++ a/mm/dmapool.c
@@ -62,8 +62,7 @@ struct dma_page { /* cacheable header f
static DEFINE_MUTEX(pools_lock);
static DEFINE_MUTEX(pools_reg_lock);
-static ssize_t
-show_pools(struct device *dev, struct device_attribute *attr, char *buf)
+static ssize_t pools_show(struct device *dev, struct device_attribute *attr, char *buf)
{
unsigned temp;
unsigned size;
@@ -103,7 +102,7 @@ show_pools(struct device *dev, struct de
return PAGE_SIZE - size;
}
-static DEVICE_ATTR(pools, 0444, show_pools, NULL);
+static DEVICE_ATTR_RO(pools);
/**
* dma_pool_create - Creates a pool of consistent memory blocks, for dma.
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 131/192] mm, tracing: unify PFN format strings
2021-06-29 2:32 incoming Andrew Morton
` (129 preceding siblings ...)
2021-06-29 2:40 ` [patch 130/192] mm/dmapool: use DEVICE_ATTR_RO macro Andrew Morton
@ 2021-06-29 2:40 ` Andrew Morton
2021-06-29 2:40 ` [patch 132/192] mm/page_alloc: add an alloc_pages_bulk_array_node() helper Andrew Morton
` (60 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:40 UTC (permalink / raw)
To: akpm, hawk, ilias.apalodimas, linux-mm, mingo, mm-commits,
rostedt, torvalds, vincent.whitchurch
From: Vincent Whitchurch <vincent.whitchurch@axis.com>
Subject: mm, tracing: unify PFN format strings
Some trace event formats print PFNs as hex while others print them as
decimal. This is rather annoying when attempting to grep through traces
to understand what's going on with a particular page.
$ git grep -ho 'pfn=[0x%lu]\+' include/trace/events/ | sort | uniq -c
11 pfn=0x%lx
12 pfn=%lu
2 pfn=%lx
Printing as hex is in the majority in the trace events, and all the normal
printks in mm/ also print PFNs as hex, so change all the PFN formats in
the trace events to use 0x%lx.
Link: https://lkml.kernel.org/r/20210602092608.1493-1-vincent.whitchurch@axis.com
Signed-off-by: Vincent Whitchurch <vincent.whitchurch@axis.com>
Cc: Steven Rostedt <rostedt@goodmis.org>
Cc: Ingo Molnar <mingo@redhat.com>
Cc: Jesper Dangaard Brouer <hawk@kernel.org>
Cc: Ilias Apalodimas <ilias.apalodimas@linaro.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/trace/events/cma.h | 4 ++--
include/trace/events/filemap.h | 2 +-
include/trace/events/kmem.h | 12 ++++++------
include/trace/events/page_pool.h | 4 ++--
include/trace/events/pagemap.h | 4 ++--
include/trace/events/vmscan.h | 2 +-
6 files changed, 14 insertions(+), 14 deletions(-)
--- a/include/trace/events/cma.h~mm-tracing-unify-pfn-format-strings
+++ a/include/trace/events/cma.h
@@ -31,7 +31,7 @@ DECLARE_EVENT_CLASS(cma_alloc_class,
__entry->align = align;
),
- TP_printk("name=%s pfn=%lx page=%p count=%lu align=%u",
+ TP_printk("name=%s pfn=0x%lx page=%p count=%lu align=%u",
__get_str(name),
__entry->pfn,
__entry->page,
@@ -60,7 +60,7 @@ TRACE_EVENT(cma_release,
__entry->count = count;
),
- TP_printk("name=%s pfn=%lx page=%p count=%lu",
+ TP_printk("name=%s pfn=0x%lx page=%p count=%lu",
__get_str(name),
__entry->pfn,
__entry->page,
--- a/include/trace/events/filemap.h~mm-tracing-unify-pfn-format-strings
+++ a/include/trace/events/filemap.h
@@ -36,7 +36,7 @@ DECLARE_EVENT_CLASS(mm_filemap_op_page_c
__entry->s_dev = page->mapping->host->i_rdev;
),
- TP_printk("dev %d:%d ino %lx page=%p pfn=%lu ofs=%lu",
+ TP_printk("dev %d:%d ino %lx page=%p pfn=0x%lx ofs=%lu",
MAJOR(__entry->s_dev), MINOR(__entry->s_dev),
__entry->i_ino,
pfn_to_page(__entry->pfn),
--- a/include/trace/events/kmem.h~mm-tracing-unify-pfn-format-strings
+++ a/include/trace/events/kmem.h
@@ -173,7 +173,7 @@ TRACE_EVENT(mm_page_free,
__entry->order = order;
),
- TP_printk("page=%p pfn=%lu order=%d",
+ TP_printk("page=%p pfn=0x%lx order=%d",
pfn_to_page(__entry->pfn),
__entry->pfn,
__entry->order)
@@ -193,7 +193,7 @@ TRACE_EVENT(mm_page_free_batched,
__entry->pfn = page_to_pfn(page);
),
- TP_printk("page=%p pfn=%lu order=0",
+ TP_printk("page=%p pfn=0x%lx order=0",
pfn_to_page(__entry->pfn),
__entry->pfn)
);
@@ -219,7 +219,7 @@ TRACE_EVENT(mm_page_alloc,
__entry->migratetype = migratetype;
),
- TP_printk("page=%p pfn=%lu order=%d migratetype=%d gfp_flags=%s",
+ TP_printk("page=%p pfn=0x%lx order=%d migratetype=%d gfp_flags=%s",
__entry->pfn != -1UL ? pfn_to_page(__entry->pfn) : NULL,
__entry->pfn != -1UL ? __entry->pfn : 0,
__entry->order,
@@ -245,7 +245,7 @@ DECLARE_EVENT_CLASS(mm_page,
__entry->migratetype = migratetype;
),
- TP_printk("page=%p pfn=%lu order=%u migratetype=%d percpu_refill=%d",
+ TP_printk("page=%p pfn=0x%lx order=%u migratetype=%d percpu_refill=%d",
__entry->pfn != -1UL ? pfn_to_page(__entry->pfn) : NULL,
__entry->pfn != -1UL ? __entry->pfn : 0,
__entry->order,
@@ -278,7 +278,7 @@ TRACE_EVENT(mm_page_pcpu_drain,
__entry->migratetype = migratetype;
),
- TP_printk("page=%p pfn=%lu order=%d migratetype=%d",
+ TP_printk("page=%p pfn=0x%lx order=%d migratetype=%d",
pfn_to_page(__entry->pfn), __entry->pfn,
__entry->order, __entry->migratetype)
);
@@ -312,7 +312,7 @@ TRACE_EVENT(mm_page_alloc_extfrag,
get_pageblock_migratetype(page));
),
- TP_printk("page=%p pfn=%lu alloc_order=%d fallback_order=%d pageblock_order=%d alloc_migratetype=%d fallback_migratetype=%d fragmenting=%d change_ownership=%d",
+ TP_printk("page=%p pfn=0x%lx alloc_order=%d fallback_order=%d pageblock_order=%d alloc_migratetype=%d fallback_migratetype=%d fragmenting=%d change_ownership=%d",
pfn_to_page(__entry->pfn),
__entry->pfn,
__entry->alloc_order,
--- a/include/trace/events/pagemap.h~mm-tracing-unify-pfn-format-strings
+++ a/include/trace/events/pagemap.h
@@ -46,7 +46,7 @@ TRACE_EVENT(mm_lru_insertion,
),
/* Flag format is based on page-types.c formatting for pagemap */
- TP_printk("page=%p pfn=%lu lru=%d flags=%s%s%s%s%s%s",
+ TP_printk("page=%p pfn=0x%lx lru=%d flags=%s%s%s%s%s%s",
__entry->page,
__entry->pfn,
__entry->lru,
@@ -75,7 +75,7 @@ TRACE_EVENT(mm_lru_activate,
),
/* Flag format is based on page-types.c formatting for pagemap */
- TP_printk("page=%p pfn=%lu", __entry->page, __entry->pfn)
+ TP_printk("page=%p pfn=0x%lx", __entry->page, __entry->pfn)
);
--- a/include/trace/events/page_pool.h~mm-tracing-unify-pfn-format-strings
+++ a/include/trace/events/page_pool.h
@@ -60,7 +60,7 @@ TRACE_EVENT(page_pool_state_release,
__entry->pfn = page_to_pfn(page);
),
- TP_printk("page_pool=%p page=%p pfn=%lu release=%u",
+ TP_printk("page_pool=%p page=%p pfn=0x%lx release=%u",
__entry->pool, __entry->page, __entry->pfn, __entry->release)
);
@@ -85,7 +85,7 @@ TRACE_EVENT(page_pool_state_hold,
__entry->pfn = page_to_pfn(page);
),
- TP_printk("page_pool=%p page=%p pfn=%lu hold=%u",
+ TP_printk("page_pool=%p page=%p pfn=0x%lx hold=%u",
__entry->pool, __entry->page, __entry->pfn, __entry->hold)
);
--- a/include/trace/events/vmscan.h~mm-tracing-unify-pfn-format-strings
+++ a/include/trace/events/vmscan.h
@@ -330,7 +330,7 @@ TRACE_EVENT(mm_vmscan_writepage,
page_is_file_lru(page));
),
- TP_printk("page=%p pfn=%lu flags=%s",
+ TP_printk("page=%p pfn=0x%lx flags=%s",
pfn_to_page(__entry->pfn),
__entry->pfn,
show_reclaim_flags(__entry->reclaim_flags))
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 132/192] mm/page_alloc: add an alloc_pages_bulk_array_node() helper
2021-06-29 2:32 incoming Andrew Morton
` (130 preceding siblings ...)
2021-06-29 2:40 ` [patch 131/192] mm, tracing: unify PFN format strings Andrew Morton
@ 2021-06-29 2:40 ` Andrew Morton
2021-06-29 2:40 ` [patch 133/192] mm/vmalloc: switch to bulk allocator in __vmalloc_area_node() Andrew Morton
` (59 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:40 UTC (permalink / raw)
To: akpm, hdanton, linux-mm, mgorman, mhocko, mm-commits, npiggin,
oleksiy.avramchenko, rostedt, torvalds, urezki, willy
From: "Uladzislau Rezki (Sony)" <urezki@gmail.com>
Subject: mm/page_alloc: add an alloc_pages_bulk_array_node() helper
Patch series "vmalloc() vs bulk allocator", v2.
This patch (of 3):
Add a "node" variant of the alloc_pages_bulk_array() function. The helper
guarantees that a __alloc_pages_bulk() is invoked with a valid NUMA node
ID.
Link: https://lkml.kernel.org/r/20210516202056.2120-1-urezki@gmail.com
Link: https://lkml.kernel.org/r/20210516202056.2120-2-urezki@gmail.com
Signed-off-by: Uladzislau Rezki (Sony) <urezki@gmail.com>
Acked-by: Mel Gorman <mgorman@suse.de>
Cc: Mel Gorman <mgorman@suse.de>
Cc: Matthew Wilcox <willy@infradead.org>
Cc: Nicholas Piggin <npiggin@gmail.com>
Cc: Hillf Danton <hdanton@sina.com>
Cc: Michal Hocko <mhocko@suse.com>
Cc: Oleksiy Avramchenko <oleksiy.avramchenko@sonymobile.com>
Cc: Steven Rostedt <rostedt@goodmis.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/linux/gfp.h | 9 +++++++++
1 file changed, 9 insertions(+)
--- a/include/linux/gfp.h~mm-page_alloc-add-an-alloc_pages_bulk_array_node-helper
+++ a/include/linux/gfp.h
@@ -536,6 +536,15 @@ alloc_pages_bulk_array(gfp_t gfp, unsign
return __alloc_pages_bulk(gfp, numa_mem_id(), NULL, nr_pages, NULL, page_array);
}
+static inline unsigned long
+alloc_pages_bulk_array_node(gfp_t gfp, int nid, unsigned long nr_pages, struct page **page_array)
+{
+ if (nid == NUMA_NO_NODE)
+ nid = numa_mem_id();
+
+ return __alloc_pages_bulk(gfp, nid, NULL, nr_pages, NULL, page_array);
+}
+
/*
* Allocate pages, preferring the node given as nid. The node must be valid and
* online. For more general interface, see alloc_pages_node().
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 133/192] mm/vmalloc: switch to bulk allocator in __vmalloc_area_node()
2021-06-29 2:32 incoming Andrew Morton
` (131 preceding siblings ...)
2021-06-29 2:40 ` [patch 132/192] mm/page_alloc: add an alloc_pages_bulk_array_node() helper Andrew Morton
@ 2021-06-29 2:40 ` Andrew Morton
2021-06-29 2:40 ` [patch 134/192] mm/vmalloc: print a warning message first on failure Andrew Morton
` (58 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:40 UTC (permalink / raw)
To: akpm, hdanton, linux-mm, mgorman, mhocko, mm-commits, npiggin,
oleksiy.avramchenko, rostedt, torvalds, urezki, willy
From: "Uladzislau Rezki (Sony)" <urezki@gmail.com>
Subject: mm/vmalloc: switch to bulk allocator in __vmalloc_area_node()
Recently there has been introduced a page bulk allocator for users which
need to get number of pages per one call request.
For order-0 pages switch to an alloc_pages_bulk_array_node() instead of
alloc_pages_node(), the reason is the former is not capable of allocating
set of pages, thus a one call is per one page.
Second, according to my tests the bulk allocator uses less cycles even for
scenarios when only one page is requested. Running the "perf" on same
test case shows below difference:
<default>
- 45.18% __vmalloc_node
- __vmalloc_node_range
- 35.60% __alloc_pages
- get_page_from_freelist
3.36% __list_del_entry_valid
3.00% check_preemption_disabled
1.42% prep_new_page
<default>
<patch>
- 31.00% __vmalloc_node
- __vmalloc_node_range
- 14.48% __alloc_pages_bulk
3.22% __list_del_entry_valid
- 0.83% __alloc_pages
get_page_from_freelist
<patch>
The "test_vmalloc.sh" also shows performance improvements:
fix_size_alloc_test_4MB loops: 1000000 avg: 89105095 usec
fix_size_alloc_test loops: 1000000 avg: 513672 usec
full_fit_alloc_test loops: 1000000 avg: 748900 usec
long_busy_list_alloc_test loops: 1000000 avg: 8043038 usec
random_size_alloc_test loops: 1000000 avg: 4028582 usec
fix_align_alloc_test loops: 1000000 avg: 1457671 usec
fix_size_alloc_test_4MB loops: 1000000 avg: 62083711 usec
fix_size_alloc_test loops: 1000000 avg: 449207 usec
full_fit_alloc_test loops: 1000000 avg: 735985 usec
long_busy_list_alloc_test loops: 1000000 avg: 5176052 usec
random_size_alloc_test loops: 1000000 avg: 2589252 usec
fix_align_alloc_test loops: 1000000 avg: 1365009 usec
For example 4MB allocations illustrates ~30% gain, all the
rest is also better.
Link: https://lkml.kernel.org/r/20210516202056.2120-3-urezki@gmail.com
Signed-off-by: Uladzislau Rezki (Sony) <urezki@gmail.com>
Acked-by: Mel Gorman <mgorman@suse.de>
Cc: Hillf Danton <hdanton@sina.com>
Cc: Matthew Wilcox <willy@infradead.org>
Cc: Michal Hocko <mhocko@suse.com>
Cc: Nicholas Piggin <npiggin@gmail.com>
Cc: Oleksiy Avramchenko <oleksiy.avramchenko@sonymobile.com>
Cc: Steven Rostedt <rostedt@goodmis.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/vmalloc.c | 76 +++++++++++++++++++++++++++----------------------
1 file changed, 42 insertions(+), 34 deletions(-)
--- a/mm/vmalloc.c~mm-vmalloc-switch-to-bulk-allocator-in-__vmalloc_area_node
+++ a/mm/vmalloc.c
@@ -2768,8 +2768,6 @@ static void *__vmalloc_area_node(struct
unsigned long array_size;
unsigned int nr_small_pages = size >> PAGE_SHIFT;
unsigned int page_order;
- struct page **pages;
- unsigned int i;
array_size = (unsigned long)nr_small_pages * sizeof(struct page *);
gfp_mask |= __GFP_NOWARN;
@@ -2778,13 +2776,13 @@ static void *__vmalloc_area_node(struct
/* Please note that the recursion is strictly bounded. */
if (array_size > PAGE_SIZE) {
- pages = __vmalloc_node(array_size, 1, nested_gfp, node,
+ area->pages = __vmalloc_node(array_size, 1, nested_gfp, node,
area->caller);
} else {
- pages = kmalloc_node(array_size, nested_gfp, node);
+ area->pages = kmalloc_node(array_size, nested_gfp, node);
}
- if (!pages) {
+ if (!area->pages) {
free_vm_area(area);
warn_alloc(gfp_mask, NULL,
"vmalloc size %lu allocation failure: "
@@ -2793,43 +2791,53 @@ static void *__vmalloc_area_node(struct
return NULL;
}
- area->pages = pages;
- area->nr_pages = nr_small_pages;
+ area->nr_pages = 0;
set_vm_area_page_order(area, page_shift - PAGE_SHIFT);
-
page_order = vm_area_page_order(area);
- /*
- * Careful, we allocate and map page_order pages, but tracking is done
- * per PAGE_SIZE page so as to keep the vm_struct APIs independent of
- * the physical/mapped size.
- */
- for (i = 0; i < area->nr_pages; i += 1U << page_order) {
- struct page *page;
- int p;
-
- /* Compound pages required for remap_vmalloc_page */
- page = alloc_pages_node(node, gfp_mask | __GFP_COMP, page_order);
- if (unlikely(!page)) {
- /* Successfully allocated i pages, free them in __vfree() */
- area->nr_pages = i;
- atomic_long_add(area->nr_pages, &nr_vmalloc_pages);
- warn_alloc(gfp_mask, NULL,
- "vmalloc size %lu allocation failure: "
- "page order %u allocation failed",
- area->nr_pages * PAGE_SIZE, page_order);
- goto fail;
- }
+ if (!page_order) {
+ area->nr_pages = alloc_pages_bulk_array_node(
+ gfp_mask, node, nr_small_pages, area->pages);
+ } else {
+ /*
+ * Careful, we allocate and map page_order pages, but tracking is done
+ * per PAGE_SIZE page so as to keep the vm_struct APIs independent of
+ * the physical/mapped size.
+ */
+ while (area->nr_pages < nr_small_pages) {
+ struct page *page;
+ int i;
+
+ /* Compound pages required for remap_vmalloc_page */
+ page = alloc_pages_node(node, gfp_mask | __GFP_COMP, page_order);
+ if (unlikely(!page))
+ break;
+
+ for (i = 0; i < (1U << page_order); i++)
+ area->pages[area->nr_pages + i] = page + i;
- for (p = 0; p < (1U << page_order); p++)
- area->pages[i + p] = page + p;
+ if (gfpflags_allow_blocking(gfp_mask))
+ cond_resched();
- if (gfpflags_allow_blocking(gfp_mask))
- cond_resched();
+ area->nr_pages += 1U << page_order;
+ }
}
+
atomic_long_add(area->nr_pages, &nr_vmalloc_pages);
- if (vmap_pages_range(addr, addr + size, prot, pages, page_shift) < 0) {
+ /*
+ * If not enough pages were obtained to accomplish an
+ * allocation request, free them via __vfree() if any.
+ */
+ if (area->nr_pages != nr_small_pages) {
+ warn_alloc(gfp_mask, NULL,
+ "vmalloc size %lu allocation failure: "
+ "page order %u allocation failed",
+ area->nr_pages * PAGE_SIZE, page_order);
+ goto fail;
+ }
+
+ if (vmap_pages_range(addr, addr + size, prot, area->pages, page_shift) < 0) {
warn_alloc(gfp_mask, NULL,
"vmalloc size %lu allocation failure: "
"failed to map pages",
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 134/192] mm/vmalloc: print a warning message first on failure
2021-06-29 2:32 incoming Andrew Morton
` (132 preceding siblings ...)
2021-06-29 2:40 ` [patch 133/192] mm/vmalloc: switch to bulk allocator in __vmalloc_area_node() Andrew Morton
@ 2021-06-29 2:40 ` Andrew Morton
2021-06-29 2:40 ` [patch 135/192] mm/vmalloc: remove quoted strings split across lines Andrew Morton
` (57 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:40 UTC (permalink / raw)
To: akpm, hdanton, linux-mm, mgorman, mhocko, mm-commits, npiggin,
oleksiy.avramchenko, rostedt, torvalds, urezki, willy
From: "Uladzislau Rezki (Sony)" <urezki@gmail.com>
Subject: mm/vmalloc: print a warning message first on failure
When a memory allocation for array of pages are not succeed emit a warning
message as a first step and then perform the further cleanup.
The reason it should be done in a right order is the clean up function
which is free_vm_area() can potentially also follow its error paths what
can lead to confusion what was broken first.
Link: https://lkml.kernel.org/r/20210516202056.2120-4-urezki@gmail.com
Signed-off-by: Uladzislau Rezki (Sony) <urezki@gmail.com>
Cc: Hillf Danton <hdanton@sina.com>
Cc: Matthew Wilcox <willy@infradead.org>
Cc: Mel Gorman <mgorman@suse.de>
Cc: Michal Hocko <mhocko@suse.com>
Cc: Nicholas Piggin <npiggin@gmail.com>
Cc: Oleksiy Avramchenko <oleksiy.avramchenko@sonymobile.com>
Cc: Steven Rostedt <rostedt@goodmis.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/vmalloc.c | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
--- a/mm/vmalloc.c~mm-vmalloc-print-a-warning-message-first-on-failure
+++ a/mm/vmalloc.c
@@ -2783,11 +2783,11 @@ static void *__vmalloc_area_node(struct
}
if (!area->pages) {
- free_vm_area(area);
warn_alloc(gfp_mask, NULL,
"vmalloc size %lu allocation failure: "
"page array size %lu allocation failed",
nr_small_pages * PAGE_SIZE, array_size);
+ free_vm_area(area);
return NULL;
}
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 135/192] mm/vmalloc: remove quoted strings split across lines
2021-06-29 2:32 incoming Andrew Morton
` (133 preceding siblings ...)
2021-06-29 2:40 ` [patch 134/192] mm/vmalloc: print a warning message first on failure Andrew Morton
@ 2021-06-29 2:40 ` Andrew Morton
2021-06-29 2:40 ` [patch 136/192] mm/vmalloc: fallback to a single page allocator Andrew Morton
` (56 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:40 UTC (permalink / raw)
To: akpm, hch, hdanton, linux-mm, mgorman, mhocko, mm-commits,
npiggin, oleksiy.avramchenko, rostedt, torvalds, urezki, willy
From: "Uladzislau Rezki (Sony)" <urezki@gmail.com>
Subject: mm/vmalloc: remove quoted strings split across lines
A checkpatch.pl script complains on splitting a text across lines. It is
because if a user wants to find an entire string he or she will not
succeeded.
<snip>
WARNING: quoted string split across lines
+ "vmalloc size %lu allocation failure: "
+ "page order %u allocation failed",
total: 0 errors, 1 warnings, 10 lines checked
<snip>
Link: https://lkml.kernel.org/r/20210521204359.19943-1-urezki@gmail.com
Signed-off-by: Uladzislau Rezki (Sony) <urezki@gmail.com>
Cc: Mel Gorman <mgorman@suse.de>
Cc: Christoph Hellwig <hch@infradead.org>
Cc: Matthew Wilcox <willy@infradead.org>
Cc: Nicholas Piggin <npiggin@gmail.com>
Cc: Hillf Danton <hdanton@sina.com>
Cc: Michal Hocko <mhocko@suse.com>
Cc: Oleksiy Avramchenko <oleksiy.avramchenko@sonymobile.com>
Cc: Steven Rostedt <rostedt@goodmis.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/vmalloc.c | 21 +++++++++------------
1 file changed, 9 insertions(+), 12 deletions(-)
--- a/mm/vmalloc.c~mm-vmalloc-remove-quoted-string-split-across-lines
+++ a/mm/vmalloc.c
@@ -2784,9 +2784,8 @@ static void *__vmalloc_area_node(struct
if (!area->pages) {
warn_alloc(gfp_mask, NULL,
- "vmalloc size %lu allocation failure: "
- "page array size %lu allocation failed",
- nr_small_pages * PAGE_SIZE, array_size);
+ "vmalloc error: size %lu, failed to allocated page array size %lu",
+ nr_small_pages * PAGE_SIZE, array_size);
free_vm_area(area);
return NULL;
}
@@ -2831,17 +2830,15 @@ static void *__vmalloc_area_node(struct
*/
if (area->nr_pages != nr_small_pages) {
warn_alloc(gfp_mask, NULL,
- "vmalloc size %lu allocation failure: "
- "page order %u allocation failed",
+ "vmalloc error: size %lu, page order %u, failed to allocate pages",
area->nr_pages * PAGE_SIZE, page_order);
goto fail;
}
if (vmap_pages_range(addr, addr + size, prot, area->pages, page_shift) < 0) {
warn_alloc(gfp_mask, NULL,
- "vmalloc size %lu allocation failure: "
- "failed to map pages",
- area->nr_pages * PAGE_SIZE);
+ "vmalloc error: size %lu, failed to map pages",
+ area->nr_pages * PAGE_SIZE);
goto fail;
}
@@ -2886,8 +2883,8 @@ void *__vmalloc_node_range(unsigned long
if ((size >> PAGE_SHIFT) > totalram_pages()) {
warn_alloc(gfp_mask, NULL,
- "vmalloc size %lu allocation failure: "
- "exceeds total pages", real_size);
+ "vmalloc error: size %lu, exceeds total pages",
+ real_size);
return NULL;
}
@@ -2918,8 +2915,8 @@ again:
gfp_mask, caller);
if (!area) {
warn_alloc(gfp_mask, NULL,
- "vmalloc size %lu allocation failure: "
- "vm_struct allocation failed", real_size);
+ "vmalloc error: size %lu, vm_struct allocation failed",
+ real_size);
goto fail;
}
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 136/192] mm/vmalloc: fallback to a single page allocator
2021-06-29 2:32 incoming Andrew Morton
` (134 preceding siblings ...)
2021-06-29 2:40 ` [patch 135/192] mm/vmalloc: remove quoted strings split across lines Andrew Morton
@ 2021-06-29 2:40 ` Andrew Morton
2021-06-29 2:40 ` [patch 137/192] mm: vmalloc: add cond_resched() in __vunmap() Andrew Morton
` (55 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:40 UTC (permalink / raw)
To: akpm, hch, hdanton, linux-mm, mgorman, mhocko, mm-commits,
npiggin, oleksiy.avramchenko, rostedt, torvalds, urezki, willy
From: Uladzislau Rezki <urezki@gmail.com>
Subject: mm/vmalloc: fallback to a single page allocator
Currently for order-0 pages we use a bulk-page allocator to get set of
pages. From the other hand not allocating all pages is something that
might occur. In that case we should fallbak to the single-page allocator
trying to get missing pages, because it is more permissive(direct reclaim,
etc).
Introduce a vm_area_alloc_pages() function where the described logic is
implemented.
Link: https://lkml.kernel.org/r/20210521130718.GA17882@pc638.lan
Signed-off-by: Uladzislau Rezki (Sony) <urezki@gmail.com>
Reviewed-by: Matthew Wilcox (Oracle) <willy@infradead.org>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Cc: Mel Gorman <mgorman@suse.de>
Cc: Nicholas Piggin <npiggin@gmail.com>
Cc: Hillf Danton <hdanton@sina.com>
Cc: Michal Hocko <mhocko@suse.com>
Cc: Oleksiy Avramchenko <oleksiy.avramchenko@sonymobile.com>
Cc: Steven Rostedt <rostedt@goodmis.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/vmalloc.c | 81 +++++++++++++++++++++++++++++++------------------
1 file changed, 52 insertions(+), 29 deletions(-)
--- a/mm/vmalloc.c~mm-vmalloc-fallback-to-a-single-page-allocator
+++ a/mm/vmalloc.c
@@ -2758,6 +2758,54 @@ void *vmap_pfn(unsigned long *pfns, unsi
EXPORT_SYMBOL_GPL(vmap_pfn);
#endif /* CONFIG_VMAP_PFN */
+static inline unsigned int
+vm_area_alloc_pages(gfp_t gfp, int nid,
+ unsigned int order, unsigned long nr_pages, struct page **pages)
+{
+ unsigned int nr_allocated = 0;
+
+ /*
+ * For order-0 pages we make use of bulk allocator, if
+ * the page array is partly or not at all populated due
+ * to fails, fallback to a single page allocator that is
+ * more permissive.
+ */
+ if (!order)
+ nr_allocated = alloc_pages_bulk_array_node(
+ gfp, nid, nr_pages, pages);
+ else
+ /*
+ * Compound pages required for remap_vmalloc_page if
+ * high-order pages.
+ */
+ gfp |= __GFP_COMP;
+
+ /* High-order pages or fallback path if "bulk" fails. */
+ while (nr_allocated < nr_pages) {
+ struct page *page;
+ int i;
+
+ page = alloc_pages_node(nid, gfp, order);
+ if (unlikely(!page))
+ break;
+
+ /*
+ * Careful, we allocate and map page-order pages, but
+ * tracking is done per PAGE_SIZE page so as to keep the
+ * vm_struct APIs independent of the physical/mapped size.
+ */
+ for (i = 0; i < (1U << order); i++)
+ pages[nr_allocated + i] = page + i;
+
+ if (gfpflags_allow_blocking(gfp))
+ cond_resched();
+
+ nr_allocated += 1U << order;
+ }
+
+ return nr_allocated;
+}
+
static void *__vmalloc_area_node(struct vm_struct *area, gfp_t gfp_mask,
pgprot_t prot, unsigned int page_shift,
int node)
@@ -2790,37 +2838,11 @@ static void *__vmalloc_area_node(struct
return NULL;
}
- area->nr_pages = 0;
set_vm_area_page_order(area, page_shift - PAGE_SHIFT);
page_order = vm_area_page_order(area);
- if (!page_order) {
- area->nr_pages = alloc_pages_bulk_array_node(
- gfp_mask, node, nr_small_pages, area->pages);
- } else {
- /*
- * Careful, we allocate and map page_order pages, but tracking is done
- * per PAGE_SIZE page so as to keep the vm_struct APIs independent of
- * the physical/mapped size.
- */
- while (area->nr_pages < nr_small_pages) {
- struct page *page;
- int i;
-
- /* Compound pages required for remap_vmalloc_page */
- page = alloc_pages_node(node, gfp_mask | __GFP_COMP, page_order);
- if (unlikely(!page))
- break;
-
- for (i = 0; i < (1U << page_order); i++)
- area->pages[area->nr_pages + i] = page + i;
-
- if (gfpflags_allow_blocking(gfp_mask))
- cond_resched();
-
- area->nr_pages += 1U << page_order;
- }
- }
+ area->nr_pages = vm_area_alloc_pages(gfp_mask, node,
+ page_order, nr_small_pages, area->pages);
atomic_long_add(area->nr_pages, &nr_vmalloc_pages);
@@ -2835,7 +2857,8 @@ static void *__vmalloc_area_node(struct
goto fail;
}
- if (vmap_pages_range(addr, addr + size, prot, area->pages, page_shift) < 0) {
+ if (vmap_pages_range(addr, addr + size, prot, area->pages,
+ page_shift) < 0) {
warn_alloc(gfp_mask, NULL,
"vmalloc error: size %lu, failed to map pages",
area->nr_pages * PAGE_SIZE);
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 137/192] mm: vmalloc: add cond_resched() in __vunmap()
2021-06-29 2:32 incoming Andrew Morton
` (135 preceding siblings ...)
2021-06-29 2:40 ` [patch 136/192] mm/vmalloc: fallback to a single page allocator Andrew Morton
@ 2021-06-29 2:40 ` Andrew Morton
2021-06-29 2:40 ` [patch 138/192] printk: introduce dump_stack_lvl() Andrew Morton
` (54 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:40 UTC (permalink / raw)
To: akpm, aquini, atomlin, linux-mm, mhocko, mm-commits, npiggin,
torvalds, urezki
From: Rafael Aquini <aquini@redhat.com>
Subject: mm: vmalloc: add cond_resched() in __vunmap()
On non-preemptible kernel builds the watchdog can complain about soft
lockups when vfree() is called against large vmalloc areas:
[ 210.851798] kvmalloc-test: vmalloc(2199023255552) succeeded
[ 238.654842] watchdog: BUG: soft lockup - CPU#181 stuck for 26s! [rmmod:5203]
[ 238.662716] Modules linked in: kvmalloc_test(OE-) ...
[ 238.772671] CPU: 181 PID: 5203 Comm: rmmod Tainted: G S OE 5.13.0-rc7+ #1
[ 238.781413] Hardware name: Intel Corporation PURLEY/PURLEY, BIOS PLYXCRB1.86B.0553.D01.1809190614 09/19/2018
[ 238.792383] RIP: 0010:free_unref_page+0x52/0x60
[ 238.797447] Code: 48 c1 fd 06 48 89 ee e8 9c d0 ff ff 84 c0 74 19 9c 41 5c fa 48 89 ee 48 89 df e8 b9 ea ff ff 41 f7 c4 00 02 00 00 74 01 fb 5b <5d> 41 5c c3 66 2e 0f 1f 84 00 00 00 00 00 0f 1f 44 00 00 f0 29 77
[ 238.818406] RSP: 0018:ffffb4d87868fe98 EFLAGS: 00000206
[ 238.824236] RAX: 0000000000000000 RBX: 000000001da0c945 RCX: ffffb4d87868fe40
[ 238.832200] RDX: ffffd79d3beed108 RSI: ffffd7998501dc08 RDI: ffff9c6fbffd7010
[ 238.840166] RBP: 000000000d518cbd R08: ffffd7998501dc08 R09: 0000000000000001
[ 238.848131] R10: 0000000000000000 R11: ffffd79d3beee088 R12: 0000000000000202
[ 238.856095] R13: ffff9e5be3eceec0 R14: 0000000000000000 R15: 0000000000000000
[ 238.864059] FS: 00007fe082c2d740(0000) GS:ffff9f4c69b40000(0000) knlGS:0000000000000000
[ 238.873089] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[ 238.879503] CR2: 000055a000611128 CR3: 000000f6094f6006 CR4: 00000000007706e0
[ 238.887467] DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
[ 238.895433] DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7: 0000000000000400
[ 238.903397] PKRU: 55555554
[ 238.906417] Call Trace:
[ 238.909149] __vunmap+0x17c/0x220
[ 238.912851] __x64_sys_delete_module+0x13a/0x250
[ 238.918008] ? syscall_trace_enter.isra.20+0x13c/0x1b0
[ 238.923746] do_syscall_64+0x39/0x80
[ 238.927740] entry_SYSCALL_64_after_hwframe+0x44/0xae
Like in other range zapping routines that iterate over a large list, lets
just add cond_resched() within __vunmap()'s page-releasing loop in order
to avoid the watchdog splats.
Link: https://lkml.kernel.org/r/20210622225030.478384-1-aquini@redhat.com
Signed-off-by: Rafael Aquini <aquini@redhat.com>
Acked-by: Nicholas Piggin <npiggin@gmail.com>
Reviewed-by: Uladzislau Rezki (Sony) <urezki@gmail.com>
Reviewed-by: Aaron Tomlin <atomlin@redhat.com>
Acked-by: Michal Hocko <mhocko@suse.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/vmalloc.c | 1 +
1 file changed, 1 insertion(+)
--- a/mm/vmalloc.c~mm-vmalloc-add-cond_resched-in-__vunmap
+++ a/mm/vmalloc.c
@@ -2567,6 +2567,7 @@ static void __vunmap(const void *addr, i
BUG_ON(!page);
__free_pages(page, page_order);
+ cond_resched();
}
atomic_long_sub(area->nr_pages, &nr_vmalloc_pages);
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 138/192] printk: introduce dump_stack_lvl()
2021-06-29 2:32 incoming Andrew Morton
` (136 preceding siblings ...)
2021-06-29 2:40 ` [patch 137/192] mm: vmalloc: add cond_resched() in __vunmap() Andrew Morton
@ 2021-06-29 2:40 ` Andrew Morton
2021-06-29 2:40 ` [patch 139/192] kasan: use dump_stack_lvl(KERN_ERR) to print stacks Andrew Morton
` (53 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:40 UTC (permalink / raw)
To: akpm, bo.he, dvyukov, elver, glider, linux-mm, mingo, mm-commits,
pmladek, psodagud, rostedt, ryabinin.a.a, senozhatsky, torvalds,
yanmin_zhang
From: Alexander Potapenko <glider@google.com>
Subject: printk: introduce dump_stack_lvl()
dump_stack() is used for many different cases, which may require a log
level consistent with other kernel messages surrounding the dump_stack()
call. Without that, certain systems that are configured to ignore the
default level messages will miss stack traces in critical error reports.
This patch introduces dump_stack_lvl() that behaves similarly to
dump_stack(), but accepts a custom log level. The old dump_stack()
becomes equal to dump_stack_lvl(KERN_DEFAULT).
A somewhat similar patch has been proposed in 2012:
https://lore.kernel.org/lkml/1332493269.2359.9.camel@hebo/ , but wasn't
merged.
[elver@google.com: add missing dump_stack_lvl() stub if CONFIG_PRINTK=n]
Link: https://lkml.kernel.org/r/YJ0KAM0hQev1AmWe@elver.google.com
Link: https://lkml.kernel.org/r/20210506105405.3535023-1-glider@google.com
Signed-off-by: Alexander Potapenko <glider@google.com>
Reviewed-by: Marco Elver <elver@google.com>
Cc: Petr Mladek <pmladek@suse.com>
Cc: Ingo Molnar <mingo@kernel.org>
Cc: he, bo <bo.he@intel.com>
Cc: Yanmin Zhang <yanmin_zhang@linux.intel.com>
Cc: Prasad Sodagudi <psodagud@quicinc.com>
Cc: Dmitry Vyukov <dvyukov@google.com>
Cc: Sergey Senozhatsky <senozhatsky@chromium.org>
Cc: Steven Rostedt <rostedt@goodmis.org>
Cc: Andrey Ryabinin <ryabinin.a.a@gmail.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/linux/printk.h | 5 +++++
lib/dump_stack.c | 20 +++++++++++++-------
2 files changed, 18 insertions(+), 7 deletions(-)
--- a/include/linux/printk.h~printk-introduce-dump_stack_lvl
+++ a/include/linux/printk.h
@@ -206,6 +206,7 @@ void __init setup_log_buf(int early);
__printf(1, 2) void dump_stack_set_arch_desc(const char *fmt, ...);
void dump_stack_print_info(const char *log_lvl);
void show_regs_print_info(const char *log_lvl);
+extern asmlinkage void dump_stack_lvl(const char *log_lvl) __cold;
extern asmlinkage void dump_stack(void) __cold;
extern void printk_safe_flush(void);
extern void printk_safe_flush_on_panic(void);
@@ -269,6 +270,10 @@ static inline void show_regs_print_info(
{
}
+static inline void dump_stack_lvl(const char *log_lvl)
+{
+}
+
static inline void dump_stack(void)
{
}
--- a/lib/dump_stack.c~printk-introduce-dump_stack_lvl
+++ a/lib/dump_stack.c
@@ -73,10 +73,10 @@ void show_regs_print_info(const char *lo
dump_stack_print_info(log_lvl);
}
-static void __dump_stack(void)
+static void __dump_stack(const char *log_lvl)
{
- dump_stack_print_info(KERN_DEFAULT);
- show_stack(NULL, NULL, KERN_DEFAULT);
+ dump_stack_print_info(log_lvl);
+ show_stack(NULL, NULL, log_lvl);
}
/**
@@ -87,7 +87,7 @@ static void __dump_stack(void)
#ifdef CONFIG_SMP
static atomic_t dump_lock = ATOMIC_INIT(-1);
-asmlinkage __visible void dump_stack(void)
+asmlinkage __visible void dump_stack_lvl(const char *log_lvl)
{
unsigned long flags;
int was_locked;
@@ -117,7 +117,7 @@ retry:
goto retry;
}
- __dump_stack();
+ __dump_stack(log_lvl);
if (!was_locked)
atomic_set(&dump_lock, -1);
@@ -125,9 +125,15 @@ retry:
local_irq_restore(flags);
}
#else
-asmlinkage __visible void dump_stack(void)
+asmlinkage __visible void dump_stack_lvl(const char *log_lvl)
{
- __dump_stack();
+ __dump_stack(log_lvl);
}
#endif
+EXPORT_SYMBOL(dump_stack_lvl);
+
+asmlinkage __visible void dump_stack(void)
+{
+ dump_stack_lvl(KERN_DEFAULT);
+}
EXPORT_SYMBOL(dump_stack);
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 139/192] kasan: use dump_stack_lvl(KERN_ERR) to print stacks
2021-06-29 2:32 incoming Andrew Morton
` (137 preceding siblings ...)
2021-06-29 2:40 ` [patch 138/192] printk: introduce dump_stack_lvl() Andrew Morton
@ 2021-06-29 2:40 ` Andrew Morton
2021-06-29 2:40 ` [patch 140/192] kasan: test: improve failure message in KUNIT_EXPECT_KASAN_FAIL() Andrew Morton
` (52 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:40 UTC (permalink / raw)
To: akpm, bo.he, dvyukov, elver, glider, linux-mm, mingo, mm-commits,
pmladek, psodagud, rostedt, ryabinin.a.a, senozhatsky, torvalds,
yanmin_zhang
From: Alexander Potapenko <glider@google.com>
Subject: kasan: use dump_stack_lvl(KERN_ERR) to print stacks
Most of the contents of KASAN reports are printed with pr_err(), so use a
consistent logging level to print the memory access stacks.
Link: https://lkml.kernel.org/r/20210506105405.3535023-2-glider@google.com
Signed-off-by: Alexander Potapenko <glider@google.com>
Reviewed-by: Marco Elver <elver@google.com>
Cc: Andrey Ryabinin <ryabinin.a.a@gmail.com>
Cc: Prasad Sodagudi <psodagud@quicinc.com>
Cc: Dmitry Vyukov <dvyukov@google.com>
Cc: he, bo <bo.he@intel.com>
Cc: Ingo Molnar <mingo@kernel.org>
Cc: Petr Mladek <pmladek@suse.com>
Cc: Sergey Senozhatsky <senozhatsky@chromium.org>
Cc: Steven Rostedt <rostedt@goodmis.org>
Cc: Yanmin Zhang <yanmin_zhang@linux.intel.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/kasan/report.c | 6 +++---
1 file changed, 3 insertions(+), 3 deletions(-)
--- a/mm/kasan/report.c~kasan-use-dump_stack_lvlkern_err-to-print-stacks
+++ a/mm/kasan/report.c
@@ -230,7 +230,7 @@ static void print_address_description(vo
{
struct page *page = kasan_addr_to_page(addr);
- dump_stack();
+ dump_stack_lvl(KERN_ERR);
pr_err("\n");
if (page && PageSlab(page)) {
@@ -375,7 +375,7 @@ void kasan_report_async(void)
pr_err("BUG: KASAN: invalid-access\n");
pr_err("Asynchronous mode enabled: no access details available\n");
pr_err("\n");
- dump_stack();
+ dump_stack_lvl(KERN_ERR);
end_report(&flags, 0);
}
#endif /* CONFIG_KASAN_HW_TAGS */
@@ -420,7 +420,7 @@ static void __kasan_report(unsigned long
pr_err("\n");
print_memory_metadata(info.first_bad_addr);
} else {
- dump_stack();
+ dump_stack_lvl(KERN_ERR);
}
end_report(&flags, addr);
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 140/192] kasan: test: improve failure message in KUNIT_EXPECT_KASAN_FAIL()
2021-06-29 2:32 incoming Andrew Morton
` (138 preceding siblings ...)
2021-06-29 2:40 ` [patch 139/192] kasan: use dump_stack_lvl(KERN_ERR) to print stacks Andrew Morton
@ 2021-06-29 2:40 ` Andrew Morton
2021-06-29 2:40 ` [patch 141/192] kasan: allow an architecture to disable inline instrumentation Andrew Morton
` (51 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:40 UTC (permalink / raw)
To: akpm, andreyknvl, brendanhiggins, corbet, davidgow, dja, dvyukov,
elver, linux-mm, mm-commits, ryabinin.a.a, torvalds
From: David Gow <davidgow@google.com>
Subject: kasan: test: improve failure message in KUNIT_EXPECT_KASAN_FAIL()
The KUNIT_EXPECT_KASAN_FAIL() macro currently uses KUNIT_EXPECT_EQ() to
compare fail_data.report_expected and fail_data.report_found. This always
gave a somewhat useless error message on failure, but the addition of
extra compile-time checking with READ_ONCE() has caused it to get much
longer, and be truncated before anything useful is displayed.
Instead, just check fail_data.report_found by hand (we've just set
report_expected to 'true'), and print a better failure message with
KUNIT_FAIL(). Because of this, report_expected is no longer used
anywhere, and can be removed.
Beforehand, a failure in:
KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)area)[3100]);
would have looked like:
[22:00:34] [FAILED] vmalloc_oob
[22:00:34] # vmalloc_oob: EXPECTATION FAILED at lib/test_kasan.c:991
[22:00:34] Expected ({ do { extern void __compiletime_assert_705(void) __attribute__((__error__("Unsupported access size for {READ,WRITE}_ONCE()."))); if (!((sizeof(fail_data.report_expected) == sizeof(char) || sizeof(fail_data.repp
[22:00:34] not ok 45 - vmalloc_oob
With this change, it instead looks like:
[22:04:04] [FAILED] vmalloc_oob
[22:04:04] # vmalloc_oob: EXPECTATION FAILED at lib/test_kasan.c:993
[22:04:04] KASAN failure expected in "((volatile char *)area)[3100]", but none occurred
[22:04:04] not ok 45 - vmalloc_oob
Also update the example failure in the documentation to reflect this.
Link: https://lkml.kernel.org/r/20210606005531.165954-1-davidgow@google.com
Signed-off-by: David Gow <davidgow@google.com>
Reviewed-by: Andrey Konovalov <andreyknvl@gmail.com>
Reviewed-by: Marco Elver <elver@google.com>
Acked-by: Brendan Higgins <brendanhiggins@google.com>
Cc: Andrey Ryabinin <ryabinin.a.a@gmail.com>
Cc: Dmitry Vyukov <dvyukov@google.com>
Cc: Daniel Axtens <dja@axtens.net>
Cc: David Gow <davidgow@google.com>
Cc: Jonathan Corbet <corbet@lwn.net>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
Documentation/dev-tools/kasan.rst | 9 ++++-----
include/linux/kasan.h | 1 -
lib/test_kasan.c | 11 +++++------
3 files changed, 9 insertions(+), 12 deletions(-)
--- a/Documentation/dev-tools/kasan.rst~kasan-test-improve-failure-message-in-kunit_expect_kasan_fail
+++ a/Documentation/dev-tools/kasan.rst
@@ -447,11 +447,10 @@ When a test fails due to a failed ``kmal
When a test fails due to a missing KASAN report::
- # kmalloc_double_kzfree: EXPECTATION FAILED at lib/test_kasan.c:629
- Expected kasan_data->report_expected == kasan_data->report_found, but
- kasan_data->report_expected == 1
- kasan_data->report_found == 0
- not ok 28 - kmalloc_double_kzfree
+ # kmalloc_double_kzfree: EXPECTATION FAILED at lib/test_kasan.c:974
+ KASAN failure expected in "kfree_sensitive(ptr)", but none occurred
+ not ok 44 - kmalloc_double_kzfree
+
At the end the cumulative status of all KASAN tests is printed. On success::
--- a/include/linux/kasan.h~kasan-test-improve-failure-message-in-kunit_expect_kasan_fail
+++ a/include/linux/kasan.h
@@ -17,7 +17,6 @@ struct task_struct;
/* kasan_data struct is used in KUnit tests for KASAN expected failures */
struct kunit_kasan_expectation {
- bool report_expected;
bool report_found;
};
--- a/lib/test_kasan.c~kasan-test-improve-failure-message-in-kunit_expect_kasan_fail
+++ a/lib/test_kasan.c
@@ -55,7 +55,6 @@ static int kasan_test_init(struct kunit
multishot = kasan_save_enable_multi_shot();
kasan_set_tagging_report_once(false);
fail_data.report_found = false;
- fail_data.report_expected = false;
kunit_add_named_resource(test, NULL, NULL, &resource,
"kasan_data", &fail_data);
return 0;
@@ -94,20 +93,20 @@ static void kasan_test_exit(struct kunit
!kasan_async_mode_enabled()) \
migrate_disable(); \
KUNIT_EXPECT_FALSE(test, READ_ONCE(fail_data.report_found)); \
- WRITE_ONCE(fail_data.report_expected, true); \
barrier(); \
expression; \
barrier(); \
- KUNIT_EXPECT_EQ(test, \
- READ_ONCE(fail_data.report_expected), \
- READ_ONCE(fail_data.report_found)); \
+ if (!READ_ONCE(fail_data.report_found)) { \
+ KUNIT_FAIL(test, KUNIT_SUBTEST_INDENT "KASAN failure " \
+ "expected in \"" #expression \
+ "\", but none occurred"); \
+ } \
if (IS_ENABLED(CONFIG_KASAN_HW_TAGS)) { \
if (READ_ONCE(fail_data.report_found)) \
kasan_enable_tagging_sync(); \
migrate_enable(); \
} \
WRITE_ONCE(fail_data.report_found, false); \
- WRITE_ONCE(fail_data.report_expected, false); \
} while (0)
#define KASAN_TEST_NEEDS_CONFIG_ON(test, config) do { \
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 141/192] kasan: allow an architecture to disable inline instrumentation
2021-06-29 2:32 incoming Andrew Morton
` (139 preceding siblings ...)
2021-06-29 2:40 ` [patch 140/192] kasan: test: improve failure message in KUNIT_EXPECT_KASAN_FAIL() Andrew Morton
@ 2021-06-29 2:40 ` Andrew Morton
2021-06-29 2:40 ` [patch 142/192] kasan: allow architectures to provide an outline readiness check Andrew Morton
` (50 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:40 UTC (permalink / raw)
To: akpm, andreyknvl, aneesh.kumar, bsingharora, christophe.leroy,
dja, dvyukov, elver, glider, linux-mm, mm-commits, ryabinin.a.a,
torvalds
From: Daniel Axtens <dja@axtens.net>
Subject: kasan: allow an architecture to disable inline instrumentation
Patch series "KASAN core changes for ppc64 radix KASAN", v16.
Building on the work of Christophe, Aneesh and Balbir, I've ported KASAN
to 64-bit Book3S kernels running on the Radix MMU. I've been trying this
for a while, but we keep having collisions between the kasan code in the
mm tree and the code I want to put in to the ppc tree.
This series just contains the kasan core changes that we need. There
should be no noticeable changes to other platforms.
This patch (of 4):
For annoying architectural reasons, it's very difficult to support inline
instrumentation on powerpc64.*
Add a Kconfig flag to allow an arch to disable inline. (It's a bit
annoying to be 'backwards', but I'm not aware of any way to have an arch
force a symbol to be 'n', rather than 'y'.)
We also disable stack instrumentation in this case as it does things that
are functionally equivalent to inline instrumentation, namely adding code
that touches the shadow directly without going through a C helper.
* on ppc64 atm, the shadow lives in virtual memory and isn't accessible in
real mode. However, before we turn on virtual memory, we parse the device
tree to determine which platform and MMU we're running under. That calls
generic DT code, which is instrumented. Inline instrumentation in DT
would unconditionally attempt to touch the shadow region, which we won't
have set up yet, and would crash. We can make outline mode wait for the
arch to be ready, but we can't change what the compiler inserts for inline
mode.
Link: https://lkml.kernel.org/r/20210624034050.511391-1-dja@axtens.net
Link: https://lkml.kernel.org/r/20210624034050.511391-2-dja@axtens.net
Signed-off-by: Daniel Axtens <dja@axtens.net>
Reviewed-by: Marco Elver <elver@google.com>
Reviewed-by: Andrey Konovalov <andreyknvl@gmail.com>
Cc: Christophe Leroy <christophe.leroy@csgroup.eu>
Cc: Aneesh Kumar K.V <aneesh.kumar@linux.ibm.com>
Cc: Balbir Singh <bsingharora@gmail.com>
Cc: Alexander Potapenko <glider@google.com>
Cc: Andrey Ryabinin <ryabinin.a.a@gmail.com>
Cc: Dmitry Vyukov <dvyukov@google.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
lib/Kconfig.kasan | 12 ++++++++++++
1 file changed, 12 insertions(+)
--- a/lib/Kconfig.kasan~kasan-allow-an-architecture-to-disable-inline-instrumentation
+++ a/lib/Kconfig.kasan
@@ -12,6 +12,13 @@ config HAVE_ARCH_KASAN_HW_TAGS
config HAVE_ARCH_KASAN_VMALLOC
bool
+config ARCH_DISABLE_KASAN_INLINE
+ bool
+ help
+ An architecture might not support inline instrumentation.
+ When this option is selected, inline and stack instrumentation are
+ disabled.
+
config CC_HAS_KASAN_GENERIC
def_bool $(cc-option, -fsanitize=kernel-address)
@@ -130,6 +137,7 @@ config KASAN_OUTLINE
config KASAN_INLINE
bool "Inline instrumentation"
+ depends on !ARCH_DISABLE_KASAN_INLINE
help
Compiler directly inserts code checking shadow memory before
memory accesses. This is faster than outline (in some workloads
@@ -141,6 +149,7 @@ endchoice
config KASAN_STACK
bool "Enable stack instrumentation (unsafe)" if CC_IS_CLANG && !COMPILE_TEST
depends on KASAN_GENERIC || KASAN_SW_TAGS
+ depends on !ARCH_DISABLE_KASAN_INLINE
default y if CC_IS_GCC
help
The LLVM stack address sanitizer has a know problem that
@@ -154,6 +163,9 @@ config KASAN_STACK
but clang users can still enable it for builds without
CONFIG_COMPILE_TEST. On gcc it is assumed to always be safe
to use and enabled by default.
+ If the architecture disables inline instrumentation, stack
+ instrumentation is also disabled as it adds inline-style
+ instrumentation that is run unconditionally.
config KASAN_SW_TAGS_IDENTIFY
bool "Enable memory corruption identification"
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 142/192] kasan: allow architectures to provide an outline readiness check
2021-06-29 2:32 incoming Andrew Morton
` (140 preceding siblings ...)
2021-06-29 2:40 ` [patch 141/192] kasan: allow an architecture to disable inline instrumentation Andrew Morton
@ 2021-06-29 2:40 ` Andrew Morton
2021-06-29 2:40 ` [patch 143/192] mm: define default MAX_PTRS_PER_* in include/pgtable.h Andrew Morton
` (49 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:40 UTC (permalink / raw)
To: akpm, andreyknvl, aneesh.kumar, bsingharora, christophe.leroy,
dja, dvyukov, elver, glider, linux-mm, mm-commits, ryabinin.a.a,
torvalds
From: Daniel Axtens <dja@axtens.net>
Subject: kasan: allow architectures to provide an outline readiness check
Allow architectures to define a kasan_arch_is_ready() hook that bails out
of any function that's about to touch the shadow unless the arch says that
it is ready for the memory to be accessed. This is fairly uninvasive and
should have a negligible performance penalty.
This will only work in outline mode, so an arch must specify
ARCH_DISABLE_KASAN_INLINE if it requires this.
Link: https://lkml.kernel.org/r/20210624034050.511391-3-dja@axtens.net
Signed-off-by: Daniel Axtens <dja@axtens.net>
Reviewed-by: Marco Elver <elver@google.com>
Suggested-by: Christophe Leroy <christophe.leroy@csgroup.eu>
Reviewed-by: Andrey Konovalov <andreyknvl@gmail.com>
Cc: Balbir Singh <bsingharora@gmail.com>
Cc: Aneesh Kumar K.V <aneesh.kumar@linux.ibm.com>
Cc: Alexander Potapenko <glider@google.com>
Cc: Andrey Ryabinin <ryabinin.a.a@gmail.com>
Cc: Dmitry Vyukov <dvyukov@google.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/kasan/common.c | 3 +++
mm/kasan/generic.c | 3 +++
mm/kasan/kasan.h | 6 ++++++
mm/kasan/shadow.c | 6 ++++++
4 files changed, 18 insertions(+)
--- a/mm/kasan/common.c~kasan-allow-architectures-to-provide-an-outline-readiness-check
+++ a/mm/kasan/common.c
@@ -331,6 +331,9 @@ static inline bool ____kasan_slab_free(s
u8 tag;
void *tagged_object;
+ if (!kasan_arch_is_ready())
+ return false;
+
tag = get_tag(object);
tagged_object = object;
object = kasan_reset_tag(object);
--- a/mm/kasan/generic.c~kasan-allow-architectures-to-provide-an-outline-readiness-check
+++ a/mm/kasan/generic.c
@@ -163,6 +163,9 @@ static __always_inline bool check_region
size_t size, bool write,
unsigned long ret_ip)
{
+ if (!kasan_arch_is_ready())
+ return true;
+
if (unlikely(size == 0))
return true;
--- a/mm/kasan/kasan.h~kasan-allow-architectures-to-provide-an-outline-readiness-check
+++ a/mm/kasan/kasan.h
@@ -449,6 +449,12 @@ static inline void kasan_poison_last_gra
#endif /* CONFIG_KASAN_GENERIC */
+#ifndef kasan_arch_is_ready
+static inline bool kasan_arch_is_ready(void) { return true; }
+#elif !defined(CONFIG_KASAN_GENERIC) || !defined(CONFIG_KASAN_OUTLINE)
+#error kasan_arch_is_ready only works in KASAN generic outline mode!
+#endif
+
/*
* Exported functions for interfaces called from assembly or from generated
* code. Declarations here to avoid warning about missing declarations.
--- a/mm/kasan/shadow.c~kasan-allow-architectures-to-provide-an-outline-readiness-check
+++ a/mm/kasan/shadow.c
@@ -73,6 +73,9 @@ void kasan_poison(const void *addr, size
{
void *shadow_start, *shadow_end;
+ if (!kasan_arch_is_ready())
+ return;
+
/*
* Perform shadow offset calculation based on untagged address, as
* some of the callers (e.g. kasan_poison_object_data) pass tagged
@@ -99,6 +102,9 @@ EXPORT_SYMBOL(kasan_poison);
#ifdef CONFIG_KASAN_GENERIC
void kasan_poison_last_granule(const void *addr, size_t size)
{
+ if (!kasan_arch_is_ready())
+ return;
+
if (size & KASAN_GRANULE_MASK) {
u8 *shadow = (u8 *)kasan_mem_to_shadow(addr + size);
*shadow = size & KASAN_GRANULE_MASK;
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 143/192] mm: define default MAX_PTRS_PER_* in include/pgtable.h
2021-06-29 2:32 incoming Andrew Morton
` (141 preceding siblings ...)
2021-06-29 2:40 ` [patch 142/192] kasan: allow architectures to provide an outline readiness check Andrew Morton
@ 2021-06-29 2:40 ` Andrew Morton
2021-06-29 2:40 ` [patch 144/192] kasan: use MAX_PTRS_PER_* for early shadow tables Andrew Morton
` (48 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:40 UTC (permalink / raw)
To: akpm, andreyknvl, aneesh.kumar, bsingharora, christophe.leroy,
dja, dvyukov, elver, glider, linux-mm, mm-commits, ryabinin.a.a,
torvalds
From: Daniel Axtens <dja@axtens.net>
Subject: mm: define default MAX_PTRS_PER_* in include/pgtable.h
Commit c65e774fb3f6 ("x86/mm: Make PGDIR_SHIFT and PTRS_PER_P4D variable")
made PTRS_PER_P4D variable on x86 and introduced MAX_PTRS_PER_P4D as a
constant for cases which need a compile-time constant (e.g. fixed-size
arrays).
powerpc likewise has boot-time selectable MMU features which can cause
other mm "constants" to vary. For KASAN, we have some static
PTE/PMD/PUD/P4D arrays so we need compile-time maximums for all these
constants. Extend the MAX_PTRS_PER_ idiom, and place default definitions
in include/pgtable.h. These define MAX_PTRS_PER_x to be PTRS_PER_x unless
an architecture has defined MAX_PTRS_PER_x in its arch headers.
Clean up pgtable-nop4d.h and s390's MAX_PTRS_PER_P4D definitions while
we're at it: both can just pick up the default now.
Link: https://lkml.kernel.org/r/20210624034050.511391-4-dja@axtens.net
Signed-off-by: Daniel Axtens <dja@axtens.net>
Acked-by: Andrey Konovalov <andreyknvl@gmail.com>
Reviewed-by: Christophe Leroy <christophe.leroy@csgroup.eu>
Reviewed-by: Marco Elver <elver@google.com>
Cc: Aneesh Kumar K.V <aneesh.kumar@linux.ibm.com>
Cc: Balbir Singh <bsingharora@gmail.com>
Cc: Alexander Potapenko <glider@google.com>
Cc: Andrey Ryabinin <ryabinin.a.a@gmail.com>
Cc: Dmitry Vyukov <dvyukov@google.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
arch/s390/include/asm/pgtable.h | 2 --
include/asm-generic/pgtable-nop4d.h | 1 -
include/linux/pgtable.h | 22 ++++++++++++++++++++++
3 files changed, 22 insertions(+), 3 deletions(-)
--- a/arch/s390/include/asm/pgtable.h~mm-define-default-max_ptrs_per_-in-include-pgtableh
+++ a/arch/s390/include/asm/pgtable.h
@@ -344,8 +344,6 @@ static inline int is_module_addr(void *a
#define PTRS_PER_P4D _CRST_ENTRIES
#define PTRS_PER_PGD _CRST_ENTRIES
-#define MAX_PTRS_PER_P4D PTRS_PER_P4D
-
/*
* Segment table and region3 table entry encoding
* (R = read-only, I = invalid, y = young bit):
--- a/include/asm-generic/pgtable-nop4d.h~mm-define-default-max_ptrs_per_-in-include-pgtableh
+++ a/include/asm-generic/pgtable-nop4d.h
@@ -9,7 +9,6 @@
typedef struct { pgd_t pgd; } p4d_t;
#define P4D_SHIFT PGDIR_SHIFT
-#define MAX_PTRS_PER_P4D 1
#define PTRS_PER_P4D 1
#define P4D_SIZE (1UL << P4D_SHIFT)
#define P4D_MASK (~(P4D_SIZE-1))
--- a/include/linux/pgtable.h~mm-define-default-max_ptrs_per_-in-include-pgtableh
+++ a/include/linux/pgtable.h
@@ -1592,4 +1592,26 @@ typedef unsigned int pgtbl_mod_mask;
#define pte_leaf_size(x) PAGE_SIZE
#endif
+/*
+ * Some architectures have MMUs that are configurable or selectable at boot
+ * time. These lead to variable PTRS_PER_x. For statically allocated arrays it
+ * helps to have a static maximum value.
+ */
+
+#ifndef MAX_PTRS_PER_PTE
+#define MAX_PTRS_PER_PTE PTRS_PER_PTE
+#endif
+
+#ifndef MAX_PTRS_PER_PMD
+#define MAX_PTRS_PER_PMD PTRS_PER_PMD
+#endif
+
+#ifndef MAX_PTRS_PER_PUD
+#define MAX_PTRS_PER_PUD PTRS_PER_PUD
+#endif
+
+#ifndef MAX_PTRS_PER_P4D
+#define MAX_PTRS_PER_P4D PTRS_PER_P4D
+#endif
+
#endif /* _LINUX_PGTABLE_H */
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 144/192] kasan: use MAX_PTRS_PER_* for early shadow tables
2021-06-29 2:32 incoming Andrew Morton
` (142 preceding siblings ...)
2021-06-29 2:40 ` [patch 143/192] mm: define default MAX_PTRS_PER_* in include/pgtable.h Andrew Morton
@ 2021-06-29 2:40 ` Andrew Morton
2021-06-29 2:40 ` [patch 145/192] kasan: rename CONFIG_KASAN_SW_TAGS_IDENTIFY to CONFIG_KASAN_TAGS_IDENTIFY Andrew Morton
` (47 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:40 UTC (permalink / raw)
To: akpm, andreyknvl, aneesh.kumar, bsingharora, christophe.leroy,
dja, dvyukov, elver, glider, linux-mm, mm-commits, ryabinin.a.a,
torvalds
From: Daniel Axtens <dja@axtens.net>
Subject: kasan: use MAX_PTRS_PER_* for early shadow tables
powerpc has a variable number of PTRS_PER_*, set at runtime based on the
MMU that the kernel is booted under.
This means the PTRS_PER_* are no longer constants, and therefore breaks
the build. Switch to using MAX_PTRS_PER_*, which are constant.
Link: https://lkml.kernel.org/r/20210624034050.511391-5-dja@axtens.net
Signed-off-by: Daniel Axtens <dja@axtens.net>
Suggested-by: Christophe Leroy <christophe.leroy@csgroup.eu>
Suggested-by: Balbir Singh <bsingharora@gmail.com>
Reviewed-by: Christophe Leroy <christophe.leroy@csgroup.eu>
Reviewed-by: Balbir Singh <bsingharora@gmail.com>
Reviewed-by: Marco Elver <elver@google.com>
Reviewed-by: Andrey Konovalov <andreyknvl@gmail.com>
Cc: Aneesh Kumar K.V <aneesh.kumar@linux.ibm.com>
Cc: Andrey Ryabinin <ryabinin.a.a@gmail.com>
Cc: Alexander Potapenko <glider@google.com>
Cc: Dmitry Vyukov <dvyukov@google.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/linux/kasan.h | 6 +++---
mm/kasan/init.c | 6 +++---
2 files changed, 6 insertions(+), 6 deletions(-)
--- a/include/linux/kasan.h~kasan-use-max_ptrs_per_-for-early-shadow-tables
+++ a/include/linux/kasan.h
@@ -40,9 +40,9 @@ struct kunit_kasan_expectation {
#endif
extern unsigned char kasan_early_shadow_page[PAGE_SIZE];
-extern pte_t kasan_early_shadow_pte[PTRS_PER_PTE + PTE_HWTABLE_PTRS];
-extern pmd_t kasan_early_shadow_pmd[PTRS_PER_PMD];
-extern pud_t kasan_early_shadow_pud[PTRS_PER_PUD];
+extern pte_t kasan_early_shadow_pte[MAX_PTRS_PER_PTE + PTE_HWTABLE_PTRS];
+extern pmd_t kasan_early_shadow_pmd[MAX_PTRS_PER_PMD];
+extern pud_t kasan_early_shadow_pud[MAX_PTRS_PER_PUD];
extern p4d_t kasan_early_shadow_p4d[MAX_PTRS_PER_P4D];
int kasan_populate_early_shadow(const void *shadow_start,
--- a/mm/kasan/init.c~kasan-use-max_ptrs_per_-for-early-shadow-tables
+++ a/mm/kasan/init.c
@@ -41,7 +41,7 @@ static inline bool kasan_p4d_table(pgd_t
}
#endif
#if CONFIG_PGTABLE_LEVELS > 3
-pud_t kasan_early_shadow_pud[PTRS_PER_PUD] __page_aligned_bss;
+pud_t kasan_early_shadow_pud[MAX_PTRS_PER_PUD] __page_aligned_bss;
static inline bool kasan_pud_table(p4d_t p4d)
{
return p4d_page(p4d) == virt_to_page(lm_alias(kasan_early_shadow_pud));
@@ -53,7 +53,7 @@ static inline bool kasan_pud_table(p4d_t
}
#endif
#if CONFIG_PGTABLE_LEVELS > 2
-pmd_t kasan_early_shadow_pmd[PTRS_PER_PMD] __page_aligned_bss;
+pmd_t kasan_early_shadow_pmd[MAX_PTRS_PER_PMD] __page_aligned_bss;
static inline bool kasan_pmd_table(pud_t pud)
{
return pud_page(pud) == virt_to_page(lm_alias(kasan_early_shadow_pmd));
@@ -64,7 +64,7 @@ static inline bool kasan_pmd_table(pud_t
return false;
}
#endif
-pte_t kasan_early_shadow_pte[PTRS_PER_PTE + PTE_HWTABLE_PTRS]
+pte_t kasan_early_shadow_pte[MAX_PTRS_PER_PTE + PTE_HWTABLE_PTRS]
__page_aligned_bss;
static inline bool kasan_pte_table(pmd_t pmd)
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 145/192] kasan: rename CONFIG_KASAN_SW_TAGS_IDENTIFY to CONFIG_KASAN_TAGS_IDENTIFY
2021-06-29 2:32 incoming Andrew Morton
` (143 preceding siblings ...)
2021-06-29 2:40 ` [patch 144/192] kasan: use MAX_PTRS_PER_* for early shadow tables Andrew Morton
@ 2021-06-29 2:40 ` Andrew Morton
2021-06-29 2:40 ` [patch 146/192] kasan: integrate the common part of two KASAN tag-based modes Andrew Morton
` (46 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:40 UTC (permalink / raw)
To: akpm, andreyknvl, chinwen.chang, dvyukov, elver, glider, gregkh,
Kuan-Ying.Lee, linux-mm, matthias.bgg, mm-commits, nicholas.tang,
ryabinin.a.a, torvalds
From: Kuan-Ying Lee <Kuan-Ying.Lee@mediatek.com>
Subject: kasan: rename CONFIG_KASAN_SW_TAGS_IDENTIFY to CONFIG_KASAN_TAGS_IDENTIFY
Patch series "kasan: add memory corruption identification support for hw tag-based kasan", v4.
Add memory corruption identification for hardware tag-based KASAN mode.
This patch (of 3):
Rename CONFIG_KASAN_SW_TAGS_IDENTIFY to CONFIG_KASAN_TAGS_IDENTIFY in
order to be compatible with hardware tag-based mode.
Link: https://lkml.kernel.org/r/20210626100931.22794-1-Kuan-Ying.Lee@mediatek.com
Link: https://lkml.kernel.org/r/20210626100931.22794-2-Kuan-Ying.Lee@mediatek.com
Signed-off-by: Kuan-Ying Lee <Kuan-Ying.Lee@mediatek.com>
Suggested-by: Marco Elver <elver@google.com>
Reviewed-by: Alexander Potapenko <glider@google.com>
Reviewed-by: Andrey Konovalov <andreyknvl@gmail.com>
Reviewed-by: Marco Elver <elver@google.com>
Cc: Andrey Ryabinin <ryabinin.a.a@gmail.com>
Cc: Dmitry Vyukov <dvyukov@google.com>
Cc: Matthias Brugger <matthias.bgg@gmail.com>
Cc: Chinwen Chang <chinwen.chang@mediatek.com>
Cc: Nicholas Tang <nicholas.tang@mediatek.com>
Cc: Kuan-Ying Lee <Kuan-Ying.Lee@mediatek.com>
Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
lib/Kconfig.kasan | 2 +-
mm/kasan/kasan.h | 4 ++--
mm/kasan/report_sw_tags.c | 2 +-
mm/kasan/sw_tags.c | 4 ++--
4 files changed, 6 insertions(+), 6 deletions(-)
--- a/lib/Kconfig.kasan~kasan-rename-config_kasan_sw_tags_identify-to-config_kasan_tags_identify
+++ a/lib/Kconfig.kasan
@@ -167,7 +167,7 @@ config KASAN_STACK
instrumentation is also disabled as it adds inline-style
instrumentation that is run unconditionally.
-config KASAN_SW_TAGS_IDENTIFY
+config KASAN_TAGS_IDENTIFY
bool "Enable memory corruption identification"
depends on KASAN_SW_TAGS
help
--- a/mm/kasan/kasan.h~kasan-rename-config_kasan_sw_tags_identify-to-config_kasan_tags_identify
+++ a/mm/kasan/kasan.h
@@ -153,7 +153,7 @@ struct kasan_track {
depot_stack_handle_t stack;
};
-#ifdef CONFIG_KASAN_SW_TAGS_IDENTIFY
+#ifdef CONFIG_KASAN_TAGS_IDENTIFY
#define KASAN_NR_FREE_STACKS 5
#else
#define KASAN_NR_FREE_STACKS 1
@@ -170,7 +170,7 @@ struct kasan_alloc_meta {
#else
struct kasan_track free_track[KASAN_NR_FREE_STACKS];
#endif
-#ifdef CONFIG_KASAN_SW_TAGS_IDENTIFY
+#ifdef CONFIG_KASAN_TAGS_IDENTIFY
u8 free_pointer_tag[KASAN_NR_FREE_STACKS];
u8 free_track_idx;
#endif
--- a/mm/kasan/report_sw_tags.c~kasan-rename-config_kasan_sw_tags_identify-to-config_kasan_tags_identify
+++ a/mm/kasan/report_sw_tags.c
@@ -31,7 +31,7 @@
const char *kasan_get_bug_type(struct kasan_access_info *info)
{
-#ifdef CONFIG_KASAN_SW_TAGS_IDENTIFY
+#ifdef CONFIG_KASAN_TAGS_IDENTIFY
struct kasan_alloc_meta *alloc_meta;
struct kmem_cache *cache;
struct page *page;
--- a/mm/kasan/sw_tags.c~kasan-rename-config_kasan_sw_tags_identify-to-config_kasan_tags_identify
+++ a/mm/kasan/sw_tags.c
@@ -177,7 +177,7 @@ void kasan_set_free_info(struct kmem_cac
if (!alloc_meta)
return;
-#ifdef CONFIG_KASAN_SW_TAGS_IDENTIFY
+#ifdef CONFIG_KASAN_TAGS_IDENTIFY
idx = alloc_meta->free_track_idx;
alloc_meta->free_pointer_tag[idx] = tag;
alloc_meta->free_track_idx = (idx + 1) % KASAN_NR_FREE_STACKS;
@@ -196,7 +196,7 @@ struct kasan_track *kasan_get_free_track
if (!alloc_meta)
return NULL;
-#ifdef CONFIG_KASAN_SW_TAGS_IDENTIFY
+#ifdef CONFIG_KASAN_TAGS_IDENTIFY
for (i = 0; i < KASAN_NR_FREE_STACKS; i++) {
if (alloc_meta->free_pointer_tag[i] == tag)
break;
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 146/192] kasan: integrate the common part of two KASAN tag-based modes
2021-06-29 2:32 incoming Andrew Morton
` (144 preceding siblings ...)
2021-06-29 2:40 ` [patch 145/192] kasan: rename CONFIG_KASAN_SW_TAGS_IDENTIFY to CONFIG_KASAN_TAGS_IDENTIFY Andrew Morton
@ 2021-06-29 2:40 ` Andrew Morton
2021-06-29 2:40 ` [patch 147/192] kasan: add memory corruption identification support for hardware tag-based mode Andrew Morton
` (45 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:40 UTC (permalink / raw)
To: akpm, andreyknvl, chinwen.chang, dvyukov, elver, glider, gregkh,
Kuan-Ying.Lee, linux-mm, matthias.bgg, mm-commits, nicholas.tang,
ryabinin.a.a, torvalds
From: Kuan-Ying Lee <Kuan-Ying.Lee@mediatek.com>
Subject: kasan: integrate the common part of two KASAN tag-based modes
1. Move kasan_get_free_track() and kasan_set_free_info() into tags.c
and combine these two functions for SW_TAGS and HW_TAGS kasan mode.
2. Move kasan_get_bug_type() to report_tags.c and make this function
compatible for SW_TAGS and HW_TAGS kasan mode.
Link: https://lkml.kernel.org/r/20210626100931.22794-3-Kuan-Ying.Lee@mediatek.com
Signed-off-by: Kuan-Ying Lee <Kuan-Ying.Lee@mediatek.com>
Suggested-by: Marco Elver <elver@google.com>
Suggested-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Reviewed-by: Andrey Konovalov <andreyknvl@gmail.com>
Reviewed-by: Marco Elver <elver@google.com>
Cc: Andrey Ryabinin <ryabinin.a.a@gmail.com>
Cc: Alexander Potapenko <glider@google.com>
Cc: Dmitry Vyukov <dvyukov@google.com>
Cc: Chinwen Chang <chinwen.chang@mediatek.com>
Cc: Matthias Brugger <matthias.bgg@gmail.com>
Cc: Nicholas Tang <nicholas.tang@mediatek.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/kasan/Makefile | 4 +-
mm/kasan/hw_tags.c | 22 -------------
mm/kasan/report_hw_tags.c | 5 ---
mm/kasan/report_sw_tags.c | 43 --------------------------
mm/kasan/report_tags.c | 51 +++++++++++++++++++++++++++++++
mm/kasan/sw_tags.c | 41 -------------------------
mm/kasan/tags.c | 59 ++++++++++++++++++++++++++++++++++++
7 files changed, 112 insertions(+), 113 deletions(-)
--- a/mm/kasan/hw_tags.c~kasan-integrate-the-common-part-of-two-kasan-tag-based-modes
+++ a/mm/kasan/hw_tags.c
@@ -216,28 +216,6 @@ void __init kasan_init_hw_tags(void)
pr_info("KernelAddressSanitizer initialized\n");
}
-void kasan_set_free_info(struct kmem_cache *cache,
- void *object, u8 tag)
-{
- struct kasan_alloc_meta *alloc_meta;
-
- alloc_meta = kasan_get_alloc_meta(cache, object);
- if (alloc_meta)
- kasan_set_track(&alloc_meta->free_track[0], GFP_NOWAIT);
-}
-
-struct kasan_track *kasan_get_free_track(struct kmem_cache *cache,
- void *object, u8 tag)
-{
- struct kasan_alloc_meta *alloc_meta;
-
- alloc_meta = kasan_get_alloc_meta(cache, object);
- if (!alloc_meta)
- return NULL;
-
- return &alloc_meta->free_track[0];
-}
-
#if IS_ENABLED(CONFIG_KASAN_KUNIT_TEST)
void kasan_set_tagging_report_once(bool state)
--- a/mm/kasan/Makefile~kasan-integrate-the-common-part-of-two-kasan-tag-based-modes
+++ a/mm/kasan/Makefile
@@ -37,5 +37,5 @@ CFLAGS_sw_tags.o := $(CC_FLAGS_KASAN_RUN
obj-$(CONFIG_KASAN) := common.o report.o
obj-$(CONFIG_KASAN_GENERIC) += init.o generic.o report_generic.o shadow.o quarantine.o
-obj-$(CONFIG_KASAN_HW_TAGS) += hw_tags.o report_hw_tags.o
-obj-$(CONFIG_KASAN_SW_TAGS) += init.o report_sw_tags.o shadow.o sw_tags.o
+obj-$(CONFIG_KASAN_HW_TAGS) += hw_tags.o report_hw_tags.o tags.o report_tags.o
+obj-$(CONFIG_KASAN_SW_TAGS) += init.o report_sw_tags.o shadow.o sw_tags.o tags.o report_tags.o
--- a/mm/kasan/report_hw_tags.c~kasan-integrate-the-common-part-of-two-kasan-tag-based-modes
+++ a/mm/kasan/report_hw_tags.c
@@ -15,11 +15,6 @@
#include "kasan.h"
-const char *kasan_get_bug_type(struct kasan_access_info *info)
-{
- return "invalid-access";
-}
-
void *kasan_find_first_bad_addr(void *addr, size_t size)
{
return kasan_reset_tag(addr);
--- a/mm/kasan/report_sw_tags.c~kasan-integrate-the-common-part-of-two-kasan-tag-based-modes
+++ a/mm/kasan/report_sw_tags.c
@@ -29,49 +29,6 @@
#include "kasan.h"
#include "../slab.h"
-const char *kasan_get_bug_type(struct kasan_access_info *info)
-{
-#ifdef CONFIG_KASAN_TAGS_IDENTIFY
- struct kasan_alloc_meta *alloc_meta;
- struct kmem_cache *cache;
- struct page *page;
- const void *addr;
- void *object;
- u8 tag;
- int i;
-
- tag = get_tag(info->access_addr);
- addr = kasan_reset_tag(info->access_addr);
- page = kasan_addr_to_page(addr);
- if (page && PageSlab(page)) {
- cache = page->slab_cache;
- object = nearest_obj(cache, page, (void *)addr);
- alloc_meta = kasan_get_alloc_meta(cache, object);
-
- if (alloc_meta) {
- for (i = 0; i < KASAN_NR_FREE_STACKS; i++) {
- if (alloc_meta->free_pointer_tag[i] == tag)
- return "use-after-free";
- }
- }
- return "out-of-bounds";
- }
-
-#endif
- /*
- * If access_size is a negative number, then it has reason to be
- * defined as out-of-bounds bug type.
- *
- * Casting negative numbers to size_t would indeed turn up as
- * a large size_t and its value will be larger than ULONG_MAX/2,
- * so that this can qualify as out-of-bounds.
- */
- if (info->access_addr + info->access_size < info->access_addr)
- return "out-of-bounds";
-
- return "invalid-access";
-}
-
void *kasan_find_first_bad_addr(void *addr, size_t size)
{
u8 tag = get_tag(addr);
--- /dev/null
+++ a/mm/kasan/report_tags.c
@@ -0,0 +1,51 @@
+// SPDX-License-Identifier: GPL-2.0
+/*
+ * Copyright (c) 2014 Samsung Electronics Co., Ltd.
+ * Copyright (c) 2020 Google, Inc.
+ */
+
+#include "kasan.h"
+#include "../slab.h"
+
+const char *kasan_get_bug_type(struct kasan_access_info *info)
+{
+#ifdef CONFIG_KASAN_TAGS_IDENTIFY
+ struct kasan_alloc_meta *alloc_meta;
+ struct kmem_cache *cache;
+ struct page *page;
+ const void *addr;
+ void *object;
+ u8 tag;
+ int i;
+
+ tag = get_tag(info->access_addr);
+ addr = kasan_reset_tag(info->access_addr);
+ page = kasan_addr_to_page(addr);
+ if (page && PageSlab(page)) {
+ cache = page->slab_cache;
+ object = nearest_obj(cache, page, (void *)addr);
+ alloc_meta = kasan_get_alloc_meta(cache, object);
+
+ if (alloc_meta) {
+ for (i = 0; i < KASAN_NR_FREE_STACKS; i++) {
+ if (alloc_meta->free_pointer_tag[i] == tag)
+ return "use-after-free";
+ }
+ }
+ return "out-of-bounds";
+ }
+#endif
+
+ /*
+ * If access_size is a negative number, then it has reason to be
+ * defined as out-of-bounds bug type.
+ *
+ * Casting negative numbers to size_t would indeed turn up as
+ * a large size_t and its value will be larger than ULONG_MAX/2,
+ * so that this can qualify as out-of-bounds.
+ */
+ if (info->access_addr + info->access_size < info->access_addr)
+ return "out-of-bounds";
+
+ return "invalid-access";
+}
--- a/mm/kasan/sw_tags.c~kasan-integrate-the-common-part-of-two-kasan-tag-based-modes
+++ a/mm/kasan/sw_tags.c
@@ -166,44 +166,3 @@ void __hwasan_tag_memory(unsigned long a
kasan_poison((void *)addr, size, tag, false);
}
EXPORT_SYMBOL(__hwasan_tag_memory);
-
-void kasan_set_free_info(struct kmem_cache *cache,
- void *object, u8 tag)
-{
- struct kasan_alloc_meta *alloc_meta;
- u8 idx = 0;
-
- alloc_meta = kasan_get_alloc_meta(cache, object);
- if (!alloc_meta)
- return;
-
-#ifdef CONFIG_KASAN_TAGS_IDENTIFY
- idx = alloc_meta->free_track_idx;
- alloc_meta->free_pointer_tag[idx] = tag;
- alloc_meta->free_track_idx = (idx + 1) % KASAN_NR_FREE_STACKS;
-#endif
-
- kasan_set_track(&alloc_meta->free_track[idx], GFP_NOWAIT);
-}
-
-struct kasan_track *kasan_get_free_track(struct kmem_cache *cache,
- void *object, u8 tag)
-{
- struct kasan_alloc_meta *alloc_meta;
- int i = 0;
-
- alloc_meta = kasan_get_alloc_meta(cache, object);
- if (!alloc_meta)
- return NULL;
-
-#ifdef CONFIG_KASAN_TAGS_IDENTIFY
- for (i = 0; i < KASAN_NR_FREE_STACKS; i++) {
- if (alloc_meta->free_pointer_tag[i] == tag)
- break;
- }
- if (i == KASAN_NR_FREE_STACKS)
- i = alloc_meta->free_track_idx;
-#endif
-
- return &alloc_meta->free_track[i];
-}
--- /dev/null
+++ a/mm/kasan/tags.c
@@ -0,0 +1,59 @@
+// SPDX-License-Identifier: GPL-2.0
+/*
+ * This file contains common tag-based KASAN code.
+ *
+ * Copyright (c) 2018 Google, Inc.
+ * Copyright (c) 2020 Google, Inc.
+ */
+
+#include <linux/init.h>
+#include <linux/kasan.h>
+#include <linux/kernel.h>
+#include <linux/memory.h>
+#include <linux/mm.h>
+#include <linux/static_key.h>
+#include <linux/string.h>
+#include <linux/types.h>
+
+#include "kasan.h"
+
+void kasan_set_free_info(struct kmem_cache *cache,
+ void *object, u8 tag)
+{
+ struct kasan_alloc_meta *alloc_meta;
+ u8 idx = 0;
+
+ alloc_meta = kasan_get_alloc_meta(cache, object);
+ if (!alloc_meta)
+ return;
+
+#ifdef CONFIG_KASAN_TAGS_IDENTIFY
+ idx = alloc_meta->free_track_idx;
+ alloc_meta->free_pointer_tag[idx] = tag;
+ alloc_meta->free_track_idx = (idx + 1) % KASAN_NR_FREE_STACKS;
+#endif
+
+ kasan_set_track(&alloc_meta->free_track[idx], GFP_NOWAIT);
+}
+
+struct kasan_track *kasan_get_free_track(struct kmem_cache *cache,
+ void *object, u8 tag)
+{
+ struct kasan_alloc_meta *alloc_meta;
+ int i = 0;
+
+ alloc_meta = kasan_get_alloc_meta(cache, object);
+ if (!alloc_meta)
+ return NULL;
+
+#ifdef CONFIG_KASAN_TAGS_IDENTIFY
+ for (i = 0; i < KASAN_NR_FREE_STACKS; i++) {
+ if (alloc_meta->free_pointer_tag[i] == tag)
+ break;
+ }
+ if (i == KASAN_NR_FREE_STACKS)
+ i = alloc_meta->free_track_idx;
+#endif
+
+ return &alloc_meta->free_track[i];
+}
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 147/192] kasan: add memory corruption identification support for hardware tag-based mode
2021-06-29 2:32 incoming Andrew Morton
` (145 preceding siblings ...)
2021-06-29 2:40 ` [patch 146/192] kasan: integrate the common part of two KASAN tag-based modes Andrew Morton
@ 2021-06-29 2:40 ` Andrew Morton
2021-06-29 2:41 ` [patch 148/192] mm: report which part of mem is being freed on initmem case Andrew Morton
` (44 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:40 UTC (permalink / raw)
To: akpm, andreyknvl, chinwen.chang, dvyukov, elver, glider, gregkh,
Kuan-Ying.Lee, linux-mm, matthias.bgg, mm-commits, nicholas.tang,
ryabinin.a.a, torvalds
From: Kuan-Ying Lee <Kuan-Ying.Lee@mediatek.com>
Subject: kasan: add memory corruption identification support for hardware tag-based mode
Add memory corruption identification support for hardware tag-based mode.
We store one old free pointer tag and free backtrace instead of five
because hardware tag-based kasan only has 16 different tags.
If we store as many stacks as SW tag-based kasan does(5 stacks), there is
high probability to find the same tag in the stacks when out-of-bound
issues happened and we will mistake out-of-bound issue for use-after-free.
Link: https://lkml.kernel.org/r/20210626100931.22794-4-Kuan-Ying.Lee@mediatek.com
Signed-off-by: Kuan-Ying Lee <Kuan-Ying.Lee@mediatek.com>
Suggested-by: Marco Elver <elver@google.com>
Reviewed-by: Alexander Potapenko <glider@google.com>
Reviewed-by: Andrey Konovalov <andreyknvl@gmail.com>
Reviewed-by: Marco Elver <elver@google.com>
Cc: Andrey Ryabinin <ryabinin.a.a@gmail.com>
Cc: Dmitry Vyukov <dvyukov@google.com>
Cc: Chinwen Chang <chinwen.chang@mediatek.com>
Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Cc: Matthias Brugger <matthias.bgg@gmail.com>
Cc: Nicholas Tang <nicholas.tang@mediatek.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
lib/Kconfig.kasan | 2 +-
mm/kasan/kasan.h | 2 +-
2 files changed, 2 insertions(+), 2 deletions(-)
--- a/lib/Kconfig.kasan~kasan-add-memory-corruption-identification-support-for-hardware-tag-based-mode
+++ a/lib/Kconfig.kasan
@@ -169,7 +169,7 @@ config KASAN_STACK
config KASAN_TAGS_IDENTIFY
bool "Enable memory corruption identification"
- depends on KASAN_SW_TAGS
+ depends on KASAN_SW_TAGS || KASAN_HW_TAGS
help
This option enables best-effort identification of bug type
(use-after-free or out-of-bounds) at the cost of increased
--- a/mm/kasan/kasan.h~kasan-add-memory-corruption-identification-support-for-hardware-tag-based-mode
+++ a/mm/kasan/kasan.h
@@ -153,7 +153,7 @@ struct kasan_track {
depot_stack_handle_t stack;
};
-#ifdef CONFIG_KASAN_TAGS_IDENTIFY
+#if defined(CONFIG_KASAN_TAGS_IDENTIFY) && defined(CONFIG_KASAN_SW_TAGS)
#define KASAN_NR_FREE_STACKS 5
#else
#define KASAN_NR_FREE_STACKS 1
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 148/192] mm: report which part of mem is being freed on initmem case
2021-06-29 2:32 incoming Andrew Morton
` (146 preceding siblings ...)
2021-06-29 2:40 ` [patch 147/192] kasan: add memory corruption identification support for hardware tag-based mode Andrew Morton
@ 2021-06-29 2:41 ` Andrew Morton
2021-06-29 2:41 ` [patch 149/192] mm/mmzone.h: simplify is_highmem_idx() Andrew Morton
` (43 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:41 UTC (permalink / raw)
To: akpm, js07.lee, linux-mm, mm-commits, torvalds
From: Jungseung Lee <js07.lee@samsung.com>
Subject: mm: report which part of mem is being freed on initmem case
Add the details for figuring out which parts of the kernel image is being
freed on initmem case.
Before:
Freeing unused kernel memory: 1024K
After:
Freeing unused kernel image (initmem) memory: 1024K
Link: https://lkml.kernel.org/r/1622706274-4533-1-git-send-email-js07.lee@samsung.com
Signed-off-by: Jungseung Lee <js07.lee@samsung.com>
Reviewed-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/linux/mm.h | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
--- a/include/linux/mm.h~mm-report-which-part-of-mem-is-being-freed-on-initmem-case
+++ a/include/linux/mm.h
@@ -2416,7 +2416,7 @@ static inline unsigned long free_initmem
extern char __init_begin[], __init_end[];
return free_reserved_area(&__init_begin, &__init_end,
- poison, "unused kernel");
+ poison, "unused kernel image (initmem)");
}
static inline unsigned long get_num_physpages(void)
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 149/192] mm/mmzone.h: simplify is_highmem_idx()
2021-06-29 2:32 incoming Andrew Morton
` (147 preceding siblings ...)
2021-06-29 2:41 ` [patch 148/192] mm: report which part of mem is being freed on initmem case Andrew Morton
@ 2021-06-29 2:41 ` Andrew Morton
2021-06-29 2:41 ` [patch 150/192] mm: make __dump_page static Andrew Morton
` (42 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:41 UTC (permalink / raw)
To: akpm, anshuman.khandual, corbet, linux-mm, mm-commits, rppt,
torvalds, willy
From: Mike Rapoport <rppt@linux.ibm.com>
Subject: mm/mmzone.h: simplify is_highmem_idx()
There is a lot of historical ifdefery in is_highmem_idx() and its helper
zone_movable_is_highmem() that was required because of two different paths
for nodes and zones initialization that were selected at compile time.
Until commit 3f08a302f533 ("mm: remove CONFIG_HAVE_MEMBLOCK_NODE_MAP
option") the movable_zone variable was only available for configurations
that had CONFIG_HAVE_MEMBLOCK_NODE_MAP enabled so the test in
zone_movable_is_highmem() used that variable only for such configurations.
For other configurations the test checked if the index of ZONE_MOVABLE
was greater by 1 than the index of ZONE_HIGMEM and then movable zone was
considered a highmem zone. Needless to say, ZONE_MOVABLE - 1 equals
ZONE_HIGHMEM by definition when CONFIG_HIGHMEM=y.
Commit 3f08a302f533 ("mm: remove CONFIG_HAVE_MEMBLOCK_NODE_MAP option")
made movable_zone variable always available. Since this variable is set
to ZONE_HIGHMEM if CONFIG_HIGHMEM is enabled and highmem zone is
populated, it is enough to check whether
zone_idx == ZONE_MOVABLE && movable_zone == ZONE_HIGMEM
to test if zone index points to a highmem zone.
Remove zone_movable_is_highmem() that is not used anywhere except
is_highmem_idx() and use the test above in is_highmem_idx() instead.
Link: https://lkml.kernel.org/r/20210426141927.1314326-3-rppt@kernel.org
Signed-off-by: Mike Rapoport <rppt@linux.ibm.com>
Reviewed-by: Anshuman Khandual <anshuman.khandual@arm.com>
Cc: Jonathan Corbet <corbet@lwn.net>
Cc: Matthew Wilcox <willy@infradead.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/linux/mmzone.h | 13 +------------
1 file changed, 1 insertion(+), 12 deletions(-)
--- a/include/linux/mmzone.h~mm-mmzoneh-simplify-is_highmem_idx
+++ a/include/linux/mmzone.h
@@ -982,22 +982,11 @@ static inline void zone_set_nid(struct z
extern int movable_zone;
-#ifdef CONFIG_HIGHMEM
-static inline int zone_movable_is_highmem(void)
-{
-#ifdef CONFIG_NEED_MULTIPLE_NODES
- return movable_zone == ZONE_HIGHMEM;
-#else
- return (ZONE_MOVABLE - 1) == ZONE_HIGHMEM;
-#endif
-}
-#endif
-
static inline int is_highmem_idx(enum zone_type idx)
{
#ifdef CONFIG_HIGHMEM
return (idx == ZONE_HIGHMEM ||
- (idx == ZONE_MOVABLE && zone_movable_is_highmem()));
+ (idx == ZONE_MOVABLE && movable_zone == ZONE_HIGHMEM));
#else
return 0;
#endif
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 150/192] mm: make __dump_page static
2021-06-29 2:32 incoming Andrew Morton
` (148 preceding siblings ...)
2021-06-29 2:41 ` [patch 149/192] mm/mmzone.h: simplify is_highmem_idx() Andrew Morton
@ 2021-06-29 2:41 ` Andrew Morton
2021-06-29 2:41 ` [patch 151/192] mm/page_alloc: bail out on fatal signal during reclaim/compaction retry attempt Andrew Morton
` (41 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:41 UTC (permalink / raw)
To: akpm, anshuman.khandual, linux-mm, mm-commits, torvalds, vbabka,
william.kucharski, willy
From: "Matthew Wilcox (Oracle)" <willy@infradead.org>
Subject: mm: make __dump_page static
Patch series "Constify struct page arguments".
While working on various solutions to the 32-bit struct page size
regression, one of the problems I found was the networking stack expects
to be able to pass const struct page pointers around, and the mm doesn't
provide a lot of const-friendly functions to call. The root tangle of
problems is that a lot of functions call VM_BUG_ON_PAGE(), which calls
dump_page(), which calls a lot of functions which don't take a const
struct page (but could be const).
This patch (of 6):
The only caller of __dump_page() now opencodes dump_page(), so remove it
as an externally visible symbol.
Link: https://lkml.kernel.org/r/20210416231531.2521383-1-willy@infradead.org
Link: https://lkml.kernel.org/r/20210416231531.2521383-2-willy@infradead.org
Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org>
Reviewed-by: Vlastimil Babka <vbabka@suse.cz>
Reviewed-by: Anshuman Khandual <anshuman.khandual@arm.com>
Reviewed-by: William Kucharski <william.kucharski@oracle.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/linux/mmdebug.h | 3 +--
mm/debug.c | 2 +-
mm/page_alloc.c | 3 +--
3 files changed, 3 insertions(+), 5 deletions(-)
--- a/include/linux/mmdebug.h~mm-make-__dump_page-static
+++ a/include/linux/mmdebug.h
@@ -9,8 +9,7 @@ struct page;
struct vm_area_struct;
struct mm_struct;
-extern void dump_page(struct page *page, const char *reason);
-extern void __dump_page(struct page *page, const char *reason);
+void dump_page(struct page *page, const char *reason);
void dump_vma(const struct vm_area_struct *vma);
void dump_mm(const struct mm_struct *mm);
--- a/mm/debug.c~mm-make-__dump_page-static
+++ a/mm/debug.c
@@ -42,7 +42,7 @@ const struct trace_print_flags vmaflag_n
{0, NULL}
};
-void __dump_page(struct page *page, const char *reason)
+static void __dump_page(struct page *page, const char *reason)
{
struct page *head = compound_head(page);
struct address_space *mapping;
--- a/mm/page_alloc.c~mm-make-__dump_page-static
+++ a/mm/page_alloc.c
@@ -658,8 +658,7 @@ static void bad_page(struct page *page,
pr_alert("BUG: Bad page state in process %s pfn:%05lx\n",
current->comm, page_to_pfn(page));
- __dump_page(page, reason);
- dump_page_owner(page);
+ dump_page(page, reason);
print_modules();
dump_stack();
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 151/192] mm/page_alloc: bail out on fatal signal during reclaim/compaction retry attempt
2021-06-29 2:32 incoming Andrew Morton
` (149 preceding siblings ...)
2021-06-29 2:41 ` [patch 150/192] mm: make __dump_page static Andrew Morton
@ 2021-06-29 2:41 ` Andrew Morton
2021-06-29 2:41 ` [patch 152/192] mm/debug: factor PagePoisoned out of __dump_page Andrew Morton
` (40 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:41 UTC (permalink / raw)
To: akpm, atomlin, linux-mm, mhocko, mm-commits, torvalds, vbabka, willy
From: Aaron Tomlin <atomlin@redhat.com>
Subject: mm/page_alloc: bail out on fatal signal during reclaim/compaction retry attempt
A customer experienced a low-memory situation and decided to issue a
SIGKILL (i.e. a fatal signal). Instead of promptly terminating as one
would expect, the aforementioned task remained unresponsive.
Further investigation indicated that the task was "stuck" in the
reclaim/compaction retry loop. Now, it does not make sense to retry
compaction when a fatal signal is pending.
In the context of try_to_compact_pages(), indeed COMPACT_SKIPPED can be
returned; albeit, not every zone, on the zone list, would be considered in
the case a fatal signal is found to be pending. Yet, in
should_compact_retry(), given the last known compaction result, each zone,
on the zone list, can be considered/or checked (see
compaction_zonelist_suitable()). For example, if a zone was found to
succeed, then reclaim/compaction would be tried again (notwithstanding the
above).
This patch ensures that compaction is not needlessly retried irrespective
of the last known compaction result e.g. if it was skipped, in the
unlikely case a fatal signal is found pending. So, OOM is at least
attempted.
Link: https://lkml.kernel.org/r/20210520142901.3371299-1-atomlin@redhat.com
Signed-off-by: Aaron Tomlin <atomlin@redhat.com>
Reviewed-by: Vlastimil Babka <vbabka@suse.cz>
Cc: Michal Hocko <mhocko@suse.com>
Cc: Matthew Wilcox <willy@infradead.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/page_alloc.c | 3 +++
1 file changed, 3 insertions(+)
--- a/mm/page_alloc.c~mm-page_alloc-bail-out-on-fatal-signal-during-reclaim-compaction-retry-attempt
+++ a/mm/page_alloc.c
@@ -4251,6 +4251,9 @@ should_compact_retry(struct alloc_contex
if (!order)
return false;
+ if (fatal_signal_pending(current))
+ return false;
+
if (compaction_made_progress(compact_result))
(*compaction_retries)++;
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 152/192] mm/debug: factor PagePoisoned out of __dump_page
2021-06-29 2:32 incoming Andrew Morton
` (150 preceding siblings ...)
2021-06-29 2:41 ` [patch 151/192] mm/page_alloc: bail out on fatal signal during reclaim/compaction retry attempt Andrew Morton
@ 2021-06-29 2:41 ` Andrew Morton
2021-06-29 2:41 ` [patch 153/192] mm/page_owner: constify dump_page_owner Andrew Morton
` (39 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:41 UTC (permalink / raw)
To: akpm, anshuman.khandual, linux-mm, mm-commits, torvalds, vbabka,
william.kucharski, willy
From: "Matthew Wilcox (Oracle)" <willy@infradead.org>
Subject: mm/debug: factor PagePoisoned out of __dump_page
Move the PagePoisoned test into dump_page(). Skip the hex print for
poisoned pages -- we know they're full of ffffffff. Move the reason
printing from __dump_page() to dump_page().
Link: https://lkml.kernel.org/r/20210416231531.2521383-3-willy@infradead.org
Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org>
Reviewed-by: Vlastimil Babka <vbabka@suse.cz>
Reviewed-by: Anshuman Khandual <anshuman.khandual@arm.com>
Reviewed-by: William Kucharski <william.kucharski@oracle.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/debug.c | 25 +++++++------------------
1 file changed, 7 insertions(+), 18 deletions(-)
--- a/mm/debug.c~mm-debug-factor-pagepoisoned-out-of-__dump_page
+++ a/mm/debug.c
@@ -42,11 +42,10 @@ const struct trace_print_flags vmaflag_n
{0, NULL}
};
-static void __dump_page(struct page *page, const char *reason)
+static void __dump_page(struct page *page)
{
struct page *head = compound_head(page);
struct address_space *mapping;
- bool page_poisoned = PagePoisoned(page);
bool compound = PageCompound(page);
/*
* Accessing the pageblock without the zone lock. It could change to
@@ -58,16 +57,6 @@ static void __dump_page(struct page *pag
int mapcount;
char *type = "";
- /*
- * If struct page is poisoned don't access Page*() functions as that
- * leads to recursive loop. Page*() check for poisoned pages, and calls
- * dump_page() when detected.
- */
- if (page_poisoned) {
- pr_warn("page:%px is uninitialized and poisoned", page);
- goto hex_only;
- }
-
if (page < head || (page >= head + MAX_ORDER_NR_PAGES)) {
/*
* Corrupt page, so we cannot call page_mapping. Instead, do a
@@ -173,8 +162,6 @@ out_mapping:
pr_warn("%sflags: %#lx(%pGp)%s\n", type, head->flags, &head->flags,
page_cma ? " CMA" : "");
-
-hex_only:
print_hex_dump(KERN_WARNING, "raw: ", DUMP_PREFIX_NONE, 32,
sizeof(unsigned long), page,
sizeof(struct page), false);
@@ -182,14 +169,16 @@ hex_only:
print_hex_dump(KERN_WARNING, "head: ", DUMP_PREFIX_NONE, 32,
sizeof(unsigned long), head,
sizeof(struct page), false);
-
- if (reason)
- pr_warn("page dumped because: %s\n", reason);
}
void dump_page(struct page *page, const char *reason)
{
- __dump_page(page, reason);
+ if (PagePoisoned(page))
+ pr_warn("page:%p is uninitialized and poisoned", page);
+ else
+ __dump_page(page);
+ if (reason)
+ pr_warn("page dumped because: %s\n", reason);
dump_page_owner(page);
}
EXPORT_SYMBOL(dump_page);
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 153/192] mm/page_owner: constify dump_page_owner
2021-06-29 2:32 incoming Andrew Morton
` (151 preceding siblings ...)
2021-06-29 2:41 ` [patch 152/192] mm/debug: factor PagePoisoned out of __dump_page Andrew Morton
@ 2021-06-29 2:41 ` Andrew Morton
2021-06-29 2:41 ` [patch 154/192] mm: make compound_head const-preserving Andrew Morton
` (38 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:41 UTC (permalink / raw)
To: akpm, anshuman.khandual, linux-mm, mm-commits, torvalds, vbabka,
william.kucharski, willy
From: "Matthew Wilcox (Oracle)" <willy@infradead.org>
Subject: mm/page_owner: constify dump_page_owner
dump_page_owner() only uses struct page to find the page_ext, and
lookup_page_ext() already takes a const argument.
Link: https://lkml.kernel.org/r/20210416231531.2521383-4-willy@infradead.org
Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org>
Reviewed-by: Vlastimil Babka <vbabka@suse.cz>
Reviewed-by: Anshuman Khandual <anshuman.khandual@arm.com>
Reviewed-by: William Kucharski <william.kucharski@oracle.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/linux/page_owner.h | 6 +++---
mm/page_owner.c | 2 +-
2 files changed, 4 insertions(+), 4 deletions(-)
--- a/include/linux/page_owner.h~mm-page_owner-constify-dump_page_owner
+++ a/include/linux/page_owner.h
@@ -14,7 +14,7 @@ extern void __set_page_owner(struct page
extern void __split_page_owner(struct page *page, unsigned int nr);
extern void __copy_page_owner(struct page *oldpage, struct page *newpage);
extern void __set_page_owner_migrate_reason(struct page *page, int reason);
-extern void __dump_page_owner(struct page *page);
+extern void __dump_page_owner(const struct page *page);
extern void pagetypeinfo_showmixedcount_print(struct seq_file *m,
pg_data_t *pgdat, struct zone *zone);
@@ -46,7 +46,7 @@ static inline void set_page_owner_migrat
if (static_branch_unlikely(&page_owner_inited))
__set_page_owner_migrate_reason(page, reason);
}
-static inline void dump_page_owner(struct page *page)
+static inline void dump_page_owner(const struct page *page)
{
if (static_branch_unlikely(&page_owner_inited))
__dump_page_owner(page);
@@ -69,7 +69,7 @@ static inline void copy_page_owner(struc
static inline void set_page_owner_migrate_reason(struct page *page, int reason)
{
}
-static inline void dump_page_owner(struct page *page)
+static inline void dump_page_owner(const struct page *page)
{
}
#endif /* CONFIG_PAGE_OWNER */
--- a/mm/page_owner.c~mm-page_owner-constify-dump_page_owner
+++ a/mm/page_owner.c
@@ -392,7 +392,7 @@ err:
return -ENOMEM;
}
-void __dump_page_owner(struct page *page)
+void __dump_page_owner(const struct page *page)
{
struct page_ext *page_ext = lookup_page_ext(page);
struct page_owner *page_owner;
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 154/192] mm: make compound_head const-preserving
2021-06-29 2:32 incoming Andrew Morton
` (152 preceding siblings ...)
2021-06-29 2:41 ` [patch 153/192] mm/page_owner: constify dump_page_owner Andrew Morton
@ 2021-06-29 2:41 ` Andrew Morton
2021-06-29 2:41 ` [patch 155/192] mm: constify get_pfnblock_flags_mask and get_pfnblock_migratetype Andrew Morton
` (37 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:41 UTC (permalink / raw)
To: akpm, anshuman.khandual, linux-mm, mm-commits, torvalds, vbabka,
william.kucharski, willy
From: "Matthew Wilcox (Oracle)" <willy@infradead.org>
Subject: mm: make compound_head const-preserving
If you pass a const pointer to compound_head(), you get a const pointer
back; if you pass a mutable pointer, you get a mutable pointer back. Also
remove an unnecessary forward definition of struct page; we're about to
dereference page->compound_head, so it must already have been defined.
Link: https://lkml.kernel.org/r/20210416231531.2521383-5-willy@infradead.org
Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org>
Reviewed-by: Vlastimil Babka <vbabka@suse.cz>
Reviewed-by: Anshuman Khandual <anshuman.khandual@arm.com>
Reviewed-by: William Kucharski <william.kucharski@oracle.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/linux/page-flags.h | 10 +++++-----
1 file changed, 5 insertions(+), 5 deletions(-)
--- a/include/linux/page-flags.h~mm-make-compound_head-const-preserving
+++ a/include/linux/page-flags.h
@@ -177,17 +177,17 @@ enum pageflags {
#ifndef __GENERATING_BOUNDS_H
-struct page; /* forward declaration */
-
-static inline struct page *compound_head(struct page *page)
+static inline unsigned long _compound_head(const struct page *page)
{
unsigned long head = READ_ONCE(page->compound_head);
if (unlikely(head & 1))
- return (struct page *) (head - 1);
- return page;
+ return head - 1;
+ return (unsigned long)page;
}
+#define compound_head(page) ((typeof(page))_compound_head(page))
+
static __always_inline int PageTail(struct page *page)
{
return READ_ONCE(page->compound_head) & 1;
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 155/192] mm: constify get_pfnblock_flags_mask and get_pfnblock_migratetype
2021-06-29 2:32 incoming Andrew Morton
` (153 preceding siblings ...)
2021-06-29 2:41 ` [patch 154/192] mm: make compound_head const-preserving Andrew Morton
@ 2021-06-29 2:41 ` Andrew Morton
2021-06-29 2:41 ` [patch 156/192] mm: constify page_count and page_ref_count Andrew Morton
` (36 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:41 UTC (permalink / raw)
To: akpm, anshuman.khandual, linux-mm, mm-commits, torvalds, vbabka,
william.kucharski, willy
From: "Matthew Wilcox (Oracle)" <willy@infradead.org>
Subject: mm: constify get_pfnblock_flags_mask and get_pfnblock_migratetype
The struct page is not modified by these routines, so it can be marked
const.
Link: https://lkml.kernel.org/r/20210416231531.2521383-6-willy@infradead.org
Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org>
Reviewed-by: Vlastimil Babka <vbabka@suse.cz>
Reviewed-by: Anshuman Khandual <anshuman.khandual@arm.com>
Reviewed-by: William Kucharski <william.kucharski@oracle.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/linux/pageblock-flags.h | 2 +-
mm/page_alloc.c | 13 +++++++------
2 files changed, 8 insertions(+), 7 deletions(-)
--- a/include/linux/pageblock-flags.h~mm-constify-get_pfnblock_flags_mask-and-get_pfnblock_migratetype
+++ a/include/linux/pageblock-flags.h
@@ -54,7 +54,7 @@ extern unsigned int pageblock_order;
/* Forward declaration */
struct page;
-unsigned long get_pfnblock_flags_mask(struct page *page,
+unsigned long get_pfnblock_flags_mask(const struct page *page,
unsigned long pfn,
unsigned long mask);
--- a/mm/page_alloc.c~mm-constify-get_pfnblock_flags_mask-and-get_pfnblock_migratetype
+++ a/mm/page_alloc.c
@@ -474,7 +474,7 @@ static inline bool defer_init(int nid, u
#endif
/* Return a pointer to the bitmap storing bits affecting a block of pages */
-static inline unsigned long *get_pageblock_bitmap(struct page *page,
+static inline unsigned long *get_pageblock_bitmap(const struct page *page,
unsigned long pfn)
{
#ifdef CONFIG_SPARSEMEM
@@ -484,7 +484,7 @@ static inline unsigned long *get_pageblo
#endif /* CONFIG_SPARSEMEM */
}
-static inline int pfn_to_bitidx(struct page *page, unsigned long pfn)
+static inline int pfn_to_bitidx(const struct page *page, unsigned long pfn)
{
#ifdef CONFIG_SPARSEMEM
pfn &= (PAGES_PER_SECTION-1);
@@ -495,7 +495,7 @@ static inline int pfn_to_bitidx(struct p
}
static __always_inline
-unsigned long __get_pfnblock_flags_mask(struct page *page,
+unsigned long __get_pfnblock_flags_mask(const struct page *page,
unsigned long pfn,
unsigned long mask)
{
@@ -520,13 +520,14 @@ unsigned long __get_pfnblock_flags_mask(
*
* Return: pageblock_bits flags
*/
-unsigned long get_pfnblock_flags_mask(struct page *page, unsigned long pfn,
- unsigned long mask)
+unsigned long get_pfnblock_flags_mask(const struct page *page,
+ unsigned long pfn, unsigned long mask)
{
return __get_pfnblock_flags_mask(page, pfn, mask);
}
-static __always_inline int get_pfnblock_migratetype(struct page *page, unsigned long pfn)
+static __always_inline int get_pfnblock_migratetype(const struct page *page,
+ unsigned long pfn)
{
return __get_pfnblock_flags_mask(page, pfn, MIGRATETYPE_MASK);
}
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 156/192] mm: constify page_count and page_ref_count
2021-06-29 2:32 incoming Andrew Morton
` (154 preceding siblings ...)
2021-06-29 2:41 ` [patch 155/192] mm: constify get_pfnblock_flags_mask and get_pfnblock_migratetype Andrew Morton
@ 2021-06-29 2:41 ` Andrew Morton
2021-06-29 2:41 ` [patch 157/192] mm: optimise nth_page for contiguous memmap Andrew Morton
` (35 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:41 UTC (permalink / raw)
To: akpm, anshuman.khandual, linux-mm, mm-commits, torvalds, vbabka,
william.kucharski, willy
From: "Matthew Wilcox (Oracle)" <willy@infradead.org>
Subject: mm: constify page_count and page_ref_count
Now that compound_head() accepts a const struct page pointer, these two
functions can be marked as not modifying the page pointer they are passed.
Link: https://lkml.kernel.org/r/20210416231531.2521383-7-willy@infradead.org
Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org>
Reviewed-by: Vlastimil Babka <vbabka@suse.cz>
Reviewed-by: Anshuman Khandual <anshuman.khandual@arm.com>
Reviewed-by: William Kucharski <william.kucharski@oracle.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/linux/page_ref.h | 4 ++--
1 file changed, 2 insertions(+), 2 deletions(-)
--- a/include/linux/page_ref.h~mm-constify-page_count-and-page_ref_count
+++ a/include/linux/page_ref.h
@@ -62,12 +62,12 @@ static inline void __page_ref_unfreeze(s
#endif
-static inline int page_ref_count(struct page *page)
+static inline int page_ref_count(const struct page *page)
{
return atomic_read(&page->_refcount);
}
-static inline int page_count(struct page *page)
+static inline int page_count(const struct page *page)
{
return atomic_read(&compound_head(page)->_refcount);
}
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 157/192] mm: optimise nth_page for contiguous memmap
2021-06-29 2:32 incoming Andrew Morton
` (155 preceding siblings ...)
2021-06-29 2:41 ` [patch 156/192] mm: constify page_count and page_ref_count Andrew Morton
@ 2021-06-29 2:41 ` Andrew Morton
2021-06-29 2:41 ` [patch 158/192] mm/page_alloc: switch to pr_debug Andrew Morton
` (34 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:41 UTC (permalink / raw)
To: akpm, chris, david, dougg, fujita.tomonori, hch, linux-mm,
mm-commits, tj, torvalds, vbabka, willy, ziy
From: "Matthew Wilcox (Oracle)" <willy@infradead.org>
Subject: mm: optimise nth_page for contiguous memmap
If the memmap is virtually contiguous (either because we're using a
virtually mapped memmap or because we don't support a discontig memmap at
all), then we can implement nth_page() by simple addition. Contrary to
popular belief, the compiler is not able to optimise this itself for a
vmemmap configuration. This reduces one example user (sg.c) by four
instructions:
struct page *page = nth_page(rsv_schp->pages[k], offset >> PAGE_SHIFT);
before:
49 8b 45 70 mov 0x70(%r13),%rax
48 63 c9 movslq %ecx,%rcx
48 c1 eb 0c shr $0xc,%rbx
48 8b 04 c8 mov (%rax,%rcx,8),%rax
48 2b 05 00 00 00 00 sub 0x0(%rip),%rax
R_X86_64_PC32 vmemmap_base-0x4
48 c1 f8 06 sar $0x6,%rax
48 01 d8 add %rbx,%rax
48 c1 e0 06 shl $0x6,%rax
48 03 05 00 00 00 00 add 0x0(%rip),%rax
R_X86_64_PC32 vmemmap_base-0x4
after:
49 8b 45 70 mov 0x70(%r13),%rax
48 63 c9 movslq %ecx,%rcx
48 c1 eb 0c shr $0xc,%rbx
48 c1 e3 06 shl $0x6,%rbx
48 03 1c c8 add (%rax,%rcx,8),%rbx
Link: https://lkml.kernel.org/r/20210413194625.1472345-1-willy@infradead.org
Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: David Hildenbrand <david@redhat.com>
Reviewed-by: Zi Yan <ziy@nvidia.com>
Reviewed-by: Vlastimil Babka <vbabka@suse.cz>
Cc: Tejun Heo <tj@kernel.org>
Cc: FUJITA Tomonori <fujita.tomonori@lab.ntt.co.jp>
Cc: Douglas Gilbert <dougg@torque.net>
Cc: Chris Wilson <chris@chris-wilson.co.uk>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/linux/mm.h | 4 ++++
1 file changed, 4 insertions(+)
--- a/include/linux/mm.h~mm-optimise-nth_page-for-contiguous-memmap
+++ a/include/linux/mm.h
@@ -234,7 +234,11 @@ int overcommit_policy_handler(struct ctl
int __add_to_page_cache_locked(struct page *page, struct address_space *mapping,
pgoff_t index, gfp_t gfp, void **shadowp);
+#if defined(CONFIG_SPARSEMEM) && !defined(CONFIG_SPARSEMEM_VMEMMAP)
#define nth_page(page,n) pfn_to_page(page_to_pfn((page)) + (n))
+#else
+#define nth_page(page,n) ((page) + (n))
+#endif
/* to align the pointer to the (next) page boundary */
#define PAGE_ALIGN(addr) ALIGN(addr, PAGE_SIZE)
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 158/192] mm/page_alloc: switch to pr_debug
2021-06-29 2:32 incoming Andrew Morton
` (156 preceding siblings ...)
2021-06-29 2:41 ` [patch 157/192] mm: optimise nth_page for contiguous memmap Andrew Morton
@ 2021-06-29 2:41 ` Andrew Morton
2021-06-29 2:41 ` [patch 159/192] kbuild: skip per-CPU BTF generation for pahole v1.18-v1.21 Andrew Morton
` (33 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:41 UTC (permalink / raw)
To: akpm, hkallweit1, linux-mm, mm-commits, torvalds
From: Heiner Kallweit <hkallweit1@gmail.com>
Subject: mm/page_alloc: switch to pr_debug
Having such debug messages in the dmesg log may confuse users. Therefore
restrict debug output to cases where DEBUG is defined or dynamic debugging
is enabled for the respective code piece.
Link: https://lkml.kernel.org/r/976adb93-3041-ce63-48fc-55a6096a51c1@gmail.com
Signed-off-by: Heiner Kallweit <hkallweit1@gmail.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/page_alloc.c | 16 ++++++----------
1 file changed, 6 insertions(+), 10 deletions(-)
--- a/mm/page_alloc.c~mm-page_alloc-switch-to-pr_debug
+++ a/mm/page_alloc.c
@@ -6770,9 +6770,8 @@ static __meminit void zone_pcp_init(stru
zone->pageset_batch = BOOT_PAGESET_BATCH;
if (populated_zone(zone))
- printk(KERN_DEBUG " %s zone: %lu pages, LIFO batch:%u\n",
- zone->name, zone->present_pages,
- zone_batchsize(zone));
+ pr_debug(" %s zone: %lu pages, LIFO batch:%u\n", zone->name,
+ zone->present_pages, zone_batchsize(zone));
}
void __meminit init_currently_empty_zone(struct zone *zone,
@@ -7042,8 +7041,7 @@ static void __init calculate_node_totalp
pgdat->node_spanned_pages = totalpages;
pgdat->node_present_pages = realtotalpages;
- printk(KERN_DEBUG "On node %d totalpages: %lu\n", pgdat->node_id,
- realtotalpages);
+ pr_debug("On node %d totalpages: %lu\n", pgdat->node_id, realtotalpages);
}
#ifndef CONFIG_SPARSEMEM
@@ -7243,9 +7241,8 @@ static void __init free_area_init_core(s
if (freesize >= memmap_pages) {
freesize -= memmap_pages;
if (memmap_pages)
- printk(KERN_DEBUG
- " %s zone: %lu pages used for memmap\n",
- zone_names[j], memmap_pages);
+ pr_debug(" %s zone: %lu pages used for memmap\n",
+ zone_names[j], memmap_pages);
} else
pr_warn(" %s zone: %lu pages exceeds freesize %lu\n",
zone_names[j], memmap_pages, freesize);
@@ -7254,8 +7251,7 @@ static void __init free_area_init_core(s
/* Account for reserved pages */
if (j == 0 && freesize > dma_reserve) {
freesize -= dma_reserve;
- printk(KERN_DEBUG " %s zone: %lu pages reserved\n",
- zone_names[0], dma_reserve);
+ pr_debug(" %s zone: %lu pages reserved\n", zone_names[0], dma_reserve);
}
if (!is_highmem_idx(j))
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 159/192] kbuild: skip per-CPU BTF generation for pahole v1.18-v1.21
2021-06-29 2:32 incoming Andrew Morton
` (157 preceding siblings ...)
2021-06-29 2:41 ` [patch 158/192] mm/page_alloc: switch to pr_debug Andrew Morton
@ 2021-06-29 2:41 ` Andrew Morton
2021-06-29 2:41 ` [patch 160/192] mm/page_alloc: split per cpu page lists and zone stats Andrew Morton
` (32 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:41 UTC (permalink / raw)
To: acme, akpm, andrii, haoluo, jolsa, linux-mm, mgorman, mm-commits,
msuchanek, torvalds
From: Andrii Nakryiko <andrii@kernel.org>
Subject: kbuild: skip per-CPU BTF generation for pahole v1.18-v1.21
Commit "mm/page_alloc: convert per-cpu list protection to local_lock" will
introduce a zero-sized per-CPU variable, which causes pahole to generate
invalid BTF. Only pahole versions 1.18 through 1.21 are impacted, as
before 1.18 pahole doesn't know anything about per-CPU variables, and 1.22
contains the proper fix for the issue.
Luckily, pahole 1.18 got --skip_encoding_btf_vars option disabling BTF
generation for per-CPU variables in anticipation of some unanticipated
problems. So use this escape hatch to disable per-CPU var BTF info on
those problematic pahole versions. Users relying on availability of
per-CPU var BTFs would need to upgrade to pahole 1.22+, but everyone won't
notice any regressions.
Link: https://lkml.kernel.org/r/20210530002536.3193829-1-andrii@kernel.org
Signed-off-by: Andrii Nakryiko <andrii@kernel.org>
Acked-by: Mel Gorman <mgorman@techsingularity.net>
Cc: Arnaldo Carvalho de Melo <acme@redhat.com>
Cc: Hao Luo <haoluo@google.com>
Cc: Michal Suchanek <msuchanek@suse.de>
Cc: Jiri Olsa <jolsa@kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
scripts/link-vmlinux.sh | 4 ++++
1 file changed, 4 insertions(+)
--- a/scripts/link-vmlinux.sh~kbuild-skip-per-cpu-btf-generation-for-pahole-v118-v121
+++ a/scripts/link-vmlinux.sh
@@ -235,6 +235,10 @@ gen_btf()
vmlinux_link ${1}
+ if [ "${pahole_ver}" -ge "118" ] && [ "${pahole_ver}" -le "121" ]; then
+ # pahole 1.18 through 1.21 can't handle zero-sized per-CPU vars
+ extra_paholeopt="${extra_paholeopt} --skip_encoding_btf_vars"
+ fi
if [ "${pahole_ver}" -ge "121" ]; then
extra_paholeopt="${extra_paholeopt} --btf_gen_floats"
fi
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 160/192] mm/page_alloc: split per cpu page lists and zone stats
2021-06-29 2:32 incoming Andrew Morton
` (158 preceding siblings ...)
2021-06-29 2:41 ` [patch 159/192] kbuild: skip per-CPU BTF generation for pahole v1.18-v1.21 Andrew Morton
@ 2021-06-29 2:41 ` Andrew Morton
2021-06-29 2:41 ` [patch 161/192] mm/page_alloc: convert per-cpu list protection to local_lock Andrew Morton
` (31 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:41 UTC (permalink / raw)
To: akpm, bigeasy, brouer, chuck.lever, linux-mm, mgorman, mhocko,
mingo, mm-commits, peterz, tglx, torvalds, vbabka
From: Mel Gorman <mgorman@techsingularity.net>
Subject: mm/page_alloc: split per cpu page lists and zone stats
The PCP (per-cpu page allocator in page_alloc.c) shares locking
requirements with vmstat and the zone lock which is inconvenient and
causes some issues. For example, the PCP list and vmstat share the same
per-cpu space meaning that it's possible that vmstat updates dirty cache
lines holding per-cpu lists across CPUs unless padding is used. Second,
PREEMPT_RT does not want to disable IRQs for too long in the page
allocator.
This series splits the locking requirements and uses locks types more
suitable for PREEMPT_RT, reduces the time when special locking is required
for stats and reduces the time when IRQs need to be disabled on
!PREEMPT_RT kernels.
Why local_lock? PREEMPT_RT considers the following sequence to be unsafe
as documented in Documentation/locking/locktypes.rst
local_irq_disable();
spin_lock(&lock);
The pcp allocator has this sequence for rmqueue_pcplist (local_irq_save)
-> __rmqueue_pcplist -> rmqueue_bulk (spin_lock). While it's possible to
separate this out, it generally means there are points where we enable
IRQs and reenable them again immediately. To prevent a migration and the
per-cpu pointer going stale, migrate_disable is also needed. That is a
custom lock that is similar, but worse, than local_lock. Furthermore, on
PREEMPT_RT, it's undesirable to leave IRQs disabled for too long. By
converting to local_lock which disables migration on PREEMPT_RT, the
locking requirements can be separated and start moving the protections for
PCP, stats and the zone lock to PREEMPT_RT-safe equivalent locking. As a
bonus, local_lock also means that PROVE_LOCKING does something useful.
After that, it's obvious that zone_statistics incurs too much overhead and
leaves IRQs disabled for longer than necessary on !PREEMPT_RT kernels.
zone_statistics uses perfectly accurate counters requiring IRQs be
disabled for parallel RMW sequences when inaccurate ones like vm_events
would do. The series makes the NUMA statistics (NUMA_HIT and friends)
inaccurate counters that then require no special protection on
!PREEMPT_RT.
The bulk page allocator can then do stat updates in bulk with IRQs enabled
which should improve the efficiency. Technically, this could have been
done without the local_lock and vmstat conversion work and the order
simply reflects the timing of when different series were implemented.
Finally, there are places where we conflate IRQs being disabled for the
PCP with the IRQ-safe zone spinlock. The remainder of the series reduces
the scope of what is protected by disabled IRQs on !PREEMPT_RT kernels.
By the end of the series, page_alloc.c does not call local_irq_save so the
locking scope is a bit clearer. The one exception is that modifying
NR_FREE_PAGES still happens in places where it's known the IRQs are
disabled as it's harmless for PREEMPT_RT and would be expensive to split
the locking there.
No performance data is included because despite the overhead of the stats,
it's within the noise for most workloads on !PREEMPT_RT. However, Jesper
Dangaard Brouer ran a page allocation microbenchmark on a E5-1650 v4 @
3.60GHz CPU on the first version of this series. Focusing on the array
variant of the bulk page allocator reveals the following.
(CPU: Intel(R) Xeon(R) CPU E5-1650 v4 @ 3.60GHz)
ARRAY variant: time_bulk_page_alloc_free_array: step=bulk size
Baseline Patched
1 56.383 54.225 (+3.83%)
2 40.047 35.492 (+11.38%)
3 37.339 32.643 (+12.58%)
4 35.578 30.992 (+12.89%)
8 33.592 29.606 (+11.87%)
16 32.362 28.532 (+11.85%)
32 31.476 27.728 (+11.91%)
64 30.633 27.252 (+11.04%)
128 30.596 27.090 (+11.46%)
While this is a positive outcome, the series is more likely to be
interesting to the RT people in terms of getting parts of the PREEMPT_RT
tree into mainline.
This patch (of 9):
The per-cpu page allocator lists and the per-cpu vmstat deltas are stored
in the same struct per_cpu_pages even though vmstats have no direct impact
on the per-cpu page lists. This is inconsistent because the vmstats for a
node are stored on a dedicated structure. The bigger issue is that the
per_cpu_pages structure is not cache-aligned and stat updates either cache
conflict with adjacent per-cpu lists incurring a runtime cost or padding
is required incurring a memory cost.
This patch splits the per-cpu pagelists and the vmstat deltas into
separate structures. It's mostly a mechanical conversion but some
variable renaming is done to clearly distinguish the per-cpu pages
structure (pcp) from the vmstats (pzstats).
Superficially, this appears to increase the size of the per_cpu_pages
structure but the movement of expire fills a structure hole so there is no
impact overall.
[mgorman@techsingularity.net: make it W=1 cleaner]
Link: https://lkml.kernel.org/r/20210514144622.GA3735@techsingularity.net
[mgorman@techsingularity.net: make it W=1 even cleaner]
Link: https://lkml.kernel.org/r/20210516140705.GB3735@techsingularity.net
[lkp@intel.com: check struct per_cpu_zonestat has a non-zero size]
[vbabka@suse.cz: Init zone->per_cpu_zonestats properly]
Link: https://lkml.kernel.org/r/20210512095458.30632-1-mgorman@techsingularity.net
Link: https://lkml.kernel.org/r/20210512095458.30632-2-mgorman@techsingularity.net
Signed-off-by: Mel Gorman <mgorman@techsingularity.net>
Acked-by: Vlastimil Babka <vbabka@suse.cz>
Acked-by: Peter Zijlstra (Intel) <peterz@infradead.org>
Cc: Chuck Lever <chuck.lever@oracle.com>
Cc: Jesper Dangaard Brouer <brouer@redhat.com>
Cc: Thomas Gleixner <tglx@linutronix.de>
Cc: Sebastian Andrzej Siewior <bigeasy@linutronix.de>
Cc: Ingo Molnar <mingo@kernel.org>
Cc: Michal Hocko <mhocko@kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/linux/mmzone.h | 18 +++----
include/linux/vmstat.h | 8 +--
mm/page_alloc.c | 85 ++++++++++++++++++---------------
mm/vmstat.c | 98 ++++++++++++++++++++-------------------
4 files changed, 113 insertions(+), 96 deletions(-)
--- a/include/linux/mmzone.h~mm-page_alloc-split-per-cpu-page-lists-and-zone-stats
+++ a/include/linux/mmzone.h
@@ -341,20 +341,21 @@ struct per_cpu_pages {
int count; /* number of pages in the list */
int high; /* high watermark, emptying needed */
int batch; /* chunk size for buddy add/remove */
+#ifdef CONFIG_NUMA
+ int expire; /* When 0, remote pagesets are drained */
+#endif
/* Lists of pages, one per migrate type stored on the pcp-lists */
struct list_head lists[MIGRATE_PCPTYPES];
};
-struct per_cpu_pageset {
- struct per_cpu_pages pcp;
-#ifdef CONFIG_NUMA
- s8 expire;
- u16 vm_numa_stat_diff[NR_VM_NUMA_STAT_ITEMS];
-#endif
+struct per_cpu_zonestat {
#ifdef CONFIG_SMP
- s8 stat_threshold;
s8 vm_stat_diff[NR_VM_ZONE_STAT_ITEMS];
+ s8 stat_threshold;
+#endif
+#ifdef CONFIG_NUMA
+ u16 vm_numa_stat_diff[NR_VM_NUMA_STAT_ITEMS];
#endif
};
@@ -484,7 +485,8 @@ struct zone {
int node;
#endif
struct pglist_data *zone_pgdat;
- struct per_cpu_pageset __percpu *pageset;
+ struct per_cpu_pages __percpu *per_cpu_pageset;
+ struct per_cpu_zonestat __percpu *per_cpu_zonestats;
/*
* the high and batch values are copied to individual pagesets for
* faster access
--- a/include/linux/vmstat.h~mm-page_alloc-split-per-cpu-page-lists-and-zone-stats
+++ a/include/linux/vmstat.h
@@ -163,7 +163,7 @@ static inline unsigned long zone_numa_st
int cpu;
for_each_online_cpu(cpu)
- x += per_cpu_ptr(zone->pageset, cpu)->vm_numa_stat_diff[item];
+ x += per_cpu_ptr(zone->per_cpu_zonestats, cpu)->vm_numa_stat_diff[item];
return x;
}
@@ -236,7 +236,7 @@ static inline unsigned long zone_page_st
#ifdef CONFIG_SMP
int cpu;
for_each_online_cpu(cpu)
- x += per_cpu_ptr(zone->pageset, cpu)->vm_stat_diff[item];
+ x += per_cpu_ptr(zone->per_cpu_zonestats, cpu)->vm_stat_diff[item];
if (x < 0)
x = 0;
@@ -291,7 +291,7 @@ struct ctl_table;
int vmstat_refresh(struct ctl_table *, int write, void *buffer, size_t *lenp,
loff_t *ppos);
-void drain_zonestat(struct zone *zone, struct per_cpu_pageset *);
+void drain_zonestat(struct zone *zone, struct per_cpu_zonestat *);
int calculate_pressure_threshold(struct zone *zone);
int calculate_normal_threshold(struct zone *zone);
@@ -399,7 +399,7 @@ static inline void cpu_vm_stats_fold(int
static inline void quiet_vmstat(void) { }
static inline void drain_zonestat(struct zone *zone,
- struct per_cpu_pageset *pset) { }
+ struct per_cpu_zonestat *pzstats) { }
#endif /* CONFIG_SMP */
static inline void __mod_zone_freepage_state(struct zone *zone, int nr_pages,
--- a/mm/page_alloc.c~mm-page_alloc-split-per-cpu-page-lists-and-zone-stats
+++ a/mm/page_alloc.c
@@ -3026,15 +3026,14 @@ void drain_zone_pages(struct zone *zone,
static void drain_pages_zone(unsigned int cpu, struct zone *zone)
{
unsigned long flags;
- struct per_cpu_pageset *pset;
struct per_cpu_pages *pcp;
local_irq_save(flags);
- pset = per_cpu_ptr(zone->pageset, cpu);
- pcp = &pset->pcp;
+ pcp = per_cpu_ptr(zone->per_cpu_pageset, cpu);
if (pcp->count)
free_pcppages_bulk(zone, pcp->count, pcp);
+
local_irq_restore(flags);
}
@@ -3133,7 +3132,7 @@ static void __drain_all_pages(struct zon
* disables preemption as part of its processing
*/
for_each_online_cpu(cpu) {
- struct per_cpu_pageset *pcp;
+ struct per_cpu_pages *pcp;
struct zone *z;
bool has_pcps = false;
@@ -3144,13 +3143,13 @@ static void __drain_all_pages(struct zon
*/
has_pcps = true;
} else if (zone) {
- pcp = per_cpu_ptr(zone->pageset, cpu);
- if (pcp->pcp.count)
+ pcp = per_cpu_ptr(zone->per_cpu_pageset, cpu);
+ if (pcp->count)
has_pcps = true;
} else {
for_each_populated_zone(z) {
- pcp = per_cpu_ptr(z->pageset, cpu);
- if (pcp->pcp.count) {
+ pcp = per_cpu_ptr(z->per_cpu_pageset, cpu);
+ if (pcp->count) {
has_pcps = true;
break;
}
@@ -3280,7 +3279,7 @@ static void free_unref_page_commit(struc
migratetype = MIGRATE_MOVABLE;
}
- pcp = &this_cpu_ptr(zone->pageset)->pcp;
+ pcp = this_cpu_ptr(zone->per_cpu_pageset);
list_add(&page->lru, &pcp->lists[migratetype]);
pcp->count++;
if (pcp->count >= READ_ONCE(pcp->high))
@@ -3496,7 +3495,7 @@ static struct page *rmqueue_pcplist(stru
unsigned long flags;
local_irq_save(flags);
- pcp = &this_cpu_ptr(zone->pageset)->pcp;
+ pcp = this_cpu_ptr(zone->per_cpu_pageset);
list = &pcp->lists[migratetype];
page = __rmqueue_pcplist(zone, migratetype, alloc_flags, pcp, list);
if (page) {
@@ -5105,7 +5104,7 @@ unsigned long __alloc_pages_bulk(gfp_t g
/* Attempt the batch allocation */
local_irq_save(flags);
- pcp = &this_cpu_ptr(zone->pageset)->pcp;
+ pcp = this_cpu_ptr(zone->per_cpu_pageset);
pcp_list = &pcp->lists[ac.migratetype];
while (nr_populated < nr_pages) {
@@ -5720,7 +5719,7 @@ void show_free_areas(unsigned int filter
continue;
for_each_online_cpu(cpu)
- free_pcp += per_cpu_ptr(zone->pageset, cpu)->pcp.count;
+ free_pcp += per_cpu_ptr(zone->per_cpu_pageset, cpu)->count;
}
printk("active_anon:%lu inactive_anon:%lu isolated_anon:%lu\n"
@@ -5812,7 +5811,7 @@ void show_free_areas(unsigned int filter
free_pcp = 0;
for_each_online_cpu(cpu)
- free_pcp += per_cpu_ptr(zone->pageset, cpu)->pcp.count;
+ free_pcp += per_cpu_ptr(zone->per_cpu_pageset, cpu)->count;
show_node(zone);
printk(KERN_CONT
@@ -5853,7 +5852,7 @@ void show_free_areas(unsigned int filter
K(zone_page_state(zone, NR_MLOCK)),
K(zone_page_state(zone, NR_BOUNCE)),
K(free_pcp),
- K(this_cpu_read(zone->pageset->pcp.count)),
+ K(this_cpu_read(zone->per_cpu_pageset->count)),
K(zone_page_state(zone, NR_FREE_CMA_PAGES)));
printk("lowmem_reserve[]:");
for (i = 0; i < MAX_NR_ZONES; i++)
@@ -6180,11 +6179,12 @@ static void build_zonelists(pg_data_t *p
* not check if the processor is online before following the pageset pointer.
* Other parts of the kernel may not check if the zone is available.
*/
-static void pageset_init(struct per_cpu_pageset *p);
+static void per_cpu_pages_init(struct per_cpu_pages *pcp, struct per_cpu_zonestat *pzstats);
/* These effectively disable the pcplists in the boot pageset completely */
#define BOOT_PAGESET_HIGH 0
#define BOOT_PAGESET_BATCH 1
-static DEFINE_PER_CPU(struct per_cpu_pageset, boot_pageset);
+static DEFINE_PER_CPU(struct per_cpu_pages, boot_pageset);
+static DEFINE_PER_CPU(struct per_cpu_zonestat, boot_zonestats);
static DEFINE_PER_CPU(struct per_cpu_nodestat, boot_nodestats);
static void __build_all_zonelists(void *data)
@@ -6251,7 +6251,7 @@ build_all_zonelists_init(void)
* (a chicken-egg dilemma).
*/
for_each_possible_cpu(cpu)
- pageset_init(&per_cpu(boot_pageset, cpu));
+ per_cpu_pages_init(&per_cpu(boot_pageset, cpu), &per_cpu(boot_zonestats, cpu));
mminit_verify_zonelist();
cpuset_init_current_mems_allowed();
@@ -6650,14 +6650,13 @@ static void pageset_update(struct per_cp
WRITE_ONCE(pcp->high, high);
}
-static void pageset_init(struct per_cpu_pageset *p)
+static void per_cpu_pages_init(struct per_cpu_pages *pcp, struct per_cpu_zonestat *pzstats)
{
- struct per_cpu_pages *pcp;
int migratetype;
- memset(p, 0, sizeof(*p));
+ memset(pcp, 0, sizeof(*pcp));
+ memset(pzstats, 0, sizeof(*pzstats));
- pcp = &p->pcp;
for (migratetype = 0; migratetype < MIGRATE_PCPTYPES; migratetype++)
INIT_LIST_HEAD(&pcp->lists[migratetype]);
@@ -6674,12 +6673,12 @@ static void pageset_init(struct per_cpu_
static void __zone_set_pageset_high_and_batch(struct zone *zone, unsigned long high,
unsigned long batch)
{
- struct per_cpu_pageset *p;
+ struct per_cpu_pages *pcp;
int cpu;
for_each_possible_cpu(cpu) {
- p = per_cpu_ptr(zone->pageset, cpu);
- pageset_update(&p->pcp, high, batch);
+ pcp = per_cpu_ptr(zone->per_cpu_pageset, cpu);
+ pageset_update(pcp, high, batch);
}
}
@@ -6714,13 +6713,20 @@ static void zone_set_pageset_high_and_ba
void __meminit setup_zone_pageset(struct zone *zone)
{
- struct per_cpu_pageset *p;
int cpu;
- zone->pageset = alloc_percpu(struct per_cpu_pageset);
+ /* Size may be 0 on !SMP && !NUMA */
+ if (sizeof(struct per_cpu_zonestat) > 0)
+ zone->per_cpu_zonestats = alloc_percpu(struct per_cpu_zonestat);
+
+ zone->per_cpu_pageset = alloc_percpu(struct per_cpu_pages);
for_each_possible_cpu(cpu) {
- p = per_cpu_ptr(zone->pageset, cpu);
- pageset_init(p);
+ struct per_cpu_pages *pcp;
+ struct per_cpu_zonestat *pzstats;
+
+ pcp = per_cpu_ptr(zone->per_cpu_pageset, cpu);
+ pzstats = per_cpu_ptr(zone->per_cpu_zonestats, cpu);
+ per_cpu_pages_init(pcp, pzstats);
}
zone_set_pageset_high_and_batch(zone);
@@ -6747,9 +6753,9 @@ void __init setup_per_cpu_pageset(void)
* the nodes these zones are associated with.
*/
for_each_possible_cpu(cpu) {
- struct per_cpu_pageset *pcp = &per_cpu(boot_pageset, cpu);
- memset(pcp->vm_numa_stat_diff, 0,
- sizeof(pcp->vm_numa_stat_diff));
+ struct per_cpu_zonestat *pzstats = &per_cpu(boot_zonestats, cpu);
+ memset(pzstats->vm_numa_stat_diff, 0,
+ sizeof(pzstats->vm_numa_stat_diff));
}
#endif
@@ -6765,7 +6771,8 @@ static __meminit void zone_pcp_init(stru
* relies on the ability of the linker to provide the
* offset of a (static) per cpu variable into the per cpu area.
*/
- zone->pageset = &boot_pageset;
+ zone->per_cpu_pageset = &boot_pageset;
+ zone->per_cpu_zonestats = &boot_zonestats;
zone->pageset_high = BOOT_PAGESET_HIGH;
zone->pageset_batch = BOOT_PAGESET_BATCH;
@@ -9046,15 +9053,17 @@ void zone_pcp_enable(struct zone *zone)
void zone_pcp_reset(struct zone *zone)
{
int cpu;
- struct per_cpu_pageset *pset;
+ struct per_cpu_zonestat *pzstats;
- if (zone->pageset != &boot_pageset) {
+ if (zone->per_cpu_pageset != &boot_pageset) {
for_each_online_cpu(cpu) {
- pset = per_cpu_ptr(zone->pageset, cpu);
- drain_zonestat(zone, pset);
+ pzstats = per_cpu_ptr(zone->per_cpu_zonestats, cpu);
+ drain_zonestat(zone, pzstats);
}
- free_percpu(zone->pageset);
- zone->pageset = &boot_pageset;
+ free_percpu(zone->per_cpu_pageset);
+ free_percpu(zone->per_cpu_zonestats);
+ zone->per_cpu_pageset = &boot_pageset;
+ zone->per_cpu_zonestats = &boot_zonestats;
}
}
--- a/mm/vmstat.c~mm-page_alloc-split-per-cpu-page-lists-and-zone-stats
+++ a/mm/vmstat.c
@@ -44,7 +44,7 @@ static void zero_zone_numa_counters(stru
for (item = 0; item < NR_VM_NUMA_STAT_ITEMS; item++) {
atomic_long_set(&zone->vm_numa_stat[item], 0);
for_each_online_cpu(cpu)
- per_cpu_ptr(zone->pageset, cpu)->vm_numa_stat_diff[item]
+ per_cpu_ptr(zone->per_cpu_zonestats, cpu)->vm_numa_stat_diff[item]
= 0;
}
}
@@ -266,7 +266,7 @@ void refresh_zone_stat_thresholds(void)
for_each_online_cpu(cpu) {
int pgdat_threshold;
- per_cpu_ptr(zone->pageset, cpu)->stat_threshold
+ per_cpu_ptr(zone->per_cpu_zonestats, cpu)->stat_threshold
= threshold;
/* Base nodestat threshold on the largest populated zone. */
@@ -303,7 +303,7 @@ void set_pgdat_percpu_threshold(pg_data_
threshold = (*calculate_pressure)(zone);
for_each_online_cpu(cpu)
- per_cpu_ptr(zone->pageset, cpu)->stat_threshold
+ per_cpu_ptr(zone->per_cpu_zonestats, cpu)->stat_threshold
= threshold;
}
}
@@ -316,7 +316,7 @@ void set_pgdat_percpu_threshold(pg_data_
void __mod_zone_page_state(struct zone *zone, enum zone_stat_item item,
long delta)
{
- struct per_cpu_pageset __percpu *pcp = zone->pageset;
+ struct per_cpu_zonestat __percpu *pcp = zone->per_cpu_zonestats;
s8 __percpu *p = pcp->vm_stat_diff + item;
long x;
long t;
@@ -389,7 +389,7 @@ EXPORT_SYMBOL(__mod_node_page_state);
*/
void __inc_zone_state(struct zone *zone, enum zone_stat_item item)
{
- struct per_cpu_pageset __percpu *pcp = zone->pageset;
+ struct per_cpu_zonestat __percpu *pcp = zone->per_cpu_zonestats;
s8 __percpu *p = pcp->vm_stat_diff + item;
s8 v, t;
@@ -435,7 +435,7 @@ EXPORT_SYMBOL(__inc_node_page_state);
void __dec_zone_state(struct zone *zone, enum zone_stat_item item)
{
- struct per_cpu_pageset __percpu *pcp = zone->pageset;
+ struct per_cpu_zonestat __percpu *pcp = zone->per_cpu_zonestats;
s8 __percpu *p = pcp->vm_stat_diff + item;
s8 v, t;
@@ -495,7 +495,7 @@ EXPORT_SYMBOL(__dec_node_page_state);
static inline void mod_zone_state(struct zone *zone,
enum zone_stat_item item, long delta, int overstep_mode)
{
- struct per_cpu_pageset __percpu *pcp = zone->pageset;
+ struct per_cpu_zonestat __percpu *pcp = zone->per_cpu_zonestats;
s8 __percpu *p = pcp->vm_stat_diff + item;
long o, n, t, z;
@@ -781,19 +781,22 @@ static int refresh_cpu_vm_stats(bool do_
int changes = 0;
for_each_populated_zone(zone) {
- struct per_cpu_pageset __percpu *p = zone->pageset;
+ struct per_cpu_zonestat __percpu *pzstats = zone->per_cpu_zonestats;
+#ifdef CONFIG_NUMA
+ struct per_cpu_pages __percpu *pcp = zone->per_cpu_pageset;
+#endif
for (i = 0; i < NR_VM_ZONE_STAT_ITEMS; i++) {
int v;
- v = this_cpu_xchg(p->vm_stat_diff[i], 0);
+ v = this_cpu_xchg(pzstats->vm_stat_diff[i], 0);
if (v) {
atomic_long_add(v, &zone->vm_stat[i]);
global_zone_diff[i] += v;
#ifdef CONFIG_NUMA
/* 3 seconds idle till flush */
- __this_cpu_write(p->expire, 3);
+ __this_cpu_write(pcp->expire, 3);
#endif
}
}
@@ -801,12 +804,12 @@ static int refresh_cpu_vm_stats(bool do_
for (i = 0; i < NR_VM_NUMA_STAT_ITEMS; i++) {
int v;
- v = this_cpu_xchg(p->vm_numa_stat_diff[i], 0);
+ v = this_cpu_xchg(pzstats->vm_numa_stat_diff[i], 0);
if (v) {
atomic_long_add(v, &zone->vm_numa_stat[i]);
global_numa_diff[i] += v;
- __this_cpu_write(p->expire, 3);
+ __this_cpu_write(pcp->expire, 3);
}
}
@@ -819,23 +822,23 @@ static int refresh_cpu_vm_stats(bool do_
* Check if there are pages remaining in this pageset
* if not then there is nothing to expire.
*/
- if (!__this_cpu_read(p->expire) ||
- !__this_cpu_read(p->pcp.count))
+ if (!__this_cpu_read(pcp->expire) ||
+ !__this_cpu_read(pcp->count))
continue;
/*
* We never drain zones local to this processor.
*/
if (zone_to_nid(zone) == numa_node_id()) {
- __this_cpu_write(p->expire, 0);
+ __this_cpu_write(pcp->expire, 0);
continue;
}
- if (__this_cpu_dec_return(p->expire))
+ if (__this_cpu_dec_return(pcp->expire))
continue;
- if (__this_cpu_read(p->pcp.count)) {
- drain_zone_pages(zone, this_cpu_ptr(&p->pcp));
+ if (__this_cpu_read(pcp->count)) {
+ drain_zone_pages(zone, this_cpu_ptr(pcp));
changes++;
}
}
@@ -882,27 +885,27 @@ void cpu_vm_stats_fold(int cpu)
int global_node_diff[NR_VM_NODE_STAT_ITEMS] = { 0, };
for_each_populated_zone(zone) {
- struct per_cpu_pageset *p;
+ struct per_cpu_zonestat *pzstats;
- p = per_cpu_ptr(zone->pageset, cpu);
+ pzstats = per_cpu_ptr(zone->per_cpu_zonestats, cpu);
for (i = 0; i < NR_VM_ZONE_STAT_ITEMS; i++)
- if (p->vm_stat_diff[i]) {
+ if (pzstats->vm_stat_diff[i]) {
int v;
- v = p->vm_stat_diff[i];
- p->vm_stat_diff[i] = 0;
+ v = pzstats->vm_stat_diff[i];
+ pzstats->vm_stat_diff[i] = 0;
atomic_long_add(v, &zone->vm_stat[i]);
global_zone_diff[i] += v;
}
#ifdef CONFIG_NUMA
for (i = 0; i < NR_VM_NUMA_STAT_ITEMS; i++)
- if (p->vm_numa_stat_diff[i]) {
+ if (pzstats->vm_numa_stat_diff[i]) {
int v;
- v = p->vm_numa_stat_diff[i];
- p->vm_numa_stat_diff[i] = 0;
+ v = pzstats->vm_numa_stat_diff[i];
+ pzstats->vm_numa_stat_diff[i] = 0;
atomic_long_add(v, &zone->vm_numa_stat[i]);
global_numa_diff[i] += v;
}
@@ -936,24 +939,24 @@ void cpu_vm_stats_fold(int cpu)
* this is only called if !populated_zone(zone), which implies no other users of
* pset->vm_stat_diff[] exist.
*/
-void drain_zonestat(struct zone *zone, struct per_cpu_pageset *pset)
+void drain_zonestat(struct zone *zone, struct per_cpu_zonestat *pzstats)
{
int i;
for (i = 0; i < NR_VM_ZONE_STAT_ITEMS; i++)
- if (pset->vm_stat_diff[i]) {
- int v = pset->vm_stat_diff[i];
- pset->vm_stat_diff[i] = 0;
+ if (pzstats->vm_stat_diff[i]) {
+ int v = pzstats->vm_stat_diff[i];
+ pzstats->vm_stat_diff[i] = 0;
atomic_long_add(v, &zone->vm_stat[i]);
atomic_long_add(v, &vm_zone_stat[i]);
}
#ifdef CONFIG_NUMA
for (i = 0; i < NR_VM_NUMA_STAT_ITEMS; i++)
- if (pset->vm_numa_stat_diff[i]) {
- int v = pset->vm_numa_stat_diff[i];
+ if (pzstats->vm_numa_stat_diff[i]) {
+ int v = pzstats->vm_numa_stat_diff[i];
- pset->vm_numa_stat_diff[i] = 0;
+ pzstats->vm_numa_stat_diff[i] = 0;
atomic_long_add(v, &zone->vm_numa_stat[i]);
atomic_long_add(v, &vm_numa_stat[i]);
}
@@ -965,8 +968,8 @@ void drain_zonestat(struct zone *zone, s
void __inc_numa_state(struct zone *zone,
enum numa_stat_item item)
{
- struct per_cpu_pageset __percpu *pcp = zone->pageset;
- u16 __percpu *p = pcp->vm_numa_stat_diff + item;
+ struct per_cpu_zonestat __percpu *pzstats = zone->per_cpu_zonestats;
+ u16 __percpu *p = pzstats->vm_numa_stat_diff + item;
u16 v;
v = __this_cpu_inc_return(*p);
@@ -1693,21 +1696,23 @@ static void zoneinfo_show_print(struct s
seq_printf(m, "\n pagesets");
for_each_online_cpu(i) {
- struct per_cpu_pageset *pageset;
+ struct per_cpu_pages *pcp;
+ struct per_cpu_zonestat __maybe_unused *pzstats;
- pageset = per_cpu_ptr(zone->pageset, i);
+ pcp = per_cpu_ptr(zone->per_cpu_pageset, i);
seq_printf(m,
"\n cpu: %i"
"\n count: %i"
"\n high: %i"
"\n batch: %i",
i,
- pageset->pcp.count,
- pageset->pcp.high,
- pageset->pcp.batch);
+ pcp->count,
+ pcp->high,
+ pcp->batch);
#ifdef CONFIG_SMP
+ pzstats = per_cpu_ptr(zone->per_cpu_zonestats, i);
seq_printf(m, "\n vm stats threshold: %d",
- pageset->stat_threshold);
+ pzstats->stat_threshold);
#endif
}
seq_printf(m,
@@ -1927,17 +1932,18 @@ static bool need_update(int cpu)
struct zone *zone;
for_each_populated_zone(zone) {
- struct per_cpu_pageset *p = per_cpu_ptr(zone->pageset, cpu);
+ struct per_cpu_zonestat *pzstats = per_cpu_ptr(zone->per_cpu_zonestats, cpu);
struct per_cpu_nodestat *n;
+
/*
* The fast way of checking if there are any vmstat diffs.
*/
- if (memchr_inv(p->vm_stat_diff, 0, NR_VM_ZONE_STAT_ITEMS *
- sizeof(p->vm_stat_diff[0])))
+ if (memchr_inv(pzstats->vm_stat_diff, 0, NR_VM_ZONE_STAT_ITEMS *
+ sizeof(pzstats->vm_stat_diff[0])))
return true;
#ifdef CONFIG_NUMA
- if (memchr_inv(p->vm_numa_stat_diff, 0, NR_VM_NUMA_STAT_ITEMS *
- sizeof(p->vm_numa_stat_diff[0])))
+ if (memchr_inv(pzstats->vm_numa_stat_diff, 0, NR_VM_NUMA_STAT_ITEMS *
+ sizeof(pzstats->vm_numa_stat_diff[0])))
return true;
#endif
if (last_pgdat == zone->zone_pgdat)
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 161/192] mm/page_alloc: convert per-cpu list protection to local_lock
2021-06-29 2:32 incoming Andrew Morton
` (159 preceding siblings ...)
2021-06-29 2:41 ` [patch 160/192] mm/page_alloc: split per cpu page lists and zone stats Andrew Morton
@ 2021-06-29 2:41 ` Andrew Morton
2021-06-29 2:41 ` [patch 162/192] mm/vmstat: convert NUMA statistics to basic NUMA counters Andrew Morton
` (30 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:41 UTC (permalink / raw)
To: akpm, bigeasy, brouer, chuck.lever, linux-mm, mgorman, mhocko,
mingo, mm-commits, peterz, tglx, torvalds, vbabka
From: Mel Gorman <mgorman@techsingularity.net>
Subject: mm/page_alloc: convert per-cpu list protection to local_lock
There is a lack of clarity of what exactly
local_irq_save/local_irq_restore protects in page_alloc.c . It conflates
the protection of per-cpu page allocation structures with per-cpu vmstat
deltas.
This patch protects the PCP structure using local_lock which for most
configurations is identical to IRQ enabling/disabling. The scope of the
lock is still wider than it should be but this is decreased later.
It is possible for the local_lock to be embedded safely within struct
per_cpu_pages but it adds complexity to free_unref_page_list.
[akpm@linux-foundation.org: coding style fixes]
[mgorman@techsingularity.net: work around a pahole limitation with zero-sized struct pagesets]
Link: https://lkml.kernel.org/r/20210526080741.GW30378@techsingularity.net
[lkp@intel.com: Make pagesets static]
Link: https://lkml.kernel.org/r/20210512095458.30632-3-mgorman@techsingularity.net
Signed-off-by: Mel Gorman <mgorman@techsingularity.net>
Acked-by: Vlastimil Babka <vbabka@suse.cz>
Acked-by: Peter Zijlstra (Intel) <peterz@infradead.org>
Cc: Chuck Lever <chuck.lever@oracle.com>
Cc: Ingo Molnar <mingo@kernel.org>
Cc: Jesper Dangaard Brouer <brouer@redhat.com>
Cc: Michal Hocko <mhocko@kernel.org>
Cc: Sebastian Andrzej Siewior <bigeasy@linutronix.de>
Cc: Thomas Gleixner <tglx@linutronix.de>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/linux/mmzone.h | 2 +
lib/Kconfig.debug | 3 +
mm/page_alloc.c | 61 +++++++++++++++++++++++++++++----------
3 files changed, 51 insertions(+), 15 deletions(-)
--- a/include/linux/mmzone.h~mm-page_alloc-convert-per-cpu-list-protection-to-local_lock
+++ a/include/linux/mmzone.h
@@ -20,6 +20,7 @@
#include <linux/atomic.h>
#include <linux/mm_types.h>
#include <linux/page-flags.h>
+#include <linux/local_lock.h>
#include <asm/page.h>
/* Free memory management - zoned buddy allocator. */
@@ -337,6 +338,7 @@ enum zone_watermarks {
#define high_wmark_pages(z) (z->_watermark[WMARK_HIGH] + z->watermark_boost)
#define wmark_pages(z, i) (z->_watermark[i] + z->watermark_boost)
+/* Fields and list protected by pagesets local_lock in page_alloc.c */
struct per_cpu_pages {
int count; /* number of pages in the list */
int high; /* high watermark, emptying needed */
--- a/lib/Kconfig.debug~mm-page_alloc-convert-per-cpu-list-protection-to-local_lock
+++ a/lib/Kconfig.debug
@@ -313,6 +313,9 @@ config DEBUG_INFO_BTF
config PAHOLE_HAS_SPLIT_BTF
def_bool $(success, test `$(PAHOLE) --version | sed -E 's/v([0-9]+)\.([0-9]+)/\1\2/'` -ge "119")
+config PAHOLE_HAS_ZEROSIZE_PERCPU_SUPPORT
+ def_bool $(success, test `$(PAHOLE) --version | sed -E 's/v([0-9]+)\.([0-9]+)/\1\2/'` -ge "122")
+
config DEBUG_INFO_BTF_MODULES
def_bool y
depends on DEBUG_INFO_BTF && MODULES && PAHOLE_HAS_SPLIT_BTF
--- a/mm/page_alloc.c~mm-page_alloc-convert-per-cpu-list-protection-to-local_lock
+++ a/mm/page_alloc.c
@@ -122,6 +122,24 @@ typedef int __bitwise fpi_t;
static DEFINE_MUTEX(pcp_batch_high_lock);
#define MIN_PERCPU_PAGELIST_FRACTION (8)
+struct pagesets {
+ local_lock_t lock;
+#if defined(CONFIG_DEBUG_INFO_BTF) && \
+ !defined(CONFIG_DEBUG_LOCK_ALLOC) && \
+ !defined(CONFIG_PAHOLE_HAS_ZEROSIZE_PERCPU_SUPPORT)
+ /*
+ * pahole 1.21 and earlier gets confused by zero-sized per-CPU
+ * variables and produces invalid BTF. Ensure that
+ * sizeof(struct pagesets) != 0 for older versions of pahole.
+ */
+ char __pahole_hack;
+ #warning "pahole too old to support zero-sized struct pagesets"
+#endif
+};
+static DEFINE_PER_CPU(struct pagesets, pagesets) = {
+ .lock = INIT_LOCAL_LOCK(lock),
+};
+
#ifdef CONFIG_USE_PERCPU_NUMA_NODE_ID
DEFINE_PER_CPU(int, numa_node);
EXPORT_PER_CPU_SYMBOL(numa_node);
@@ -1453,6 +1471,10 @@ static void free_pcppages_bulk(struct zo
} while (--count && --batch_free && !list_empty(list));
}
+ /*
+ * local_lock_irq held so equivalent to spin_lock_irqsave for
+ * both PREEMPT_RT and non-PREEMPT_RT configurations.
+ */
spin_lock(&zone->lock);
isolated_pageblocks = has_isolate_pageblock(zone);
@@ -1573,6 +1595,11 @@ static void __free_pages_ok(struct page
return;
migratetype = get_pfnblock_migratetype(page, pfn);
+
+ /*
+ * TODO FIX: Disable IRQs before acquiring IRQ-safe zone->lock
+ * and protect vmstat updates.
+ */
local_irq_save(flags);
__count_vm_events(PGFREE, 1 << order);
free_one_page(page_zone(page), page, pfn, order, migratetype,
@@ -2955,6 +2982,10 @@ static int rmqueue_bulk(struct zone *zon
{
int i, allocated = 0;
+ /*
+ * local_lock_irq held so equivalent to spin_lock_irqsave for
+ * both PREEMPT_RT and non-PREEMPT_RT configurations.
+ */
spin_lock(&zone->lock);
for (i = 0; i < count; ++i) {
struct page *page = __rmqueue(zone, order, migratetype,
@@ -3007,12 +3038,12 @@ void drain_zone_pages(struct zone *zone,
unsigned long flags;
int to_drain, batch;
- local_irq_save(flags);
+ local_lock_irqsave(&pagesets.lock, flags);
batch = READ_ONCE(pcp->batch);
to_drain = min(pcp->count, batch);
if (to_drain > 0)
free_pcppages_bulk(zone, to_drain, pcp);
- local_irq_restore(flags);
+ local_unlock_irqrestore(&pagesets.lock, flags);
}
#endif
@@ -3028,13 +3059,13 @@ static void drain_pages_zone(unsigned in
unsigned long flags;
struct per_cpu_pages *pcp;
- local_irq_save(flags);
+ local_lock_irqsave(&pagesets.lock, flags);
pcp = per_cpu_ptr(zone->per_cpu_pageset, cpu);
if (pcp->count)
free_pcppages_bulk(zone, pcp->count, pcp);
- local_irq_restore(flags);
+ local_unlock_irqrestore(&pagesets.lock, flags);
}
/*
@@ -3297,9 +3328,9 @@ void free_unref_page(struct page *page)
if (!free_unref_page_prepare(page, pfn))
return;
- local_irq_save(flags);
+ local_lock_irqsave(&pagesets.lock, flags);
free_unref_page_commit(page, pfn);
- local_irq_restore(flags);
+ local_unlock_irqrestore(&pagesets.lock, flags);
}
/*
@@ -3319,7 +3350,7 @@ void free_unref_page_list(struct list_he
set_page_private(page, pfn);
}
- local_irq_save(flags);
+ local_lock_irqsave(&pagesets.lock, flags);
list_for_each_entry_safe(page, next, list, lru) {
unsigned long pfn = page_private(page);
@@ -3332,12 +3363,12 @@ void free_unref_page_list(struct list_he
* a large list of pages to free.
*/
if (++batch_count == SWAP_CLUSTER_MAX) {
- local_irq_restore(flags);
+ local_unlock_irqrestore(&pagesets.lock, flags);
batch_count = 0;
- local_irq_save(flags);
+ local_lock_irqsave(&pagesets.lock, flags);
}
}
- local_irq_restore(flags);
+ local_unlock_irqrestore(&pagesets.lock, flags);
}
/*
@@ -3494,7 +3525,7 @@ static struct page *rmqueue_pcplist(stru
struct page *page;
unsigned long flags;
- local_irq_save(flags);
+ local_lock_irqsave(&pagesets.lock, flags);
pcp = this_cpu_ptr(zone->per_cpu_pageset);
list = &pcp->lists[migratetype];
page = __rmqueue_pcplist(zone, migratetype, alloc_flags, pcp, list);
@@ -3502,7 +3533,7 @@ static struct page *rmqueue_pcplist(stru
__count_zid_vm_events(PGALLOC, page_zonenum(page), 1);
zone_statistics(preferred_zone, zone);
}
- local_irq_restore(flags);
+ local_unlock_irqrestore(&pagesets.lock, flags);
return page;
}
@@ -5103,7 +5134,7 @@ unsigned long __alloc_pages_bulk(gfp_t g
goto failed;
/* Attempt the batch allocation */
- local_irq_save(flags);
+ local_lock_irqsave(&pagesets.lock, flags);
pcp = this_cpu_ptr(zone->per_cpu_pageset);
pcp_list = &pcp->lists[ac.migratetype];
@@ -5141,12 +5172,12 @@ unsigned long __alloc_pages_bulk(gfp_t g
nr_populated++;
}
- local_irq_restore(flags);
+ local_unlock_irqrestore(&pagesets.lock, flags);
return nr_populated;
failed_irq:
- local_irq_restore(flags);
+ local_unlock_irqrestore(&pagesets.lock, flags);
failed:
page = __alloc_pages(gfp, 0, preferred_nid, nodemask);
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 162/192] mm/vmstat: convert NUMA statistics to basic NUMA counters
2021-06-29 2:32 incoming Andrew Morton
` (160 preceding siblings ...)
2021-06-29 2:41 ` [patch 161/192] mm/page_alloc: convert per-cpu list protection to local_lock Andrew Morton
@ 2021-06-29 2:41 ` Andrew Morton
2021-06-29 2:41 ` [patch 163/192] mm/vmstat: inline NUMA event counter updates Andrew Morton
` (29 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:41 UTC (permalink / raw)
To: akpm, bigeasy, brouer, chuck.lever, linux-mm, mgorman, mhocko,
mingo, mm-commits, peterz, tglx, torvalds, vbabka
From: Mel Gorman <mgorman@techsingularity.net>
Subject: mm/vmstat: convert NUMA statistics to basic NUMA counters
NUMA statistics are maintained on the zone level for hits, misses, foreign
etc but nothing relies on them being perfectly accurate for functional
correctness. The counters are used by userspace to get a general overview
of a workloads NUMA behaviour but the page allocator incurs a high cost to
maintain perfect accuracy similar to what is required for a vmstat like
NR_FREE_PAGES. There even is a sysctl vm.numa_stat to allow userspace to
turn off the collection of NUMA statistics like NUMA_HIT.
This patch converts NUMA_HIT and friends to be NUMA events with similar
accuracy to VM events. There is a possibility that slight errors will be
introduced but the overall trend as seen by userspace will be similar.
The counters are no longer updated from vmstat_refresh context as it is
unnecessary overhead for counters that may never be read by userspace.
Note that counters could be maintained at the node level to save space but
it would have a user-visible impact due to /proc/zoneinfo.
[lkp@intel.com: Fix misplaced closing brace for !CONFIG_NUMA]
Link: https://lkml.kernel.org/r/20210512095458.30632-4-mgorman@techsingularity.net
Signed-off-by: Mel Gorman <mgorman@techsingularity.net>
Acked-by: Vlastimil Babka <vbabka@suse.cz>
Acked-by: Peter Zijlstra (Intel) <peterz@infradead.org>
Cc: Chuck Lever <chuck.lever@oracle.com>
Cc: Ingo Molnar <mingo@kernel.org>
Cc: Jesper Dangaard Brouer <brouer@redhat.com>
Cc: Michal Hocko <mhocko@kernel.org>
Cc: Sebastian Andrzej Siewior <bigeasy@linutronix.de>
Cc: Thomas Gleixner <tglx@linutronix.de>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
drivers/base/node.c | 18 ++--
include/linux/mmzone.h | 13 ++
include/linux/vmstat.h | 43 ++++-----
mm/mempolicy.c | 2
mm/page_alloc.c | 12 +-
mm/vmstat.c | 175 +++++++++++++++------------------------
6 files changed, 114 insertions(+), 149 deletions(-)
--- a/drivers/base/node.c~mm-vmstat-convert-numa-statistics-to-basic-numa-counters
+++ a/drivers/base/node.c
@@ -482,6 +482,7 @@ static DEVICE_ATTR(meminfo, 0444, node_r
static ssize_t node_read_numastat(struct device *dev,
struct device_attribute *attr, char *buf)
{
+ fold_vm_numa_events();
return sysfs_emit(buf,
"numa_hit %lu\n"
"numa_miss %lu\n"
@@ -489,12 +490,12 @@ static ssize_t node_read_numastat(struct
"interleave_hit %lu\n"
"local_node %lu\n"
"other_node %lu\n",
- sum_zone_numa_state(dev->id, NUMA_HIT),
- sum_zone_numa_state(dev->id, NUMA_MISS),
- sum_zone_numa_state(dev->id, NUMA_FOREIGN),
- sum_zone_numa_state(dev->id, NUMA_INTERLEAVE_HIT),
- sum_zone_numa_state(dev->id, NUMA_LOCAL),
- sum_zone_numa_state(dev->id, NUMA_OTHER));
+ sum_zone_numa_event_state(dev->id, NUMA_HIT),
+ sum_zone_numa_event_state(dev->id, NUMA_MISS),
+ sum_zone_numa_event_state(dev->id, NUMA_FOREIGN),
+ sum_zone_numa_event_state(dev->id, NUMA_INTERLEAVE_HIT),
+ sum_zone_numa_event_state(dev->id, NUMA_LOCAL),
+ sum_zone_numa_event_state(dev->id, NUMA_OTHER));
}
static DEVICE_ATTR(numastat, 0444, node_read_numastat, NULL);
@@ -512,10 +513,11 @@ static ssize_t node_read_vmstat(struct d
sum_zone_node_page_state(nid, i));
#ifdef CONFIG_NUMA
- for (i = 0; i < NR_VM_NUMA_STAT_ITEMS; i++)
+ fold_vm_numa_events();
+ for (i = 0; i < NR_VM_NUMA_EVENT_ITEMS; i++)
len += sysfs_emit_at(buf, len, "%s %lu\n",
numa_stat_name(i),
- sum_zone_numa_state(nid, i));
+ sum_zone_numa_event_state(nid, i));
#endif
for (i = 0; i < NR_VM_NODE_STAT_ITEMS; i++) {
--- a/include/linux/mmzone.h~mm-vmstat-convert-numa-statistics-to-basic-numa-counters
+++ a/include/linux/mmzone.h
@@ -135,10 +135,10 @@ enum numa_stat_item {
NUMA_INTERLEAVE_HIT, /* interleaver preferred this zone */
NUMA_LOCAL, /* allocation from local node */
NUMA_OTHER, /* allocation from other node */
- NR_VM_NUMA_STAT_ITEMS
+ NR_VM_NUMA_EVENT_ITEMS
};
#else
-#define NR_VM_NUMA_STAT_ITEMS 0
+#define NR_VM_NUMA_EVENT_ITEMS 0
#endif
enum zone_stat_item {
@@ -357,7 +357,12 @@ struct per_cpu_zonestat {
s8 stat_threshold;
#endif
#ifdef CONFIG_NUMA
- u16 vm_numa_stat_diff[NR_VM_NUMA_STAT_ITEMS];
+ /*
+ * Low priority inaccurate counters that are only folded
+ * on demand. Use a large type to avoid the overhead of
+ * folding during refresh_cpu_vm_stats.
+ */
+ unsigned long vm_numa_event[NR_VM_NUMA_EVENT_ITEMS];
#endif
};
@@ -623,7 +628,7 @@ struct zone {
ZONE_PADDING(_pad3_)
/* Zone statistics */
atomic_long_t vm_stat[NR_VM_ZONE_STAT_ITEMS];
- atomic_long_t vm_numa_stat[NR_VM_NUMA_STAT_ITEMS];
+ atomic_long_t vm_numa_event[NR_VM_NUMA_EVENT_ITEMS];
} ____cacheline_internodealigned_in_smp;
enum pgdat_flags {
--- a/include/linux/vmstat.h~mm-vmstat-convert-numa-statistics-to-basic-numa-counters
+++ a/include/linux/vmstat.h
@@ -138,34 +138,27 @@ static inline void vm_events_fold_cpu(in
* Zone and node-based page accounting with per cpu differentials.
*/
extern atomic_long_t vm_zone_stat[NR_VM_ZONE_STAT_ITEMS];
-extern atomic_long_t vm_numa_stat[NR_VM_NUMA_STAT_ITEMS];
extern atomic_long_t vm_node_stat[NR_VM_NODE_STAT_ITEMS];
+extern atomic_long_t vm_numa_event[NR_VM_NUMA_EVENT_ITEMS];
#ifdef CONFIG_NUMA
-static inline void zone_numa_state_add(long x, struct zone *zone,
- enum numa_stat_item item)
+static inline void zone_numa_event_add(long x, struct zone *zone,
+ enum numa_stat_item item)
{
- atomic_long_add(x, &zone->vm_numa_stat[item]);
- atomic_long_add(x, &vm_numa_stat[item]);
+ atomic_long_add(x, &zone->vm_numa_event[item]);
+ atomic_long_add(x, &vm_numa_event[item]);
}
-static inline unsigned long global_numa_state(enum numa_stat_item item)
+static inline unsigned long zone_numa_event_state(struct zone *zone,
+ enum numa_stat_item item)
{
- long x = atomic_long_read(&vm_numa_stat[item]);
-
- return x;
+ return atomic_long_read(&zone->vm_numa_event[item]);
}
-static inline unsigned long zone_numa_state_snapshot(struct zone *zone,
- enum numa_stat_item item)
+static inline unsigned long
+global_numa_event_state(enum numa_stat_item item)
{
- long x = atomic_long_read(&zone->vm_numa_stat[item]);
- int cpu;
-
- for_each_online_cpu(cpu)
- x += per_cpu_ptr(zone->per_cpu_zonestats, cpu)->vm_numa_stat_diff[item];
-
- return x;
+ return atomic_long_read(&vm_numa_event[item]);
}
#endif /* CONFIG_NUMA */
@@ -245,18 +238,22 @@ static inline unsigned long zone_page_st
}
#ifdef CONFIG_NUMA
-extern void __inc_numa_state(struct zone *zone, enum numa_stat_item item);
+extern void __count_numa_event(struct zone *zone, enum numa_stat_item item);
extern unsigned long sum_zone_node_page_state(int node,
enum zone_stat_item item);
-extern unsigned long sum_zone_numa_state(int node, enum numa_stat_item item);
+extern unsigned long sum_zone_numa_event_state(int node, enum numa_stat_item item);
extern unsigned long node_page_state(struct pglist_data *pgdat,
enum node_stat_item item);
extern unsigned long node_page_state_pages(struct pglist_data *pgdat,
enum node_stat_item item);
+extern void fold_vm_numa_events(void);
#else
#define sum_zone_node_page_state(node, item) global_zone_page_state(item)
#define node_page_state(node, item) global_node_page_state(item)
#define node_page_state_pages(node, item) global_node_page_state_pages(item)
+static inline void fold_vm_numa_events(void)
+{
+}
#endif /* CONFIG_NUMA */
#ifdef CONFIG_SMP
@@ -428,7 +425,7 @@ static inline const char *numa_stat_name
static inline const char *node_stat_name(enum node_stat_item item)
{
return vmstat_text[NR_VM_ZONE_STAT_ITEMS +
- NR_VM_NUMA_STAT_ITEMS +
+ NR_VM_NUMA_EVENT_ITEMS +
item];
}
@@ -440,7 +437,7 @@ static inline const char *lru_list_name(
static inline const char *writeback_stat_name(enum writeback_stat_item item)
{
return vmstat_text[NR_VM_ZONE_STAT_ITEMS +
- NR_VM_NUMA_STAT_ITEMS +
+ NR_VM_NUMA_EVENT_ITEMS +
NR_VM_NODE_STAT_ITEMS +
item];
}
@@ -449,7 +446,7 @@ static inline const char *writeback_stat
static inline const char *vm_event_name(enum vm_event_item item)
{
return vmstat_text[NR_VM_ZONE_STAT_ITEMS +
- NR_VM_NUMA_STAT_ITEMS +
+ NR_VM_NUMA_EVENT_ITEMS +
NR_VM_NODE_STAT_ITEMS +
NR_VM_WRITEBACK_STAT_ITEMS +
item];
--- a/mm/mempolicy.c~mm-vmstat-convert-numa-statistics-to-basic-numa-counters
+++ a/mm/mempolicy.c
@@ -2150,7 +2150,7 @@ static struct page *alloc_page_interleav
return page;
if (page && page_to_nid(page) == nid) {
preempt_disable();
- __inc_numa_state(page_zone(page), NUMA_INTERLEAVE_HIT);
+ __count_numa_event(page_zone(page), NUMA_INTERLEAVE_HIT);
preempt_enable();
}
return page;
--- a/mm/page_alloc.c~mm-vmstat-convert-numa-statistics-to-basic-numa-counters
+++ a/mm/page_alloc.c
@@ -3480,12 +3480,12 @@ static inline void zone_statistics(struc
local_stat = NUMA_OTHER;
if (zone_to_nid(z) == zone_to_nid(preferred_zone))
- __inc_numa_state(z, NUMA_HIT);
+ __count_numa_event(z, NUMA_HIT);
else {
- __inc_numa_state(z, NUMA_MISS);
- __inc_numa_state(preferred_zone, NUMA_FOREIGN);
+ __count_numa_event(z, NUMA_MISS);
+ __count_numa_event(preferred_zone, NUMA_FOREIGN);
}
- __inc_numa_state(z, local_stat);
+ __count_numa_event(z, local_stat);
#endif
}
@@ -6785,8 +6785,8 @@ void __init setup_per_cpu_pageset(void)
*/
for_each_possible_cpu(cpu) {
struct per_cpu_zonestat *pzstats = &per_cpu(boot_zonestats, cpu);
- memset(pzstats->vm_numa_stat_diff, 0,
- sizeof(pzstats->vm_numa_stat_diff));
+ memset(pzstats->vm_numa_event, 0,
+ sizeof(pzstats->vm_numa_event));
}
#endif
--- a/mm/vmstat.c~mm-vmstat-convert-numa-statistics-to-basic-numa-counters
+++ a/mm/vmstat.c
@@ -31,8 +31,6 @@
#include "internal.h"
-#define NUMA_STATS_THRESHOLD (U16_MAX - 2)
-
#ifdef CONFIG_NUMA
int sysctl_vm_numa_stat = ENABLE_NUMA_STAT;
@@ -41,11 +39,12 @@ static void zero_zone_numa_counters(stru
{
int item, cpu;
- for (item = 0; item < NR_VM_NUMA_STAT_ITEMS; item++) {
- atomic_long_set(&zone->vm_numa_stat[item], 0);
- for_each_online_cpu(cpu)
- per_cpu_ptr(zone->per_cpu_zonestats, cpu)->vm_numa_stat_diff[item]
+ for (item = 0; item < NR_VM_NUMA_EVENT_ITEMS; item++) {
+ atomic_long_set(&zone->vm_numa_event[item], 0);
+ for_each_online_cpu(cpu) {
+ per_cpu_ptr(zone->per_cpu_zonestats, cpu)->vm_numa_event[item]
= 0;
+ }
}
}
@@ -63,8 +62,8 @@ static void zero_global_numa_counters(vo
{
int item;
- for (item = 0; item < NR_VM_NUMA_STAT_ITEMS; item++)
- atomic_long_set(&vm_numa_stat[item], 0);
+ for (item = 0; item < NR_VM_NUMA_EVENT_ITEMS; item++)
+ atomic_long_set(&vm_numa_event[item], 0);
}
static void invalid_numa_statistics(void)
@@ -161,10 +160,9 @@ void vm_events_fold_cpu(int cpu)
* vm_stat contains the global counters
*/
atomic_long_t vm_zone_stat[NR_VM_ZONE_STAT_ITEMS] __cacheline_aligned_in_smp;
-atomic_long_t vm_numa_stat[NR_VM_NUMA_STAT_ITEMS] __cacheline_aligned_in_smp;
atomic_long_t vm_node_stat[NR_VM_NODE_STAT_ITEMS] __cacheline_aligned_in_smp;
+atomic_long_t vm_numa_event[NR_VM_NUMA_EVENT_ITEMS] __cacheline_aligned_in_smp;
EXPORT_SYMBOL(vm_zone_stat);
-EXPORT_SYMBOL(vm_numa_stat);
EXPORT_SYMBOL(vm_node_stat);
#ifdef CONFIG_SMP
@@ -706,8 +704,7 @@ EXPORT_SYMBOL(dec_node_page_state);
* Fold a differential into the global counters.
* Returns the number of counters updated.
*/
-#ifdef CONFIG_NUMA
-static int fold_diff(int *zone_diff, int *numa_diff, int *node_diff)
+static int fold_diff(int *zone_diff, int *node_diff)
{
int i;
int changes = 0;
@@ -718,12 +715,6 @@ static int fold_diff(int *zone_diff, int
changes++;
}
- for (i = 0; i < NR_VM_NUMA_STAT_ITEMS; i++)
- if (numa_diff[i]) {
- atomic_long_add(numa_diff[i], &vm_numa_stat[i]);
- changes++;
- }
-
for (i = 0; i < NR_VM_NODE_STAT_ITEMS; i++)
if (node_diff[i]) {
atomic_long_add(node_diff[i], &vm_node_stat[i]);
@@ -731,26 +722,34 @@ static int fold_diff(int *zone_diff, int
}
return changes;
}
-#else
-static int fold_diff(int *zone_diff, int *node_diff)
+
+#ifdef CONFIG_NUMA
+static void fold_vm_zone_numa_events(struct zone *zone)
{
- int i;
- int changes = 0;
+ unsigned long zone_numa_events[NR_VM_NUMA_EVENT_ITEMS] = { 0, };
+ int cpu;
+ enum numa_stat_item item;
- for (i = 0; i < NR_VM_ZONE_STAT_ITEMS; i++)
- if (zone_diff[i]) {
- atomic_long_add(zone_diff[i], &vm_zone_stat[i]);
- changes++;
- }
+ for_each_online_cpu(cpu) {
+ struct per_cpu_zonestat *pzstats;
- for (i = 0; i < NR_VM_NODE_STAT_ITEMS; i++)
- if (node_diff[i]) {
- atomic_long_add(node_diff[i], &vm_node_stat[i]);
- changes++;
+ pzstats = per_cpu_ptr(zone->per_cpu_zonestats, cpu);
+ for (item = 0; item < NR_VM_NUMA_EVENT_ITEMS; item++)
+ zone_numa_events[item] += xchg(&pzstats->vm_numa_event[item], 0);
}
- return changes;
+
+ for (item = 0; item < NR_VM_NUMA_EVENT_ITEMS; item++)
+ zone_numa_event_add(zone_numa_events[item], zone, item);
}
-#endif /* CONFIG_NUMA */
+
+void fold_vm_numa_events(void)
+{
+ struct zone *zone;
+
+ for_each_populated_zone(zone)
+ fold_vm_zone_numa_events(zone);
+}
+#endif
/*
* Update the zone counters for the current cpu.
@@ -774,9 +773,6 @@ static int refresh_cpu_vm_stats(bool do_
struct zone *zone;
int i;
int global_zone_diff[NR_VM_ZONE_STAT_ITEMS] = { 0, };
-#ifdef CONFIG_NUMA
- int global_numa_diff[NR_VM_NUMA_STAT_ITEMS] = { 0, };
-#endif
int global_node_diff[NR_VM_NODE_STAT_ITEMS] = { 0, };
int changes = 0;
@@ -801,17 +797,6 @@ static int refresh_cpu_vm_stats(bool do_
}
}
#ifdef CONFIG_NUMA
- for (i = 0; i < NR_VM_NUMA_STAT_ITEMS; i++) {
- int v;
-
- v = this_cpu_xchg(pzstats->vm_numa_stat_diff[i], 0);
- if (v) {
-
- atomic_long_add(v, &zone->vm_numa_stat[i]);
- global_numa_diff[i] += v;
- __this_cpu_write(pcp->expire, 3);
- }
- }
if (do_pagesets) {
cond_resched();
@@ -859,12 +844,7 @@ static int refresh_cpu_vm_stats(bool do_
}
}
-#ifdef CONFIG_NUMA
- changes += fold_diff(global_zone_diff, global_numa_diff,
- global_node_diff);
-#else
changes += fold_diff(global_zone_diff, global_node_diff);
-#endif
return changes;
}
@@ -879,9 +859,6 @@ void cpu_vm_stats_fold(int cpu)
struct zone *zone;
int i;
int global_zone_diff[NR_VM_ZONE_STAT_ITEMS] = { 0, };
-#ifdef CONFIG_NUMA
- int global_numa_diff[NR_VM_NUMA_STAT_ITEMS] = { 0, };
-#endif
int global_node_diff[NR_VM_NODE_STAT_ITEMS] = { 0, };
for_each_populated_zone(zone) {
@@ -889,7 +866,7 @@ void cpu_vm_stats_fold(int cpu)
pzstats = per_cpu_ptr(zone->per_cpu_zonestats, cpu);
- for (i = 0; i < NR_VM_ZONE_STAT_ITEMS; i++)
+ for (i = 0; i < NR_VM_ZONE_STAT_ITEMS; i++) {
if (pzstats->vm_stat_diff[i]) {
int v;
@@ -898,17 +875,17 @@ void cpu_vm_stats_fold(int cpu)
atomic_long_add(v, &zone->vm_stat[i]);
global_zone_diff[i] += v;
}
-
+ }
#ifdef CONFIG_NUMA
- for (i = 0; i < NR_VM_NUMA_STAT_ITEMS; i++)
- if (pzstats->vm_numa_stat_diff[i]) {
- int v;
-
- v = pzstats->vm_numa_stat_diff[i];
- pzstats->vm_numa_stat_diff[i] = 0;
- atomic_long_add(v, &zone->vm_numa_stat[i]);
- global_numa_diff[i] += v;
+ for (i = 0; i < NR_VM_NUMA_EVENT_ITEMS; i++) {
+ if (pzstats->vm_numa_event[i]) {
+ unsigned long v;
+
+ v = pzstats->vm_numa_event[i];
+ pzstats->vm_numa_event[i] = 0;
+ zone_numa_event_add(v, zone, i);
}
+ }
#endif
}
@@ -928,11 +905,7 @@ void cpu_vm_stats_fold(int cpu)
}
}
-#ifdef CONFIG_NUMA
- fold_diff(global_zone_diff, global_numa_diff, global_node_diff);
-#else
fold_diff(global_zone_diff, global_node_diff);
-#endif
}
/*
@@ -941,43 +914,37 @@ void cpu_vm_stats_fold(int cpu)
*/
void drain_zonestat(struct zone *zone, struct per_cpu_zonestat *pzstats)
{
+ unsigned long v;
int i;
- for (i = 0; i < NR_VM_ZONE_STAT_ITEMS; i++)
+ for (i = 0; i < NR_VM_ZONE_STAT_ITEMS; i++) {
if (pzstats->vm_stat_diff[i]) {
- int v = pzstats->vm_stat_diff[i];
+ v = pzstats->vm_stat_diff[i];
pzstats->vm_stat_diff[i] = 0;
- atomic_long_add(v, &zone->vm_stat[i]);
- atomic_long_add(v, &vm_zone_stat[i]);
+ zone_page_state_add(v, zone, i);
}
+ }
#ifdef CONFIG_NUMA
- for (i = 0; i < NR_VM_NUMA_STAT_ITEMS; i++)
- if (pzstats->vm_numa_stat_diff[i]) {
- int v = pzstats->vm_numa_stat_diff[i];
-
- pzstats->vm_numa_stat_diff[i] = 0;
- atomic_long_add(v, &zone->vm_numa_stat[i]);
- atomic_long_add(v, &vm_numa_stat[i]);
+ for (i = 0; i < NR_VM_NUMA_EVENT_ITEMS; i++) {
+ if (pzstats->vm_numa_event[i]) {
+ v = pzstats->vm_numa_event[i];
+ pzstats->vm_numa_event[i] = 0;
+ zone_numa_event_add(v, zone, i);
}
+ }
#endif
}
#endif
#ifdef CONFIG_NUMA
-void __inc_numa_state(struct zone *zone,
+/* See __count_vm_event comment on why raw_cpu_inc is used. */
+void __count_numa_event(struct zone *zone,
enum numa_stat_item item)
{
struct per_cpu_zonestat __percpu *pzstats = zone->per_cpu_zonestats;
- u16 __percpu *p = pzstats->vm_numa_stat_diff + item;
- u16 v;
-
- v = __this_cpu_inc_return(*p);
- if (unlikely(v > NUMA_STATS_THRESHOLD)) {
- zone_numa_state_add(v, zone, item);
- __this_cpu_write(*p, 0);
- }
+ raw_cpu_inc(pzstats->vm_numa_event[item]);
}
/*
@@ -998,19 +965,16 @@ unsigned long sum_zone_node_page_state(i
return count;
}
-/*
- * Determine the per node value of a numa stat item. To avoid deviation,
- * the per cpu stat number in vm_numa_stat_diff[] is also included.
- */
-unsigned long sum_zone_numa_state(int node,
+/* Determine the per node value of a numa stat item. */
+unsigned long sum_zone_numa_event_state(int node,
enum numa_stat_item item)
{
struct zone *zones = NODE_DATA(node)->node_zones;
- int i;
unsigned long count = 0;
+ int i;
for (i = 0; i < MAX_NR_ZONES; i++)
- count += zone_numa_state_snapshot(zones + i, item);
+ count += zone_numa_event_state(zones + i, item);
return count;
}
@@ -1689,9 +1653,9 @@ static void zoneinfo_show_print(struct s
zone_page_state(zone, i));
#ifdef CONFIG_NUMA
- for (i = 0; i < NR_VM_NUMA_STAT_ITEMS; i++)
+ for (i = 0; i < NR_VM_NUMA_EVENT_ITEMS; i++)
seq_printf(m, "\n %-12s %lu", numa_stat_name(i),
- zone_numa_state_snapshot(zone, i));
+ zone_numa_event_state(zone, i));
#endif
seq_printf(m, "\n pagesets");
@@ -1745,7 +1709,7 @@ static const struct seq_operations zonei
};
#define NR_VMSTAT_ITEMS (NR_VM_ZONE_STAT_ITEMS + \
- NR_VM_NUMA_STAT_ITEMS + \
+ NR_VM_NUMA_EVENT_ITEMS + \
NR_VM_NODE_STAT_ITEMS + \
NR_VM_WRITEBACK_STAT_ITEMS + \
(IS_ENABLED(CONFIG_VM_EVENT_COUNTERS) ? \
@@ -1760,6 +1724,7 @@ static void *vmstat_start(struct seq_fil
return NULL;
BUILD_BUG_ON(ARRAY_SIZE(vmstat_text) < NR_VMSTAT_ITEMS);
+ fold_vm_numa_events();
v = kmalloc_array(NR_VMSTAT_ITEMS, sizeof(unsigned long), GFP_KERNEL);
m->private = v;
if (!v)
@@ -1769,9 +1734,9 @@ static void *vmstat_start(struct seq_fil
v += NR_VM_ZONE_STAT_ITEMS;
#ifdef CONFIG_NUMA
- for (i = 0; i < NR_VM_NUMA_STAT_ITEMS; i++)
- v[i] = global_numa_state(i);
- v += NR_VM_NUMA_STAT_ITEMS;
+ for (i = 0; i < NR_VM_NUMA_EVENT_ITEMS; i++)
+ v[i] = global_numa_event_state(i);
+ v += NR_VM_NUMA_EVENT_ITEMS;
#endif
for (i = 0; i < NR_VM_NODE_STAT_ITEMS; i++) {
@@ -1941,11 +1906,7 @@ static bool need_update(int cpu)
if (memchr_inv(pzstats->vm_stat_diff, 0, NR_VM_ZONE_STAT_ITEMS *
sizeof(pzstats->vm_stat_diff[0])))
return true;
-#ifdef CONFIG_NUMA
- if (memchr_inv(pzstats->vm_numa_stat_diff, 0, NR_VM_NUMA_STAT_ITEMS *
- sizeof(pzstats->vm_numa_stat_diff[0])))
- return true;
-#endif
+
if (last_pgdat == zone->zone_pgdat)
continue;
last_pgdat = zone->zone_pgdat;
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 163/192] mm/vmstat: inline NUMA event counter updates
2021-06-29 2:32 incoming Andrew Morton
` (161 preceding siblings ...)
2021-06-29 2:41 ` [patch 162/192] mm/vmstat: convert NUMA statistics to basic NUMA counters Andrew Morton
@ 2021-06-29 2:41 ` Andrew Morton
2021-06-29 2:41 ` [patch 164/192] mm/page_alloc: batch the accounting updates in the bulk allocator Andrew Morton
` (28 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:41 UTC (permalink / raw)
To: akpm, bigeasy, brouer, chuck.lever, linux-mm, mgorman, mhocko,
mingo, mm-commits, peterz, tglx, torvalds, vbabka
From: Mel Gorman <mgorman@techsingularity.net>
Subject: mm/vmstat: inline NUMA event counter updates
__count_numa_event is small enough to be treated similarly to
__count_vm_event so inline it.
Link: https://lkml.kernel.org/r/20210512095458.30632-5-mgorman@techsingularity.net
Signed-off-by: Mel Gorman <mgorman@techsingularity.net>
Acked-by: Vlastimil Babka <vbabka@suse.cz>
Acked-by: Peter Zijlstra (Intel) <peterz@infradead.org>
Cc: Chuck Lever <chuck.lever@oracle.com>
Cc: Ingo Molnar <mingo@kernel.org>
Cc: Jesper Dangaard Brouer <brouer@redhat.com>
Cc: Michal Hocko <mhocko@kernel.org>
Cc: Sebastian Andrzej Siewior <bigeasy@linutronix.de>
Cc: Thomas Gleixner <tglx@linutronix.de>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/linux/vmstat.h | 10 +++++++++-
mm/vmstat.c | 9 ---------
2 files changed, 9 insertions(+), 10 deletions(-)
--- a/include/linux/vmstat.h~mm-vmstat-inline-numa-event-counter-updates
+++ a/include/linux/vmstat.h
@@ -238,7 +238,15 @@ static inline unsigned long zone_page_st
}
#ifdef CONFIG_NUMA
-extern void __count_numa_event(struct zone *zone, enum numa_stat_item item);
+/* See __count_vm_event comment on why raw_cpu_inc is used. */
+static inline void
+__count_numa_event(struct zone *zone, enum numa_stat_item item)
+{
+ struct per_cpu_zonestat __percpu *pzstats = zone->per_cpu_zonestats;
+
+ raw_cpu_inc(pzstats->vm_numa_event[item]);
+}
+
extern unsigned long sum_zone_node_page_state(int node,
enum zone_stat_item item);
extern unsigned long sum_zone_numa_event_state(int node, enum numa_stat_item item);
--- a/mm/vmstat.c~mm-vmstat-inline-numa-event-counter-updates
+++ a/mm/vmstat.c
@@ -938,15 +938,6 @@ void drain_zonestat(struct zone *zone, s
#endif
#ifdef CONFIG_NUMA
-/* See __count_vm_event comment on why raw_cpu_inc is used. */
-void __count_numa_event(struct zone *zone,
- enum numa_stat_item item)
-{
- struct per_cpu_zonestat __percpu *pzstats = zone->per_cpu_zonestats;
-
- raw_cpu_inc(pzstats->vm_numa_event[item]);
-}
-
/*
* Determine the per node value of a stat item. This function
* is called frequently in a NUMA machine, so try to be as
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 164/192] mm/page_alloc: batch the accounting updates in the bulk allocator
2021-06-29 2:32 incoming Andrew Morton
` (162 preceding siblings ...)
2021-06-29 2:41 ` [patch 163/192] mm/vmstat: inline NUMA event counter updates Andrew Morton
@ 2021-06-29 2:41 ` Andrew Morton
2021-06-29 2:41 ` [patch 165/192] mm/page_alloc: reduce duration that IRQs are disabled for VM counters Andrew Morton
` (27 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:41 UTC (permalink / raw)
To: akpm, bigeasy, brouer, chuck.lever, linux-mm, mgorman, mhocko,
mingo, mm-commits, peterz, tglx, torvalds, vbabka
From: Mel Gorman <mgorman@techsingularity.net>
Subject: mm/page_alloc: batch the accounting updates in the bulk allocator
Now that the zone_statistics are simple counters that do not require
special protection, the bulk allocator accounting updates can be batch
updated without adding too much complexity with protected RMW updates or
using xchg.
Link: https://lkml.kernel.org/r/20210512095458.30632-6-mgorman@techsingularity.net
Signed-off-by: Mel Gorman <mgorman@techsingularity.net>
Acked-by: Vlastimil Babka <vbabka@suse.cz>
Acked-by: Peter Zijlstra (Intel) <peterz@infradead.org>
Cc: Chuck Lever <chuck.lever@oracle.com>
Cc: Ingo Molnar <mingo@kernel.org>
Cc: Jesper Dangaard Brouer <brouer@redhat.com>
Cc: Michal Hocko <mhocko@kernel.org>
Cc: Sebastian Andrzej Siewior <bigeasy@linutronix.de>
Cc: Thomas Gleixner <tglx@linutronix.de>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/linux/vmstat.h | 8 ++++++++
mm/page_alloc.c | 30 +++++++++++++-----------------
2 files changed, 21 insertions(+), 17 deletions(-)
--- a/include/linux/vmstat.h~mm-page_alloc-batch-the-accounting-updates-in-the-bulk-allocator
+++ a/include/linux/vmstat.h
@@ -247,6 +247,14 @@ __count_numa_event(struct zone *zone, en
raw_cpu_inc(pzstats->vm_numa_event[item]);
}
+static inline void
+__count_numa_events(struct zone *zone, enum numa_stat_item item, long delta)
+{
+ struct per_cpu_zonestat __percpu *pzstats = zone->per_cpu_zonestats;
+
+ raw_cpu_add(pzstats->vm_numa_event[item], delta);
+}
+
extern unsigned long sum_zone_node_page_state(int node,
enum zone_stat_item item);
extern unsigned long sum_zone_numa_event_state(int node, enum numa_stat_item item);
--- a/mm/page_alloc.c~mm-page_alloc-batch-the-accounting-updates-in-the-bulk-allocator
+++ a/mm/page_alloc.c
@@ -3467,7 +3467,8 @@ void __putback_isolated_page(struct page
*
* Must be called with interrupts disabled.
*/
-static inline void zone_statistics(struct zone *preferred_zone, struct zone *z)
+static inline void zone_statistics(struct zone *preferred_zone, struct zone *z,
+ long nr_account)
{
#ifdef CONFIG_NUMA
enum numa_stat_item local_stat = NUMA_LOCAL;
@@ -3480,12 +3481,12 @@ static inline void zone_statistics(struc
local_stat = NUMA_OTHER;
if (zone_to_nid(z) == zone_to_nid(preferred_zone))
- __count_numa_event(z, NUMA_HIT);
+ __count_numa_events(z, NUMA_HIT, nr_account);
else {
- __count_numa_event(z, NUMA_MISS);
- __count_numa_event(preferred_zone, NUMA_FOREIGN);
+ __count_numa_events(z, NUMA_MISS, nr_account);
+ __count_numa_events(preferred_zone, NUMA_FOREIGN, nr_account);
}
- __count_numa_event(z, local_stat);
+ __count_numa_events(z, local_stat, nr_account);
#endif
}
@@ -3531,7 +3532,7 @@ static struct page *rmqueue_pcplist(stru
page = __rmqueue_pcplist(zone, migratetype, alloc_flags, pcp, list);
if (page) {
__count_zid_vm_events(PGALLOC, page_zonenum(page), 1);
- zone_statistics(preferred_zone, zone);
+ zone_statistics(preferred_zone, zone, 1);
}
local_unlock_irqrestore(&pagesets.lock, flags);
return page;
@@ -3592,7 +3593,7 @@ struct page *rmqueue(struct zone *prefer
get_pcppage_migratetype(page));
__count_zid_vm_events(PGALLOC, page_zonenum(page), 1 << order);
- zone_statistics(preferred_zone, zone);
+ zone_statistics(preferred_zone, zone, 1);
local_irq_restore(flags);
out:
@@ -5077,7 +5078,7 @@ unsigned long __alloc_pages_bulk(gfp_t g
struct alloc_context ac;
gfp_t alloc_gfp;
unsigned int alloc_flags = ALLOC_WMARK_LOW;
- int nr_populated = 0;
+ int nr_populated = 0, nr_account = 0;
if (unlikely(nr_pages <= 0))
return 0;
@@ -5154,15 +5155,7 @@ unsigned long __alloc_pages_bulk(gfp_t g
goto failed_irq;
break;
}
-
- /*
- * Ideally this would be batched but the best way to do
- * that cheaply is to first convert zone_statistics to
- * be inaccurate per-cpu counter like vm_events to avoid
- * a RMW cycle then do the accounting with IRQs enabled.
- */
- __count_zid_vm_events(PGALLOC, zone_idx(zone), 1);
- zone_statistics(ac.preferred_zoneref->zone, zone);
+ nr_account++;
prep_new_page(page, 0, gfp, 0);
if (page_list)
@@ -5172,6 +5165,9 @@ unsigned long __alloc_pages_bulk(gfp_t g
nr_populated++;
}
+ __count_zid_vm_events(PGALLOC, zone_idx(zone), nr_account);
+ zone_statistics(ac.preferred_zoneref->zone, zone, nr_account);
+
local_unlock_irqrestore(&pagesets.lock, flags);
return nr_populated;
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 165/192] mm/page_alloc: reduce duration that IRQs are disabled for VM counters
2021-06-29 2:32 incoming Andrew Morton
` (163 preceding siblings ...)
2021-06-29 2:41 ` [patch 164/192] mm/page_alloc: batch the accounting updates in the bulk allocator Andrew Morton
@ 2021-06-29 2:41 ` Andrew Morton
2021-06-29 2:41 ` [patch 166/192] mm/page_alloc: explicitly acquire the zone lock in __free_pages_ok Andrew Morton
` (26 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:41 UTC (permalink / raw)
To: akpm, bigeasy, brouer, chuck.lever, linux-mm, mgorman, mhocko,
mingo, mm-commits, peterz, tglx, torvalds, vbabka
From: Mel Gorman <mgorman@techsingularity.net>
Subject: mm/page_alloc: reduce duration that IRQs are disabled for VM counters
IRQs are left disabled for the zone and node VM event counters. This is
unnecessary as the affected counters are allowed to race for preemmption
and IRQs.
This patch reduces the scope of IRQs being disabled via
local_[lock|unlock]_irq on !PREEMPT_RT kernels. One
__mod_zone_freepage_state is still called with IRQs disabled. While this
could be moved out, it's not free on all architectures as some require
IRQs to be disabled for mod_zone_page_state on !PREEMPT_RT kernels.
Link: https://lkml.kernel.org/r/20210512095458.30632-7-mgorman@techsingularity.net
Signed-off-by: Mel Gorman <mgorman@techsingularity.net>
Acked-by: Vlastimil Babka <vbabka@suse.cz>
Acked-by: Peter Zijlstra (Intel) <peterz@infradead.org>
Cc: Chuck Lever <chuck.lever@oracle.com>
Cc: Ingo Molnar <mingo@kernel.org>
Cc: Jesper Dangaard Brouer <brouer@redhat.com>
Cc: Michal Hocko <mhocko@kernel.org>
Cc: Sebastian Andrzej Siewior <bigeasy@linutronix.de>
Cc: Thomas Gleixner <tglx@linutronix.de>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/page_alloc.c | 12 ++++++------
1 file changed, 6 insertions(+), 6 deletions(-)
--- a/mm/page_alloc.c~mm-page_alloc-reduce-duration-that-irqs-are-disabled-for-vm-counters
+++ a/mm/page_alloc.c
@@ -3530,11 +3530,11 @@ static struct page *rmqueue_pcplist(stru
pcp = this_cpu_ptr(zone->per_cpu_pageset);
list = &pcp->lists[migratetype];
page = __rmqueue_pcplist(zone, migratetype, alloc_flags, pcp, list);
+ local_unlock_irqrestore(&pagesets.lock, flags);
if (page) {
__count_zid_vm_events(PGALLOC, page_zonenum(page), 1);
zone_statistics(preferred_zone, zone, 1);
}
- local_unlock_irqrestore(&pagesets.lock, flags);
return page;
}
@@ -3586,15 +3586,15 @@ struct page *rmqueue(struct zone *prefer
if (!page)
page = __rmqueue(zone, order, migratetype, alloc_flags);
} while (page && check_new_pages(page, order));
- spin_unlock(&zone->lock);
if (!page)
goto failed;
+
__mod_zone_freepage_state(zone, -(1 << order),
get_pcppage_migratetype(page));
+ spin_unlock_irqrestore(&zone->lock, flags);
__count_zid_vm_events(PGALLOC, page_zonenum(page), 1 << order);
zone_statistics(preferred_zone, zone, 1);
- local_irq_restore(flags);
out:
/* Separate test+clear to avoid unnecessary atomics */
@@ -3607,7 +3607,7 @@ out:
return page;
failed:
- local_irq_restore(flags);
+ spin_unlock_irqrestore(&zone->lock, flags);
return NULL;
}
@@ -5165,11 +5165,11 @@ unsigned long __alloc_pages_bulk(gfp_t g
nr_populated++;
}
+ local_unlock_irqrestore(&pagesets.lock, flags);
+
__count_zid_vm_events(PGALLOC, zone_idx(zone), nr_account);
zone_statistics(ac.preferred_zoneref->zone, zone, nr_account);
- local_unlock_irqrestore(&pagesets.lock, flags);
-
return nr_populated;
failed_irq:
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 166/192] mm/page_alloc: explicitly acquire the zone lock in __free_pages_ok
2021-06-29 2:32 incoming Andrew Morton
` (164 preceding siblings ...)
2021-06-29 2:41 ` [patch 165/192] mm/page_alloc: reduce duration that IRQs are disabled for VM counters Andrew Morton
@ 2021-06-29 2:41 ` Andrew Morton
2021-06-29 2:42 ` [patch 167/192] mm/page_alloc: avoid conflating IRQs disabled with zone->lock Andrew Morton
` (25 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:41 UTC (permalink / raw)
To: akpm, bigeasy, brouer, chuck.lever, linux-mm, mgorman, mhocko,
mingo, mm-commits, peterz, tglx, torvalds, vbabka
From: Mel Gorman <mgorman@techsingularity.net>
Subject: mm/page_alloc: explicitly acquire the zone lock in __free_pages_ok
__free_pages_ok() disables IRQs before calling a common helper
free_one_page() that acquires the zone lock. This is not safe according
to Documentation/locking/locktypes.rst and in this context, IRQ disabling
is not protecting a per_cpu_pages structure either or a local_lock would
be used.
This patch explicitly acquires the lock with spin_lock_irqsave instead of
relying on a helper. This removes the last instance of local_irq_save()
in page_alloc.c.
Link: https://lkml.kernel.org/r/20210512095458.30632-8-mgorman@techsingularity.net
Signed-off-by: Mel Gorman <mgorman@techsingularity.net>
Acked-by: Vlastimil Babka <vbabka@suse.cz>
Acked-by: Peter Zijlstra (Intel) <peterz@infradead.org>
Cc: Chuck Lever <chuck.lever@oracle.com>
Cc: Ingo Molnar <mingo@kernel.org>
Cc: Jesper Dangaard Brouer <brouer@redhat.com>
Cc: Michal Hocko <mhocko@kernel.org>
Cc: Sebastian Andrzej Siewior <bigeasy@linutronix.de>
Cc: Thomas Gleixner <tglx@linutronix.de>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/page_alloc.c | 16 ++++++++--------
1 file changed, 8 insertions(+), 8 deletions(-)
--- a/mm/page_alloc.c~mm-page_alloc-explicitly-acquire-the-zone-lock-in-__free_pages_ok
+++ a/mm/page_alloc.c
@@ -1590,21 +1590,21 @@ static void __free_pages_ok(struct page
unsigned long flags;
int migratetype;
unsigned long pfn = page_to_pfn(page);
+ struct zone *zone = page_zone(page);
if (!free_pages_prepare(page, order, true, fpi_flags))
return;
migratetype = get_pfnblock_migratetype(page, pfn);
- /*
- * TODO FIX: Disable IRQs before acquiring IRQ-safe zone->lock
- * and protect vmstat updates.
- */
- local_irq_save(flags);
+ spin_lock_irqsave(&zone->lock, flags);
__count_vm_events(PGFREE, 1 << order);
- free_one_page(page_zone(page), page, pfn, order, migratetype,
- fpi_flags);
- local_irq_restore(flags);
+ if (unlikely(has_isolate_pageblock(zone) ||
+ is_migrate_isolate(migratetype))) {
+ migratetype = get_pfnblock_migratetype(page, pfn);
+ }
+ __free_one_page(page, pfn, zone, order, migratetype, fpi_flags);
+ spin_unlock_irqrestore(&zone->lock, flags);
}
void __free_pages_core(struct page *page, unsigned int order)
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 167/192] mm/page_alloc: avoid conflating IRQs disabled with zone->lock
2021-06-29 2:32 incoming Andrew Morton
` (165 preceding siblings ...)
2021-06-29 2:41 ` [patch 166/192] mm/page_alloc: explicitly acquire the zone lock in __free_pages_ok Andrew Morton
@ 2021-06-29 2:42 ` Andrew Morton
2021-06-29 2:42 ` [patch 168/192] mm/page_alloc: update PGFREE outside the zone lock in __free_pages_ok Andrew Morton
` (24 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:42 UTC (permalink / raw)
To: akpm, bigeasy, brouer, chuck.lever, linux-mm, mgorman, mhocko,
mingo, mm-commits, peterz, tglx, torvalds, vbabka
From: Mel Gorman <mgorman@techsingularity.net>
Subject: mm/page_alloc: avoid conflating IRQs disabled with zone->lock
Historically when freeing pages, free_one_page() assumed that callers had
IRQs disabled and the zone->lock could be acquired with spin_lock(). This
confuses the scope of what local_lock_irq is protecting and what
zone->lock is protecting in free_unref_page_list in particular.
This patch uses spin_lock_irqsave() for the zone->lock in free_one_page()
instead of relying on callers to have disabled IRQs.
free_unref_page_commit() is changed to only deal with PCP pages protected
by the local lock. free_unref_page_list() then first frees isolated pages
to the buddy lists with free_one_page() and frees the rest of the pages to
the PCP via free_unref_page_commit(). The end result is that
free_one_page() is no longer depending on side-effects of local_lock to be
correct.
Note that this may incur a performance penalty while memory hot-remove is
running but that is not a common operation.
[lkp@intel.com: Ensure CMA pages get addded to correct pcp list]
Link: https://lkml.kernel.org/r/20210512095458.30632-9-mgorman@techsingularity.net
Signed-off-by: Mel Gorman <mgorman@techsingularity.net>
Acked-by: Vlastimil Babka <vbabka@suse.cz>
Acked-by: Peter Zijlstra (Intel) <peterz@infradead.org>
Cc: Chuck Lever <chuck.lever@oracle.com>
Cc: Ingo Molnar <mingo@kernel.org>
Cc: Jesper Dangaard Brouer <brouer@redhat.com>
Cc: Michal Hocko <mhocko@kernel.org>
Cc: Sebastian Andrzej Siewior <bigeasy@linutronix.de>
Cc: Thomas Gleixner <tglx@linutronix.de>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/page_alloc.c | 75 ++++++++++++++++++++++++++++++----------------
1 file changed, 49 insertions(+), 26 deletions(-)
--- a/mm/page_alloc.c~mm-page_alloc-avoid-conflating-irqs-disabled-with-zone-lock
+++ a/mm/page_alloc.c
@@ -1501,13 +1501,15 @@ static void free_one_page(struct zone *z
unsigned int order,
int migratetype, fpi_t fpi_flags)
{
- spin_lock(&zone->lock);
+ unsigned long flags;
+
+ spin_lock_irqsave(&zone->lock, flags);
if (unlikely(has_isolate_pageblock(zone) ||
is_migrate_isolate(migratetype))) {
migratetype = get_pfnblock_migratetype(page, pfn);
}
__free_one_page(page, pfn, zone, order, migratetype, fpi_flags);
- spin_unlock(&zone->lock);
+ spin_unlock_irqrestore(&zone->lock, flags);
}
static void __meminit __init_single_page(struct page *page, unsigned long pfn,
@@ -3285,31 +3287,13 @@ static bool free_unref_page_prepare(stru
return true;
}
-static void free_unref_page_commit(struct page *page, unsigned long pfn)
+static void free_unref_page_commit(struct page *page, unsigned long pfn,
+ int migratetype)
{
struct zone *zone = page_zone(page);
struct per_cpu_pages *pcp;
- int migratetype;
- migratetype = get_pcppage_migratetype(page);
__count_vm_event(PGFREE);
-
- /*
- * We only track unmovable, reclaimable and movable on pcp lists.
- * Free ISOLATE pages back to the allocator because they are being
- * offlined but treat HIGHATOMIC as movable pages so we can get those
- * areas back if necessary. Otherwise, we may have to free
- * excessively into the page allocator
- */
- if (migratetype >= MIGRATE_PCPTYPES) {
- if (unlikely(is_migrate_isolate(migratetype))) {
- free_one_page(zone, page, pfn, 0, migratetype,
- FPI_NONE);
- return;
- }
- migratetype = MIGRATE_MOVABLE;
- }
-
pcp = this_cpu_ptr(zone->per_cpu_pageset);
list_add(&page->lru, &pcp->lists[migratetype]);
pcp->count++;
@@ -3324,12 +3308,29 @@ void free_unref_page(struct page *page)
{
unsigned long flags;
unsigned long pfn = page_to_pfn(page);
+ int migratetype;
if (!free_unref_page_prepare(page, pfn))
return;
+ /*
+ * We only track unmovable, reclaimable and movable on pcp lists.
+ * Place ISOLATE pages on the isolated list because they are being
+ * offlined but treat HIGHATOMIC as movable pages so we can get those
+ * areas back if necessary. Otherwise, we may have to free
+ * excessively into the page allocator
+ */
+ migratetype = get_pcppage_migratetype(page);
+ if (unlikely(migratetype >= MIGRATE_PCPTYPES)) {
+ if (unlikely(is_migrate_isolate(migratetype))) {
+ free_one_page(page_zone(page), page, pfn, 0, migratetype, FPI_NONE);
+ return;
+ }
+ migratetype = MIGRATE_MOVABLE;
+ }
+
local_lock_irqsave(&pagesets.lock, flags);
- free_unref_page_commit(page, pfn);
+ free_unref_page_commit(page, pfn, migratetype);
local_unlock_irqrestore(&pagesets.lock, flags);
}
@@ -3341,22 +3342,44 @@ void free_unref_page_list(struct list_he
struct page *page, *next;
unsigned long flags, pfn;
int batch_count = 0;
+ int migratetype;
/* Prepare pages for freeing */
list_for_each_entry_safe(page, next, list, lru) {
pfn = page_to_pfn(page);
if (!free_unref_page_prepare(page, pfn))
list_del(&page->lru);
+
+ /*
+ * Free isolated pages directly to the allocator, see
+ * comment in free_unref_page.
+ */
+ migratetype = get_pcppage_migratetype(page);
+ if (unlikely(migratetype >= MIGRATE_PCPTYPES)) {
+ if (unlikely(is_migrate_isolate(migratetype))) {
+ list_del(&page->lru);
+ free_one_page(page_zone(page), page, pfn, 0,
+ migratetype, FPI_NONE);
+ continue;
+ }
+
+ /*
+ * Non-isolated types over MIGRATE_PCPTYPES get added
+ * to the MIGRATE_MOVABLE pcp list.
+ */
+ set_pcppage_migratetype(page, MIGRATE_MOVABLE);
+ }
+
set_page_private(page, pfn);
}
local_lock_irqsave(&pagesets.lock, flags);
list_for_each_entry_safe(page, next, list, lru) {
- unsigned long pfn = page_private(page);
-
+ pfn = page_private(page);
set_page_private(page, 0);
+ migratetype = get_pcppage_migratetype(page);
trace_mm_page_free_batched(page);
- free_unref_page_commit(page, pfn);
+ free_unref_page_commit(page, pfn, migratetype);
/*
* Guard against excessive IRQ disabled times when we get
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 168/192] mm/page_alloc: update PGFREE outside the zone lock in __free_pages_ok
2021-06-29 2:32 incoming Andrew Morton
` (166 preceding siblings ...)
2021-06-29 2:42 ` [patch 167/192] mm/page_alloc: avoid conflating IRQs disabled with zone->lock Andrew Morton
@ 2021-06-29 2:42 ` Andrew Morton
2021-06-29 2:42 ` [patch 169/192] mm: page_alloc: dump migrate-failed pages only at -EBUSY Andrew Morton
` (23 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:42 UTC (permalink / raw)
To: akpm, bigeasy, brouer, chuck.lever, linux-mm, mgorman, mhocko,
mingo, mm-commits, peterz, tglx, torvalds, vbabka
From: Mel Gorman <mgorman@techsingularity.net>
Subject: mm/page_alloc: update PGFREE outside the zone lock in __free_pages_ok
VM events do not need explicit protection by disabling IRQs so update the
counter with IRQs enabled in __free_pages_ok.
Link: https://lkml.kernel.org/r/20210512095458.30632-10-mgorman@techsingularity.net
Signed-off-by: Mel Gorman <mgorman@techsingularity.net>
Acked-by: Vlastimil Babka <vbabka@suse.cz>
Acked-by: Peter Zijlstra (Intel) <peterz@infradead.org>
Cc: Chuck Lever <chuck.lever@oracle.com>
Cc: Ingo Molnar <mingo@kernel.org>
Cc: Jesper Dangaard Brouer <brouer@redhat.com>
Cc: Michal Hocko <mhocko@kernel.org>
Cc: Sebastian Andrzej Siewior <bigeasy@linutronix.de>
Cc: Thomas Gleixner <tglx@linutronix.de>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/page_alloc.c | 3 ++-
1 file changed, 2 insertions(+), 1 deletion(-)
--- a/mm/page_alloc.c~mm-page_alloc-update-pgfree-outside-the-zone-lock-in-__free_pages_ok
+++ a/mm/page_alloc.c
@@ -1600,13 +1600,14 @@ static void __free_pages_ok(struct page
migratetype = get_pfnblock_migratetype(page, pfn);
spin_lock_irqsave(&zone->lock, flags);
- __count_vm_events(PGFREE, 1 << order);
if (unlikely(has_isolate_pageblock(zone) ||
is_migrate_isolate(migratetype))) {
migratetype = get_pfnblock_migratetype(page, pfn);
}
__free_one_page(page, pfn, zone, order, migratetype, fpi_flags);
spin_unlock_irqrestore(&zone->lock, flags);
+
+ __count_vm_events(PGFREE, 1 << order);
}
void __free_pages_core(struct page *page, unsigned int order)
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 169/192] mm: page_alloc: dump migrate-failed pages only at -EBUSY
2021-06-29 2:32 incoming Andrew Morton
` (167 preceding siblings ...)
2021-06-29 2:42 ` [patch 168/192] mm/page_alloc: update PGFREE outside the zone lock in __free_pages_ok Andrew Morton
@ 2021-06-29 2:42 ` Andrew Morton
2021-06-29 2:42 ` [patch 170/192] mm/page_alloc: delete vm.percpu_pagelist_fraction Andrew Morton
` (22 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:42 UTC (permalink / raw)
To: akpm, david, joaodias, linux-mm, mhocko, minchan, mm-commits,
surenb, torvalds
From: Minchan Kim <minchan@kernel.org>
Subject: mm: page_alloc: dump migrate-failed pages only at -EBUSY
alloc_contig_dump_pages() aims for helping debugging page migration
failure by elevated page refcount compared to expected_count. (for the
detail, please look at migrate_page_move_mapping)
However, -ENOMEM is just the case that system is under memory pressure
state, not relevant with page refcount at all. Thus, the dumping page
list is not helpful for the debugging point of view.
Link: https://lkml.kernel.org/r/YKa2Wyo9xqIErpfa@google.com
Signed-off-by: Minchan Kim <minchan@kernel.org>
Reviewed-by: David Hildenbrand <david@redhat.com>
Cc: Suren Baghdasaryan <surenb@google.com>
Cc: John Dias <joaodias@google.com>
Cc: Michal Hocko <mhocko@suse.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/page_alloc.c | 3 ++-
1 file changed, 2 insertions(+), 1 deletion(-)
--- a/mm/page_alloc.c~mm-page_alloc-dump-migrate-failed-pages-only-at-ebusy
+++ a/mm/page_alloc.c
@@ -8800,7 +8800,8 @@ static int __alloc_contig_migrate_range(
lru_cache_enable();
if (ret < 0) {
- alloc_contig_dump_pages(&cc->migratepages);
+ if (ret == -EBUSY)
+ alloc_contig_dump_pages(&cc->migratepages);
putback_movable_pages(&cc->migratepages);
return ret;
}
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 170/192] mm/page_alloc: delete vm.percpu_pagelist_fraction
2021-06-29 2:32 incoming Andrew Morton
` (168 preceding siblings ...)
2021-06-29 2:42 ` [patch 169/192] mm: page_alloc: dump migrate-failed pages only at -EBUSY Andrew Morton
@ 2021-06-29 2:42 ` Andrew Morton
2021-06-29 2:42 ` [patch 171/192] mm/page_alloc: disassociate the pcp->high from pcp->batch Andrew Morton
` (21 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:42 UTC (permalink / raw)
To: akpm, dave.hansen, hdanton, linux-mm, mgorman, mhocko,
mm-commits, torvalds, vbabka
From: Mel Gorman <mgorman@techsingularity.net>
Subject: mm/page_alloc: delete vm.percpu_pagelist_fraction
Patch series "Calculate pcp->high based on zone sizes and active CPUs", v2.
The per-cpu page allocator (PCP) is meant to reduce contention on the zone
lock but the sizing of batch and high is archaic and neither takes the
zone size into account or the number of CPUs local to a zone. With larger
zones and more CPUs per node, the contention is getting worse.
Furthermore, the fact that vm.percpu_pagelist_fraction adjusts both batch
and high values means that the sysctl can reduce zone lock contention but
also increase allocation latencies.
This series disassociates pcp->high from pcp->batch and then scales
pcp->high based on the size of the local zone with limited impact to
reclaim and accounting for active CPUs but leaves pcp->batch static. It
also adapts the number of pages that can be on the pcp list based on
recent freeing patterns.
The motivation is partially to adjust to larger memory sizes but is also
driven by the fact that large batches of page freeing via release_pages()
often shows zone contention as a major part of the problem. Another is a
bug report based on an older kernel where a multi-terabyte process can
takes several minutes to exit. A workaround was to use
vm.percpu_pagelist_fraction to increase the pcp->high value but testing
indicated that a production workload could not use the same values because
of an increase in allocation latencies. Unfortunately, I cannot reproduce
this test case myself as the multi-terabyte machines are in active use but
it should alleviate the problem.
The series aims to address both and partially acts as a pre-requisite.
pcp only works with order-0 which is useless for SLUB (when using high
orders) and THP (unconditionally). To store high-order pages on PCP, the
pcp->high values need to be increased first.
This patch (of 6):
The vm.percpu_pagelist_fraction is used to increase the batch and high
limits for the per-cpu page allocator (PCP). The intent behind the sysctl
is to reduce zone lock acquisition when allocating/freeing pages but it
has a problem. While it can decrease contention, it can also increase
latency on the allocation side due to unreasonably large batch sizes.
This leads to games where an administrator adjusts
percpu_pagelist_fraction on the fly to work around contention and
allocation latency problems.
This series aims to alleviate the problems with zone lock contention while
avoiding the allocation-side latency problems. For the purposes of
review, it's easier to remove this sysctl now and reintroduce a similar
sysctl later in the series that deals only with pcp->high.
Link: https://lkml.kernel.org/r/20210525080119.5455-1-mgorman@techsingularity.net
Link: https://lkml.kernel.org/r/20210525080119.5455-2-mgorman@techsingularity.net
Signed-off-by: Mel Gorman <mgorman@techsingularity.net>
Acked-by: Dave Hansen <dave.hansen@linux.intel.com>
Acked-by: Vlastimil Babka <vbabka@suse.cz>
Cc: Hillf Danton <hdanton@sina.com>
Cc: Michal Hocko <mhocko@kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
Documentation/admin-guide/sysctl/vm.rst | 19 -------
include/linux/mmzone.h | 3 -
kernel/sysctl.c | 8 ---
mm/page_alloc.c | 55 +---------------------
4 files changed, 4 insertions(+), 81 deletions(-)
--- a/Documentation/admin-guide/sysctl/vm.rst~mm-page_alloc-delete-vmpercpu_pagelist_fraction
+++ a/Documentation/admin-guide/sysctl/vm.rst
@@ -64,7 +64,6 @@ Currently, these files are in /proc/sys/
- overcommit_ratio
- page-cluster
- panic_on_oom
-- percpu_pagelist_fraction
- stat_interval
- stat_refresh
- numa_stat
@@ -790,24 +789,6 @@ panic_on_oom=2+kdump gives you very stro
why oom happens. You can get snapshot.
-percpu_pagelist_fraction
-========================
-
-This is the fraction of pages at most (high mark pcp->high) in each zone that
-are allocated for each per cpu page list. The min value for this is 8. It
-means that we don't allow more than 1/8th of pages in each zone to be
-allocated in any single per_cpu_pagelist. This entry only changes the value
-of hot per cpu pagelists. User can specify a number like 100 to allocate
-1/100th of each zone to each per cpu page list.
-
-The batch value of each per cpu pagelist is also updated as a result. It is
-set to pcp->high/4. The upper limit of batch is (PAGE_SHIFT * 8)
-
-The initial value is zero. Kernel does not use this value at boot time to set
-the high water marks for each per cpu page list. If the user writes '0' to this
-sysctl, it will revert to this default behavior.
-
-
stat_interval
=============
--- a/include/linux/mmzone.h~mm-page_alloc-delete-vmpercpu_pagelist_fraction
+++ a/include/linux/mmzone.h
@@ -1027,15 +1027,12 @@ int watermark_scale_factor_sysctl_handle
extern int sysctl_lowmem_reserve_ratio[MAX_NR_ZONES];
int lowmem_reserve_ratio_sysctl_handler(struct ctl_table *, int, void *,
size_t *, loff_t *);
-int percpu_pagelist_fraction_sysctl_handler(struct ctl_table *, int,
- void *, size_t *, loff_t *);
int sysctl_min_unmapped_ratio_sysctl_handler(struct ctl_table *, int,
void *, size_t *, loff_t *);
int sysctl_min_slab_ratio_sysctl_handler(struct ctl_table *, int,
void *, size_t *, loff_t *);
int numa_zonelist_order_handler(struct ctl_table *, int,
void *, size_t *, loff_t *);
-extern int percpu_pagelist_fraction;
extern char numa_zonelist_order[];
#define NUMA_ZONELIST_ORDER_LEN 16
--- a/kernel/sysctl.c~mm-page_alloc-delete-vmpercpu_pagelist_fraction
+++ a/kernel/sysctl.c
@@ -2909,14 +2909,6 @@ static struct ctl_table vm_table[] = {
.extra2 = &one_thousand,
},
{
- .procname = "percpu_pagelist_fraction",
- .data = &percpu_pagelist_fraction,
- .maxlen = sizeof(percpu_pagelist_fraction),
- .mode = 0644,
- .proc_handler = percpu_pagelist_fraction_sysctl_handler,
- .extra1 = SYSCTL_ZERO,
- },
- {
.procname = "page_lock_unfairness",
.data = &sysctl_page_lock_unfairness,
.maxlen = sizeof(sysctl_page_lock_unfairness),
--- a/mm/page_alloc.c~mm-page_alloc-delete-vmpercpu_pagelist_fraction
+++ a/mm/page_alloc.c
@@ -120,7 +120,6 @@ typedef int __bitwise fpi_t;
/* prevent >1 _updater_ of zone percpu pageset ->high and ->batch fields */
static DEFINE_MUTEX(pcp_batch_high_lock);
-#define MIN_PERCPU_PAGELIST_FRACTION (8)
struct pagesets {
local_lock_t lock;
@@ -193,7 +192,6 @@ EXPORT_SYMBOL(_totalram_pages);
unsigned long totalreserve_pages __read_mostly;
unsigned long totalcma_pages __read_mostly;
-int percpu_pagelist_fraction;
gfp_t gfp_allowed_mask __read_mostly = GFP_BOOT_MASK;
DEFINE_STATIC_KEY_MAYBE(CONFIG_INIT_ON_ALLOC_DEFAULT_ON, init_on_alloc);
EXPORT_SYMBOL(init_on_alloc);
@@ -6735,22 +6733,15 @@ static void __zone_set_pageset_high_and_
/*
* Calculate and set new high and batch values for all per-cpu pagesets of a
- * zone, based on the zone's size and the percpu_pagelist_fraction sysctl.
+ * zone based on the zone's size.
*/
static void zone_set_pageset_high_and_batch(struct zone *zone)
{
unsigned long new_high, new_batch;
- if (percpu_pagelist_fraction) {
- new_high = zone_managed_pages(zone) / percpu_pagelist_fraction;
- new_batch = max(1UL, new_high / 4);
- if ((new_high / 4) > (PAGE_SHIFT * 8))
- new_batch = PAGE_SHIFT * 8;
- } else {
- new_batch = zone_batchsize(zone);
- new_high = 6 * new_batch;
- new_batch = max(1UL, 1 * new_batch);
- }
+ new_batch = zone_batchsize(zone);
+ new_high = 6 * new_batch;
+ new_batch = max(1UL, 1 * new_batch);
if (zone->pageset_high == new_high &&
zone->pageset_batch == new_batch)
@@ -8413,44 +8404,6 @@ int lowmem_reserve_ratio_sysctl_handler(
return 0;
}
-/*
- * percpu_pagelist_fraction - changes the pcp->high for each zone on each
- * cpu. It is the fraction of total pages in each zone that a hot per cpu
- * pagelist can have before it gets flushed back to buddy allocator.
- */
-int percpu_pagelist_fraction_sysctl_handler(struct ctl_table *table, int write,
- void *buffer, size_t *length, loff_t *ppos)
-{
- struct zone *zone;
- int old_percpu_pagelist_fraction;
- int ret;
-
- mutex_lock(&pcp_batch_high_lock);
- old_percpu_pagelist_fraction = percpu_pagelist_fraction;
-
- ret = proc_dointvec_minmax(table, write, buffer, length, ppos);
- if (!write || ret < 0)
- goto out;
-
- /* Sanity checking to avoid pcp imbalance */
- if (percpu_pagelist_fraction &&
- percpu_pagelist_fraction < MIN_PERCPU_PAGELIST_FRACTION) {
- percpu_pagelist_fraction = old_percpu_pagelist_fraction;
- ret = -EINVAL;
- goto out;
- }
-
- /* No change? */
- if (percpu_pagelist_fraction == old_percpu_pagelist_fraction)
- goto out;
-
- for_each_populated_zone(zone)
- zone_set_pageset_high_and_batch(zone);
-out:
- mutex_unlock(&pcp_batch_high_lock);
- return ret;
-}
-
#ifndef __HAVE_ARCH_RESERVED_KERNEL_PAGES
/*
* Returns the number of pages that arch has reserved but
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 171/192] mm/page_alloc: disassociate the pcp->high from pcp->batch
2021-06-29 2:32 incoming Andrew Morton
` (169 preceding siblings ...)
2021-06-29 2:42 ` [patch 170/192] mm/page_alloc: delete vm.percpu_pagelist_fraction Andrew Morton
@ 2021-06-29 2:42 ` Andrew Morton
2021-06-29 2:42 ` [patch 172/192] mm/page_alloc: adjust pcp->high after CPU hotplug events Andrew Morton
` (20 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:42 UTC (permalink / raw)
To: akpm, dave.hansen, hdanton, linux-mm, mgorman, mhocko,
mm-commits, torvalds, vbabka
From: Mel Gorman <mgorman@techsingularity.net>
Subject: mm/page_alloc: disassociate the pcp->high from pcp->batch
The pcp high watermark is based on the batch size but there is no
relationship between them other than it is convenient to use early in
boot.
This patch takes the first step and bases pcp->high on the zone low
watermark split across the number of CPUs local to a zone while the batch
size remains the same to avoid increasing allocation latencies. The
intent behind the default pcp->high is "set the number of PCP pages such
that if they are all full that background reclaim is not started
prematurely".
Note that in this patch the pcp->high values are adjusted after memory
hotplug events, min_free_kbytes adjustments and watermark scale factor
adjustments but not CPU hotplug events which is handled later in the
series.
On a test KVM instance;
Before grep -E "high:|batch" /proc/zoneinfo | tail -2
high: 378
batch: 63
After grep -E "high:|batch" /proc/zoneinfo | tail -2
high: 649
batch: 63
[mgorman@techsingularity.net: fix __setup_per_zone_wmarks for parallel memory
hotplug]
Link: https://lkml.kernel.org/r/20210528105925.GN30378@techsingularity.net
Link: https://lkml.kernel.org/r/20210525080119.5455-3-mgorman@techsingularity.net
Signed-off-by: Mel Gorman <mgorman@techsingularity.net>
Acked-by: Vlastimil Babka <vbabka@suse.cz>
Cc: Dave Hansen <dave.hansen@linux.intel.com>
Cc: Hillf Danton <hdanton@sina.com>
Cc: Michal Hocko <mhocko@kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/memory_hotplug.c | 6 ++--
mm/page_alloc.c | 62 +++++++++++++++++++++++++++++-------------
2 files changed, 47 insertions(+), 21 deletions(-)
--- a/mm/memory_hotplug.c~mm-page_alloc-disassociate-the-pcp-high-from-pcp-batch
+++ a/mm/memory_hotplug.c
@@ -961,7 +961,6 @@ int __ref online_pages(unsigned long pfn
node_states_set_node(nid, &arg);
if (need_zonelists_rebuild)
build_all_zonelists(NULL);
- zone_pcp_update(zone);
/* Basic onlining is complete, allow allocation of onlined pages. */
undo_isolate_page_range(pfn, pfn + nr_pages, MIGRATE_MOVABLE);
@@ -974,6 +973,7 @@ int __ref online_pages(unsigned long pfn
*/
shuffle_zone(zone);
+ /* reinitialise watermarks and update pcp limits */
init_per_zone_wmark_min();
kswapd_run(nid);
@@ -1829,13 +1829,13 @@ int __ref offline_pages(unsigned long st
adjust_managed_page_count(pfn_to_page(start_pfn), -nr_pages);
adjust_present_page_count(zone, -nr_pages);
+ /* reinitialise watermarks and update pcp limits */
init_per_zone_wmark_min();
if (!populated_zone(zone)) {
zone_pcp_reset(zone);
build_all_zonelists(NULL);
- } else
- zone_pcp_update(zone);
+ }
node_states_clear_node(node, &arg);
if (arg.status_change_nid >= 0) {
--- a/mm/page_alloc.c~mm-page_alloc-disassociate-the-pcp-high-from-pcp-batch
+++ a/mm/page_alloc.c
@@ -2175,14 +2175,6 @@ void __init page_alloc_init_late(void)
wait_for_completion(&pgdat_init_all_done_comp);
/*
- * The number of managed pages has changed due to the initialisation
- * so the pcpu batch and high limits needs to be updated or the limits
- * will be artificially small.
- */
- for_each_populated_zone(zone)
- zone_pcp_update(zone);
-
- /*
* We initialized the rest of the deferred pages. Permanently disable
* on-demand struct page initialization.
*/
@@ -6633,13 +6625,12 @@ static int zone_batchsize(struct zone *z
int batch;
/*
- * The per-cpu-pages pools are set to around 1000th of the
- * size of the zone.
+ * The number of pages to batch allocate is either ~0.1%
+ * of the zone or 1MB, whichever is smaller. The batch
+ * size is striking a balance between allocation latency
+ * and zone lock contention.
*/
- batch = zone_managed_pages(zone) / 1024;
- /* But no more than a meg. */
- if (batch * PAGE_SIZE > 1024 * 1024)
- batch = (1024 * 1024) / PAGE_SIZE;
+ batch = min(zone_managed_pages(zone) >> 10, (1024 * 1024) / PAGE_SIZE);
batch /= 4; /* We effectively *= 4 below */
if (batch < 1)
batch = 1;
@@ -6676,6 +6667,34 @@ static int zone_batchsize(struct zone *z
#endif
}
+static int zone_highsize(struct zone *zone, int batch)
+{
+#ifdef CONFIG_MMU
+ int high;
+ int nr_local_cpus;
+
+ /*
+ * The high value of the pcp is based on the zone low watermark
+ * so that if they are full then background reclaim will not be
+ * started prematurely. The value is split across all online CPUs
+ * local to the zone. Note that early in boot that CPUs may not be
+ * online yet.
+ */
+ nr_local_cpus = max(1U, cpumask_weight(cpumask_of_node(zone_to_nid(zone))));
+ high = low_wmark_pages(zone) / nr_local_cpus;
+
+ /*
+ * Ensure high is at least batch*4. The multiple is based on the
+ * historical relationship between high and batch.
+ */
+ high = max(high, batch << 2);
+
+ return high;
+#else
+ return 0;
+#endif
+}
+
/*
* pcp->high and pcp->batch values are related and generally batch is lower
* than high. They are also related to pcp->count such that count is lower
@@ -6737,11 +6756,10 @@ static void __zone_set_pageset_high_and_
*/
static void zone_set_pageset_high_and_batch(struct zone *zone)
{
- unsigned long new_high, new_batch;
+ int new_high, new_batch;
- new_batch = zone_batchsize(zone);
- new_high = 6 * new_batch;
- new_batch = max(1UL, 1 * new_batch);
+ new_batch = max(1, zone_batchsize(zone));
+ new_high = zone_highsize(zone, new_batch);
if (zone->pageset_high == new_high &&
zone->pageset_batch == new_batch)
@@ -8222,11 +8240,19 @@ static void __setup_per_zone_wmarks(void
*/
void setup_per_zone_wmarks(void)
{
+ struct zone *zone;
static DEFINE_SPINLOCK(lock);
spin_lock(&lock);
__setup_per_zone_wmarks();
spin_unlock(&lock);
+
+ /*
+ * The watermark size have changed so update the pcpu batch
+ * and high limits or the limits may be inappropriate.
+ */
+ for_each_zone(zone)
+ zone_pcp_update(zone);
}
/*
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 172/192] mm/page_alloc: adjust pcp->high after CPU hotplug events
2021-06-29 2:32 incoming Andrew Morton
` (170 preceding siblings ...)
2021-06-29 2:42 ` [patch 171/192] mm/page_alloc: disassociate the pcp->high from pcp->batch Andrew Morton
@ 2021-06-29 2:42 ` Andrew Morton
2021-06-29 2:42 ` [patch 173/192] mm/page_alloc: scale the number of pages that are batch freed Andrew Morton
` (19 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:42 UTC (permalink / raw)
To: akpm, dave.hansen, hdanton, linux-mm, mgorman, mhocko,
mm-commits, torvalds, vbabka
From: Mel Gorman <mgorman@techsingularity.net>
Subject: mm/page_alloc: adjust pcp->high after CPU hotplug events
The PCP high watermark is based on the number of online CPUs so the
watermarks must be adjusted during CPU hotplug. At the time of
hot-remove, the number of online CPUs is already adjusted but during
hot-add, a delta needs to be applied to update PCP to the correct value.
After this patch is applied, the high watermarks are adjusted correctly.
# grep high: /proc/zoneinfo | tail -1
high: 649
# echo 0 > /sys/devices/system/cpu/cpu4/online
# grep high: /proc/zoneinfo | tail -1
high: 664
# echo 1 > /sys/devices/system/cpu/cpu4/online
# grep high: /proc/zoneinfo | tail -1
high: 649
Link: https://lkml.kernel.org/r/20210525080119.5455-4-mgorman@techsingularity.net
Signed-off-by: Mel Gorman <mgorman@techsingularity.net>
Acked-by: Vlastimil Babka <vbabka@suse.cz>
Cc: Dave Hansen <dave.hansen@linux.intel.com>
Cc: Hillf Danton <hdanton@sina.com>
Cc: Michal Hocko <mhocko@kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/linux/cpuhotplug.h | 2 -
mm/internal.h | 2 -
mm/page_alloc.c | 38 ++++++++++++++++++++++++-----------
3 files changed, 29 insertions(+), 13 deletions(-)
--- a/include/linux/cpuhotplug.h~mm-page_alloc-adjust-pcp-high-after-cpu-hotplug-events
+++ a/include/linux/cpuhotplug.h
@@ -54,7 +54,7 @@ enum cpuhp_state {
CPUHP_MM_MEMCQ_DEAD,
CPUHP_PERCPU_CNT_DEAD,
CPUHP_RADIX_DEAD,
- CPUHP_PAGE_ALLOC_DEAD,
+ CPUHP_PAGE_ALLOC,
CPUHP_NET_DEV_DEAD,
CPUHP_PCI_XGENE_DEAD,
CPUHP_IOMMU_IOVA_DEAD,
--- a/mm/internal.h~mm-page_alloc-adjust-pcp-high-after-cpu-hotplug-events
+++ a/mm/internal.h
@@ -206,7 +206,7 @@ extern int user_min_free_kbytes;
extern void free_unref_page(struct page *page);
extern void free_unref_page_list(struct list_head *list);
-extern void zone_pcp_update(struct zone *zone);
+extern void zone_pcp_update(struct zone *zone, int cpu_online);
extern void zone_pcp_reset(struct zone *zone);
extern void zone_pcp_disable(struct zone *zone);
extern void zone_pcp_enable(struct zone *zone);
--- a/mm/page_alloc.c~mm-page_alloc-adjust-pcp-high-after-cpu-hotplug-events
+++ a/mm/page_alloc.c
@@ -6667,7 +6667,7 @@ static int zone_batchsize(struct zone *z
#endif
}
-static int zone_highsize(struct zone *zone, int batch)
+static int zone_highsize(struct zone *zone, int batch, int cpu_online)
{
#ifdef CONFIG_MMU
int high;
@@ -6678,9 +6678,10 @@ static int zone_highsize(struct zone *zo
* so that if they are full then background reclaim will not be
* started prematurely. The value is split across all online CPUs
* local to the zone. Note that early in boot that CPUs may not be
- * online yet.
+ * online yet and that during CPU hotplug that the cpumask is not
+ * yet updated when a CPU is being onlined.
*/
- nr_local_cpus = max(1U, cpumask_weight(cpumask_of_node(zone_to_nid(zone))));
+ nr_local_cpus = max(1U, cpumask_weight(cpumask_of_node(zone_to_nid(zone)))) + cpu_online;
high = low_wmark_pages(zone) / nr_local_cpus;
/*
@@ -6754,12 +6755,12 @@ static void __zone_set_pageset_high_and_
* Calculate and set new high and batch values for all per-cpu pagesets of a
* zone based on the zone's size.
*/
-static void zone_set_pageset_high_and_batch(struct zone *zone)
+static void zone_set_pageset_high_and_batch(struct zone *zone, int cpu_online)
{
int new_high, new_batch;
new_batch = max(1, zone_batchsize(zone));
- new_high = zone_highsize(zone, new_batch);
+ new_high = zone_highsize(zone, new_batch, cpu_online);
if (zone->pageset_high == new_high &&
zone->pageset_batch == new_batch)
@@ -6789,7 +6790,7 @@ void __meminit setup_zone_pageset(struct
per_cpu_pages_init(pcp, pzstats);
}
- zone_set_pageset_high_and_batch(zone);
+ zone_set_pageset_high_and_batch(zone, 0);
}
/*
@@ -8044,6 +8045,7 @@ void __init set_dma_reserve(unsigned lon
static int page_alloc_cpu_dead(unsigned int cpu)
{
+ struct zone *zone;
lru_add_drain_cpu(cpu);
drain_pages(cpu);
@@ -8064,6 +8066,19 @@ static int page_alloc_cpu_dead(unsigned
* race with what we are doing.
*/
cpu_vm_stats_fold(cpu);
+
+ for_each_populated_zone(zone)
+ zone_pcp_update(zone, 0);
+
+ return 0;
+}
+
+static int page_alloc_cpu_online(unsigned int cpu)
+{
+ struct zone *zone;
+
+ for_each_populated_zone(zone)
+ zone_pcp_update(zone, 1);
return 0;
}
@@ -8089,8 +8104,9 @@ void __init page_alloc_init(void)
hashdist = 0;
#endif
- ret = cpuhp_setup_state_nocalls(CPUHP_PAGE_ALLOC_DEAD,
- "mm/page_alloc:dead", NULL,
+ ret = cpuhp_setup_state_nocalls(CPUHP_PAGE_ALLOC,
+ "mm/page_alloc:pcp",
+ page_alloc_cpu_online,
page_alloc_cpu_dead);
WARN_ON(ret < 0);
}
@@ -8252,7 +8268,7 @@ void setup_per_zone_wmarks(void)
* and high limits or the limits may be inappropriate.
*/
for_each_zone(zone)
- zone_pcp_update(zone);
+ zone_pcp_update(zone, 0);
}
/*
@@ -9053,10 +9069,10 @@ EXPORT_SYMBOL(free_contig_range);
* The zone indicated has a new number of managed_pages; batch sizes and percpu
* page high values need to be recalculated.
*/
-void __meminit zone_pcp_update(struct zone *zone)
+void zone_pcp_update(struct zone *zone, int cpu_online)
{
mutex_lock(&pcp_batch_high_lock);
- zone_set_pageset_high_and_batch(zone);
+ zone_set_pageset_high_and_batch(zone, cpu_online);
mutex_unlock(&pcp_batch_high_lock);
}
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 173/192] mm/page_alloc: scale the number of pages that are batch freed
2021-06-29 2:32 incoming Andrew Morton
` (171 preceding siblings ...)
2021-06-29 2:42 ` [patch 172/192] mm/page_alloc: adjust pcp->high after CPU hotplug events Andrew Morton
@ 2021-06-29 2:42 ` Andrew Morton
2021-06-29 2:42 ` [patch 174/192] mm/page_alloc: limit the number of pages on PCP lists when reclaim is active Andrew Morton
` (18 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:42 UTC (permalink / raw)
To: akpm, dave.hansen, hdanton, linux-mm, mgorman, mhocko,
mm-commits, torvalds, vbabka
From: Mel Gorman <mgorman@techsingularity.net>
Subject: mm/page_alloc: scale the number of pages that are batch freed
When a task is freeing a large number of order-0 pages, it may acquire the
zone->lock multiple times freeing pages in batches. This may
unnecessarily contend on the zone lock when freeing very large number of
pages. This patch adapts the size of the batch based on the recent
pattern to scale the batch size for subsequent frees.
As the machines I used were not large enough to test this are not large
enough to illustrate a problem, a debugging patch shows patterns like the
following (slightly editted for clarity)
Baseline vanilla kernel
time-unmap-14426 [...] free_pcppages_bulk: free 63 count 378 high 378
time-unmap-14426 [...] free_pcppages_bulk: free 63 count 378 high 378
time-unmap-14426 [...] free_pcppages_bulk: free 63 count 378 high 378
time-unmap-14426 [...] free_pcppages_bulk: free 63 count 378 high 378
time-unmap-14426 [...] free_pcppages_bulk: free 63 count 378 high 378
With patches
time-unmap-7724 [...] free_pcppages_bulk: free 126 count 814 high 814
time-unmap-7724 [...] free_pcppages_bulk: free 252 count 814 high 814
time-unmap-7724 [...] free_pcppages_bulk: free 504 count 814 high 814
time-unmap-7724 [...] free_pcppages_bulk: free 751 count 814 high 814
time-unmap-7724 [...] free_pcppages_bulk: free 751 count 814 high 814
Link: https://lkml.kernel.org/r/20210525080119.5455-5-mgorman@techsingularity.net
Signed-off-by: Mel Gorman <mgorman@techsingularity.net>
Acked-by: Dave Hansen <dave.hansen@linux.intel.com>
Acked-by: Vlastimil Babka <vbabka@suse.cz>
Cc: Hillf Danton <hdanton@sina.com>
Cc: Michal Hocko <mhocko@kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/linux/mmzone.h | 3 +-
mm/page_alloc.c | 41 +++++++++++++++++++++++++++++++++++++--
2 files changed, 41 insertions(+), 3 deletions(-)
--- a/include/linux/mmzone.h~mm-page_alloc-scale-the-number-of-pages-that-are-batch-freed
+++ a/include/linux/mmzone.h
@@ -343,8 +343,9 @@ struct per_cpu_pages {
int count; /* number of pages in the list */
int high; /* high watermark, emptying needed */
int batch; /* chunk size for buddy add/remove */
+ short free_factor; /* batch scaling factor during free */
#ifdef CONFIG_NUMA
- int expire; /* When 0, remote pagesets are drained */
+ short expire; /* When 0, remote pagesets are drained */
#endif
/* Lists of pages, one per migrate type stored on the pcp-lists */
--- a/mm/page_alloc.c~mm-page_alloc-scale-the-number-of-pages-that-are-batch-freed
+++ a/mm/page_alloc.c
@@ -3278,18 +3278,47 @@ static bool free_unref_page_prepare(stru
return true;
}
+static int nr_pcp_free(struct per_cpu_pages *pcp, int high, int batch)
+{
+ int min_nr_free, max_nr_free;
+
+ /* Check for PCP disabled or boot pageset */
+ if (unlikely(high < batch))
+ return 1;
+
+ /* Leave at least pcp->batch pages on the list */
+ min_nr_free = batch;
+ max_nr_free = high - batch;
+
+ /*
+ * Double the number of pages freed each time there is subsequent
+ * freeing of pages without any allocation.
+ */
+ batch <<= pcp->free_factor;
+ if (batch < max_nr_free)
+ pcp->free_factor++;
+ batch = clamp(batch, min_nr_free, max_nr_free);
+
+ return batch;
+}
+
static void free_unref_page_commit(struct page *page, unsigned long pfn,
int migratetype)
{
struct zone *zone = page_zone(page);
struct per_cpu_pages *pcp;
+ int high;
__count_vm_event(PGFREE);
pcp = this_cpu_ptr(zone->per_cpu_pageset);
list_add(&page->lru, &pcp->lists[migratetype]);
pcp->count++;
- if (pcp->count >= READ_ONCE(pcp->high))
- free_pcppages_bulk(zone, READ_ONCE(pcp->batch), pcp);
+ high = READ_ONCE(pcp->high);
+ if (pcp->count >= high) {
+ int batch = READ_ONCE(pcp->batch);
+
+ free_pcppages_bulk(zone, nr_pcp_free(pcp, high, batch), pcp);
+ }
}
/*
@@ -3541,7 +3570,14 @@ static struct page *rmqueue_pcplist(stru
unsigned long flags;
local_lock_irqsave(&pagesets.lock, flags);
+
+ /*
+ * On allocation, reduce the number of pages that are batch freed.
+ * See nr_pcp_free() where free_factor is increased for subsequent
+ * frees.
+ */
pcp = this_cpu_ptr(zone->per_cpu_pageset);
+ pcp->free_factor >>= 1;
list = &pcp->lists[migratetype];
page = __rmqueue_pcplist(zone, migratetype, alloc_flags, pcp, list);
local_unlock_irqrestore(&pagesets.lock, flags);
@@ -6737,6 +6773,7 @@ static void per_cpu_pages_init(struct pe
*/
pcp->high = BOOT_PAGESET_HIGH;
pcp->batch = BOOT_PAGESET_BATCH;
+ pcp->free_factor = 0;
}
static void __zone_set_pageset_high_and_batch(struct zone *zone, unsigned long high,
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 174/192] mm/page_alloc: limit the number of pages on PCP lists when reclaim is active
2021-06-29 2:32 incoming Andrew Morton
` (172 preceding siblings ...)
2021-06-29 2:42 ` [patch 173/192] mm/page_alloc: scale the number of pages that are batch freed Andrew Morton
@ 2021-06-29 2:42 ` Andrew Morton
2021-06-29 2:42 ` [patch 175/192] mm/page_alloc: introduce vm.percpu_pagelist_high_fraction Andrew Morton
` (17 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:42 UTC (permalink / raw)
To: akpm, dave.hansen, hdanton, linux-mm, mgorman, mhocko,
mm-commits, torvalds, vbabka
From: Mel Gorman <mgorman@techsingularity.net>
Subject: mm/page_alloc: limit the number of pages on PCP lists when reclaim is active
When kswapd is active then direct reclaim is potentially active. In
either case, it is possible that a zone would be balanced if pages were
not trapped on PCP lists. Instead of draining remote pages, simply limit
the size of the PCP lists while kswapd is active.
Link: https://lkml.kernel.org/r/20210525080119.5455-6-mgorman@techsingularity.net
Signed-off-by: Mel Gorman <mgorman@techsingularity.net>
Acked-by: Vlastimil Babka <vbabka@suse.cz>
Cc: Dave Hansen <dave.hansen@linux.intel.com>
Cc: Hillf Danton <hdanton@sina.com>
Cc: Michal Hocko <mhocko@kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/linux/mmzone.h | 1 +
mm/page_alloc.c | 19 ++++++++++++++++++-
mm/vmscan.c | 35 +++++++++++++++++++++++++++++++++++
3 files changed, 54 insertions(+), 1 deletion(-)
--- a/include/linux/mmzone.h~mm-page_alloc-limit-the-number-of-pages-on-pcp-lists-when-reclaim-is-active
+++ a/include/linux/mmzone.h
@@ -647,6 +647,7 @@ enum zone_flags {
ZONE_BOOSTED_WATERMARK, /* zone recently boosted watermarks.
* Cleared when kswapd is woken.
*/
+ ZONE_RECLAIM_ACTIVE, /* kswapd may be scanning the zone. */
};
static inline unsigned long zone_managed_pages(struct zone *zone)
--- a/mm/page_alloc.c~mm-page_alloc-limit-the-number-of-pages-on-pcp-lists-when-reclaim-is-active
+++ a/mm/page_alloc.c
@@ -3302,6 +3302,23 @@ static int nr_pcp_free(struct per_cpu_pa
return batch;
}
+static int nr_pcp_high(struct per_cpu_pages *pcp, struct zone *zone)
+{
+ int high = READ_ONCE(pcp->high);
+
+ if (unlikely(!high))
+ return 0;
+
+ if (!test_bit(ZONE_RECLAIM_ACTIVE, &zone->flags))
+ return high;
+
+ /*
+ * If reclaim is active, limit the number of pages that can be
+ * stored on pcp lists
+ */
+ return min(READ_ONCE(pcp->batch) << 2, high);
+}
+
static void free_unref_page_commit(struct page *page, unsigned long pfn,
int migratetype)
{
@@ -3313,7 +3330,7 @@ static void free_unref_page_commit(struc
pcp = this_cpu_ptr(zone->per_cpu_pageset);
list_add(&page->lru, &pcp->lists[migratetype]);
pcp->count++;
- high = READ_ONCE(pcp->high);
+ high = nr_pcp_high(pcp, zone);
if (pcp->count >= high) {
int batch = READ_ONCE(pcp->batch);
--- a/mm/vmscan.c~mm-page_alloc-limit-the-number-of-pages-on-pcp-lists-when-reclaim-is-active
+++ a/mm/vmscan.c
@@ -3722,6 +3722,38 @@ static bool kswapd_shrink_node(pg_data_t
return sc->nr_scanned >= sc->nr_to_reclaim;
}
+/* Page allocator PCP high watermark is lowered if reclaim is active. */
+static inline void
+update_reclaim_active(pg_data_t *pgdat, int highest_zoneidx, bool active)
+{
+ int i;
+ struct zone *zone;
+
+ for (i = 0; i <= highest_zoneidx; i++) {
+ zone = pgdat->node_zones + i;
+
+ if (!managed_zone(zone))
+ continue;
+
+ if (active)
+ set_bit(ZONE_RECLAIM_ACTIVE, &zone->flags);
+ else
+ clear_bit(ZONE_RECLAIM_ACTIVE, &zone->flags);
+ }
+}
+
+static inline void
+set_reclaim_active(pg_data_t *pgdat, int highest_zoneidx)
+{
+ update_reclaim_active(pgdat, highest_zoneidx, true);
+}
+
+static inline void
+clear_reclaim_active(pg_data_t *pgdat, int highest_zoneidx)
+{
+ update_reclaim_active(pgdat, highest_zoneidx, false);
+}
+
/*
* For kswapd, balance_pgdat() will reclaim pages across a node from zones
* that are eligible for use by the caller until at least one zone is
@@ -3774,6 +3806,7 @@ static int balance_pgdat(pg_data_t *pgda
boosted = nr_boost_reclaim;
restart:
+ set_reclaim_active(pgdat, highest_zoneidx);
sc.priority = DEF_PRIORITY;
do {
unsigned long nr_reclaimed = sc.nr_reclaimed;
@@ -3907,6 +3940,8 @@ restart:
pgdat->kswapd_failures++;
out:
+ clear_reclaim_active(pgdat, highest_zoneidx);
+
/* If reclaim was boosted, account for the reclaim done in this pass */
if (boosted) {
unsigned long flags;
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 175/192] mm/page_alloc: introduce vm.percpu_pagelist_high_fraction
2021-06-29 2:32 incoming Andrew Morton
` (173 preceding siblings ...)
2021-06-29 2:42 ` [patch 174/192] mm/page_alloc: limit the number of pages on PCP lists when reclaim is active Andrew Morton
@ 2021-06-29 2:42 ` Andrew Morton
2021-06-29 2:42 ` [patch 176/192] mm: drop SECTION_SHIFT in code comments Andrew Morton
` (16 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:42 UTC (permalink / raw)
To: akpm, dave.hansen, hdanton, linux-mm, mgorman, mhocko,
mm-commits, torvalds, vbabka
From: Mel Gorman <mgorman@techsingularity.net>
Subject: mm/page_alloc: introduce vm.percpu_pagelist_high_fraction
This introduces a new sysctl vm.percpu_pagelist_high_fraction. It is
similar to the old vm.percpu_pagelist_fraction. The old sysctl increased
both pcp->batch and pcp->high with the higher pcp->high potentially
reducing zone->lock contention. However, the higher pcp->batch value also
potentially increased allocation latency while the PCP was refilled. This
sysctl only adjusts pcp->high so that zone->lock contention is potentially
reduced but allocation latency during a PCP refill remains the same.
# grep -E "high:|batch" /proc/zoneinfo | tail -2
high: 649
batch: 63
# sysctl vm.percpu_pagelist_high_fraction=8
# grep -E "high:|batch" /proc/zoneinfo | tail -2
high: 35071
batch: 63
# sysctl vm.percpu_pagelist_high_fraction=64
high: 4383
batch: 63
# sysctl vm.percpu_pagelist_high_fraction=0
high: 649
batch: 63
[mgorman@techsingularity.net: fix documentation]
Link: https://lkml.kernel.org/r/20210528151010.GQ30378@techsingularity.net
Link: https://lkml.kernel.org/r/20210525080119.5455-7-mgorman@techsingularity.net
Signed-off-by: Mel Gorman <mgorman@techsingularity.net>
Acked-by: Dave Hansen <dave.hansen@linux.intel.com>
Acked-by: Vlastimil Babka <vbabka@suse.cz>
Cc: Hillf Danton <hdanton@sina.com>
Cc: Michal Hocko <mhocko@kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
Documentation/admin-guide/sysctl/vm.rst | 21 ++++++
include/linux/mmzone.h | 3
kernel/sysctl.c | 8 ++
mm/page_alloc.c | 69 +++++++++++++++++++---
4 files changed, 94 insertions(+), 7 deletions(-)
--- a/Documentation/admin-guide/sysctl/vm.rst~mm-page_alloc-introduce-vmpercpu_pagelist_high_fraction
+++ a/Documentation/admin-guide/sysctl/vm.rst
@@ -64,6 +64,7 @@ Currently, these files are in /proc/sys/
- overcommit_ratio
- page-cluster
- panic_on_oom
+- percpu_pagelist_high_fraction
- stat_interval
- stat_refresh
- numa_stat
@@ -789,6 +790,26 @@ panic_on_oom=2+kdump gives you very stro
why oom happens. You can get snapshot.
+percpu_pagelist_high_fraction
+=============================
+
+This is the fraction of pages in each zone that are can be stored to
+per-cpu page lists. It is an upper boundary that is divided depending
+on the number of online CPUs. The min value for this is 8 which means
+that we do not allow more than 1/8th of pages in each zone to be stored
+on per-cpu page lists. This entry only changes the value of hot per-cpu
+page lists. A user can specify a number like 100 to allocate 1/100th of
+each zone between per-cpu lists.
+
+The batch value of each per-cpu page list remains the same regardless of
+the value of the high fraction so allocation latencies are unaffected.
+
+The initial value is zero. Kernel uses this value to set the high pcp->high
+mark based on the low watermark for the zone and the number of local
+online CPUs. If the user writes '0' to this sysctl, it will revert to
+this default behavior.
+
+
stat_interval
=============
--- a/include/linux/mmzone.h~mm-page_alloc-introduce-vmpercpu_pagelist_high_fraction
+++ a/include/linux/mmzone.h
@@ -1029,12 +1029,15 @@ int watermark_scale_factor_sysctl_handle
extern int sysctl_lowmem_reserve_ratio[MAX_NR_ZONES];
int lowmem_reserve_ratio_sysctl_handler(struct ctl_table *, int, void *,
size_t *, loff_t *);
+int percpu_pagelist_high_fraction_sysctl_handler(struct ctl_table *, int,
+ void *, size_t *, loff_t *);
int sysctl_min_unmapped_ratio_sysctl_handler(struct ctl_table *, int,
void *, size_t *, loff_t *);
int sysctl_min_slab_ratio_sysctl_handler(struct ctl_table *, int,
void *, size_t *, loff_t *);
int numa_zonelist_order_handler(struct ctl_table *, int,
void *, size_t *, loff_t *);
+extern int percpu_pagelist_high_fraction;
extern char numa_zonelist_order[];
#define NUMA_ZONELIST_ORDER_LEN 16
--- a/kernel/sysctl.c~mm-page_alloc-introduce-vmpercpu_pagelist_high_fraction
+++ a/kernel/sysctl.c
@@ -2909,6 +2909,14 @@ static struct ctl_table vm_table[] = {
.extra2 = &one_thousand,
},
{
+ .procname = "percpu_pagelist_high_fraction",
+ .data = &percpu_pagelist_high_fraction,
+ .maxlen = sizeof(percpu_pagelist_high_fraction),
+ .mode = 0644,
+ .proc_handler = percpu_pagelist_high_fraction_sysctl_handler,
+ .extra1 = SYSCTL_ZERO,
+ },
+ {
.procname = "page_lock_unfairness",
.data = &sysctl_page_lock_unfairness,
.maxlen = sizeof(sysctl_page_lock_unfairness),
--- a/mm/page_alloc.c~mm-page_alloc-introduce-vmpercpu_pagelist_high_fraction
+++ a/mm/page_alloc.c
@@ -120,6 +120,7 @@ typedef int __bitwise fpi_t;
/* prevent >1 _updater_ of zone percpu pageset ->high and ->batch fields */
static DEFINE_MUTEX(pcp_batch_high_lock);
+#define MIN_PERCPU_PAGELIST_HIGH_FRACTION (8)
struct pagesets {
local_lock_t lock;
@@ -192,6 +193,7 @@ EXPORT_SYMBOL(_totalram_pages);
unsigned long totalreserve_pages __read_mostly;
unsigned long totalcma_pages __read_mostly;
+int percpu_pagelist_high_fraction;
gfp_t gfp_allowed_mask __read_mostly = GFP_BOOT_MASK;
DEFINE_STATIC_KEY_MAYBE(CONFIG_INIT_ON_ALLOC_DEFAULT_ON, init_on_alloc);
EXPORT_SYMBOL(init_on_alloc);
@@ -6725,17 +6727,32 @@ static int zone_highsize(struct zone *zo
#ifdef CONFIG_MMU
int high;
int nr_local_cpus;
+ unsigned long total_pages;
+
+ if (!percpu_pagelist_high_fraction) {
+ /*
+ * By default, the high value of the pcp is based on the zone
+ * low watermark so that if they are full then background
+ * reclaim will not be started prematurely.
+ */
+ total_pages = low_wmark_pages(zone);
+ } else {
+ /*
+ * If percpu_pagelist_high_fraction is configured, the high
+ * value is based on a fraction of the managed pages in the
+ * zone.
+ */
+ total_pages = zone_managed_pages(zone) / percpu_pagelist_high_fraction;
+ }
/*
- * The high value of the pcp is based on the zone low watermark
- * so that if they are full then background reclaim will not be
- * started prematurely. The value is split across all online CPUs
- * local to the zone. Note that early in boot that CPUs may not be
- * online yet and that during CPU hotplug that the cpumask is not
- * yet updated when a CPU is being onlined.
+ * Split the high value across all online CPUs local to the zone. Note
+ * that early in boot that CPUs may not be online yet and that during
+ * CPU hotplug that the cpumask is not yet updated when a CPU is being
+ * onlined.
*/
nr_local_cpus = max(1U, cpumask_weight(cpumask_of_node(zone_to_nid(zone)))) + cpu_online;
- high = low_wmark_pages(zone) / nr_local_cpus;
+ high = total_pages / nr_local_cpus;
/*
* Ensure high is at least batch*4. The multiple is based on the
@@ -8500,6 +8517,44 @@ int lowmem_reserve_ratio_sysctl_handler(
return 0;
}
+/*
+ * percpu_pagelist_high_fraction - changes the pcp->high for each zone on each
+ * cpu. It is the fraction of total pages in each zone that a hot per cpu
+ * pagelist can have before it gets flushed back to buddy allocator.
+ */
+int percpu_pagelist_high_fraction_sysctl_handler(struct ctl_table *table,
+ int write, void *buffer, size_t *length, loff_t *ppos)
+{
+ struct zone *zone;
+ int old_percpu_pagelist_high_fraction;
+ int ret;
+
+ mutex_lock(&pcp_batch_high_lock);
+ old_percpu_pagelist_high_fraction = percpu_pagelist_high_fraction;
+
+ ret = proc_dointvec_minmax(table, write, buffer, length, ppos);
+ if (!write || ret < 0)
+ goto out;
+
+ /* Sanity checking to avoid pcp imbalance */
+ if (percpu_pagelist_high_fraction &&
+ percpu_pagelist_high_fraction < MIN_PERCPU_PAGELIST_HIGH_FRACTION) {
+ percpu_pagelist_high_fraction = old_percpu_pagelist_high_fraction;
+ ret = -EINVAL;
+ goto out;
+ }
+
+ /* No change? */
+ if (percpu_pagelist_high_fraction == old_percpu_pagelist_high_fraction)
+ goto out;
+
+ for_each_populated_zone(zone)
+ zone_set_pageset_high_and_batch(zone, 0);
+out:
+ mutex_unlock(&pcp_batch_high_lock);
+ return ret;
+}
+
#ifndef __HAVE_ARCH_RESERVED_KERNEL_PAGES
/*
* Returns the number of pages that arch has reserved but
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 176/192] mm: drop SECTION_SHIFT in code comments
2021-06-29 2:32 incoming Andrew Morton
` (174 preceding siblings ...)
2021-06-29 2:42 ` [patch 175/192] mm/page_alloc: introduce vm.percpu_pagelist_high_fraction Andrew Morton
@ 2021-06-29 2:42 ` Andrew Morton
2021-06-29 2:42 ` [patch 177/192] mm/page_alloc: improve memmap_pages dbg msg Andrew Morton
` (15 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:42 UTC (permalink / raw)
To: aisheng.dong, akpm, andreyknvl, catalin.marinas, keescook,
linux-mm, mm-commits, torvalds, yuzhao
From: Dong Aisheng <aisheng.dong@nxp.com>
Subject: mm: drop SECTION_SHIFT in code comments
Actually SECTIONS_SHIFT is used in the kernel code, so the code comments
is strictly incorrect. And since commit bbeae5b05ef6 ("mm: move page
flags layout to separate header"), SECTIONS_SHIFT definition has been
moved to include/linux/page-flags-layout.h, since code itself looks quite
straighforward, instead of moving the code comment into the new place as
well, we just simply remove it.
This also fixed a checkpatch complain derived from the original code:
WARNING: please, no space before tabs
+ * SECTIONS_SHIFT ^I^I#bits space required to store a section #$
Link: https://lkml.kernel.org/r/20210531091908.1738465-2-aisheng.dong@nxp.com
Signed-off-by: Dong Aisheng <aisheng.dong@nxp.com>
Suggested-by: Yu Zhao <yuzhao@google.com>
Reviewed-by: Yu Zhao <yuzhao@google.com>
Cc: Andrey Konovalov <andreyknvl@gmail.com>
Cc: Catalin Marinas <catalin.marinas@arm.com>
Cc: Kees Cook <keescook@chromium.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/linux/mmzone.h | 2 --
1 file changed, 2 deletions(-)
--- a/include/linux/mmzone.h~mm-drop-section_shift-in-code-comments
+++ a/include/linux/mmzone.h
@@ -1200,8 +1200,6 @@ static inline struct zoneref *first_zone
#ifdef CONFIG_SPARSEMEM
/*
- * SECTION_SHIFT #bits space required to store a section #
- *
* PA_SECTION_SHIFT physical address to/from section number
* PFN_SECTION_SHIFT pfn to/from section number
*/
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 177/192] mm/page_alloc: improve memmap_pages dbg msg
2021-06-29 2:32 incoming Andrew Morton
` (175 preceding siblings ...)
2021-06-29 2:42 ` [patch 176/192] mm: drop SECTION_SHIFT in code comments Andrew Morton
@ 2021-06-29 2:42 ` Andrew Morton
2021-06-29 2:42 ` [patch 178/192] mm/page_alloc: fix counting of managed_pages Andrew Morton
` (14 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:42 UTC (permalink / raw)
To: aisheng.dong, akpm, david, linux-mm, mm-commits, torvalds
From: Dong Aisheng <aisheng.dong@nxp.com>
Subject: mm/page_alloc: improve memmap_pages dbg msg
Make debug message more accurate.
Link: https://lkml.kernel.org/r/20210531091908.1738465-6-aisheng.dong@nxp.com
Signed-off-by: Dong Aisheng <aisheng.dong@nxp.com>
Reviewed-by: David Hildenbrand <david@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/page_alloc.c | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
--- a/mm/page_alloc.c~mm-page_alloc-improve-memmap_pages-dbg-msg
+++ a/mm/page_alloc.c
@@ -7383,7 +7383,7 @@ static void __init free_area_init_core(s
pr_debug(" %s zone: %lu pages used for memmap\n",
zone_names[j], memmap_pages);
} else
- pr_warn(" %s zone: %lu pages exceeds freesize %lu\n",
+ pr_warn(" %s zone: %lu memmap pages exceeds freesize %lu\n",
zone_names[j], memmap_pages, freesize);
}
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 178/192] mm/page_alloc: fix counting of managed_pages
2021-06-29 2:32 incoming Andrew Morton
` (176 preceding siblings ...)
2021-06-29 2:42 ` [patch 177/192] mm/page_alloc: improve memmap_pages dbg msg Andrew Morton
@ 2021-06-29 2:42 ` Andrew Morton
2021-06-29 2:42 ` [patch 179/192] mm/page_alloc: move free_the_page Andrew Morton
` (13 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:42 UTC (permalink / raw)
To: akpm, bhe, david, linux-mm, liushixin2, mm-commits, torvalds, yangerkun
From: Liu Shixin <liushixin2@huawei.com>
Subject: mm/page_alloc: fix counting of managed_pages
commit f63661566fad ("mm/page_alloc.c: clear out zone->lowmem_reserve[] if
the zone is empty") clears out zone->lowmem_reserve[] if zone is empty.
But when zone is not empty and sysctl_lowmem_reserve_ratio[i] is set to
zero, zone_managed_pages(zone) is not counted in the managed_pages either.
This is inconsistent with the description of lowmem_reserve, so fix it.
Link: https://lkml.kernel.org/r/20210527125707.3760259-1-liushixin2@huawei.com
Fixes: f63661566fad ("mm/page_alloc.c: clear out zone->lowmem_reserve[] if the zone is empty")
Signed-off-by: Liu Shixin <liushixin2@huawei.com>
Reported-by: yangerkun <yangerkun@huawei.com>
Reviewed-by: Baoquan He <bhe@redhat.com>
Acked-by: David Hildenbrand <david@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/page_alloc.c | 12 ++++++------
1 file changed, 6 insertions(+), 6 deletions(-)
--- a/mm/page_alloc.c~mm-page_alloc-fix-counting-of-managed_pages
+++ a/mm/page_alloc.c
@@ -8240,14 +8240,14 @@ static void setup_per_zone_lowmem_reserv
unsigned long managed_pages = 0;
for (j = i + 1; j < MAX_NR_ZONES; j++) {
- if (clear) {
- zone->lowmem_reserve[j] = 0;
- } else {
- struct zone *upper_zone = &pgdat->node_zones[j];
+ struct zone *upper_zone = &pgdat->node_zones[j];
+
+ managed_pages += zone_managed_pages(upper_zone);
- managed_pages += zone_managed_pages(upper_zone);
+ if (clear)
+ zone->lowmem_reserve[j] = 0;
+ else
zone->lowmem_reserve[j] = managed_pages / ratio;
- }
}
}
}
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 179/192] mm/page_alloc: move free_the_page
2021-06-29 2:32 incoming Andrew Morton
` (177 preceding siblings ...)
2021-06-29 2:42 ` [patch 178/192] mm/page_alloc: fix counting of managed_pages Andrew Morton
@ 2021-06-29 2:42 ` Andrew Morton
2021-06-29 2:42 ` [patch 180/192] alpha: remove DISCONTIGMEM and NUMA Andrew Morton
` (12 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:42 UTC (permalink / raw)
To: akpm, brouer, dave.hansen, linux-mm, mgorman, mhocko, mm-commits,
torvalds, vbabka
From: Mel Gorman <mgorman@techsingularity.net>
Subject: mm/page_alloc: move free_the_page
Patch series "Allow high order pages to be stored on PCP", v2.
The per-cpu page allocator (PCP) only handles order-0 pages. With the
series "Use local_lock for pcp protection and reduce stat overhead" and
"Calculate pcp->high based on zone sizes and active CPUs", it's now
feasible to store high-order pages on PCP lists.
This small series allows PCP to store "cheap" orders where cheap is
determined by PAGE_ALLOC_COSTLY_ORDER and THP-sized allocations.
This patch (of 2):
In the next page, free_compount_page is going to use the common helper
free_the_page. This patch moves the definition to ease review. No
functional change.
Link: https://lkml.kernel.org/r/20210603142220.10851-1-mgorman@techsingularity.net
Link: https://lkml.kernel.org/r/20210603142220.10851-2-mgorman@techsingularity.net
Signed-off-by: Mel Gorman <mgorman@techsingularity.net>
Acked-by: Vlastimil Babka <vbabka@suse.cz>
Cc: Dave Hansen <dave.hansen@linux.intel.com>
Cc: Jesper Dangaard Brouer <brouer@redhat.com>
Cc: Michal Hocko <mhocko@kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/page_alloc.c | 16 ++++++++--------
1 file changed, 8 insertions(+), 8 deletions(-)
--- a/mm/page_alloc.c~mm-page_alloc-move-free_the_page
+++ a/mm/page_alloc.c
@@ -687,6 +687,14 @@ out:
add_taint(TAINT_BAD_PAGE, LOCKDEP_NOW_UNRELIABLE);
}
+static inline void free_the_page(struct page *page, unsigned int order)
+{
+ if (order == 0) /* Via pcp? */
+ free_unref_page(page);
+ else
+ __free_pages_ok(page, order, FPI_NONE);
+}
+
/*
* Higher-order pages are called "compound pages". They are structured thusly:
*
@@ -5349,14 +5357,6 @@ unsigned long get_zeroed_page(gfp_t gfp_
}
EXPORT_SYMBOL(get_zeroed_page);
-static inline void free_the_page(struct page *page, unsigned int order)
-{
- if (order == 0) /* Via pcp? */
- free_unref_page(page);
- else
- __free_pages_ok(page, order, FPI_NONE);
-}
-
/**
* __free_pages - Free pages allocated with alloc_pages().
* @page: The page pointer returned from alloc_pages().
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 180/192] alpha: remove DISCONTIGMEM and NUMA
2021-06-29 2:32 incoming Andrew Morton
` (178 preceding siblings ...)
2021-06-29 2:42 ` [patch 179/192] mm/page_alloc: move free_the_page Andrew Morton
@ 2021-06-29 2:42 ` Andrew Morton
2021-06-29 2:42 ` [patch 181/192] arc: update comment about HIGHMEM implementation Andrew Morton
` (11 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:42 UTC (permalink / raw)
To: akpm, arnd, corbet, david, geert, ink, linux-mm, mattst88,
mm-commits, rppt, rth, torvalds, vgupta
From: Mike Rapoport <rppt@linux.ibm.com>
Subject: alpha: remove DISCONTIGMEM and NUMA
Patch series "Remove DISCONTIGMEM memory model", v3.
SPARSEMEM memory model was supposed to entirely replace DISCONTIGMEM a
(long) while ago. The last architectures that used DISCONTIGMEM were
updated to use other memory models in v5.11 and it is about the time to
entirely remove DISCONTIGMEM from the kernel.
This set removes DISCONTIGMEM from alpha, arc and m68k, simplifies memory
model selection in mm/Kconfig and replaces usage of redundant
CONFIG_NEED_MULTIPLE_NODES and CONFIG_FLAT_NODE_MEM_MAP with CONFIG_NUMA
and CONFIG_FLATMEM respectively.
I've also removed NUMA support on alpha that was BROKEN for more than 15
years.
There were also minor updates all over arch/ to remove mentions of
DISCONTIGMEM in comments and #ifdefs.
This patch (of 9):
NUMA is marked broken on alpha for more than 15 years and DISCONTIGMEM was
replaced with SPARSEMEM in v5.11.
Remove both NUMA and DISCONTIGMEM support from alpha.
Link: https://lkml.kernel.org/r/20210608091316.3622-1-rppt@kernel.org
Link: https://lkml.kernel.org/r/20210608091316.3622-2-rppt@kernel.org
Signed-off-by: Mike Rapoport <rppt@linux.ibm.com>
Acked-by: Arnd Bergmann <arnd@arndb.de>
Acked-by: David Hildenbrand <david@redhat.com>
Cc: Geert Uytterhoeven <geert@linux-m68k.org>
Cc: Ivan Kokshaysky <ink@jurassic.park.msu.ru>
Cc: Jonathan Corbet <corbet@lwn.net>
Cc: Matt Turner <mattst88@gmail.com>
Cc: Richard Henderson <rth@twiddle.net>
Cc: Vineet Gupta <vgupta@synopsys.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
arch/alpha/Kconfig | 22 --
arch/alpha/include/asm/machvec.h | 6
arch/alpha/include/asm/mmzone.h | 100 ------------
arch/alpha/include/asm/pgtable.h | 4
arch/alpha/include/asm/topology.h | 39 ----
arch/alpha/kernel/core_marvel.c | 53 ------
arch/alpha/kernel/core_wildfire.c | 29 ---
arch/alpha/kernel/pci_iommu.c | 29 ---
arch/alpha/kernel/proto.h | 8 -
arch/alpha/kernel/setup.c | 16 --
arch/alpha/kernel/sys_marvel.c | 5
arch/alpha/kernel/sys_wildfire.c | 5
arch/alpha/mm/Makefile | 2
arch/alpha/mm/init.c | 3
arch/alpha/mm/numa.c | 223 ----------------------------
15 files changed, 4 insertions(+), 540 deletions(-)
--- a/arch/alpha/include/asm/machvec.h~alpha-remove-discontigmem-and-numa
+++ a/arch/alpha/include/asm/machvec.h
@@ -99,12 +99,6 @@ struct alpha_machine_vector
const char *vector_name;
- /* NUMA information */
- int (*pa_to_nid)(unsigned long);
- int (*cpuid_to_nid)(int);
- unsigned long (*node_mem_start)(int);
- unsigned long (*node_mem_size)(int);
-
/* System specific parameters. */
union {
struct {
--- a/arch/alpha/include/asm/mmzone.h
+++ /dev/null
@@ -1,100 +0,0 @@
-/* SPDX-License-Identifier: GPL-2.0 */
-/*
- * Written by Kanoj Sarcar (kanoj@sgi.com) Aug 99
- * Adapted for the alpha wildfire architecture Jan 2001.
- */
-#ifndef _ASM_MMZONE_H_
-#define _ASM_MMZONE_H_
-
-#ifdef CONFIG_DISCONTIGMEM
-
-#include <asm/smp.h>
-
-/*
- * Following are macros that are specific to this numa platform.
- */
-
-extern pg_data_t node_data[];
-
-#define alpha_pa_to_nid(pa) \
- (alpha_mv.pa_to_nid \
- ? alpha_mv.pa_to_nid(pa) \
- : (0))
-#define node_mem_start(nid) \
- (alpha_mv.node_mem_start \
- ? alpha_mv.node_mem_start(nid) \
- : (0UL))
-#define node_mem_size(nid) \
- (alpha_mv.node_mem_size \
- ? alpha_mv.node_mem_size(nid) \
- : ((nid) ? (0UL) : (~0UL)))
-
-#define pa_to_nid(pa) alpha_pa_to_nid(pa)
-#define NODE_DATA(nid) (&node_data[(nid)])
-
-#define node_localnr(pfn, nid) ((pfn) - NODE_DATA(nid)->node_start_pfn)
-
-#if 1
-#define PLAT_NODE_DATA_LOCALNR(p, n) \
- (((p) >> PAGE_SHIFT) - PLAT_NODE_DATA(n)->gendata.node_start_pfn)
-#else
-static inline unsigned long
-PLAT_NODE_DATA_LOCALNR(unsigned long p, int n)
-{
- unsigned long temp;
- temp = p >> PAGE_SHIFT;
- return temp - PLAT_NODE_DATA(n)->gendata.node_start_pfn;
-}
-#endif
-
-/*
- * Following are macros that each numa implementation must define.
- */
-
-/*
- * Given a kernel address, find the home node of the underlying memory.
- */
-#define kvaddr_to_nid(kaddr) pa_to_nid(__pa(kaddr))
-
-/*
- * Given a kaddr, LOCAL_BASE_ADDR finds the owning node of the memory
- * and returns the kaddr corresponding to first physical page in the
- * node's mem_map.
- */
-#define LOCAL_BASE_ADDR(kaddr) \
- ((unsigned long)__va(NODE_DATA(kvaddr_to_nid(kaddr))->node_start_pfn \
- << PAGE_SHIFT))
-
-/* XXX: FIXME -- nyc */
-#define kern_addr_valid(kaddr) (0)
-
-#define mk_pte(page, pgprot) \
-({ \
- pte_t pte; \
- unsigned long pfn; \
- \
- pfn = page_to_pfn(page) << 32; \
- pte_val(pte) = pfn | pgprot_val(pgprot); \
- \
- pte; \
-})
-
-#define pte_page(x) \
-({ \
- unsigned long kvirt; \
- struct page * __xx; \
- \
- kvirt = (unsigned long)__va(pte_val(x) >> (32-PAGE_SHIFT)); \
- __xx = virt_to_page(kvirt); \
- \
- __xx; \
-})
-
-#define pfn_to_nid(pfn) pa_to_nid(((u64)(pfn) << PAGE_SHIFT))
-#define pfn_valid(pfn) \
- (((pfn) - node_start_pfn(pfn_to_nid(pfn))) < \
- node_spanned_pages(pfn_to_nid(pfn))) \
-
-#endif /* CONFIG_DISCONTIGMEM */
-
-#endif /* _ASM_MMZONE_H_ */
--- a/arch/alpha/include/asm/pgtable.h~alpha-remove-discontigmem-and-numa
+++ a/arch/alpha/include/asm/pgtable.h
@@ -206,7 +206,6 @@ extern unsigned long __zero_page(void);
#define page_to_pa(page) (page_to_pfn(page) << PAGE_SHIFT)
#define pte_pfn(pte) (pte_val(pte) >> 32)
-#ifndef CONFIG_DISCONTIGMEM
#define pte_page(pte) pfn_to_page(pte_pfn(pte))
#define mk_pte(page, pgprot) \
({ \
@@ -215,7 +214,6 @@ extern unsigned long __zero_page(void);
pte_val(pte) = (page_to_pfn(page) << 32) | pgprot_val(pgprot); \
pte; \
})
-#endif
extern inline pte_t pfn_pte(unsigned long physpfn, pgprot_t pgprot)
{ pte_t pte; pte_val(pte) = (PHYS_TWIDDLE(physpfn) << 32) | pgprot_val(pgprot); return pte; }
@@ -330,9 +328,7 @@ extern inline pte_t mk_swap_pte(unsigned
#define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) })
#define __swp_entry_to_pte(x) ((pte_t) { (x).val })
-#ifndef CONFIG_DISCONTIGMEM
#define kern_addr_valid(addr) (1)
-#endif
#define pte_ERROR(e) \
printk("%s:%d: bad pte %016lx.\n", __FILE__, __LINE__, pte_val(e))
--- a/arch/alpha/include/asm/topology.h~alpha-remove-discontigmem-and-numa
+++ a/arch/alpha/include/asm/topology.h
@@ -7,45 +7,6 @@
#include <linux/numa.h>
#include <asm/machvec.h>
-#ifdef CONFIG_NUMA
-static inline int cpu_to_node(int cpu)
-{
- int node;
-
- if (!alpha_mv.cpuid_to_nid)
- return 0;
-
- node = alpha_mv.cpuid_to_nid(cpu);
-
-#ifdef DEBUG_NUMA
- BUG_ON(node < 0);
-#endif
-
- return node;
-}
-
-extern struct cpumask node_to_cpumask_map[];
-/* FIXME: This is dumb, recalculating every time. But simple. */
-static const struct cpumask *cpumask_of_node(int node)
-{
- int cpu;
-
- if (node == NUMA_NO_NODE)
- return cpu_all_mask;
-
- cpumask_clear(&node_to_cpumask_map[node]);
-
- for_each_online_cpu(cpu) {
- if (cpu_to_node(cpu) == node)
- cpumask_set_cpu(cpu, node_to_cpumask_map[node]);
- }
-
- return &node_to_cpumask_map[node];
-}
-
-#define cpumask_of_pcibus(bus) (cpu_online_mask)
-
-#endif /* !CONFIG_NUMA */
# include <asm-generic/topology.h>
#endif /* _ASM_ALPHA_TOPOLOGY_H */
--- a/arch/alpha/Kconfig~alpha-remove-discontigmem-and-numa
+++ a/arch/alpha/Kconfig
@@ -549,29 +549,12 @@ config NR_CPUS
MARVEL support can handle a maximum of 32 CPUs, all the others
with working support have a maximum of 4 CPUs.
-config ARCH_DISCONTIGMEM_ENABLE
- bool "Discontiguous Memory Support"
- depends on BROKEN
- help
- Say Y to support efficient handling of discontiguous physical memory,
- for architectures which are either NUMA (Non-Uniform Memory Access)
- or have huge holes in the physical address space for other reasons.
- See <file:Documentation/vm/numa.rst> for more.
-
config ARCH_SPARSEMEM_ENABLE
bool "Sparse Memory Support"
help
Say Y to support efficient handling of discontiguous physical memory,
for systems that have huge holes in the physical address space.
-config NUMA
- bool "NUMA Support (EXPERIMENTAL)"
- depends on DISCONTIGMEM && BROKEN
- help
- Say Y to compile the kernel to support NUMA (Non-Uniform Memory
- Access). This option is for configuring high-end multiprocessor
- server machines. If in doubt, say N.
-
config ALPHA_WTINT
bool "Use WTINT" if ALPHA_SRM || ALPHA_GENERIC
default y if ALPHA_QEMU
@@ -596,11 +579,6 @@ config ALPHA_WTINT
If unsure, say N.
-config NODES_SHIFT
- int
- default "7"
- depends on NEED_MULTIPLE_NODES
-
# LARGE_VMALLOC is racy, if you *really* need it then fix it first
config ALPHA_LARGE_VMALLOC
bool
--- a/arch/alpha/kernel/core_marvel.c~alpha-remove-discontigmem-and-numa
+++ a/arch/alpha/kernel/core_marvel.c
@@ -287,8 +287,7 @@ io7_init_hose(struct io7 *io7, int port)
/*
* Set up window 0 for scatter-gather 8MB at 8MB.
*/
- hose->sg_isa = iommu_arena_new_node(marvel_cpuid_to_nid(io7->pe),
- hose, 0x00800000, 0x00800000, 0);
+ hose->sg_isa = iommu_arena_new_node(0, hose, 0x00800000, 0x00800000, 0);
hose->sg_isa->align_entry = 8; /* cache line boundary */
csrs->POx_WBASE[0].csr =
hose->sg_isa->dma_base | wbase_m_ena | wbase_m_sg;
@@ -305,8 +304,7 @@ io7_init_hose(struct io7 *io7, int port)
/*
* Set up window 2 for scatter-gather (up-to) 1GB at 3GB.
*/
- hose->sg_pci = iommu_arena_new_node(marvel_cpuid_to_nid(io7->pe),
- hose, 0xc0000000, 0x40000000, 0);
+ hose->sg_pci = iommu_arena_new_node(0, hose, 0xc0000000, 0x40000000, 0);
hose->sg_pci->align_entry = 8; /* cache line boundary */
csrs->POx_WBASE[2].csr =
hose->sg_pci->dma_base | wbase_m_ena | wbase_m_sg;
@@ -843,53 +841,8 @@ EXPORT_SYMBOL(marvel_ioportmap);
EXPORT_SYMBOL(marvel_ioread8);
EXPORT_SYMBOL(marvel_iowrite8);
#endif
-\f
-/*
- * NUMA Support
- */
-/**********
- * FIXME - for now each cpu is a node by itself
- * -- no real support for striped mode
- **********
- */
-int
-marvel_pa_to_nid(unsigned long pa)
-{
- int cpuid;
- if ((pa >> 43) & 1) /* I/O */
- cpuid = (~(pa >> 35) & 0xff);
- else /* mem */
- cpuid = ((pa >> 34) & 0x3) | ((pa >> (37 - 2)) & (0x1f << 2));
-
- return marvel_cpuid_to_nid(cpuid);
-}
-
-int
-marvel_cpuid_to_nid(int cpuid)
-{
- return cpuid;
-}
-
-unsigned long
-marvel_node_mem_start(int nid)
-{
- unsigned long pa;
-
- pa = (nid & 0x3) | ((nid & (0x1f << 2)) << 1);
- pa <<= 34;
-
- return pa;
-}
-
-unsigned long
-marvel_node_mem_size(int nid)
-{
- return 16UL * 1024 * 1024 * 1024; /* 16GB */
-}
-
-\f
-/*
+/*
* AGP GART Support.
*/
#include <linux/agp_backend.h>
--- a/arch/alpha/kernel/core_wildfire.c~alpha-remove-discontigmem-and-numa
+++ a/arch/alpha/kernel/core_wildfire.c
@@ -434,39 +434,12 @@ wildfire_write_config(struct pci_bus *bu
return PCIBIOS_SUCCESSFUL;
}
-struct pci_ops wildfire_pci_ops =
+struct pci_ops wildfire_pci_ops =
{
.read = wildfire_read_config,
.write = wildfire_write_config,
};
-\f
-/*
- * NUMA Support
- */
-int wildfire_pa_to_nid(unsigned long pa)
-{
- return pa >> 36;
-}
-
-int wildfire_cpuid_to_nid(int cpuid)
-{
- /* assume 4 CPUs per node */
- return cpuid >> 2;
-}
-
-unsigned long wildfire_node_mem_start(int nid)
-{
- /* 64GB per node */
- return (unsigned long)nid * (64UL * 1024 * 1024 * 1024);
-}
-
-unsigned long wildfire_node_mem_size(int nid)
-{
- /* 64GB per node */
- return 64UL * 1024 * 1024 * 1024;
-}
-
#if DEBUG_DUMP_REGS
static void __init
--- a/arch/alpha/kernel/pci_iommu.c~alpha-remove-discontigmem-and-numa
+++ a/arch/alpha/kernel/pci_iommu.c
@@ -71,33 +71,6 @@ iommu_arena_new_node(int nid, struct pci
if (align < mem_size)
align = mem_size;
-
-#ifdef CONFIG_DISCONTIGMEM
-
- arena = memblock_alloc_node(sizeof(*arena), align, nid);
- if (!NODE_DATA(nid) || !arena) {
- printk("%s: couldn't allocate arena from node %d\n"
- " falling back to system-wide allocation\n",
- __func__, nid);
- arena = memblock_alloc(sizeof(*arena), SMP_CACHE_BYTES);
- if (!arena)
- panic("%s: Failed to allocate %zu bytes\n", __func__,
- sizeof(*arena));
- }
-
- arena->ptes = memblock_alloc_node(sizeof(*arena), align, nid);
- if (!NODE_DATA(nid) || !arena->ptes) {
- printk("%s: couldn't allocate arena ptes from node %d\n"
- " falling back to system-wide allocation\n",
- __func__, nid);
- arena->ptes = memblock_alloc(mem_size, align);
- if (!arena->ptes)
- panic("%s: Failed to allocate %lu bytes align=0x%lx\n",
- __func__, mem_size, align);
- }
-
-#else /* CONFIG_DISCONTIGMEM */
-
arena = memblock_alloc(sizeof(*arena), SMP_CACHE_BYTES);
if (!arena)
panic("%s: Failed to allocate %zu bytes\n", __func__,
@@ -107,8 +80,6 @@ iommu_arena_new_node(int nid, struct pci
panic("%s: Failed to allocate %lu bytes align=0x%lx\n",
__func__, mem_size, align);
-#endif /* CONFIG_DISCONTIGMEM */
-
spin_lock_init(&arena->lock);
arena->hose = hose;
arena->dma_base = base;
--- a/arch/alpha/kernel/proto.h~alpha-remove-discontigmem-and-numa
+++ a/arch/alpha/kernel/proto.h
@@ -49,10 +49,6 @@ extern void marvel_init_arch(void);
extern void marvel_kill_arch(int);
extern void marvel_machine_check(unsigned long, unsigned long);
extern void marvel_pci_tbi(struct pci_controller *, dma_addr_t, dma_addr_t);
-extern int marvel_pa_to_nid(unsigned long);
-extern int marvel_cpuid_to_nid(int);
-extern unsigned long marvel_node_mem_start(int);
-extern unsigned long marvel_node_mem_size(int);
extern struct _alpha_agp_info *marvel_agp_info(void);
struct io7 *marvel_find_io7(int pe);
struct io7 *marvel_next_io7(struct io7 *prev);
@@ -101,10 +97,6 @@ extern void wildfire_init_arch(void);
extern void wildfire_kill_arch(int);
extern void wildfire_machine_check(unsigned long vector, unsigned long la_ptr);
extern void wildfire_pci_tbi(struct pci_controller *, dma_addr_t, dma_addr_t);
-extern int wildfire_pa_to_nid(unsigned long);
-extern int wildfire_cpuid_to_nid(int);
-extern unsigned long wildfire_node_mem_start(int);
-extern unsigned long wildfire_node_mem_size(int);
/* console.c */
#ifdef CONFIG_VGA_HOSE
--- a/arch/alpha/kernel/setup.c~alpha-remove-discontigmem-and-numa
+++ a/arch/alpha/kernel/setup.c
@@ -79,11 +79,6 @@ int alpha_l3_cacheshape;
unsigned long alpha_verbose_mcheck = CONFIG_VERBOSE_MCHECK_ON;
#endif
-#ifdef CONFIG_NUMA
-struct cpumask node_to_cpumask_map[MAX_NUMNODES] __read_mostly;
-EXPORT_SYMBOL(node_to_cpumask_map);
-#endif
-
/* Which processor we booted from. */
int boot_cpuid;
@@ -305,7 +300,6 @@ move_initrd(unsigned long mem_limit)
}
#endif
-#ifndef CONFIG_DISCONTIGMEM
static void __init
setup_memory(void *kernel_end)
{
@@ -389,9 +383,6 @@ setup_memory(void *kernel_end)
}
#endif /* CONFIG_BLK_DEV_INITRD */
}
-#else
-extern void setup_memory(void *);
-#endif /* !CONFIG_DISCONTIGMEM */
int __init
page_is_ram(unsigned long pfn)
@@ -618,13 +609,6 @@ setup_arch(char **cmdline_p)
"VERBOSE_MCHECK "
#endif
-#ifdef CONFIG_DISCONTIGMEM
- "DISCONTIGMEM "
-#ifdef CONFIG_NUMA
- "NUMA "
-#endif
-#endif
-
#ifdef CONFIG_DEBUG_SPINLOCK
"DEBUG_SPINLOCK "
#endif
--- a/arch/alpha/kernel/sys_marvel.c~alpha-remove-discontigmem-and-numa
+++ a/arch/alpha/kernel/sys_marvel.c
@@ -461,10 +461,5 @@ struct alpha_machine_vector marvel_ev7_m
.kill_arch = marvel_kill_arch,
.pci_map_irq = marvel_map_irq,
.pci_swizzle = common_swizzle,
-
- .pa_to_nid = marvel_pa_to_nid,
- .cpuid_to_nid = marvel_cpuid_to_nid,
- .node_mem_start = marvel_node_mem_start,
- .node_mem_size = marvel_node_mem_size,
};
ALIAS_MV(marvel_ev7)
--- a/arch/alpha/kernel/sys_wildfire.c~alpha-remove-discontigmem-and-numa
+++ a/arch/alpha/kernel/sys_wildfire.c
@@ -337,10 +337,5 @@ struct alpha_machine_vector wildfire_mv
.kill_arch = wildfire_kill_arch,
.pci_map_irq = wildfire_map_irq,
.pci_swizzle = common_swizzle,
-
- .pa_to_nid = wildfire_pa_to_nid,
- .cpuid_to_nid = wildfire_cpuid_to_nid,
- .node_mem_start = wildfire_node_mem_start,
- .node_mem_size = wildfire_node_mem_size,
};
ALIAS_MV(wildfire)
--- a/arch/alpha/mm/init.c~alpha-remove-discontigmem-and-numa
+++ a/arch/alpha/mm/init.c
@@ -235,8 +235,6 @@ callback_init(void * kernel_end)
return kernel_end;
}
-
-#ifndef CONFIG_DISCONTIGMEM
/*
* paging_init() sets up the memory map.
*/
@@ -257,7 +255,6 @@ void __init paging_init(void)
/* Initialize the kernel's ZERO_PGE. */
memset((void *)ZERO_PGE, 0, PAGE_SIZE);
}
-#endif /* CONFIG_DISCONTIGMEM */
#if defined(CONFIG_ALPHA_GENERIC) || defined(CONFIG_ALPHA_SRM)
void
--- a/arch/alpha/mm/Makefile~alpha-remove-discontigmem-and-numa
+++ a/arch/alpha/mm/Makefile
@@ -6,5 +6,3 @@
ccflags-y := -Werror
obj-y := init.o fault.o
-
-obj-$(CONFIG_DISCONTIGMEM) += numa.o
--- a/arch/alpha/mm/numa.c
+++ /dev/null
@@ -1,223 +0,0 @@
-// SPDX-License-Identifier: GPL-2.0
-/*
- * linux/arch/alpha/mm/numa.c
- *
- * DISCONTIGMEM NUMA alpha support.
- *
- * Copyright (C) 2001 Andrea Arcangeli <andrea@suse.de> SuSE
- */
-
-#include <linux/types.h>
-#include <linux/kernel.h>
-#include <linux/mm.h>
-#include <linux/memblock.h>
-#include <linux/swap.h>
-#include <linux/initrd.h>
-#include <linux/pfn.h>
-#include <linux/module.h>
-
-#include <asm/hwrpb.h>
-#include <asm/sections.h>
-
-pg_data_t node_data[MAX_NUMNODES];
-EXPORT_SYMBOL(node_data);
-
-#undef DEBUG_DISCONTIG
-#ifdef DEBUG_DISCONTIG
-#define DBGDCONT(args...) printk(args)
-#else
-#define DBGDCONT(args...)
-#endif
-
-#define for_each_mem_cluster(memdesc, _cluster, i) \
- for ((_cluster) = (memdesc)->cluster, (i) = 0; \
- (i) < (memdesc)->numclusters; (i)++, (_cluster)++)
-
-static void __init show_mem_layout(void)
-{
- struct memclust_struct * cluster;
- struct memdesc_struct * memdesc;
- int i;
-
- /* Find free clusters, and init and free the bootmem accordingly. */
- memdesc = (struct memdesc_struct *)
- (hwrpb->mddt_offset + (unsigned long) hwrpb);
-
- printk("Raw memory layout:\n");
- for_each_mem_cluster(memdesc, cluster, i) {
- printk(" memcluster %2d, usage %1lx, start %8lu, end %8lu\n",
- i, cluster->usage, cluster->start_pfn,
- cluster->start_pfn + cluster->numpages);
- }
-}
-
-static void __init
-setup_memory_node(int nid, void *kernel_end)
-{
- extern unsigned long mem_size_limit;
- struct memclust_struct * cluster;
- struct memdesc_struct * memdesc;
- unsigned long start_kernel_pfn, end_kernel_pfn;
- unsigned long start, end;
- unsigned long node_pfn_start, node_pfn_end;
- unsigned long node_min_pfn, node_max_pfn;
- int i;
- int show_init = 0;
-
- /* Find the bounds of current node */
- node_pfn_start = (node_mem_start(nid)) >> PAGE_SHIFT;
- node_pfn_end = node_pfn_start + (node_mem_size(nid) >> PAGE_SHIFT);
-
- /* Find free clusters, and init and free the bootmem accordingly. */
- memdesc = (struct memdesc_struct *)
- (hwrpb->mddt_offset + (unsigned long) hwrpb);
-
- /* find the bounds of this node (node_min_pfn/node_max_pfn) */
- node_min_pfn = ~0UL;
- node_max_pfn = 0UL;
- for_each_mem_cluster(memdesc, cluster, i) {
- /* Bit 0 is console/PALcode reserved. Bit 1 is
- non-volatile memory -- we might want to mark
- this for later. */
- if (cluster->usage & 3)
- continue;
-
- start = cluster->start_pfn;
- end = start + cluster->numpages;
-
- if (start >= node_pfn_end || end <= node_pfn_start)
- continue;
-
- if (!show_init) {
- show_init = 1;
- printk("Initializing bootmem allocator on Node ID %d\n", nid);
- }
- printk(" memcluster %2d, usage %1lx, start %8lu, end %8lu\n",
- i, cluster->usage, cluster->start_pfn,
- cluster->start_pfn + cluster->numpages);
-
- if (start < node_pfn_start)
- start = node_pfn_start;
- if (end > node_pfn_end)
- end = node_pfn_end;
-
- if (start < node_min_pfn)
- node_min_pfn = start;
- if (end > node_max_pfn)
- node_max_pfn = end;
- }
-
- if (mem_size_limit && node_max_pfn > mem_size_limit) {
- static int msg_shown = 0;
- if (!msg_shown) {
- msg_shown = 1;
- printk("setup: forcing memory size to %ldK (from %ldK).\n",
- mem_size_limit << (PAGE_SHIFT - 10),
- node_max_pfn << (PAGE_SHIFT - 10));
- }
- node_max_pfn = mem_size_limit;
- }
-
- if (node_min_pfn >= node_max_pfn)
- return;
-
- /* Update global {min,max}_low_pfn from node information. */
- if (node_min_pfn < min_low_pfn)
- min_low_pfn = node_min_pfn;
- if (node_max_pfn > max_low_pfn)
- max_pfn = max_low_pfn = node_max_pfn;
-
-#if 0 /* we'll try this one again in a little while */
- /* Cute trick to make sure our local node data is on local memory */
- node_data[nid] = (pg_data_t *)(__va(node_min_pfn << PAGE_SHIFT));
-#endif
- printk(" Detected node memory: start %8lu, end %8lu\n",
- node_min_pfn, node_max_pfn);
-
- DBGDCONT(" DISCONTIG: node_data[%d] is at 0x%p\n", nid, NODE_DATA(nid));
-
- /* Find the bounds of kernel memory. */
- start_kernel_pfn = PFN_DOWN(KERNEL_START_PHYS);
- end_kernel_pfn = PFN_UP(virt_to_phys(kernel_end));
-
- if (!nid && (node_max_pfn < end_kernel_pfn || node_min_pfn > start_kernel_pfn))
- panic("kernel loaded out of ram");
-
- memblock_add_node(PFN_PHYS(node_min_pfn),
- (node_max_pfn - node_min_pfn) << PAGE_SHIFT, nid);
-
- /* Zone start phys-addr must be 2^(MAX_ORDER-1) aligned.
- Note that we round this down, not up - node memory
- has much larger alignment than 8Mb, so it's safe. */
- node_min_pfn &= ~((1UL << (MAX_ORDER-1))-1);
-
- NODE_DATA(nid)->node_start_pfn = node_min_pfn;
- NODE_DATA(nid)->node_present_pages = node_max_pfn - node_min_pfn;
-
- node_set_online(nid);
-}
-
-void __init
-setup_memory(void *kernel_end)
-{
- unsigned long kernel_size;
- int nid;
-
- show_mem_layout();
-
- nodes_clear(node_online_map);
-
- min_low_pfn = ~0UL;
- max_low_pfn = 0UL;
- for (nid = 0; nid < MAX_NUMNODES; nid++)
- setup_memory_node(nid, kernel_end);
-
- kernel_size = virt_to_phys(kernel_end) - KERNEL_START_PHYS;
- memblock_reserve(KERNEL_START_PHYS, kernel_size);
-
-#ifdef CONFIG_BLK_DEV_INITRD
- initrd_start = INITRD_START;
- if (initrd_start) {
- extern void *move_initrd(unsigned long);
-
- initrd_end = initrd_start+INITRD_SIZE;
- printk("Initial ramdisk at: 0x%p (%lu bytes)\n",
- (void *) initrd_start, INITRD_SIZE);
-
- if ((void *)initrd_end > phys_to_virt(PFN_PHYS(max_low_pfn))) {
- if (!move_initrd(PFN_PHYS(max_low_pfn)))
- printk("initrd extends beyond end of memory "
- "(0x%08lx > 0x%p)\ndisabling initrd\n",
- initrd_end,
- phys_to_virt(PFN_PHYS(max_low_pfn)));
- } else {
- nid = kvaddr_to_nid(initrd_start);
- memblock_reserve(virt_to_phys((void *)initrd_start),
- INITRD_SIZE);
- }
- }
-#endif /* CONFIG_BLK_DEV_INITRD */
-}
-
-void __init paging_init(void)
-{
- unsigned long max_zone_pfn[MAX_NR_ZONES] = {0, };
- unsigned long dma_local_pfn;
-
- /*
- * The old global MAX_DMA_ADDRESS per-arch API doesn't fit
- * in the NUMA model, for now we convert it to a pfn and
- * we interpret this pfn as a local per-node information.
- * This issue isn't very important since none of these machines
- * have legacy ISA slots anyways.
- */
- dma_local_pfn = virt_to_phys((char *)MAX_DMA_ADDRESS) >> PAGE_SHIFT;
-
- max_zone_pfn[ZONE_DMA] = dma_local_pfn;
- max_zone_pfn[ZONE_NORMAL] = max_pfn;
-
- free_area_init(max_zone_pfn);
-
- /* Initialize the kernel's ZERO_PGE. */
- memset((void *)ZERO_PGE, 0, PAGE_SIZE);
-}
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 181/192] arc: update comment about HIGHMEM implementation
2021-06-29 2:32 incoming Andrew Morton
` (179 preceding siblings ...)
2021-06-29 2:42 ` [patch 180/192] alpha: remove DISCONTIGMEM and NUMA Andrew Morton
@ 2021-06-29 2:42 ` Andrew Morton
2021-06-29 2:42 ` [patch 182/192] arc: remove support for DISCONTIGMEM Andrew Morton
` (10 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:42 UTC (permalink / raw)
To: akpm, arnd, corbet, david, geert, ink, linux-mm, mattst88,
mm-commits, rppt, rth, torvalds, vgupta
From: Mike Rapoport <rppt@linux.ibm.com>
Subject: arc: update comment about HIGHMEM implementation
Arc does not use DISCONTIGMEM to implement high memory, update the comment
describing how high memory works to reflect this.
Link: https://lkml.kernel.org/r/20210608091316.3622-3-rppt@kernel.org
Signed-off-by: Mike Rapoport <rppt@linux.ibm.com>
Acked-by: Vineet Gupta <vgupta@synopsys.com>
Acked-by: Arnd Bergmann <arnd@arndb.de>
Reviewed-by: David Hildenbrand <david@redhat.com>
Cc: Geert Uytterhoeven <geert@linux-m68k.org>
Cc: Ivan Kokshaysky <ink@jurassic.park.msu.ru>
Cc: Jonathan Corbet <corbet@lwn.net>
Cc: Matt Turner <mattst88@gmail.com>
Cc: Richard Henderson <rth@twiddle.net>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
arch/arc/mm/init.c | 13 +++++--------
1 file changed, 5 insertions(+), 8 deletions(-)
--- a/arch/arc/mm/init.c~arc-update-comment-about-highmem-implementation
+++ a/arch/arc/mm/init.c
@@ -139,16 +139,13 @@ void __init setup_arch_memory(void)
#ifdef CONFIG_HIGHMEM
/*
- * Populate a new node with highmem
- *
* On ARC (w/o PAE) HIGHMEM addresses are actually smaller (0 based)
- * than addresses in normal ala low memory (0x8000_0000 based).
+ * than addresses in normal aka low memory (0x8000_0000 based).
* Even with PAE, the huge peripheral space hole would waste a lot of
- * mem with single mem_map[]. This warrants a mem_map per region design.
- * Thus HIGHMEM on ARC is imlemented with DISCONTIGMEM.
- *
- * DISCONTIGMEM in turns requires multiple nodes. node 0 above is
- * populated with normal memory zone while node 1 only has highmem
+ * mem with single contiguous mem_map[].
+ * Thus when HIGHMEM on ARC is enabled the memory map corresponding
+ * to the hole is freed and ARC specific version of pfn_valid()
+ * handles the hole in the memory map.
*/
#ifdef CONFIG_DISCONTIGMEM
node_set_online(1);
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 182/192] arc: remove support for DISCONTIGMEM
2021-06-29 2:32 incoming Andrew Morton
` (180 preceding siblings ...)
2021-06-29 2:42 ` [patch 181/192] arc: update comment about HIGHMEM implementation Andrew Morton
@ 2021-06-29 2:42 ` Andrew Morton
2021-06-29 2:42 ` [patch 183/192] m68k: " Andrew Morton
` (9 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:42 UTC (permalink / raw)
To: akpm, arnd, corbet, david, geert, ink, linux-mm, mattst88,
mm-commits, rppt, rth, torvalds, vgupta
From: Mike Rapoport <rppt@linux.ibm.com>
Subject: arc: remove support for DISCONTIGMEM
DISCONTIGMEM was replaced by FLATMEM with freeing of the unused memory map
in v5.11.
Remove the support for DISCONTIGMEM entirely.
Link: https://lkml.kernel.org/r/20210608091316.3622-4-rppt@kernel.org
Signed-off-by: Mike Rapoport <rppt@linux.ibm.com>
Acked-by: Vineet Gupta <vgupta@synopsys.com>
Acked-by: Arnd Bergmann <arnd@arndb.de>
Acked-by: David Hildenbrand <david@redhat.com>
Cc: Geert Uytterhoeven <geert@linux-m68k.org>
Cc: Ivan Kokshaysky <ink@jurassic.park.msu.ru>
Cc: Jonathan Corbet <corbet@lwn.net>
Cc: Matt Turner <mattst88@gmail.com>
Cc: Richard Henderson <rth@twiddle.net>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
arch/arc/Kconfig | 13 ----------
arch/arc/include/asm/mmzone.h | 40 --------------------------------
arch/arc/mm/init.c | 8 ------
3 files changed, 61 deletions(-)
--- a/arch/arc/include/asm/mmzone.h
+++ /dev/null
@@ -1,40 +0,0 @@
-/* SPDX-License-Identifier: GPL-2.0-only */
-/*
- * Copyright (C) 2016 Synopsys, Inc. (www.synopsys.com)
- */
-
-#ifndef _ASM_ARC_MMZONE_H
-#define _ASM_ARC_MMZONE_H
-
-#ifdef CONFIG_DISCONTIGMEM
-
-extern struct pglist_data node_data[];
-#define NODE_DATA(nid) (&node_data[nid])
-
-static inline int pfn_to_nid(unsigned long pfn)
-{
- int is_end_low = 1;
-
- if (IS_ENABLED(CONFIG_ARC_HAS_PAE40))
- is_end_low = pfn <= virt_to_pfn(0xFFFFFFFFUL);
-
- /*
- * node 0: lowmem: 0x8000_0000 to 0xFFFF_FFFF
- * node 1: HIGHMEM w/o PAE40: 0x0 to 0x7FFF_FFFF
- * HIGHMEM with PAE40: 0x1_0000_0000 to ...
- */
- if (pfn >= ARCH_PFN_OFFSET && is_end_low)
- return 0;
-
- return 1;
-}
-
-static inline int pfn_valid(unsigned long pfn)
-{
- int nid = pfn_to_nid(pfn);
-
- return (pfn <= node_end_pfn(nid));
-}
-#endif /* CONFIG_DISCONTIGMEM */
-
-#endif
--- a/arch/arc/Kconfig~arc-remove-support-for-discontigmem
+++ a/arch/arc/Kconfig
@@ -62,10 +62,6 @@ config SCHED_OMIT_FRAME_POINTER
config GENERIC_CSUM
def_bool y
-config ARCH_DISCONTIGMEM_ENABLE
- def_bool n
- depends on BROKEN
-
config ARCH_FLATMEM_ENABLE
def_bool y
@@ -344,15 +340,6 @@ config ARC_HUGEPAGE_16M
endchoice
-config NODES_SHIFT
- int "Maximum NUMA Nodes (as a power of 2)"
- default "0" if !DISCONTIGMEM
- default "1" if DISCONTIGMEM
- depends on NEED_MULTIPLE_NODES
- help
- Accessing memory beyond 1GB (with or w/o PAE) requires 2 memory
- zones.
-
config ARC_COMPACT_IRQ_LEVELS
depends on ISA_ARCOMPACT
bool "Setup Timer IRQ as high Priority"
--- a/arch/arc/mm/init.c~arc-remove-support-for-discontigmem
+++ a/arch/arc/mm/init.c
@@ -32,11 +32,6 @@ unsigned long arch_pfn_offset;
EXPORT_SYMBOL(arch_pfn_offset);
#endif
-#ifdef CONFIG_DISCONTIGMEM
-struct pglist_data node_data[MAX_NUMNODES] __read_mostly;
-EXPORT_SYMBOL(node_data);
-#endif
-
long __init arc_get_mem_sz(void)
{
return low_mem_sz;
@@ -147,9 +142,6 @@ void __init setup_arch_memory(void)
* to the hole is freed and ARC specific version of pfn_valid()
* handles the hole in the memory map.
*/
-#ifdef CONFIG_DISCONTIGMEM
- node_set_online(1);
-#endif
min_high_pfn = PFN_DOWN(high_mem_start);
max_high_pfn = PFN_DOWN(high_mem_start + high_mem_sz);
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 183/192] m68k: remove support for DISCONTIGMEM
2021-06-29 2:32 incoming Andrew Morton
` (181 preceding siblings ...)
2021-06-29 2:42 ` [patch 182/192] arc: remove support for DISCONTIGMEM Andrew Morton
@ 2021-06-29 2:42 ` Andrew Morton
2021-06-29 2:42 ` [patch 184/192] mm: remove CONFIG_DISCONTIGMEM Andrew Morton
` (8 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:42 UTC (permalink / raw)
To: akpm, arnd, corbet, geert, ink, linux-mm, mattst88, mm-commits,
rppt, rth, torvalds, vgupta
From: Mike Rapoport <rppt@linux.ibm.com>
Subject: m68k: remove support for DISCONTIGMEM
DISCONTIGMEM was replaced by FLATMEM with freeing of the unused memory map
in v5.11.
Remove the support for DISCONTIGMEM entirely.
Link: https://lkml.kernel.org/r/20210608091316.3622-5-rppt@kernel.org
Signed-off-by: Mike Rapoport <rppt@linux.ibm.com>
Reviewed-by: Geert Uytterhoeven <geert@linux-m68k.org>
Acked-by: Geert Uytterhoeven <geert@linux-m68k.org>
Acked-by: Arnd Bergmann <arnd@arndb.de>
Cc: Ivan Kokshaysky <ink@jurassic.park.msu.ru>
Cc: Jonathan Corbet <corbet@lwn.net>
Cc: Matt Turner <mattst88@gmail.com>
Cc: Richard Henderson <rth@twiddle.net>
Cc: Vineet Gupta <vgupta@synopsys.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
arch/m68k/Kconfig.cpu | 10 --------
arch/m68k/include/asm/mmzone.h | 10 --------
arch/m68k/include/asm/page.h | 2 -
arch/m68k/include/asm/page_mm.h | 35 ------------------------------
arch/m68k/mm/init.c | 20 -----------------
5 files changed, 1 insertion(+), 76 deletions(-)
--- a/arch/m68k/include/asm/mmzone.h
+++ /dev/null
@@ -1,10 +0,0 @@
-/* SPDX-License-Identifier: GPL-2.0 */
-#ifndef _ASM_M68K_MMZONE_H_
-#define _ASM_M68K_MMZONE_H_
-
-extern pg_data_t pg_data_map[];
-
-#define NODE_DATA(nid) (&pg_data_map[nid])
-#define NODE_MEM_MAP(nid) (NODE_DATA(nid)->node_mem_map)
-
-#endif /* _ASM_M68K_MMZONE_H_ */
--- a/arch/m68k/include/asm/page.h~m68k-remove-support-for-discontigmem
+++ a/arch/m68k/include/asm/page.h
@@ -62,7 +62,7 @@ extern unsigned long _ramend;
#include <asm/page_no.h>
#endif
-#if !defined(CONFIG_MMU) || defined(CONFIG_DISCONTIGMEM)
+#ifndef CONFIG_MMU
#define __phys_to_pfn(paddr) ((unsigned long)((paddr) >> PAGE_SHIFT))
#define __pfn_to_phys(pfn) PFN_PHYS(pfn)
#endif
--- a/arch/m68k/include/asm/page_mm.h~m68k-remove-support-for-discontigmem
+++ a/arch/m68k/include/asm/page_mm.h
@@ -126,26 +126,6 @@ static inline void *__va(unsigned long x
extern int m68k_virt_to_node_shift;
-#ifndef CONFIG_DISCONTIGMEM
-#define __virt_to_node(addr) (&pg_data_map[0])
-#else
-extern struct pglist_data *pg_data_table[];
-
-static inline __attribute_const__ int __virt_to_node_shift(void)
-{
- int shift;
-
- asm (
- "1: moveq #0,%0\n"
- m68k_fixup(%c1, 1b)
- : "=d" (shift)
- : "i" (m68k_fixup_vnode_shift));
- return shift;
-}
-
-#define __virt_to_node(addr) (pg_data_table[(unsigned long)(addr) >> __virt_to_node_shift()])
-#endif
-
#define virt_to_page(addr) ({ \
pfn_to_page(virt_to_pfn(addr)); \
})
@@ -153,23 +133,8 @@ static inline __attribute_const__ int __
pfn_to_virt(page_to_pfn(page)); \
})
-#ifdef CONFIG_DISCONTIGMEM
-#define pfn_to_page(pfn) ({ \
- unsigned long __pfn = (pfn); \
- struct pglist_data *pgdat; \
- pgdat = __virt_to_node((unsigned long)pfn_to_virt(__pfn)); \
- pgdat->node_mem_map + (__pfn - pgdat->node_start_pfn); \
-})
-#define page_to_pfn(_page) ({ \
- const struct page *__p = (_page); \
- struct pglist_data *pgdat; \
- pgdat = &pg_data_map[page_to_nid(__p)]; \
- ((__p) - pgdat->node_mem_map) + pgdat->node_start_pfn; \
-})
-#else
#define ARCH_PFN_OFFSET (m68k_memory[0].addr >> PAGE_SHIFT)
#include <asm-generic/memory_model.h>
-#endif
#define virt_addr_valid(kaddr) ((unsigned long)(kaddr) >= PAGE_OFFSET && (unsigned long)(kaddr) < (unsigned long)high_memory)
#define pfn_valid(pfn) virt_addr_valid(pfn_to_virt(pfn))
--- a/arch/m68k/Kconfig.cpu~m68k-remove-support-for-discontigmem
+++ a/arch/m68k/Kconfig.cpu
@@ -408,10 +408,6 @@ config SINGLE_MEMORY_CHUNK
order" to save memory that could be wasted for unused memory map.
Say N if not sure.
-config ARCH_DISCONTIGMEM_ENABLE
- depends on BROKEN
- def_bool MMU && !SINGLE_MEMORY_CHUNK
-
config FORCE_MAX_ZONEORDER
int "Maximum zone order" if ADVANCED
depends on !SINGLE_MEMORY_CHUNK
@@ -451,11 +447,6 @@ config M68K_L2_CACHE
depends on MAC
default y
-config NODES_SHIFT
- int
- default "3"
- depends on DISCONTIGMEM
-
config CPU_HAS_NO_BITFIELDS
bool
@@ -553,4 +544,3 @@ config CACHE_COPYBACK
The ColdFire CPU cache is set into Copy-back mode.
endchoice
endif
-
--- a/arch/m68k/mm/init.c~m68k-remove-support-for-discontigmem
+++ a/arch/m68k/mm/init.c
@@ -44,28 +44,8 @@ EXPORT_SYMBOL(empty_zero_page);
int m68k_virt_to_node_shift;
-#ifdef CONFIG_DISCONTIGMEM
-pg_data_t pg_data_map[MAX_NUMNODES];
-EXPORT_SYMBOL(pg_data_map);
-
-pg_data_t *pg_data_table[65];
-EXPORT_SYMBOL(pg_data_table);
-#endif
-
void __init m68k_setup_node(int node)
{
-#ifdef CONFIG_DISCONTIGMEM
- struct m68k_mem_info *info = m68k_memory + node;
- int i, end;
-
- i = (unsigned long)phys_to_virt(info->addr) >> __virt_to_node_shift();
- end = (unsigned long)phys_to_virt(info->addr + info->size - 1) >> __virt_to_node_shift();
- for (; i <= end; i++) {
- if (pg_data_table[i])
- pr_warn("overlap at %u for chunk %u\n", i, node);
- pg_data_table[i] = pg_data_map + node;
- }
-#endif
node_set_online(node);
}
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 184/192] mm: remove CONFIG_DISCONTIGMEM
2021-06-29 2:32 incoming Andrew Morton
` (182 preceding siblings ...)
2021-06-29 2:42 ` [patch 183/192] m68k: " Andrew Morton
@ 2021-06-29 2:42 ` Andrew Morton
2021-06-29 2:42 ` [patch 185/192] arch, mm: remove stale mentions of DISCONIGMEM Andrew Morton
` (7 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:42 UTC (permalink / raw)
To: akpm, arnd, corbet, david, geert, ink, linux-mm, mattst88,
mm-commits, rppt, rth, torvalds, vgupta
From: Mike Rapoport <rppt@linux.ibm.com>
Subject: mm: remove CONFIG_DISCONTIGMEM
There are no architectures that support DISCONTIGMEM left.
Remove the configuration option and the dead code it was guarding in the
generic memory management code.
Link: https://lkml.kernel.org/r/20210608091316.3622-6-rppt@kernel.org
Signed-off-by: Mike Rapoport <rppt@linux.ibm.com>
Acked-by: Arnd Bergmann <arnd@arndb.de>
Acked-by: David Hildenbrand <david@redhat.com>
Cc: Geert Uytterhoeven <geert@linux-m68k.org>
Cc: Ivan Kokshaysky <ink@jurassic.park.msu.ru>
Cc: Jonathan Corbet <corbet@lwn.net>
Cc: Matt Turner <mattst88@gmail.com>
Cc: Richard Henderson <rth@twiddle.net>
Cc: Vineet Gupta <vgupta@synopsys.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/asm-generic/memory_model.h | 37 ++-------------------------
include/linux/mmzone.h | 8 +++--
mm/Kconfig | 25 ++----------------
mm/page_alloc.c | 13 ---------
4 files changed, 12 insertions(+), 71 deletions(-)
--- a/include/asm-generic/memory_model.h~mm-remove-config_discontigmem
+++ a/include/asm-generic/memory_model.h
@@ -6,47 +6,18 @@
#ifndef __ASSEMBLY__
+/*
+ * supports 3 memory models.
+ */
#if defined(CONFIG_FLATMEM)
#ifndef ARCH_PFN_OFFSET
#define ARCH_PFN_OFFSET (0UL)
#endif
-#elif defined(CONFIG_DISCONTIGMEM)
-
-#ifndef arch_pfn_to_nid
-#define arch_pfn_to_nid(pfn) pfn_to_nid(pfn)
-#endif
-
-#ifndef arch_local_page_offset
-#define arch_local_page_offset(pfn, nid) \
- ((pfn) - NODE_DATA(nid)->node_start_pfn)
-#endif
-
-#endif /* CONFIG_DISCONTIGMEM */
-
-/*
- * supports 3 memory models.
- */
-#if defined(CONFIG_FLATMEM)
-
#define __pfn_to_page(pfn) (mem_map + ((pfn) - ARCH_PFN_OFFSET))
#define __page_to_pfn(page) ((unsigned long)((page) - mem_map) + \
ARCH_PFN_OFFSET)
-#elif defined(CONFIG_DISCONTIGMEM)
-
-#define __pfn_to_page(pfn) \
-({ unsigned long __pfn = (pfn); \
- unsigned long __nid = arch_pfn_to_nid(__pfn); \
- NODE_DATA(__nid)->node_mem_map + arch_local_page_offset(__pfn, __nid);\
-})
-
-#define __page_to_pfn(pg) \
-({ const struct page *__pg = (pg); \
- struct pglist_data *__pgdat = NODE_DATA(page_to_nid(__pg)); \
- (unsigned long)(__pg - __pgdat->node_mem_map) + \
- __pgdat->node_start_pfn; \
-})
#elif defined(CONFIG_SPARSEMEM_VMEMMAP)
@@ -70,7 +41,7 @@
struct mem_section *__sec = __pfn_to_section(__pfn); \
__section_mem_map_addr(__sec) + __pfn; \
})
-#endif /* CONFIG_FLATMEM/DISCONTIGMEM/SPARSEMEM */
+#endif /* CONFIG_FLATMEM/SPARSEMEM */
/*
* Convert a physical address to a Page Frame Number and back
--- a/include/linux/mmzone.h~mm-remove-config_discontigmem
+++ a/include/linux/mmzone.h
@@ -749,10 +749,12 @@ struct zonelist {
struct zoneref _zonerefs[MAX_ZONES_PER_ZONELIST + 1];
};
-#ifndef CONFIG_DISCONTIGMEM
-/* The array of struct pages - for discontigmem use pgdat->lmem_map */
+/*
+ * The array of struct pages for flatmem.
+ * It must be declared for SPARSEMEM as well because there are configurations
+ * that rely on that.
+ */
extern struct page *mem_map;
-#endif
#ifdef CONFIG_TRANSPARENT_HUGEPAGE
struct deferred_split {
--- a/mm/Kconfig~mm-remove-config_discontigmem
+++ a/mm/Kconfig
@@ -19,7 +19,7 @@ choice
config FLATMEM_MANUAL
bool "Flat Memory"
- depends on !(ARCH_DISCONTIGMEM_ENABLE || ARCH_SPARSEMEM_ENABLE) || ARCH_FLATMEM_ENABLE
+ depends on !ARCH_SPARSEMEM_ENABLE || ARCH_FLATMEM_ENABLE
help
This option is best suited for non-NUMA systems with
flat address space. The FLATMEM is the most efficient
@@ -32,21 +32,6 @@ config FLATMEM_MANUAL
If unsure, choose this option (Flat Memory) over any other.
-config DISCONTIGMEM_MANUAL
- bool "Discontiguous Memory"
- depends on ARCH_DISCONTIGMEM_ENABLE
- help
- This option provides enhanced support for discontiguous
- memory systems, over FLATMEM. These systems have holes
- in their physical address spaces, and this option provides
- more efficient handling of these holes.
-
- Although "Discontiguous Memory" is still used by several
- architectures, it is considered deprecated in favor of
- "Sparse Memory".
-
- If unsure, choose "Sparse Memory" over this option.
-
config SPARSEMEM_MANUAL
bool "Sparse Memory"
depends on ARCH_SPARSEMEM_ENABLE
@@ -62,17 +47,13 @@ config SPARSEMEM_MANUAL
endchoice
-config DISCONTIGMEM
- def_bool y
- depends on (!SELECT_MEMORY_MODEL && ARCH_DISCONTIGMEM_ENABLE) || DISCONTIGMEM_MANUAL
-
config SPARSEMEM
def_bool y
depends on (!SELECT_MEMORY_MODEL && ARCH_SPARSEMEM_ENABLE) || SPARSEMEM_MANUAL
config FLATMEM
def_bool y
- depends on (!DISCONTIGMEM && !SPARSEMEM) || FLATMEM_MANUAL
+ depends on !SPARSEMEM || FLATMEM_MANUAL
config FLAT_NODE_MEM_MAP
def_bool y
@@ -85,7 +66,7 @@ config FLAT_NODE_MEM_MAP
#
config NEED_MULTIPLE_NODES
def_bool y
- depends on DISCONTIGMEM || NUMA
+ depends on NUMA
#
# SPARSEMEM_EXTREME (which is the default) does some bootmem
--- a/mm/page_alloc.c~mm-remove-config_discontigmem
+++ a/mm/page_alloc.c
@@ -349,20 +349,7 @@ compound_page_dtor * const compound_page
int min_free_kbytes = 1024;
int user_min_free_kbytes = -1;
-#ifdef CONFIG_DISCONTIGMEM
-/*
- * DiscontigMem defines memory ranges as separate pg_data_t even if the ranges
- * are not on separate NUMA nodes. Functionally this works but with
- * watermark_boost_factor, it can reclaim prematurely as the ranges can be
- * quite small. By default, do not boost watermarks on discontigmem as in
- * many cases very high-order allocations like THP are likely to be
- * unsupported and the premature reclaim offsets the advantage of long-term
- * fragmentation avoidance.
- */
-int watermark_boost_factor __read_mostly;
-#else
int watermark_boost_factor __read_mostly = 15000;
-#endif
int watermark_scale_factor = 10;
static unsigned long nr_kernel_pages __initdata;
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 185/192] arch, mm: remove stale mentions of DISCONIGMEM
2021-06-29 2:32 incoming Andrew Morton
` (183 preceding siblings ...)
2021-06-29 2:42 ` [patch 184/192] mm: remove CONFIG_DISCONTIGMEM Andrew Morton
@ 2021-06-29 2:42 ` Andrew Morton
2021-06-29 2:42 ` [patch 186/192] docs: remove description of DISCONTIGMEM Andrew Morton
` (6 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:42 UTC (permalink / raw)
To: akpm, arnd, corbet, david, geert, ink, linux-mm, mattst88,
mm-commits, rppt, rth, torvalds, vgupta
From: Mike Rapoport <rppt@linux.ibm.com>
Subject: arch, mm: remove stale mentions of DISCONIGMEM
There are several places that mention DISCONIGMEM in comments or have
stale code guarded by CONFIG_DISCONTIGMEM.
Remove the dead code and update the comments.
Link: https://lkml.kernel.org/r/20210608091316.3622-7-rppt@kernel.org
Signed-off-by: Mike Rapoport <rppt@linux.ibm.com>
Acked-by: Arnd Bergmann <arnd@arndb.de>
Reviewed-by: David Hildenbrand <david@redhat.com>
Cc: Geert Uytterhoeven <geert@linux-m68k.org>
Cc: Ivan Kokshaysky <ink@jurassic.park.msu.ru>
Cc: Jonathan Corbet <corbet@lwn.net>
Cc: Matt Turner <mattst88@gmail.com>
Cc: Richard Henderson <rth@twiddle.net>
Cc: Vineet Gupta <vgupta@synopsys.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
arch/ia64/kernel/topology.c | 5 ++---
arch/ia64/mm/numa.c | 5 ++---
arch/mips/include/asm/mmzone.h | 6 ------
arch/mips/mm/init.c | 3 ---
arch/nds32/include/asm/memory.h | 6 ------
arch/xtensa/include/asm/page.h | 4 ----
include/linux/gfp.h | 4 ++--
7 files changed, 6 insertions(+), 27 deletions(-)
--- a/arch/ia64/kernel/topology.c~arch-mm-remove-stale-mentions-of-disconigmem
+++ a/arch/ia64/kernel/topology.c
@@ -3,9 +3,8 @@
* License. See the file "COPYING" in the main directory of this archive
* for more details.
*
- * This file contains NUMA specific variables and functions which can
- * be split away from DISCONTIGMEM and are used on NUMA machines with
- * contiguous memory.
+ * This file contains NUMA specific variables and functions which are used on
+ * NUMA machines with contiguous memory.
* 2002/08/07 Erich Focht <efocht@ess.nec.de>
* Populate cpu entries in sysfs for non-numa systems as well
* Intel Corporation - Ashok Raj
--- a/arch/ia64/mm/numa.c~arch-mm-remove-stale-mentions-of-disconigmem
+++ a/arch/ia64/mm/numa.c
@@ -3,9 +3,8 @@
* License. See the file "COPYING" in the main directory of this archive
* for more details.
*
- * This file contains NUMA specific variables and functions which can
- * be split away from DISCONTIGMEM and are used on NUMA machines with
- * contiguous memory.
+ * This file contains NUMA specific variables and functions which are used on
+ * NUMA machines with contiguous memory.
*
* 2002/08/07 Erich Focht <efocht@ess.nec.de>
*/
--- a/arch/mips/include/asm/mmzone.h~arch-mm-remove-stale-mentions-of-disconigmem
+++ a/arch/mips/include/asm/mmzone.h
@@ -20,10 +20,4 @@
#define nid_to_addrbase(nid) 0
#endif
-#ifdef CONFIG_DISCONTIGMEM
-
-#define pfn_to_nid(pfn) pa_to_nid((pfn) << PAGE_SHIFT)
-
-#endif /* CONFIG_DISCONTIGMEM */
-
#endif /* _ASM_MMZONE_H_ */
--- a/arch/mips/mm/init.c~arch-mm-remove-stale-mentions-of-disconigmem
+++ a/arch/mips/mm/init.c
@@ -454,9 +454,6 @@ void __init mem_init(void)
BUILD_BUG_ON(IS_ENABLED(CONFIG_32BIT) && (_PFN_SHIFT > PAGE_SHIFT));
#ifdef CONFIG_HIGHMEM
-#ifdef CONFIG_DISCONTIGMEM
-#error "CONFIG_HIGHMEM and CONFIG_DISCONTIGMEM dont work together yet"
-#endif
max_mapnr = highend_pfn ? highend_pfn : max_low_pfn;
#else
max_mapnr = max_low_pfn;
--- a/arch/nds32/include/asm/memory.h~arch-mm-remove-stale-mentions-of-disconigmem
+++ a/arch/nds32/include/asm/memory.h
@@ -76,18 +76,12 @@
* virt_to_page(k) convert a _valid_ virtual address to struct page *
* virt_addr_valid(k) indicates whether a virtual address is valid
*/
-#ifndef CONFIG_DISCONTIGMEM
-
#define ARCH_PFN_OFFSET PHYS_PFN_OFFSET
#define pfn_valid(pfn) ((pfn) >= PHYS_PFN_OFFSET && (pfn) < (PHYS_PFN_OFFSET + max_mapnr))
#define virt_to_page(kaddr) (pfn_to_page(__pa(kaddr) >> PAGE_SHIFT))
#define virt_addr_valid(kaddr) ((unsigned long)(kaddr) >= PAGE_OFFSET && (unsigned long)(kaddr) < (unsigned long)high_memory)
-#else /* CONFIG_DISCONTIGMEM */
-#error CONFIG_DISCONTIGMEM is not supported yet.
-#endif /* !CONFIG_DISCONTIGMEM */
-
#define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
#endif
--- a/arch/xtensa/include/asm/page.h~arch-mm-remove-stale-mentions-of-disconigmem
+++ a/arch/xtensa/include/asm/page.h
@@ -192,10 +192,6 @@ static inline unsigned long ___pa(unsign
#define pfn_valid(pfn) \
((pfn) >= ARCH_PFN_OFFSET && ((pfn) - ARCH_PFN_OFFSET) < max_mapnr)
-#ifdef CONFIG_DISCONTIGMEM
-# error CONFIG_DISCONTIGMEM not supported
-#endif
-
#define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
#define page_to_virt(page) __va(page_to_pfn(page) << PAGE_SHIFT)
#define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
--- a/include/linux/gfp.h~arch-mm-remove-stale-mentions-of-disconigmem
+++ a/include/linux/gfp.h
@@ -494,8 +494,8 @@ static inline int gfp_zonelist(gfp_t fla
* There are two zonelists per node, one for all zones with memory and
* one containing just zones from the node the zonelist belongs to.
*
- * For the normal case of non-DISCONTIGMEM systems the NODE_DATA() gets
- * optimized to &contig_page_data at compile-time.
+ * For the case of non-NUMA systems the NODE_DATA() gets optimized to
+ * &contig_page_data at compile-time.
*/
static inline struct zonelist *node_zonelist(int nid, gfp_t flags)
{
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 186/192] docs: remove description of DISCONTIGMEM
2021-06-29 2:32 incoming Andrew Morton
` (184 preceding siblings ...)
2021-06-29 2:42 ` [patch 185/192] arch, mm: remove stale mentions of DISCONIGMEM Andrew Morton
@ 2021-06-29 2:42 ` Andrew Morton
2021-06-29 2:43 ` [patch 187/192] mm: replace CONFIG_NEED_MULTIPLE_NODES with CONFIG_NUMA Andrew Morton
` (5 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:42 UTC (permalink / raw)
To: akpm, arnd, corbet, david, geert, ink, linux-mm, mattst88,
mm-commits, rppt, rth, torvalds, vgupta
From: Mike Rapoport <rppt@linux.ibm.com>
Subject: docs: remove description of DISCONTIGMEM
Remove description of DISCONTIGMEM from the "Memory Models" document and
update VM sysctl description so that it won't mention DISCONIGMEM.
Link: https://lkml.kernel.org/r/20210608091316.3622-8-rppt@kernel.org
Signed-off-by: Mike Rapoport <rppt@linux.ibm.com>
Acked-by: Arnd Bergmann <arnd@arndb.de>
Reviewed-by: David Hildenbrand <david@redhat.com>
Cc: Geert Uytterhoeven <geert@linux-m68k.org>
Cc: Ivan Kokshaysky <ink@jurassic.park.msu.ru>
Cc: Jonathan Corbet <corbet@lwn.net>
Cc: Matt Turner <mattst88@gmail.com>
Cc: Richard Henderson <rth@twiddle.net>
Cc: Vineet Gupta <vgupta@synopsys.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
Documentation/admin-guide/sysctl/vm.rst | 12 ++---
Documentation/vm/memory-model.rst | 45 ----------------------
2 files changed, 8 insertions(+), 49 deletions(-)
--- a/Documentation/admin-guide/sysctl/vm.rst~docs-remove-description-of-discontigmem
+++ a/Documentation/admin-guide/sysctl/vm.rst
@@ -938,12 +938,12 @@ allocations, THP and hugetlbfs pages.
To make it sensible with respect to the watermark_scale_factor
parameter, the unit is in fractions of 10,000. The default value of
-15,000 on !DISCONTIGMEM configurations means that up to 150% of the high
-watermark will be reclaimed in the event of a pageblock being mixed due
-to fragmentation. The level of reclaim is determined by the number of
-fragmentation events that occurred in the recent past. If this value is
-smaller than a pageblock then a pageblocks worth of pages will be reclaimed
-(e.g. 2MB on 64-bit x86). A boost factor of 0 will disable the feature.
+15,000 means that up to 150% of the high watermark will be reclaimed in the
+event of a pageblock being mixed due to fragmentation. The level of reclaim
+is determined by the number of fragmentation events that occurred in the
+recent past. If this value is smaller than a pageblock then a pageblocks
+worth of pages will be reclaimed (e.g. 2MB on 64-bit x86). A boost factor
+of 0 will disable the feature.
watermark_scale_factor
--- a/Documentation/vm/memory-model.rst~docs-remove-description-of-discontigmem
+++ a/Documentation/vm/memory-model.rst
@@ -14,15 +14,11 @@ for the CPU. Then there could be several
completely distinct addresses. And, don't forget about NUMA, where
different memory banks are attached to different CPUs.
-Linux abstracts this diversity using one of the three memory models:
-FLATMEM, DISCONTIGMEM and SPARSEMEM. Each architecture defines what
+Linux abstracts this diversity using one of the two memory models:
+FLATMEM and SPARSEMEM. Each architecture defines what
memory models it supports, what the default memory model is and
whether it is possible to manually override that default.
-.. note::
- At time of this writing, DISCONTIGMEM is considered deprecated,
- although it is still in use by several architectures.
-
All the memory models track the status of physical page frames using
struct page arranged in one or more arrays.
@@ -63,43 +59,6 @@ straightforward: `PFN - ARCH_PFN_OFFSET`
The `ARCH_PFN_OFFSET` defines the first page frame number for
systems with physical memory starting at address different from 0.
-DISCONTIGMEM
-============
-
-The DISCONTIGMEM model treats the physical memory as a collection of
-`nodes` similarly to how Linux NUMA support does. For each node Linux
-constructs an independent memory management subsystem represented by
-`struct pglist_data` (or `pg_data_t` for short). Among other
-things, `pg_data_t` holds the `node_mem_map` array that maps
-physical pages belonging to that node. The `node_start_pfn` field of
-`pg_data_t` is the number of the first page frame belonging to that
-node.
-
-The architecture setup code should call :c:func:`free_area_init_node` for
-each node in the system to initialize the `pg_data_t` object and its
-`node_mem_map`.
-
-Every `node_mem_map` behaves exactly as FLATMEM's `mem_map` -
-every physical page frame in a node has a `struct page` entry in the
-`node_mem_map` array. When DISCONTIGMEM is enabled, a portion of the
-`flags` field of the `struct page` encodes the node number of the
-node hosting that page.
-
-The conversion between a PFN and the `struct page` in the
-DISCONTIGMEM model became slightly more complex as it has to determine
-which node hosts the physical page and which `pg_data_t` object
-holds the `struct page`.
-
-Architectures that support DISCONTIGMEM provide :c:func:`pfn_to_nid`
-to convert PFN to the node number. The opposite conversion helper
-:c:func:`page_to_nid` is generic as it uses the node number encoded in
-page->flags.
-
-Once the node number is known, the PFN can be used to index
-appropriate `node_mem_map` array to access the `struct page` and
-the offset of the `struct page` from the `node_mem_map` plus
-`node_start_pfn` is the PFN of that page.
-
SPARSEMEM
=========
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 187/192] mm: replace CONFIG_NEED_MULTIPLE_NODES with CONFIG_NUMA
2021-06-29 2:32 incoming Andrew Morton
` (185 preceding siblings ...)
2021-06-29 2:42 ` [patch 186/192] docs: remove description of DISCONTIGMEM Andrew Morton
@ 2021-06-29 2:43 ` Andrew Morton
2021-06-29 2:43 ` [patch 188/192] mm: replace CONFIG_FLAT_NODE_MEM_MAP with CONFIG_FLATMEM Andrew Morton
` (4 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:43 UTC (permalink / raw)
To: akpm, arnd, corbet, david, geert, ink, linux-mm, mattst88,
mm-commits, rppt, rth, torvalds, vgupta
From: Mike Rapoport <rppt@linux.ibm.com>
Subject: mm: replace CONFIG_NEED_MULTIPLE_NODES with CONFIG_NUMA
After removal of DISCINTIGMEM the NEED_MULTIPLE_NODES and NUMA
configuration options are equivalent.
Drop CONFIG_NEED_MULTIPLE_NODES and use CONFIG_NUMA instead.
Done with
$ sed -i 's/CONFIG_NEED_MULTIPLE_NODES/CONFIG_NUMA/' \
$(git grep -wl CONFIG_NEED_MULTIPLE_NODES)
$ sed -i 's/NEED_MULTIPLE_NODES/NUMA/' \
$(git grep -wl NEED_MULTIPLE_NODES)
with manual tweaks afterwards.
[rppt@linux.ibm.com: fix arm boot crash]
Link: https://lkml.kernel.org/r/YMj9vHhHOiCVN4BF@linux.ibm.com
Link: https://lkml.kernel.org/r/20210608091316.3622-9-rppt@kernel.org
Signed-off-by: Mike Rapoport <rppt@linux.ibm.com>
Acked-by: Arnd Bergmann <arnd@arndb.de>
Acked-by: David Hildenbrand <david@redhat.com>
Cc: Geert Uytterhoeven <geert@linux-m68k.org>
Cc: Ivan Kokshaysky <ink@jurassic.park.msu.ru>
Cc: Jonathan Corbet <corbet@lwn.net>
Cc: Matt Turner <mattst88@gmail.com>
Cc: Richard Henderson <rth@twiddle.net>
Cc: Vineet Gupta <vgupta@synopsys.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
arch/arm64/Kconfig | 2 +-
arch/ia64/Kconfig | 2 +-
arch/mips/Kconfig | 2 +-
arch/mips/include/asm/mmzone.h | 2 +-
arch/mips/include/asm/page.h | 2 +-
arch/mips/mm/init.c | 4 ++--
arch/powerpc/Kconfig | 2 +-
arch/powerpc/include/asm/mmzone.h | 4 ++--
arch/powerpc/kernel/setup_64.c | 2 +-
arch/powerpc/kernel/smp.c | 2 +-
arch/powerpc/kexec/core.c | 4 ++--
arch/powerpc/mm/Makefile | 2 +-
arch/powerpc/mm/mem.c | 4 ++--
arch/riscv/Kconfig | 2 +-
arch/s390/Kconfig | 2 +-
arch/sh/include/asm/mmzone.h | 4 ++--
arch/sh/kernel/topology.c | 2 +-
arch/sh/mm/Kconfig | 2 +-
arch/sh/mm/init.c | 2 +-
arch/sparc/Kconfig | 2 +-
arch/sparc/include/asm/mmzone.h | 4 ++--
arch/sparc/kernel/smp_64.c | 2 +-
arch/sparc/mm/init_64.c | 12 ++++++------
arch/x86/Kconfig | 2 +-
arch/x86/kernel/setup_percpu.c | 6 +++---
arch/x86/mm/init_32.c | 4 ++--
include/asm-generic/topology.h | 2 +-
include/linux/memblock.h | 6 +++---
include/linux/mm.h | 4 ++--
include/linux/mmzone.h | 6 +++---
kernel/crash_core.c | 2 +-
mm/Kconfig | 9 ---------
mm/memblock.c | 8 ++++----
mm/memory.c | 3 +--
mm/page_alloc.c | 6 +++---
mm/sparse.c | 2 +-
36 files changed, 59 insertions(+), 69 deletions(-)
--- a/arch/arm64/Kconfig~mm-replace-config_need_multiple_nodes-with-config_numa
+++ a/arch/arm64/Kconfig
@@ -1035,7 +1035,7 @@ config NODES_SHIFT
int "Maximum NUMA Nodes (as a power of 2)"
range 1 10
default "4"
- depends on NEED_MULTIPLE_NODES
+ depends on NUMA
help
Specify the maximum number of NUMA Nodes available on the target
system. Increases memory reserved to accommodate various tables.
--- a/arch/ia64/Kconfig~mm-replace-config_need_multiple_nodes-with-config_numa
+++ a/arch/ia64/Kconfig
@@ -302,7 +302,7 @@ config NODES_SHIFT
int "Max num nodes shift(3-10)"
range 3 10
default "10"
- depends on NEED_MULTIPLE_NODES
+ depends on NUMA
help
This option specifies the maximum number of nodes in your SSI system.
MAX_NUMNODES will be 2^(This value).
--- a/arch/mips/include/asm/mmzone.h~mm-replace-config_need_multiple_nodes-with-config_numa
+++ a/arch/mips/include/asm/mmzone.h
@@ -8,7 +8,7 @@
#include <asm/page.h>
-#ifdef CONFIG_NEED_MULTIPLE_NODES
+#ifdef CONFIG_NUMA
# include <mmzone.h>
#endif
--- a/arch/mips/include/asm/page.h~mm-replace-config_need_multiple_nodes-with-config_numa
+++ a/arch/mips/include/asm/page.h
@@ -239,7 +239,7 @@ static inline int pfn_valid(unsigned lon
/* pfn_valid is defined in linux/mmzone.h */
-#elif defined(CONFIG_NEED_MULTIPLE_NODES)
+#elif defined(CONFIG_NUMA)
#define pfn_valid(pfn) \
({ \
--- a/arch/mips/Kconfig~mm-replace-config_need_multiple_nodes-with-config_numa
+++ a/arch/mips/Kconfig
@@ -2867,7 +2867,7 @@ config RANDOMIZE_BASE_MAX_OFFSET
config NODES_SHIFT
int
default "6"
- depends on NEED_MULTIPLE_NODES
+ depends on NUMA
config HW_PERF_EVENTS
bool "Enable hardware performance counter support for perf events"
--- a/arch/mips/mm/init.c~mm-replace-config_need_multiple_nodes-with-config_numa
+++ a/arch/mips/mm/init.c
@@ -394,7 +394,7 @@ void maar_init(void)
}
}
-#ifndef CONFIG_NEED_MULTIPLE_NODES
+#ifndef CONFIG_NUMA
void __init paging_init(void)
{
unsigned long max_zone_pfns[MAX_NR_ZONES];
@@ -473,7 +473,7 @@ void __init mem_init(void)
0x80000000 - 4, KCORE_TEXT);
#endif
}
-#endif /* !CONFIG_NEED_MULTIPLE_NODES */
+#endif /* !CONFIG_NUMA */
void free_init_pages(const char *what, unsigned long begin, unsigned long end)
{
--- a/arch/powerpc/include/asm/mmzone.h~mm-replace-config_need_multiple_nodes-with-config_numa
+++ a/arch/powerpc/include/asm/mmzone.h
@@ -18,7 +18,7 @@
* flags field of the struct page
*/
-#ifdef CONFIG_NEED_MULTIPLE_NODES
+#ifdef CONFIG_NUMA
extern struct pglist_data *node_data[];
/*
@@ -41,7 +41,7 @@ u64 memory_hotplug_max(void);
#else
#define memory_hotplug_max() memblock_end_of_DRAM()
-#endif /* CONFIG_NEED_MULTIPLE_NODES */
+#endif /* CONFIG_NUMA */
#ifdef CONFIG_FA_DUMP
#define __HAVE_ARCH_RESERVED_KERNEL_PAGES
#endif
--- a/arch/powerpc/Kconfig~mm-replace-config_need_multiple_nodes-with-config_numa
+++ a/arch/powerpc/Kconfig
@@ -671,7 +671,7 @@ config NODES_SHIFT
int
default "8" if PPC64
default "4"
- depends on NEED_MULTIPLE_NODES
+ depends on NUMA
config USE_PERCPU_NUMA_NODE_ID
def_bool y
--- a/arch/powerpc/kernel/setup_64.c~mm-replace-config_need_multiple_nodes-with-config_numa
+++ a/arch/powerpc/kernel/setup_64.c
@@ -788,7 +788,7 @@ static void * __init pcpu_alloc_bootmem(
size_t align)
{
const unsigned long goal = __pa(MAX_DMA_ADDRESS);
-#ifdef CONFIG_NEED_MULTIPLE_NODES
+#ifdef CONFIG_NUMA
int node = early_cpu_to_node(cpu);
void *ptr;
--- a/arch/powerpc/kernel/smp.c~mm-replace-config_need_multiple_nodes-with-config_numa
+++ a/arch/powerpc/kernel/smp.c
@@ -1047,7 +1047,7 @@ void __init smp_prepare_cpus(unsigned in
zalloc_cpumask_var_node(&per_cpu(cpu_coregroup_map, cpu),
GFP_KERNEL, cpu_to_node(cpu));
-#ifdef CONFIG_NEED_MULTIPLE_NODES
+#ifdef CONFIG_NUMA
/*
* numa_node_id() works after this.
*/
--- a/arch/powerpc/kexec/core.c~mm-replace-config_need_multiple_nodes-with-config_numa
+++ a/arch/powerpc/kexec/core.c
@@ -68,11 +68,11 @@ void machine_kexec_cleanup(struct kimage
void arch_crash_save_vmcoreinfo(void)
{
-#ifdef CONFIG_NEED_MULTIPLE_NODES
+#ifdef CONFIG_NUMA
VMCOREINFO_SYMBOL(node_data);
VMCOREINFO_LENGTH(node_data, MAX_NUMNODES);
#endif
-#ifndef CONFIG_NEED_MULTIPLE_NODES
+#ifndef CONFIG_NUMA
VMCOREINFO_SYMBOL(contig_page_data);
#endif
#if defined(CONFIG_PPC64) && defined(CONFIG_SPARSEMEM_VMEMMAP)
--- a/arch/powerpc/mm/Makefile~mm-replace-config_need_multiple_nodes-with-config_numa
+++ a/arch/powerpc/mm/Makefile
@@ -13,7 +13,7 @@ obj-y := fault.o mem.o pgtable.o mmap
obj-$(CONFIG_PPC_MMU_NOHASH) += nohash/
obj-$(CONFIG_PPC_BOOK3S_32) += book3s32/
obj-$(CONFIG_PPC_BOOK3S_64) += book3s64/
-obj-$(CONFIG_NEED_MULTIPLE_NODES) += numa.o
+obj-$(CONFIG_NUMA) += numa.o
obj-$(CONFIG_PPC_MM_SLICES) += slice.o
obj-$(CONFIG_HUGETLB_PAGE) += hugetlbpage.o
obj-$(CONFIG_NOT_COHERENT_CACHE) += dma-noncoherent.o
--- a/arch/powerpc/mm/mem.c~mm-replace-config_need_multiple_nodes-with-config_numa
+++ a/arch/powerpc/mm/mem.c
@@ -127,7 +127,7 @@ void __ref arch_remove_memory(int nid, u
}
#endif
-#ifndef CONFIG_NEED_MULTIPLE_NODES
+#ifndef CONFIG_NUMA
void __init mem_topology_setup(void)
{
max_low_pfn = max_pfn = memblock_end_of_DRAM() >> PAGE_SHIFT;
@@ -162,7 +162,7 @@ static int __init mark_nonram_nosave(voi
return 0;
}
-#else /* CONFIG_NEED_MULTIPLE_NODES */
+#else /* CONFIG_NUMA */
static int __init mark_nonram_nosave(void)
{
return 0;
--- a/arch/riscv/Kconfig~mm-replace-config_need_multiple_nodes-with-config_numa
+++ a/arch/riscv/Kconfig
@@ -332,7 +332,7 @@ config NODES_SHIFT
int "Maximum NUMA Nodes (as a power of 2)"
range 1 10
default "2"
- depends on NEED_MULTIPLE_NODES
+ depends on NUMA
help
Specify the maximum number of NUMA Nodes available on the target
system. Increases memory reserved to accommodate various tables.
--- a/arch/s390/Kconfig~mm-replace-config_need_multiple_nodes-with-config_numa
+++ a/arch/s390/Kconfig
@@ -475,7 +475,7 @@ config NUMA
config NODES_SHIFT
int
- depends on NEED_MULTIPLE_NODES
+ depends on NUMA
default "1"
config SCHED_SMT
--- a/arch/sh/include/asm/mmzone.h~mm-replace-config_need_multiple_nodes-with-config_numa
+++ a/arch/sh/include/asm/mmzone.h
@@ -2,7 +2,7 @@
#ifndef __ASM_SH_MMZONE_H
#define __ASM_SH_MMZONE_H
-#ifdef CONFIG_NEED_MULTIPLE_NODES
+#ifdef CONFIG_NUMA
#include <linux/numa.h>
extern struct pglist_data *node_data[];
@@ -31,7 +31,7 @@ static inline void
setup_bootmem_node(int nid, unsigned long start, unsigned long end)
{
}
-#endif /* CONFIG_NEED_MULTIPLE_NODES */
+#endif /* CONFIG_NUMA */
/* Platform specific mem init */
void __init plat_mem_setup(void);
--- a/arch/sh/kernel/topology.c~mm-replace-config_need_multiple_nodes-with-config_numa
+++ a/arch/sh/kernel/topology.c
@@ -46,7 +46,7 @@ static int __init topology_init(void)
{
int i, ret;
-#ifdef CONFIG_NEED_MULTIPLE_NODES
+#ifdef CONFIG_NUMA
for_each_online_node(i)
register_one_node(i);
#endif
--- a/arch/sh/mm/init.c~mm-replace-config_need_multiple_nodes-with-config_numa
+++ a/arch/sh/mm/init.c
@@ -211,7 +211,7 @@ void __init allocate_pgdat(unsigned int
get_pfn_range_for_nid(nid, &start_pfn, &end_pfn);
-#ifdef CONFIG_NEED_MULTIPLE_NODES
+#ifdef CONFIG_NUMA
NODE_DATA(nid) = memblock_alloc_try_nid(
sizeof(struct pglist_data),
SMP_CACHE_BYTES, MEMBLOCK_LOW_LIMIT,
--- a/arch/sh/mm/Kconfig~mm-replace-config_need_multiple_nodes-with-config_numa
+++ a/arch/sh/mm/Kconfig
@@ -120,7 +120,7 @@ config NODES_SHIFT
int
default "3" if CPU_SUBTYPE_SHX3
default "1"
- depends on NEED_MULTIPLE_NODES
+ depends on NUMA
config ARCH_FLATMEM_ENABLE
def_bool y
--- a/arch/sparc/include/asm/mmzone.h~mm-replace-config_need_multiple_nodes-with-config_numa
+++ a/arch/sparc/include/asm/mmzone.h
@@ -2,7 +2,7 @@
#ifndef _SPARC64_MMZONE_H
#define _SPARC64_MMZONE_H
-#ifdef CONFIG_NEED_MULTIPLE_NODES
+#ifdef CONFIG_NUMA
#include <linux/cpumask.h>
@@ -13,6 +13,6 @@ extern struct pglist_data *node_data[];
extern int numa_cpu_lookup_table[];
extern cpumask_t numa_cpumask_lookup_table[];
-#endif /* CONFIG_NEED_MULTIPLE_NODES */
+#endif /* CONFIG_NUMA */
#endif /* _SPARC64_MMZONE_H */
--- a/arch/sparc/Kconfig~mm-replace-config_need_multiple_nodes-with-config_numa
+++ a/arch/sparc/Kconfig
@@ -265,7 +265,7 @@ config NODES_SHIFT
int "Maximum NUMA Nodes (as a power of 2)"
range 4 5 if SPARC64
default "5"
- depends on NEED_MULTIPLE_NODES
+ depends on NUMA
help
Specify the maximum number of NUMA Nodes available on the target
system. Increases memory reserved to accommodate various tables.
--- a/arch/sparc/kernel/smp_64.c~mm-replace-config_need_multiple_nodes-with-config_numa
+++ a/arch/sparc/kernel/smp_64.c
@@ -1546,7 +1546,7 @@ static void * __init pcpu_alloc_bootmem(
size_t align)
{
const unsigned long goal = __pa(MAX_DMA_ADDRESS);
-#ifdef CONFIG_NEED_MULTIPLE_NODES
+#ifdef CONFIG_NUMA
int node = cpu_to_node(cpu);
void *ptr;
--- a/arch/sparc/mm/init_64.c~mm-replace-config_need_multiple_nodes-with-config_numa
+++ a/arch/sparc/mm/init_64.c
@@ -903,7 +903,7 @@ struct node_mem_mask {
static struct node_mem_mask node_masks[MAX_NUMNODES];
static int num_node_masks;
-#ifdef CONFIG_NEED_MULTIPLE_NODES
+#ifdef CONFIG_NUMA
struct mdesc_mlgroup {
u64 node;
@@ -1059,7 +1059,7 @@ static void __init allocate_node_data(in
{
struct pglist_data *p;
unsigned long start_pfn, end_pfn;
-#ifdef CONFIG_NEED_MULTIPLE_NODES
+#ifdef CONFIG_NUMA
NODE_DATA(nid) = memblock_alloc_node(sizeof(struct pglist_data),
SMP_CACHE_BYTES, nid);
@@ -1080,7 +1080,7 @@ static void __init allocate_node_data(in
static void init_node_masks_nonnuma(void)
{
-#ifdef CONFIG_NEED_MULTIPLE_NODES
+#ifdef CONFIG_NUMA
int i;
#endif
@@ -1090,7 +1090,7 @@ static void init_node_masks_nonnuma(void
node_masks[0].match = 0;
num_node_masks = 1;
-#ifdef CONFIG_NEED_MULTIPLE_NODES
+#ifdef CONFIG_NUMA
for (i = 0; i < NR_CPUS; i++)
numa_cpu_lookup_table[i] = 0;
@@ -1098,7 +1098,7 @@ static void init_node_masks_nonnuma(void
#endif
}
-#ifdef CONFIG_NEED_MULTIPLE_NODES
+#ifdef CONFIG_NUMA
struct pglist_data *node_data[MAX_NUMNODES];
EXPORT_SYMBOL(numa_cpu_lookup_table);
@@ -2487,7 +2487,7 @@ int page_in_phys_avail(unsigned long pad
static void __init register_page_bootmem_info(void)
{
-#ifdef CONFIG_NEED_MULTIPLE_NODES
+#ifdef CONFIG_NUMA
int i;
for_each_online_node(i)
--- a/arch/x86/Kconfig~mm-replace-config_need_multiple_nodes-with-config_numa
+++ a/arch/x86/Kconfig
@@ -1597,7 +1597,7 @@ config NODES_SHIFT
default "10" if MAXSMP
default "6" if X86_64
default "3"
- depends on NEED_MULTIPLE_NODES
+ depends on NUMA
help
Specify the maximum number of NUMA Nodes available on the target
system. Increases memory reserved to accommodate various tables.
--- a/arch/x86/kernel/setup_percpu.c~mm-replace-config_need_multiple_nodes-with-config_numa
+++ a/arch/x86/kernel/setup_percpu.c
@@ -66,7 +66,7 @@ EXPORT_SYMBOL(__per_cpu_offset);
*/
static bool __init pcpu_need_numa(void)
{
-#ifdef CONFIG_NEED_MULTIPLE_NODES
+#ifdef CONFIG_NUMA
pg_data_t *last = NULL;
unsigned int cpu;
@@ -101,7 +101,7 @@ static void * __init pcpu_alloc_bootmem(
unsigned long align)
{
const unsigned long goal = __pa(MAX_DMA_ADDRESS);
-#ifdef CONFIG_NEED_MULTIPLE_NODES
+#ifdef CONFIG_NUMA
int node = early_cpu_to_node(cpu);
void *ptr;
@@ -140,7 +140,7 @@ static void __init pcpu_fc_free(void *pt
static int __init pcpu_cpu_distance(unsigned int from, unsigned int to)
{
-#ifdef CONFIG_NEED_MULTIPLE_NODES
+#ifdef CONFIG_NUMA
if (early_cpu_to_node(from) == early_cpu_to_node(to))
return LOCAL_DISTANCE;
else
--- a/arch/x86/mm/init_32.c~mm-replace-config_need_multiple_nodes-with-config_numa
+++ a/arch/x86/mm/init_32.c
@@ -651,7 +651,7 @@ void __init find_low_pfn_range(void)
highmem_pfn_init();
}
-#ifndef CONFIG_NEED_MULTIPLE_NODES
+#ifndef CONFIG_NUMA
void __init initmem_init(void)
{
#ifdef CONFIG_HIGHMEM
@@ -677,7 +677,7 @@ void __init initmem_init(void)
setup_bootmem_allocator();
}
-#endif /* !CONFIG_NEED_MULTIPLE_NODES */
+#endif /* !CONFIG_NUMA */
void __init setup_bootmem_allocator(void)
{
--- a/include/asm-generic/topology.h~mm-replace-config_need_multiple_nodes-with-config_numa
+++ a/include/asm-generic/topology.h
@@ -45,7 +45,7 @@
#endif
#ifndef cpumask_of_node
- #ifdef CONFIG_NEED_MULTIPLE_NODES
+ #ifdef CONFIG_NUMA
#define cpumask_of_node(node) ((node) == 0 ? cpu_online_mask : cpu_none_mask)
#else
#define cpumask_of_node(node) ((void)(node), cpu_online_mask)
--- a/include/linux/memblock.h~mm-replace-config_need_multiple_nodes-with-config_numa
+++ a/include/linux/memblock.h
@@ -50,7 +50,7 @@ struct memblock_region {
phys_addr_t base;
phys_addr_t size;
enum memblock_flags flags;
-#ifdef CONFIG_NEED_MULTIPLE_NODES
+#ifdef CONFIG_NUMA
int nid;
#endif
};
@@ -347,7 +347,7 @@ int __init deferred_page_init_max_thread
int memblock_set_node(phys_addr_t base, phys_addr_t size,
struct memblock_type *type, int nid);
-#ifdef CONFIG_NEED_MULTIPLE_NODES
+#ifdef CONFIG_NUMA
static inline void memblock_set_region_node(struct memblock_region *r, int nid)
{
r->nid = nid;
@@ -366,7 +366,7 @@ static inline int memblock_get_region_no
{
return 0;
}
-#endif /* CONFIG_NEED_MULTIPLE_NODES */
+#endif /* CONFIG_NUMA */
/* Flags for memblock allocation APIs */
#define MEMBLOCK_ALLOC_ANYWHERE (~(phys_addr_t)0)
--- a/include/linux/mm.h~mm-replace-config_need_multiple_nodes-with-config_numa
+++ a/include/linux/mm.h
@@ -46,7 +46,7 @@ extern int sysctl_page_lock_unfairness;
void init_mm_internals(void);
-#ifndef CONFIG_NEED_MULTIPLE_NODES /* Don't use mapnrs, do it properly */
+#ifndef CONFIG_NUMA /* Don't use mapnrs, do it properly */
extern unsigned long max_mapnr;
static inline void set_max_mapnr(unsigned long limit)
@@ -2460,7 +2460,7 @@ extern void get_pfn_range_for_nid(unsign
unsigned long *start_pfn, unsigned long *end_pfn);
extern unsigned long find_min_pfn_with_active_regions(void);
-#ifndef CONFIG_NEED_MULTIPLE_NODES
+#ifndef CONFIG_NUMA
static inline int early_pfn_to_nid(unsigned long pfn)
{
return 0;
--- a/include/linux/mmzone.h~mm-replace-config_need_multiple_nodes-with-config_numa
+++ a/include/linux/mmzone.h
@@ -1043,17 +1043,17 @@ extern int percpu_pagelist_high_fraction
extern char numa_zonelist_order[];
#define NUMA_ZONELIST_ORDER_LEN 16
-#ifndef CONFIG_NEED_MULTIPLE_NODES
+#ifndef CONFIG_NUMA
extern struct pglist_data contig_page_data;
#define NODE_DATA(nid) (&contig_page_data)
#define NODE_MEM_MAP(nid) mem_map
-#else /* CONFIG_NEED_MULTIPLE_NODES */
+#else /* CONFIG_NUMA */
#include <asm/mmzone.h>
-#endif /* !CONFIG_NEED_MULTIPLE_NODES */
+#endif /* !CONFIG_NUMA */
extern struct pglist_data *first_online_pgdat(void);
extern struct pglist_data *next_online_pgdat(struct pglist_data *pgdat);
--- a/kernel/crash_core.c~mm-replace-config_need_multiple_nodes-with-config_numa
+++ a/kernel/crash_core.c
@@ -455,7 +455,7 @@ static int __init crash_save_vmcoreinfo_
VMCOREINFO_SYMBOL(_stext);
VMCOREINFO_SYMBOL(vmap_area_list);
-#ifndef CONFIG_NEED_MULTIPLE_NODES
+#ifndef CONFIG_NUMA
VMCOREINFO_SYMBOL(mem_map);
VMCOREINFO_SYMBOL(contig_page_data);
#endif
--- a/mm/Kconfig~mm-replace-config_need_multiple_nodes-with-config_numa
+++ a/mm/Kconfig
@@ -60,15 +60,6 @@ config FLAT_NODE_MEM_MAP
depends on !SPARSEMEM
#
-# Both the NUMA code and DISCONTIGMEM use arrays of pg_data_t's
-# to represent different areas of memory. This variable allows
-# those dependencies to exist individually.
-#
-config NEED_MULTIPLE_NODES
- def_bool y
- depends on NUMA
-
-#
# SPARSEMEM_EXTREME (which is the default) does some bootmem
# allocations when sparse_init() is called. If this cannot
# be done on your architecture, select this option. However,
--- a/mm/memblock.c~mm-replace-config_need_multiple_nodes-with-config_numa
+++ a/mm/memblock.c
@@ -92,7 +92,7 @@
* system initialization completes.
*/
-#ifndef CONFIG_NEED_MULTIPLE_NODES
+#ifndef CONFIG_NUMA
struct pglist_data __refdata contig_page_data;
EXPORT_SYMBOL(contig_page_data);
#endif
@@ -607,7 +607,7 @@ repeat:
* area, insert that portion.
*/
if (rbase > base) {
-#ifdef CONFIG_NEED_MULTIPLE_NODES
+#ifdef CONFIG_NUMA
WARN_ON(nid != memblock_get_region_node(rgn));
#endif
WARN_ON(flags != rgn->flags);
@@ -1205,7 +1205,7 @@ void __init_memblock __next_mem_pfn_rang
int __init_memblock memblock_set_node(phys_addr_t base, phys_addr_t size,
struct memblock_type *type, int nid)
{
-#ifdef CONFIG_NEED_MULTIPLE_NODES
+#ifdef CONFIG_NUMA
int start_rgn, end_rgn;
int i, ret;
@@ -1849,7 +1849,7 @@ static void __init_memblock memblock_dum
size = rgn->size;
end = base + size - 1;
flags = rgn->flags;
-#ifdef CONFIG_NEED_MULTIPLE_NODES
+#ifdef CONFIG_NUMA
if (memblock_get_region_node(rgn) != MAX_NUMNODES)
snprintf(nid_buf, sizeof(nid_buf), " on node %d",
memblock_get_region_node(rgn));
--- a/mm/memory.c~mm-replace-config_need_multiple_nodes-with-config_numa
+++ a/mm/memory.c
@@ -90,8 +90,7 @@
#warning Unfortunate NUMA and NUMA Balancing config, growing page-frame for last_cpupid.
#endif
-#ifndef CONFIG_NEED_MULTIPLE_NODES
-/* use the per-pgdat data instead for discontigmem - mbligh */
+#ifndef CONFIG_NUMA
unsigned long max_mapnr;
EXPORT_SYMBOL(max_mapnr);
--- a/mm/page_alloc.c~mm-replace-config_need_multiple_nodes-with-config_numa
+++ a/mm/page_alloc.c
@@ -1634,7 +1634,7 @@ void __free_pages_core(struct page *page
__free_pages_ok(page, order, FPI_TO_TAIL | FPI_SKIP_KASAN_POISON);
}
-#ifdef CONFIG_NEED_MULTIPLE_NODES
+#ifdef CONFIG_NUMA
/*
* During memory init memblocks map pfns to nids. The search is expensive and
@@ -1684,7 +1684,7 @@ int __meminit early_pfn_to_nid(unsigned
return nid;
}
-#endif /* CONFIG_NEED_MULTIPLE_NODES */
+#endif /* CONFIG_NUMA */
void __init memblock_free_pages(struct page *page, unsigned long pfn,
unsigned int order)
@@ -7438,7 +7438,7 @@ static void __ref alloc_node_mem_map(str
pr_debug("%s: node %d, pgdat %08lx, node_mem_map %08lx\n",
__func__, pgdat->node_id, (unsigned long)pgdat,
(unsigned long)pgdat->node_mem_map);
-#ifndef CONFIG_NEED_MULTIPLE_NODES
+#ifndef CONFIG_NUMA
/*
* With no DISCONTIG, the global mem_map is just set as node 0's
*/
--- a/mm/sparse.c~mm-replace-config_need_multiple_nodes-with-config_numa
+++ a/mm/sparse.c
@@ -346,7 +346,7 @@ size_t mem_section_usage_size(void)
static inline phys_addr_t pgdat_to_phys(struct pglist_data *pgdat)
{
-#ifndef CONFIG_NEED_MULTIPLE_NODES
+#ifndef CONFIG_NUMA
return __pa_symbol(pgdat);
#else
return __pa(pgdat);
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 188/192] mm: replace CONFIG_FLAT_NODE_MEM_MAP with CONFIG_FLATMEM
2021-06-29 2:32 incoming Andrew Morton
` (186 preceding siblings ...)
2021-06-29 2:43 ` [patch 187/192] mm: replace CONFIG_NEED_MULTIPLE_NODES with CONFIG_NUMA Andrew Morton
@ 2021-06-29 2:43 ` Andrew Morton
2021-06-29 2:43 ` [patch 189/192] mm/page_alloc: allow high-order pages to be stored on the per-cpu lists Andrew Morton
` (3 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:43 UTC (permalink / raw)
To: akpm, arnd, corbet, david, geert, ink, linux-mm, mattst88,
mm-commits, rppt, rth, torvalds, vgupta
From: Mike Rapoport <rppt@linux.ibm.com>
Subject: mm: replace CONFIG_FLAT_NODE_MEM_MAP with CONFIG_FLATMEM
After removal of the DISCONTIGMEM memory model the FLAT_NODE_MEM_MAP
configuration option is equivalent to FLATMEM.
Drop CONFIG_FLAT_NODE_MEM_MAP and use CONFIG_FLATMEM instead.
Link: https://lkml.kernel.org/r/20210608091316.3622-10-rppt@kernel.org
Signed-off-by: Mike Rapoport <rppt@linux.ibm.com>
Acked-by: Arnd Bergmann <arnd@arndb.de>
Acked-by: David Hildenbrand <david@redhat.com>
Cc: Geert Uytterhoeven <geert@linux-m68k.org>
Cc: Ivan Kokshaysky <ink@jurassic.park.msu.ru>
Cc: Jonathan Corbet <corbet@lwn.net>
Cc: Matt Turner <mattst88@gmail.com>
Cc: Richard Henderson <rth@twiddle.net>
Cc: Vineet Gupta <vgupta@synopsys.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/linux/mmzone.h | 4 ++--
kernel/crash_core.c | 2 +-
mm/Kconfig | 4 ----
mm/page_alloc.c | 6 +++---
mm/page_ext.c | 2 +-
5 files changed, 7 insertions(+), 11 deletions(-)
--- a/include/linux/mmzone.h~mm-replace-config_flat_node_mem_map-with-config_flatmem
+++ a/include/linux/mmzone.h
@@ -788,7 +788,7 @@ typedef struct pglist_data {
struct zonelist node_zonelists[MAX_ZONELISTS];
int nr_zones; /* number of populated zones in this node */
-#ifdef CONFIG_FLAT_NODE_MEM_MAP /* means !SPARSEMEM */
+#ifdef CONFIG_FLATMEM /* means !SPARSEMEM */
struct page *node_mem_map;
#ifdef CONFIG_PAGE_EXTENSION
struct page_ext *node_page_ext;
@@ -878,7 +878,7 @@ typedef struct pglist_data {
#define node_present_pages(nid) (NODE_DATA(nid)->node_present_pages)
#define node_spanned_pages(nid) (NODE_DATA(nid)->node_spanned_pages)
-#ifdef CONFIG_FLAT_NODE_MEM_MAP
+#ifdef CONFIG_FLATMEM
#define pgdat_page_nr(pgdat, pagenr) ((pgdat)->node_mem_map + (pagenr))
#else
#define pgdat_page_nr(pgdat, pagenr) pfn_to_page((pgdat)->node_start_pfn + (pagenr))
--- a/kernel/crash_core.c~mm-replace-config_flat_node_mem_map-with-config_flatmem
+++ a/kernel/crash_core.c
@@ -484,7 +484,7 @@ static int __init crash_save_vmcoreinfo_
VMCOREINFO_OFFSET(page, compound_head);
VMCOREINFO_OFFSET(pglist_data, node_zones);
VMCOREINFO_OFFSET(pglist_data, nr_zones);
-#ifdef CONFIG_FLAT_NODE_MEM_MAP
+#ifdef CONFIG_FLATMEM
VMCOREINFO_OFFSET(pglist_data, node_mem_map);
#endif
VMCOREINFO_OFFSET(pglist_data, node_start_pfn);
--- a/mm/Kconfig~mm-replace-config_flat_node_mem_map-with-config_flatmem
+++ a/mm/Kconfig
@@ -55,10 +55,6 @@ config FLATMEM
def_bool y
depends on !SPARSEMEM || FLATMEM_MANUAL
-config FLAT_NODE_MEM_MAP
- def_bool y
- depends on !SPARSEMEM
-
#
# SPARSEMEM_EXTREME (which is the default) does some bootmem
# allocations when sparse_init() is called. If this cannot
--- a/mm/page_alloc.c~mm-replace-config_flat_node_mem_map-with-config_flatmem
+++ a/mm/page_alloc.c
@@ -6547,7 +6547,7 @@ static void __meminit zone_init_free_lis
}
}
-#if !defined(CONFIG_FLAT_NODE_MEM_MAP)
+#if !defined(CONFIG_FLATMEM)
/*
* Only struct pages that correspond to ranges defined by memblock.memory
* are zeroed and initialized by going through __init_single_page() during
@@ -7403,7 +7403,7 @@ static void __init free_area_init_core(s
}
}
-#ifdef CONFIG_FLAT_NODE_MEM_MAP
+#ifdef CONFIG_FLATMEM
static void __ref alloc_node_mem_map(struct pglist_data *pgdat)
{
unsigned long __maybe_unused start = 0;
@@ -7451,7 +7451,7 @@ static void __ref alloc_node_mem_map(str
}
#else
static void __ref alloc_node_mem_map(struct pglist_data *pgdat) { }
-#endif /* CONFIG_FLAT_NODE_MEM_MAP */
+#endif /* CONFIG_FLATMEM */
#ifdef CONFIG_DEFERRED_STRUCT_PAGE_INIT
static inline void pgdat_set_deferred_range(pg_data_t *pgdat)
--- a/mm/page_ext.c~mm-replace-config_flat_node_mem_map-with-config_flatmem
+++ a/mm/page_ext.c
@@ -191,7 +191,7 @@ fail:
panic("Out of memory");
}
-#else /* CONFIG_FLAT_NODE_MEM_MAP */
+#else /* CONFIG_FLATMEM */
struct page_ext *lookup_page_ext(const struct page *page)
{
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 189/192] mm/page_alloc: allow high-order pages to be stored on the per-cpu lists
2021-06-29 2:32 incoming Andrew Morton
` (187 preceding siblings ...)
2021-06-29 2:43 ` [patch 188/192] mm: replace CONFIG_FLAT_NODE_MEM_MAP with CONFIG_FLATMEM Andrew Morton
@ 2021-06-29 2:43 ` Andrew Morton
2021-06-29 2:43 ` [patch 190/192] mm/page_alloc: split pcp->high across all online CPUs for cpuless nodes Andrew Morton
` (2 subsequent siblings)
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:43 UTC (permalink / raw)
To: akpm, brouer, dave.hansen, linux-mm, mgorman, mhocko, mm-commits,
torvalds, vbabka, ziy
From: Mel Gorman <mgorman@techsingularity.net>
Subject: mm/page_alloc: allow high-order pages to be stored on the per-cpu lists
The per-cpu page allocator (PCP) only stores order-0 pages. This means
that all THP and "cheap" high-order allocations including SLUB contends on
the zone->lock. This patch extends the PCP allocator to store THP and
"cheap" high-order pages. Note that struct per_cpu_pages increases in
size to 256 bytes (4 cache lines) on x86-64.
Note that this is not necessarily a universal performance win because of
how it is implemented. High-order pages can cause pcp->high to be
exceeded prematurely for lower-orders so for example, a large number of
THP pages being freed could release order-0 pages from the PCP lists.
Hence, much depends on the allocation/free pattern as observed by a single
CPU to determine if caching helps or hurts a particular workload.
That said, basic performance testing passed. The following is a netperf
UDP_STREAM test which hits the relevant patches as some of the network
allocations are high-order.
netperf-udp
5.13.0-rc2 5.13.0-rc2
mm-pcpburst-v3r4 mm-pcphighorder-v1r7
Hmean send-64 261.46 ( 0.00%) 266.30 * 1.85%*
Hmean send-128 516.35 ( 0.00%) 536.78 * 3.96%*
Hmean send-256 1014.13 ( 0.00%) 1034.63 * 2.02%*
Hmean send-1024 3907.65 ( 0.00%) 4046.11 * 3.54%*
Hmean send-2048 7492.93 ( 0.00%) 7754.85 * 3.50%*
Hmean send-3312 11410.04 ( 0.00%) 11772.32 * 3.18%*
Hmean send-4096 13521.95 ( 0.00%) 13912.34 * 2.89%*
Hmean send-8192 21660.50 ( 0.00%) 22730.72 * 4.94%*
Hmean send-16384 31902.32 ( 0.00%) 32637.50 * 2.30%*
Functionally, a patch like this is necessary to make bulk allocation of
high-order pages work with similar performance to order-0 bulk
allocations. The bulk allocator is not updated in this series as it would
have to be determined by bulk allocation users how they want to track the
order of pages allocated with the bulk allocator.
Link: https://lkml.kernel.org/r/20210611135753.GC30378@techsingularity.net
Signed-off-by: Mel Gorman <mgorman@techsingularity.net>
Acked-by: Vlastimil Babka <vbabka@suse.cz>
Cc: Zi Yan <ziy@nvidia.com>
Cc: Dave Hansen <dave.hansen@linux.intel.com>
Cc: Michal Hocko <mhocko@kernel.org>
Cc: Jesper Dangaard Brouer <brouer@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
include/linux/mmzone.h | 20 ++++
mm/internal.h | 2
mm/page_alloc.c | 169 ++++++++++++++++++++++++++++-----------
mm/swap.c | 2
4 files changed, 144 insertions(+), 49 deletions(-)
--- a/include/linux/mmzone.h~mm-page_alloc-allow-high-order-pages-to-be-stored-on-the-per-cpu-lists
+++ a/include/linux/mmzone.h
@@ -333,6 +333,24 @@ enum zone_watermarks {
NR_WMARK
};
+/*
+ * One per migratetype for each PAGE_ALLOC_COSTLY_ORDER plus one additional
+ * for pageblock size for THP if configured.
+ */
+#ifdef CONFIG_TRANSPARENT_HUGEPAGE
+#define NR_PCP_THP 1
+#else
+#define NR_PCP_THP 0
+#endif
+#define NR_PCP_LISTS (MIGRATE_PCPTYPES * (PAGE_ALLOC_COSTLY_ORDER + 1 + NR_PCP_THP))
+
+/*
+ * Shift to encode migratetype and order in the same integer, with order
+ * in the least significant bits.
+ */
+#define NR_PCP_ORDER_WIDTH 8
+#define NR_PCP_ORDER_MASK ((1<<NR_PCP_ORDER_WIDTH) - 1)
+
#define min_wmark_pages(z) (z->_watermark[WMARK_MIN] + z->watermark_boost)
#define low_wmark_pages(z) (z->_watermark[WMARK_LOW] + z->watermark_boost)
#define high_wmark_pages(z) (z->_watermark[WMARK_HIGH] + z->watermark_boost)
@@ -349,7 +367,7 @@ struct per_cpu_pages {
#endif
/* Lists of pages, one per migrate type stored on the pcp-lists */
- struct list_head lists[MIGRATE_PCPTYPES];
+ struct list_head lists[NR_PCP_LISTS];
};
struct per_cpu_zonestat {
--- a/mm/internal.h~mm-page_alloc-allow-high-order-pages-to-be-stored-on-the-per-cpu-lists
+++ a/mm/internal.h
@@ -203,7 +203,7 @@ extern void post_alloc_hook(struct page
gfp_t gfp_flags);
extern int user_min_free_kbytes;
-extern void free_unref_page(struct page *page);
+extern void free_unref_page(struct page *page, unsigned int order);
extern void free_unref_page_list(struct list_head *list);
extern void zone_pcp_update(struct zone *zone, int cpu_online);
--- a/mm/page_alloc.c~mm-page_alloc-allow-high-order-pages-to-be-stored-on-the-per-cpu-lists
+++ a/mm/page_alloc.c
@@ -674,10 +674,53 @@ out:
add_taint(TAINT_BAD_PAGE, LOCKDEP_NOW_UNRELIABLE);
}
+static inline unsigned int order_to_pindex(int migratetype, int order)
+{
+ int base = order;
+
+#ifdef CONFIG_TRANSPARENT_HUGEPAGE
+ if (order > PAGE_ALLOC_COSTLY_ORDER) {
+ VM_BUG_ON(order != pageblock_order);
+ base = PAGE_ALLOC_COSTLY_ORDER + 1;
+ }
+#else
+ VM_BUG_ON(order > PAGE_ALLOC_COSTLY_ORDER);
+#endif
+
+ return (MIGRATE_PCPTYPES * base) + migratetype;
+}
+
+static inline int pindex_to_order(unsigned int pindex)
+{
+ int order = pindex / MIGRATE_PCPTYPES;
+
+#ifdef CONFIG_TRANSPARENT_HUGEPAGE
+ if (order > PAGE_ALLOC_COSTLY_ORDER) {
+ order = pageblock_order;
+ VM_BUG_ON(order != pageblock_order);
+ }
+#else
+ VM_BUG_ON(order > PAGE_ALLOC_COSTLY_ORDER);
+#endif
+
+ return order;
+}
+
+static inline bool pcp_allowed_order(unsigned int order)
+{
+ if (order <= PAGE_ALLOC_COSTLY_ORDER)
+ return true;
+#ifdef CONFIG_TRANSPARENT_HUGEPAGE
+ if (order == pageblock_order)
+ return true;
+#endif
+ return false;
+}
+
static inline void free_the_page(struct page *page, unsigned int order)
{
- if (order == 0) /* Via pcp? */
- free_unref_page(page);
+ if (pcp_allowed_order(order)) /* Via pcp? */
+ free_unref_page(page, order);
else
__free_pages_ok(page, order, FPI_NONE);
}
@@ -700,7 +743,7 @@ static inline void free_the_page(struct
void free_compound_page(struct page *page)
{
mem_cgroup_uncharge(page);
- __free_pages_ok(page, compound_order(page), FPI_NONE);
+ free_the_page(page, compound_order(page));
}
void prep_compound_page(struct page *page, unsigned int order)
@@ -1350,9 +1393,9 @@ static __always_inline bool free_pages_p
* to pcp lists. With debug_pagealloc also enabled, they are also rechecked when
* moved from pcp lists to free lists.
*/
-static bool free_pcp_prepare(struct page *page)
+static bool free_pcp_prepare(struct page *page, unsigned int order)
{
- return free_pages_prepare(page, 0, true, FPI_NONE);
+ return free_pages_prepare(page, order, true, FPI_NONE);
}
static bool bulkfree_pcp_prepare(struct page *page)
@@ -1369,12 +1412,12 @@ static bool bulkfree_pcp_prepare(struct
* debug_pagealloc enabled, they are checked also immediately when being freed
* to the pcp lists.
*/
-static bool free_pcp_prepare(struct page *page)
+static bool free_pcp_prepare(struct page *page, unsigned int order)
{
if (debug_pagealloc_enabled_static())
- return free_pages_prepare(page, 0, true, FPI_NONE);
+ return free_pages_prepare(page, order, true, FPI_NONE);
else
- return free_pages_prepare(page, 0, false, FPI_NONE);
+ return free_pages_prepare(page, order, false, FPI_NONE);
}
static bool bulkfree_pcp_prepare(struct page *page)
@@ -1406,8 +1449,10 @@ static inline void prefetch_buddy(struct
static void free_pcppages_bulk(struct zone *zone, int count,
struct per_cpu_pages *pcp)
{
- int migratetype = 0;
+ int pindex = 0;
int batch_free = 0;
+ int nr_freed = 0;
+ unsigned int order;
int prefetch_nr = READ_ONCE(pcp->batch);
bool isolated_pageblocks;
struct page *page, *tmp;
@@ -1418,7 +1463,7 @@ static void free_pcppages_bulk(struct zo
* below while (list_empty(list)) loop.
*/
count = min(pcp->count, count);
- while (count) {
+ while (count > 0) {
struct list_head *list;
/*
@@ -1430,24 +1475,31 @@ static void free_pcppages_bulk(struct zo
*/
do {
batch_free++;
- if (++migratetype == MIGRATE_PCPTYPES)
- migratetype = 0;
- list = &pcp->lists[migratetype];
+ if (++pindex == NR_PCP_LISTS)
+ pindex = 0;
+ list = &pcp->lists[pindex];
} while (list_empty(list));
/* This is the only non-empty list. Free them all. */
- if (batch_free == MIGRATE_PCPTYPES)
+ if (batch_free == NR_PCP_LISTS)
batch_free = count;
+ order = pindex_to_order(pindex);
+ BUILD_BUG_ON(MAX_ORDER >= (1<<NR_PCP_ORDER_WIDTH));
do {
page = list_last_entry(list, struct page, lru);
/* must delete to avoid corrupting pcp list */
list_del(&page->lru);
- pcp->count--;
+ nr_freed += 1 << order;
+ count -= 1 << order;
if (bulkfree_pcp_prepare(page))
continue;
+ /* Encode order with the migratetype */
+ page->index <<= NR_PCP_ORDER_WIDTH;
+ page->index |= order;
+
list_add_tail(&page->lru, &head);
/*
@@ -1463,8 +1515,9 @@ static void free_pcppages_bulk(struct zo
prefetch_buddy(page);
prefetch_nr--;
}
- } while (--count && --batch_free && !list_empty(list));
+ } while (count > 0 && --batch_free && !list_empty(list));
}
+ pcp->count -= nr_freed;
/*
* local_lock_irq held so equivalent to spin_lock_irqsave for
@@ -1479,14 +1532,19 @@ static void free_pcppages_bulk(struct zo
*/
list_for_each_entry_safe(page, tmp, &head, lru) {
int mt = get_pcppage_migratetype(page);
+
+ /* mt has been encoded with the order (see above) */
+ order = mt & NR_PCP_ORDER_MASK;
+ mt >>= NR_PCP_ORDER_WIDTH;
+
/* MIGRATE_ISOLATE page should not go to pcplists */
VM_BUG_ON_PAGE(is_migrate_isolate(mt), page);
/* Pageblock could have been isolated meanwhile */
if (unlikely(isolated_pageblocks))
mt = get_pageblock_migratetype(page);
- __free_one_page(page, page_to_pfn(page), zone, 0, mt, FPI_NONE);
- trace_mm_page_pcpu_drain(page, 0, mt);
+ __free_one_page(page, page_to_pfn(page), zone, order, mt, FPI_NONE);
+ trace_mm_page_pcpu_drain(page, order, mt);
}
spin_unlock(&zone->lock);
}
@@ -3263,11 +3321,12 @@ void mark_free_pages(struct zone *zone)
}
#endif /* CONFIG_PM */
-static bool free_unref_page_prepare(struct page *page, unsigned long pfn)
+static bool free_unref_page_prepare(struct page *page, unsigned long pfn,
+ unsigned int order)
{
int migratetype;
- if (!free_pcp_prepare(page))
+ if (!free_pcp_prepare(page, order))
return false;
migratetype = get_pfnblock_migratetype(page, pfn);
@@ -3317,16 +3376,18 @@ static int nr_pcp_high(struct per_cpu_pa
}
static void free_unref_page_commit(struct page *page, unsigned long pfn,
- int migratetype)
+ int migratetype, unsigned int order)
{
struct zone *zone = page_zone(page);
struct per_cpu_pages *pcp;
int high;
+ int pindex;
__count_vm_event(PGFREE);
pcp = this_cpu_ptr(zone->per_cpu_pageset);
- list_add(&page->lru, &pcp->lists[migratetype]);
- pcp->count++;
+ pindex = order_to_pindex(migratetype, order);
+ list_add(&page->lru, &pcp->lists[pindex]);
+ pcp->count += 1 << order;
high = nr_pcp_high(pcp, zone);
if (pcp->count >= high) {
int batch = READ_ONCE(pcp->batch);
@@ -3336,15 +3397,15 @@ static void free_unref_page_commit(struc
}
/*
- * Free a 0-order page
+ * Free a pcp page
*/
-void free_unref_page(struct page *page)
+void free_unref_page(struct page *page, unsigned int order)
{
unsigned long flags;
unsigned long pfn = page_to_pfn(page);
int migratetype;
- if (!free_unref_page_prepare(page, pfn))
+ if (!free_unref_page_prepare(page, pfn, order))
return;
/*
@@ -3357,14 +3418,14 @@ void free_unref_page(struct page *page)
migratetype = get_pcppage_migratetype(page);
if (unlikely(migratetype >= MIGRATE_PCPTYPES)) {
if (unlikely(is_migrate_isolate(migratetype))) {
- free_one_page(page_zone(page), page, pfn, 0, migratetype, FPI_NONE);
+ free_one_page(page_zone(page), page, pfn, order, migratetype, FPI_NONE);
return;
}
migratetype = MIGRATE_MOVABLE;
}
local_lock_irqsave(&pagesets.lock, flags);
- free_unref_page_commit(page, pfn, migratetype);
+ free_unref_page_commit(page, pfn, migratetype, order);
local_unlock_irqrestore(&pagesets.lock, flags);
}
@@ -3381,7 +3442,7 @@ void free_unref_page_list(struct list_he
/* Prepare pages for freeing */
list_for_each_entry_safe(page, next, list, lru) {
pfn = page_to_pfn(page);
- if (!free_unref_page_prepare(page, pfn))
+ if (!free_unref_page_prepare(page, pfn, 0))
list_del(&page->lru);
/*
@@ -3413,7 +3474,7 @@ void free_unref_page_list(struct list_he
set_page_private(page, 0);
migratetype = get_pcppage_migratetype(page);
trace_mm_page_free_batched(page);
- free_unref_page_commit(page, pfn, migratetype);
+ free_unref_page_commit(page, pfn, migratetype, 0);
/*
* Guard against excessive IRQ disabled times when we get
@@ -3549,7 +3610,8 @@ static inline void zone_statistics(struc
/* Remove page from the per-cpu list, caller must protect the list */
static inline
-struct page *__rmqueue_pcplist(struct zone *zone, int migratetype,
+struct page *__rmqueue_pcplist(struct zone *zone, unsigned int order,
+ int migratetype,
unsigned int alloc_flags,
struct per_cpu_pages *pcp,
struct list_head *list)
@@ -3558,16 +3620,30 @@ struct page *__rmqueue_pcplist(struct zo
do {
if (list_empty(list)) {
- pcp->count += rmqueue_bulk(zone, 0,
- READ_ONCE(pcp->batch), list,
+ int batch = READ_ONCE(pcp->batch);
+ int alloced;
+
+ /*
+ * Scale batch relative to order if batch implies
+ * free pages can be stored on the PCP. Batch can
+ * be 1 for small zones or for boot pagesets which
+ * should never store free pages as the pages may
+ * belong to arbitrary zones.
+ */
+ if (batch > 1)
+ batch = max(batch >> order, 2);
+ alloced = rmqueue_bulk(zone, order,
+ batch, list,
migratetype, alloc_flags);
+
+ pcp->count += alloced << order;
if (unlikely(list_empty(list)))
return NULL;
}
page = list_first_entry(list, struct page, lru);
list_del(&page->lru);
- pcp->count--;
+ pcp->count -= 1 << order;
} while (check_new_pcp(page));
return page;
@@ -3575,8 +3651,9 @@ struct page *__rmqueue_pcplist(struct zo
/* Lock and remove page from the per-cpu list */
static struct page *rmqueue_pcplist(struct zone *preferred_zone,
- struct zone *zone, gfp_t gfp_flags,
- int migratetype, unsigned int alloc_flags)
+ struct zone *zone, unsigned int order,
+ gfp_t gfp_flags, int migratetype,
+ unsigned int alloc_flags)
{
struct per_cpu_pages *pcp;
struct list_head *list;
@@ -3592,8 +3669,8 @@ static struct page *rmqueue_pcplist(stru
*/
pcp = this_cpu_ptr(zone->per_cpu_pageset);
pcp->free_factor >>= 1;
- list = &pcp->lists[migratetype];
- page = __rmqueue_pcplist(zone, migratetype, alloc_flags, pcp, list);
+ list = &pcp->lists[order_to_pindex(migratetype, order)];
+ page = __rmqueue_pcplist(zone, order, migratetype, alloc_flags, pcp, list);
local_unlock_irqrestore(&pagesets.lock, flags);
if (page) {
__count_zid_vm_events(PGALLOC, page_zonenum(page), 1);
@@ -3614,15 +3691,15 @@ struct page *rmqueue(struct zone *prefer
unsigned long flags;
struct page *page;
- if (likely(order == 0)) {
+ if (likely(pcp_allowed_order(order))) {
/*
* MIGRATE_MOVABLE pcplist could have the pages on CMA area and
* we need to skip it when CMA area isn't allowed.
*/
if (!IS_ENABLED(CONFIG_CMA) || alloc_flags & ALLOC_CMA ||
migratetype != MIGRATE_MOVABLE) {
- page = rmqueue_pcplist(preferred_zone, zone, gfp_flags,
- migratetype, alloc_flags);
+ page = rmqueue_pcplist(preferred_zone, zone, order,
+ gfp_flags, migratetype, alloc_flags);
goto out;
}
}
@@ -5201,7 +5278,7 @@ unsigned long __alloc_pages_bulk(gfp_t g
/* Attempt the batch allocation */
local_lock_irqsave(&pagesets.lock, flags);
pcp = this_cpu_ptr(zone->per_cpu_pageset);
- pcp_list = &pcp->lists[ac.migratetype];
+ pcp_list = &pcp->lists[order_to_pindex(ac.migratetype, 0)];
while (nr_populated < nr_pages) {
@@ -5211,7 +5288,7 @@ unsigned long __alloc_pages_bulk(gfp_t g
continue;
}
- page = __rmqueue_pcplist(zone, ac.migratetype, alloc_flags,
+ page = __rmqueue_pcplist(zone, 0, ac.migratetype, alloc_flags,
pcp, pcp_list);
if (unlikely(!page)) {
/* Try and get at least one page */
@@ -6778,13 +6855,13 @@ static void pageset_update(struct per_cp
static void per_cpu_pages_init(struct per_cpu_pages *pcp, struct per_cpu_zonestat *pzstats)
{
- int migratetype;
+ int pindex;
memset(pcp, 0, sizeof(*pcp));
memset(pzstats, 0, sizeof(*pzstats));
- for (migratetype = 0; migratetype < MIGRATE_PCPTYPES; migratetype++)
- INIT_LIST_HEAD(&pcp->lists[migratetype]);
+ for (pindex = 0; pindex < NR_PCP_LISTS; pindex++)
+ INIT_LIST_HEAD(&pcp->lists[pindex]);
/*
* Set batch and high values safe for a boot pageset. A true percpu
--- a/mm/swap.c~mm-page_alloc-allow-high-order-pages-to-be-stored-on-the-per-cpu-lists
+++ a/mm/swap.c
@@ -95,7 +95,7 @@ static void __put_single_page(struct pag
{
__page_cache_release(page);
mem_cgroup_uncharge(page);
- free_unref_page(page);
+ free_unref_page(page, 0);
}
static void __put_compound_page(struct page *page)
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 190/192] mm/page_alloc: split pcp->high across all online CPUs for cpuless nodes
2021-06-29 2:32 incoming Andrew Morton
` (188 preceding siblings ...)
2021-06-29 2:43 ` [patch 189/192] mm/page_alloc: allow high-order pages to be stored on the per-cpu lists Andrew Morton
@ 2021-06-29 2:43 ` Andrew Morton
2021-06-29 2:43 ` [patch 191/192] mm,hwpoison: send SIGBUS with error virutal address Andrew Morton
2021-06-29 2:43 ` [patch 192/192] mm,hwpoison: make get_hwpoison_page() call get_any_page() Andrew Morton
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:43 UTC (permalink / raw)
To: akpm, dave.hansen, feng.tang, hdanton, linux-mm, mgorman, mhocko,
mm-commits, torvalds, vbabka
From: Mel Gorman <mgorman@techsingularity.net>
Subject: mm/page_alloc: split pcp->high across all online CPUs for cpuless nodes
Dave Hansen reported the following about Feng Tang's tests on a machine
with persistent memory onlined as a DRAM-like device.
Feng Tang tossed these on a "Cascade Lake" system with 96 threads and
~512G of persistent memory and 128G of DRAM. The PMEM is in "volatile
use" mode and being managed via the buddy just like the normal RAM.
The PMEM zones are big ones:
present 65011712 = 248 G
high 134595 = 525 M
The PMEM nodes, of course, don't have any CPUs in them.
With your series, the pcp->high value per-cpu is 69584 pages or about
270MB per CPU. Scaled up by the 96 CPU threads, that's ~26GB of
worst-case memory in the pcps per zone, or roughly 10% of the size of
the zone.
This should not cause a problem as such although it could trigger reclaim
due to pages being stored on per-cpu lists for CPUs remote to a node. It
is not possible to treat cpuless nodes exactly the same as normal nodes
but the worst-case scenario can be mitigated by splitting pcp->high across
all online CPUs for cpuless memory nodes.
Link: https://lkml.kernel.org/r/20210616110743.GK30378@techsingularity.net
Suggested-by: Dave Hansen <dave.hansen@intel.com>
Signed-off-by: Mel Gorman <mgorman@techsingularity.net>
Acked-by: Vlastimil Babka <vbabka@suse.cz>
Acked-by: Dave Hansen <dave.hansen@intel.com>
Cc: Hillf Danton <hdanton@sina.com>
Cc: Michal Hocko <mhocko@kernel.org>
Cc: "Tang, Feng" <feng.tang@intel.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/page_alloc.c | 12 ++++++++----
1 file changed, 8 insertions(+), 4 deletions(-)
--- a/mm/page_alloc.c~mm-page_alloc-split-pcp-high-across-all-online-cpus-for-cpuless-nodes
+++ a/mm/page_alloc.c
@@ -6790,7 +6790,7 @@ static int zone_highsize(struct zone *zo
{
#ifdef CONFIG_MMU
int high;
- int nr_local_cpus;
+ int nr_split_cpus;
unsigned long total_pages;
if (!percpu_pagelist_high_fraction) {
@@ -6813,10 +6813,14 @@ static int zone_highsize(struct zone *zo
* Split the high value across all online CPUs local to the zone. Note
* that early in boot that CPUs may not be online yet and that during
* CPU hotplug that the cpumask is not yet updated when a CPU is being
- * onlined.
+ * onlined. For memory nodes that have no CPUs, split pcp->high across
+ * all online CPUs to mitigate the risk that reclaim is triggered
+ * prematurely due to pages stored on pcp lists.
*/
- nr_local_cpus = max(1U, cpumask_weight(cpumask_of_node(zone_to_nid(zone)))) + cpu_online;
- high = total_pages / nr_local_cpus;
+ nr_split_cpus = cpumask_weight(cpumask_of_node(zone_to_nid(zone))) + cpu_online;
+ if (!nr_split_cpus)
+ nr_split_cpus = num_online_cpus();
+ high = total_pages / nr_split_cpus;
/*
* Ensure high is at least batch*4. The multiple is based on the
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 191/192] mm,hwpoison: send SIGBUS with error virutal address
2021-06-29 2:32 incoming Andrew Morton
` (189 preceding siblings ...)
2021-06-29 2:43 ` [patch 190/192] mm/page_alloc: split pcp->high across all online CPUs for cpuless nodes Andrew Morton
@ 2021-06-29 2:43 ` Andrew Morton
2021-06-29 2:43 ` [patch 192/192] mm,hwpoison: make get_hwpoison_page() call get_any_page() Andrew Morton
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:43 UTC (permalink / raw)
To: akpm, bp, bp, david, juew, linux-mm, luto, mm-commits,
naoya.horiguchi, osalvador, tony.luck, torvalds, yaoaili
From: Naoya Horiguchi <naoya.horiguchi@nec.com>
Subject: mm,hwpoison: send SIGBUS with error virutal address
Now an action required MCE in already hwpoisoned address surely sends a
SIGBUS to current process, but the SIGBUS doesn't convey error virtual
address. That's not optimal for hwpoison-aware applications.
To fix the issue, make memory_failure() call kill_accessing_process(),
that does pagetable walk to find the error virtual address. It could find
multiple virtual addresses for the same error page, and it seems hard to
tell which virtual address is correct one. But that's rare and sending
incorrect virtual address could be better than no address. So let's
report the first found virtual address for now.
[naoya.horiguchi@nec.com: fix walk_page_range() return]
Link: https://lkml.kernel.org/r/20210603051055.GA244241@hori.linux.bs1.fc.nec.co.jp
Link: https://lkml.kernel.org/r/20210521030156.2612074-4-nao.horiguchi@gmail.com
Signed-off-by: Naoya Horiguchi <naoya.horiguchi@nec.com>
Cc: Tony Luck <tony.luck@intel.com>
Cc: Aili Yao <yaoaili@kingsoft.com>
Cc: Oscar Salvador <osalvador@suse.de>
Cc: David Hildenbrand <david@redhat.com>
Cc: Borislav Petkov <bp@alien8.de>
Cc: Andy Lutomirski <luto@kernel.org>
Cc: Jue Wang <juew@google.com>
Cc: Borislav Petkov <bp@suse.de>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
arch/x86/kernel/cpu/mce/core.c | 13 ++
include/linux/swapops.h | 5 +
mm/memory-failure.c | 150 ++++++++++++++++++++++++++++++-
3 files changed, 165 insertions(+), 3 deletions(-)
--- a/arch/x86/kernel/cpu/mce/core.c~mmhwpoison-send-sigbus-with-error-virutal-address
+++ a/arch/x86/kernel/cpu/mce/core.c
@@ -1257,19 +1257,28 @@ static void kill_me_maybe(struct callbac
{
struct task_struct *p = container_of(cb, struct task_struct, mce_kill_me);
int flags = MF_ACTION_REQUIRED;
+ int ret;
pr_err("Uncorrected hardware memory error in user-access at %llx", p->mce_addr);
if (!p->mce_ripv)
flags |= MF_MUST_KILL;
- if (!memory_failure(p->mce_addr >> PAGE_SHIFT, flags) &&
- !(p->mce_kflags & MCE_IN_KERNEL_COPYIN)) {
+ ret = memory_failure(p->mce_addr >> PAGE_SHIFT, flags);
+ if (!ret && !(p->mce_kflags & MCE_IN_KERNEL_COPYIN)) {
set_mce_nospec(p->mce_addr >> PAGE_SHIFT, p->mce_whole_page);
sync_core();
return;
}
+ /*
+ * -EHWPOISON from memory_failure() means that it already sent SIGBUS
+ * to the current process with the proper error info, so no need to
+ * send SIGBUS here again.
+ */
+ if (ret == -EHWPOISON)
+ return;
+
if (p->mce_vaddr != (void __user *)-1l) {
force_sig_mceerr(BUS_MCEERR_AR, p->mce_vaddr, PAGE_SHIFT);
} else {
--- a/include/linux/swapops.h~mmhwpoison-send-sigbus-with-error-virutal-address
+++ a/include/linux/swapops.h
@@ -330,6 +330,11 @@ static inline int is_hwpoison_entry(swp_
return swp_type(entry) == SWP_HWPOISON;
}
+static inline unsigned long hwpoison_entry_to_pfn(swp_entry_t entry)
+{
+ return swp_offset(entry);
+}
+
static inline void num_poisoned_pages_inc(void)
{
atomic_long_inc(&num_poisoned_pages);
--- a/mm/memory-failure.c~mmhwpoison-send-sigbus-with-error-virutal-address
+++ a/mm/memory-failure.c
@@ -56,6 +56,7 @@
#include <linux/kfifo.h>
#include <linux/ratelimit.h>
#include <linux/page-isolation.h>
+#include <linux/pagewalk.h>
#include "internal.h"
#include "ras/ras_event.h"
@@ -554,6 +555,148 @@ static void collect_procs(struct page *p
collect_procs_file(page, tokill, force_early);
}
+struct hwp_walk {
+ struct to_kill tk;
+ unsigned long pfn;
+ int flags;
+};
+
+static void set_to_kill(struct to_kill *tk, unsigned long addr, short shift)
+{
+ tk->addr = addr;
+ tk->size_shift = shift;
+}
+
+static int check_hwpoisoned_entry(pte_t pte, unsigned long addr, short shift,
+ unsigned long poisoned_pfn, struct to_kill *tk)
+{
+ unsigned long pfn = 0;
+
+ if (pte_present(pte)) {
+ pfn = pte_pfn(pte);
+ } else {
+ swp_entry_t swp = pte_to_swp_entry(pte);
+
+ if (is_hwpoison_entry(swp))
+ pfn = hwpoison_entry_to_pfn(swp);
+ }
+
+ if (!pfn || pfn != poisoned_pfn)
+ return 0;
+
+ set_to_kill(tk, addr, shift);
+ return 1;
+}
+
+#ifdef CONFIG_TRANSPARENT_HUGEPAGE
+static int check_hwpoisoned_pmd_entry(pmd_t *pmdp, unsigned long addr,
+ struct hwp_walk *hwp)
+{
+ pmd_t pmd = *pmdp;
+ unsigned long pfn;
+ unsigned long hwpoison_vaddr;
+
+ if (!pmd_present(pmd))
+ return 0;
+ pfn = pmd_pfn(pmd);
+ if (pfn <= hwp->pfn && hwp->pfn < pfn + HPAGE_PMD_NR) {
+ hwpoison_vaddr = addr + ((hwp->pfn - pfn) << PAGE_SHIFT);
+ set_to_kill(&hwp->tk, hwpoison_vaddr, PAGE_SHIFT);
+ return 1;
+ }
+ return 0;
+}
+#else
+static int check_hwpoisoned_pmd_entry(pmd_t *pmdp, unsigned long addr,
+ struct hwp_walk *hwp)
+{
+ return 0;
+}
+#endif
+
+static int hwpoison_pte_range(pmd_t *pmdp, unsigned long addr,
+ unsigned long end, struct mm_walk *walk)
+{
+ struct hwp_walk *hwp = (struct hwp_walk *)walk->private;
+ int ret = 0;
+ pte_t *ptep;
+ spinlock_t *ptl;
+
+ ptl = pmd_trans_huge_lock(pmdp, walk->vma);
+ if (ptl) {
+ ret = check_hwpoisoned_pmd_entry(pmdp, addr, hwp);
+ spin_unlock(ptl);
+ goto out;
+ }
+
+ if (pmd_trans_unstable(pmdp))
+ goto out;
+
+ ptep = pte_offset_map_lock(walk->vma->vm_mm, pmdp, addr, &ptl);
+ for (; addr != end; ptep++, addr += PAGE_SIZE) {
+ ret = check_hwpoisoned_entry(*ptep, addr, PAGE_SHIFT,
+ hwp->pfn, &hwp->tk);
+ if (ret == 1)
+ break;
+ }
+ pte_unmap_unlock(ptep - 1, ptl);
+out:
+ cond_resched();
+ return ret;
+}
+
+#ifdef CONFIG_HUGETLB_PAGE
+static int hwpoison_hugetlb_range(pte_t *ptep, unsigned long hmask,
+ unsigned long addr, unsigned long end,
+ struct mm_walk *walk)
+{
+ struct hwp_walk *hwp = (struct hwp_walk *)walk->private;
+ pte_t pte = huge_ptep_get(ptep);
+ struct hstate *h = hstate_vma(walk->vma);
+
+ return check_hwpoisoned_entry(pte, addr, huge_page_shift(h),
+ hwp->pfn, &hwp->tk);
+}
+#else
+#define hwpoison_hugetlb_range NULL
+#endif
+
+static struct mm_walk_ops hwp_walk_ops = {
+ .pmd_entry = hwpoison_pte_range,
+ .hugetlb_entry = hwpoison_hugetlb_range,
+};
+
+/*
+ * Sends SIGBUS to the current process with error info.
+ *
+ * This function is intended to handle "Action Required" MCEs on already
+ * hardware poisoned pages. They could happen, for example, when
+ * memory_failure() failed to unmap the error page at the first call, or
+ * when multiple local machine checks happened on different CPUs.
+ *
+ * MCE handler currently has no easy access to the error virtual address,
+ * so this function walks page table to find it. The returned virtual address
+ * is proper in most cases, but it could be wrong when the application
+ * process has multiple entries mapping the error page.
+ */
+static int kill_accessing_process(struct task_struct *p, unsigned long pfn,
+ int flags)
+{
+ int ret;
+ struct hwp_walk priv = {
+ .pfn = pfn,
+ };
+ priv.tk.tsk = p;
+
+ mmap_read_lock(p->mm);
+ ret = walk_page_range(p->mm, 0, TASK_SIZE, &hwp_walk_ops,
+ (void *)&priv);
+ if (ret == 1 && priv.tk.addr)
+ kill_proc(&priv.tk, pfn, flags);
+ mmap_read_unlock(p->mm);
+ return ret ? -EFAULT : -EHWPOISON;
+}
+
static const char *action_name[] = {
[MF_IGNORED] = "Ignored",
[MF_FAILED] = "Failed",
@@ -1267,7 +1410,10 @@ static int memory_failure_hugetlb(unsign
if (TestSetPageHWPoison(head)) {
pr_err("Memory failure: %#lx: already hardware poisoned\n",
pfn);
- return -EHWPOISON;
+ res = -EHWPOISON;
+ if (flags & MF_ACTION_REQUIRED)
+ res = kill_accessing_process(current, page_to_pfn(head), flags);
+ return res;
}
num_poisoned_pages_inc();
@@ -1476,6 +1622,8 @@ try_again:
pr_err("Memory failure: %#lx: already hardware poisoned\n",
pfn);
res = -EHWPOISON;
+ if (flags & MF_ACTION_REQUIRED)
+ res = kill_accessing_process(current, pfn, flags);
goto unlock_mutex;
}
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* [patch 192/192] mm,hwpoison: make get_hwpoison_page() call get_any_page()
2021-06-29 2:32 incoming Andrew Morton
` (190 preceding siblings ...)
2021-06-29 2:43 ` [patch 191/192] mm,hwpoison: send SIGBUS with error virutal address Andrew Morton
@ 2021-06-29 2:43 ` Andrew Morton
191 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-29 2:43 UTC (permalink / raw)
To: akpm, linux-mm, mhocko, mike.kravetz, mm-commits,
naoya.horiguchi, osalvador, songmuchun, tony.luck, torvalds
From: Naoya Horiguchi <naoya.horiguchi@nec.com>
Subject: mm,hwpoison: make get_hwpoison_page() call get_any_page()
__get_hwpoison_page() could fail to grab refcount by some race condition,
so it's helpful if we can handle it by retrying. We already have retry
logic, so make get_hwpoison_page() call get_any_page() when called from
memory_failure().
As a result, get_hwpoison_page() can return negative values (i.e. error
code), so some callers are also changed to handle error cases.
soft_offline_page() does nothing for -EBUSY because that's enough and
users in userspace can easily handle it. unpoison_memory() is also
unchanged because it's broken and need thorough fixes (will be done
later).
Link: https://lkml.kernel.org/r/20210603233632.2964832-3-nao.horiguchi@gmail.com
Signed-off-by: Naoya Horiguchi <naoya.horiguchi@nec.com>
Cc: Oscar Salvador <osalvador@suse.de>
Cc: Muchun Song <songmuchun@bytedance.com>
Cc: Mike Kravetz <mike.kravetz@oracle.com>
Cc: Michal Hocko <mhocko@suse.com>
Cc: Tony Luck <tony.luck@intel.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/hugetlb.c | 2
mm/memory-failure.c | 194 +++++++++++++++++++++++-------------------
2 files changed, 111 insertions(+), 85 deletions(-)
--- a/mm/hugetlb.c~mmhwpoison-make-get_hwpoison_page-call-get_any_page
+++ a/mm/hugetlb.c
@@ -5938,6 +5938,8 @@ int get_hwpoison_huge_page(struct page *
*hugetlb = true;
if (HPageFreed(page) || HPageMigratable(page))
ret = get_page_unless_zero(page);
+ else
+ ret = -EBUSY;
}
spin_unlock_irq(&hugetlb_lock);
return ret;
--- a/mm/memory-failure.c~mmhwpoison-make-get_hwpoison_page-call-get_any_page
+++ a/mm/memory-failure.c
@@ -1117,13 +1117,6 @@ static inline bool HWPoisonHandlable(str
return PageLRU(page) || __PageMovable(page);
}
-/**
- * __get_hwpoison_page() - Get refcount for memory error handling:
- * @page: raw error page (hit by memory error)
- *
- * Return: return 0 if failed to grab the refcount, otherwise true (some
- * non-zero value.)
- */
static int __get_hwpoison_page(struct page *page)
{
struct page *head = compound_head(page);
@@ -1168,15 +1161,6 @@ static int __get_hwpoison_page(struct pa
return 0;
}
-/*
- * Safely get reference count of an arbitrary page.
- *
- * Returns 0 for a free page, 1 for an in-use page,
- * -EIO for a page-type we cannot handle and -EBUSY if we raced with an
- * allocation.
- * We only incremented refcount in case the page was already in-use and it
- * is a known type we can handle.
- */
static int get_any_page(struct page *p, unsigned long flags)
{
int ret = 0, pass = 0;
@@ -1186,50 +1170,77 @@ static int get_any_page(struct page *p,
count_increased = true;
try_again:
- if (!count_increased && !__get_hwpoison_page(p)) {
- if (page_count(p)) {
- /* We raced with an allocation, retry. */
- if (pass++ < 3)
- goto try_again;
- ret = -EBUSY;
- } else if (!PageHuge(p) && !is_free_buddy_page(p)) {
- /* We raced with put_page, retry. */
+ if (!count_increased) {
+ ret = __get_hwpoison_page(p);
+ if (!ret) {
+ if (page_count(p)) {
+ /* We raced with an allocation, retry. */
+ if (pass++ < 3)
+ goto try_again;
+ ret = -EBUSY;
+ } else if (!PageHuge(p) && !is_free_buddy_page(p)) {
+ /* We raced with put_page, retry. */
+ if (pass++ < 3)
+ goto try_again;
+ ret = -EIO;
+ }
+ goto out;
+ } else if (ret == -EBUSY) {
+ /* We raced with freeing huge page to buddy, retry. */
if (pass++ < 3)
goto try_again;
- ret = -EIO;
+ goto out;
}
+ }
+
+ if (PageHuge(p) || HWPoisonHandlable(p)) {
+ ret = 1;
} else {
- if (PageHuge(p) || HWPoisonHandlable(p)) {
- ret = 1;
- } else {
- /*
- * A page we cannot handle. Check whether we can turn
- * it into something we can handle.
- */
- if (pass++ < 3) {
- put_page(p);
- shake_page(p, 1);
- count_increased = false;
- goto try_again;
- }
+ /*
+ * A page we cannot handle. Check whether we can turn
+ * it into something we can handle.
+ */
+ if (pass++ < 3) {
put_page(p);
- ret = -EIO;
+ shake_page(p, 1);
+ count_increased = false;
+ goto try_again;
}
+ put_page(p);
+ ret = -EIO;
}
-
+out:
return ret;
}
-static int get_hwpoison_page(struct page *p, unsigned long flags,
- enum mf_flags ctxt)
+/**
+ * get_hwpoison_page() - Get refcount for memory error handling
+ * @p: Raw error page (hit by memory error)
+ * @flags: Flags controlling behavior of error handling
+ *
+ * get_hwpoison_page() takes a page refcount of an error page to handle memory
+ * error on it, after checking that the error page is in a well-defined state
+ * (defined as a page-type we can successfully handle the memor error on it,
+ * such as LRU page and hugetlb page).
+ *
+ * Memory error handling could be triggered at any time on any type of page,
+ * so it's prone to race with typical memory management lifecycle (like
+ * allocation and free). So to avoid such races, get_hwpoison_page() takes
+ * extra care for the error page's state (as done in __get_hwpoison_page()),
+ * and has some retry logic in get_any_page().
+ *
+ * Return: 0 on failure,
+ * 1 on success for in-use pages in a well-defined state,
+ * -EIO for pages on which we can not handle memory errors,
+ * -EBUSY when get_hwpoison_page() has raced with page lifecycle
+ * operations like allocation and free.
+ */
+static int get_hwpoison_page(struct page *p, unsigned long flags)
{
int ret;
zone_pcp_disable(page_zone(p));
- if (ctxt == MF_SOFT_OFFLINE)
- ret = get_any_page(p, flags);
- else
- ret = __get_hwpoison_page(p);
+ ret = get_any_page(p, flags);
zone_pcp_enable(page_zone(p));
return ret;
@@ -1418,27 +1429,33 @@ static int memory_failure_hugetlb(unsign
num_poisoned_pages_inc();
- if (!(flags & MF_COUNT_INCREASED) && !get_hwpoison_page(p, flags, 0)) {
- /*
- * Check "filter hit" and "race with other subpage."
- */
- lock_page(head);
- if (PageHWPoison(head)) {
- if ((hwpoison_filter(p) && TestClearPageHWPoison(p))
- || (p != head && TestSetPageHWPoison(head))) {
- num_poisoned_pages_dec();
- unlock_page(head);
- return 0;
+ if (!(flags & MF_COUNT_INCREASED)) {
+ res = get_hwpoison_page(p, flags);
+ if (!res) {
+ /*
+ * Check "filter hit" and "race with other subpage."
+ */
+ lock_page(head);
+ if (PageHWPoison(head)) {
+ if ((hwpoison_filter(p) && TestClearPageHWPoison(p))
+ || (p != head && TestSetPageHWPoison(head))) {
+ num_poisoned_pages_dec();
+ unlock_page(head);
+ return 0;
+ }
}
+ unlock_page(head);
+ res = MF_FAILED;
+ if (!dissolve_free_huge_page(p) && take_page_off_buddy(p)) {
+ page_ref_inc(p);
+ res = MF_RECOVERED;
+ }
+ action_result(pfn, MF_MSG_FREE_HUGE, res);
+ return res == MF_RECOVERED ? 0 : -EBUSY;
+ } else if (res < 0) {
+ action_result(pfn, MF_MSG_UNKNOWN, MF_IGNORED);
+ return -EBUSY;
}
- unlock_page(head);
- res = MF_FAILED;
- if (!dissolve_free_huge_page(p) && take_page_off_buddy(p)) {
- page_ref_inc(p);
- res = MF_RECOVERED;
- }
- action_result(pfn, MF_MSG_FREE_HUGE, res);
- return res == MF_RECOVERED ? 0 : -EBUSY;
}
lock_page(head);
@@ -1641,28 +1658,35 @@ try_again:
* In fact it's dangerous to directly bump up page count from 0,
* that may make page_ref_freeze()/page_ref_unfreeze() mismatch.
*/
- if (!(flags & MF_COUNT_INCREASED) && !get_hwpoison_page(p, flags, 0)) {
- if (is_free_buddy_page(p)) {
- if (take_page_off_buddy(p)) {
- page_ref_inc(p);
- res = MF_RECOVERED;
- } else {
- /* We lost the race, try again */
- if (retry) {
- ClearPageHWPoison(p);
- num_poisoned_pages_dec();
- retry = false;
- goto try_again;
+ if (!(flags & MF_COUNT_INCREASED)) {
+ res = get_hwpoison_page(p, flags);
+ if (!res) {
+ if (is_free_buddy_page(p)) {
+ if (take_page_off_buddy(p)) {
+ page_ref_inc(p);
+ res = MF_RECOVERED;
+ } else {
+ /* We lost the race, try again */
+ if (retry) {
+ ClearPageHWPoison(p);
+ num_poisoned_pages_dec();
+ retry = false;
+ goto try_again;
+ }
+ res = MF_FAILED;
}
- res = MF_FAILED;
+ action_result(pfn, MF_MSG_BUDDY, res);
+ res = res == MF_RECOVERED ? 0 : -EBUSY;
+ } else {
+ action_result(pfn, MF_MSG_KERNEL_HIGH_ORDER, MF_IGNORED);
+ res = -EBUSY;
}
- action_result(pfn, MF_MSG_BUDDY, res);
- res = res == MF_RECOVERED ? 0 : -EBUSY;
- } else {
- action_result(pfn, MF_MSG_KERNEL_HIGH_ORDER, MF_IGNORED);
+ goto unlock_mutex;
+ } else if (res < 0) {
+ action_result(pfn, MF_MSG_UNKNOWN, MF_IGNORED);
res = -EBUSY;
+ goto unlock_mutex;
}
- goto unlock_mutex;
}
if (PageTransHuge(hpage)) {
@@ -1940,7 +1964,7 @@ int unpoison_memory(unsigned long pfn)
return 0;
}
- if (!get_hwpoison_page(p, flags, 0)) {
+ if (!get_hwpoison_page(p, flags)) {
if (TestClearPageHWPoison(p))
num_poisoned_pages_dec();
unpoison_pr_info("Unpoison: Software-unpoisoned free page %#lx\n",
@@ -2156,7 +2180,7 @@ int soft_offline_page(unsigned long pfn,
retry:
get_online_mems();
- ret = get_hwpoison_page(page, flags, MF_SOFT_OFFLINE);
+ ret = get_hwpoison_page(page, flags);
put_online_mems();
if (ret > 0) {
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: [patch 128/192] mm: improve mprotect(R|W) efficiency on pages referenced once
2021-06-29 2:39 ` [patch 128/192] mm: improve mprotect(R|W) efficiency on pages referenced once Andrew Morton
@ 2021-06-29 17:50 ` Linus Torvalds
2021-06-30 0:12 ` Peter Xu
0 siblings, 1 reply; 409+ messages in thread
From: Linus Torvalds @ 2021-06-29 17:50 UTC (permalink / raw)
To: Andrew Morton
Cc: Andrea Arcangeli, Evgeniy Stepanov, kostyak, Linux-MM,
mm-commits, Peter Collingbourne, Peter Xu
On Mon, Jun 28, 2021 at 7:40 PM Andrew Morton <akpm@linux-foundation.org> wrote:
>
>
> - /* Avoid taking write faults for known dirty pages */
> - if (dirty_accountable && pte_dirty(ptent) &&
> - (pte_soft_dirty(ptent) ||
> - !(vma->vm_flags & VM_SOFTDIRTY))) {
> + if (may_avoid_write_fault(ptent, vma, cp_flags))
> ptent = pte_mkwrite(ptent);
> - }
Hmm. I don't think this is correct.
As fat as I can tell, may_avoid_write_fault() doesn't even check if
the vma is writable!
Am I misreading it? Because I think you just made even a shared mmap
with "mprotect(PROT_READ)" turn the pte's writable.
Which is a "slight" security issue.
Maybe the new code is fine, and I'm missing something. The old code
looks strange too, which makes me think that the MM_CP_DIRTY_ACCT test
ends up saving us and depend on VM_WRITE. But it's very much not
obvious.
And even if I _am_ missing something, I really would like a very
obvious and direct test for "this vma is writable", ie maybe a
if (!(vma->vm_flags & VM_WRITE))
return false;
at the very top of the function.
And no, "pte_dirty()" is not a reason to make something writable, it
might have started out as a writable mapping, and we dirtied the page,
and we made it read-only. The page stays dirty, but it shouldn't
become writable just because of that.
So please make me get the warm and fuzzies about this code. Because
as-is, it just looks scary.
Linus
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: [patch 128/192] mm: improve mprotect(R|W) efficiency on pages referenced once
2021-06-29 17:50 ` Linus Torvalds
@ 2021-06-30 0:12 ` Peter Xu
2021-06-30 1:39 ` Peter Xu
0 siblings, 1 reply; 409+ messages in thread
From: Peter Xu @ 2021-06-30 0:12 UTC (permalink / raw)
To: Linus Torvalds
Cc: Andrew Morton, Andrea Arcangeli, Evgeniy Stepanov, kostyak,
Linux-MM, mm-commits, Peter Collingbourne
On Tue, Jun 29, 2021 at 10:50:12AM -0700, Linus Torvalds wrote:
> On Mon, Jun 28, 2021 at 7:40 PM Andrew Morton <akpm@linux-foundation.org> wrote:
> >
> >
> > - /* Avoid taking write faults for known dirty pages */
> > - if (dirty_accountable && pte_dirty(ptent) &&
> > - (pte_soft_dirty(ptent) ||
> > - !(vma->vm_flags & VM_SOFTDIRTY))) {
> > + if (may_avoid_write_fault(ptent, vma, cp_flags))
> > ptent = pte_mkwrite(ptent);
> > - }
>
> Hmm. I don't think this is correct.
>
> As fat as I can tell, may_avoid_write_fault() doesn't even check if
> the vma is writable!
>
> Am I misreading it? Because I think you just made even a shared mmap
> with "mprotect(PROT_READ)" turn the pte's writable.
>
> Which is a "slight" security issue.
>
> Maybe the new code is fine, and I'm missing something. The old code
> looks strange too, which makes me think that the MM_CP_DIRTY_ACCT test
> ends up saving us and depend on VM_WRITE. But it's very much not
> obvious.
vma_wants_writenotify() checks first VM_WRITE|VM_SHARED, otherwise
MM_CP_DIRTY_ACCT will not be set. While for anonymous vmas the newly
introduced may_avoid_write_fault() checks VM_WRITE explicitly.
Agreed even if it's checked it's not straightforward. Maybe it'll be a bonus
to have a comment above may_avoid_write_fault() about it in a follow up.
>
> And even if I _am_ missing something, I really would like a very
> obvious and direct test for "this vma is writable", ie maybe a
>
> if (!(vma->vm_flags & VM_WRITE))
> return false;
>
> at the very top of the function.
Yes looks okay too; I think using MM_CP_DIRTY_ACCT flag has a slight advantage
in that it checks VM_WRITE only once before calling change_protection(), rather
than doing the check for every pte even if we know it'll have the same result.
However it indeed hides the facts deeper..
>
> And no, "pte_dirty()" is not a reason to make something writable, it
> might have started out as a writable mapping, and we dirtied the page,
> and we made it read-only. The page stays dirty, but it shouldn't
> become writable just because of that.
I think the dirty bit checks are only to make sure we don't need those extra
write faults. It should definitely be based on the fact that VM_WRITE being
set already.
Thanks,
--
Peter Xu
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: [patch 128/192] mm: improve mprotect(R|W) efficiency on pages referenced once
2021-06-30 0:12 ` Peter Xu
@ 2021-06-30 1:39 ` Peter Xu
2021-06-30 2:25 ` Linus Torvalds
0 siblings, 1 reply; 409+ messages in thread
From: Peter Xu @ 2021-06-30 1:39 UTC (permalink / raw)
To: Linus Torvalds
Cc: Andrew Morton, Andrea Arcangeli, Evgeniy Stepanov, kostyak,
Linux-MM, mm-commits, Peter Collingbourne
On Tue, Jun 29, 2021 at 08:12:12PM -0400, Peter Xu wrote:
> On Tue, Jun 29, 2021 at 10:50:12AM -0700, Linus Torvalds wrote:
> > On Mon, Jun 28, 2021 at 7:40 PM Andrew Morton <akpm@linux-foundation.org> wrote:
> > >
> > >
> > > - /* Avoid taking write faults for known dirty pages */
> > > - if (dirty_accountable && pte_dirty(ptent) &&
> > > - (pte_soft_dirty(ptent) ||
> > > - !(vma->vm_flags & VM_SOFTDIRTY))) {
> > > + if (may_avoid_write_fault(ptent, vma, cp_flags))
> > > ptent = pte_mkwrite(ptent);
> > > - }
> >
> > Hmm. I don't think this is correct.
> >
> > As fat as I can tell, may_avoid_write_fault() doesn't even check if
> > the vma is writable!
> >
> > Am I misreading it? Because I think you just made even a shared mmap
> > with "mprotect(PROT_READ)" turn the pte's writable.
> >
> > Which is a "slight" security issue.
> >
> > Maybe the new code is fine, and I'm missing something. The old code
> > looks strange too, which makes me think that the MM_CP_DIRTY_ACCT test
> > ends up saving us and depend on VM_WRITE. But it's very much not
> > obvious.
>
> vma_wants_writenotify() checks first VM_WRITE|VM_SHARED, otherwise
> MM_CP_DIRTY_ACCT will not be set. While for anonymous vmas the newly
> introduced may_avoid_write_fault() checks VM_WRITE explicitly.
Sorry, this statement is unclear. It's not about anonymous or not, it's just
that a hidden check against VM_WRITE is there already..
Say, below chunk of the patch:
if (!(cp_flags & MM_CP_DIRTY_ACCT)) {
/* Otherwise, we must have exclusive access to the page. */
if (!(vma_is_anonymous(vma) && (vma->vm_flags & VM_WRITE)))
return false;
if (page_count(pte_page(pte)) != 1)
return false;
}
Should be the same as:
if (!(cp_flags & MM_CP_DIRTY_ACCT)) {
if (!vma_is_anonymous(vma))
return false;
if (!(vma->vm_flags & VM_WRITE))
return false;
if (page_count(pte_page(pte)) != 1)
return false;
}
And since MM_CP_DIRTY_ACCT implies "VM_WRITE|VM_SHARED" all set, above should
be a slightly faster version of below:
/* This just never trigger if MM_CP_DIRTY_ACCT set */
if (!(vma->vm_flags & VM_WRITE))
return false;
if (!(cp_flags & MM_CP_DIRTY_ACCT)) {
if (!vma_is_anonymous(vma))
return false;
if (page_count(pte_page(pte)) != 1)
return false;
}
It's just that we avoid checking "vma->vm_flags & VM_WRITE" when
MM_CP_DIRTY_ACCT set.
Again, I think in all cases some more comment should be good indeed..
>
> Agreed even if it's checked it's not straightforward. Maybe it'll be a bonus
> to have a comment above may_avoid_write_fault() about it in a follow up.
>
> >
> > And even if I _am_ missing something, I really would like a very
> > obvious and direct test for "this vma is writable", ie maybe a
> >
> > if (!(vma->vm_flags & VM_WRITE))
> > return false;
> >
> > at the very top of the function.
>
> Yes looks okay too; I think using MM_CP_DIRTY_ACCT flag has a slight advantage
> in that it checks VM_WRITE only once before calling change_protection(), rather
> than doing the check for every pte even if we know it'll have the same result.
> However it indeed hides the facts deeper..
>
> >
> > And no, "pte_dirty()" is not a reason to make something writable, it
> > might have started out as a writable mapping, and we dirtied the page,
> > and we made it read-only. The page stays dirty, but it shouldn't
> > become writable just because of that.
>
> I think the dirty bit checks are only to make sure we don't need those extra
> write faults. It should definitely be based on the fact that VM_WRITE being
> set already.
>
> Thanks,
>
> --
> Peter Xu
--
Peter Xu
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: [patch 128/192] mm: improve mprotect(R|W) efficiency on pages referenced once
2021-06-30 1:39 ` Peter Xu
@ 2021-06-30 2:25 ` Linus Torvalds
2021-06-30 16:42 ` Peter Xu
0 siblings, 1 reply; 409+ messages in thread
From: Linus Torvalds @ 2021-06-30 2:25 UTC (permalink / raw)
To: Peter Xu
Cc: Andrew Morton, Andrea Arcangeli, Evgeniy Stepanov, kostyak,
Linux-MM, mm-commits, Peter Collingbourne
On Tue, Jun 29, 2021 at 6:39 PM Peter Xu <peterx@redhat.com> wrote:
>
> And since MM_CP_DIRTY_ACCT implies "VM_WRITE|VM_SHARED" all set, above should
> be a slightly faster version of below:
That's way too subtle, particularly since the MM_CP_DIRTY_ACCT logic
comes from another file entirely.
I don't think it's even faster, considering that presumably the
anonymous mapping case is the common one, and that's the one that
needs all the extra tests, it's likely better to _not_ test that very
subtle flag at all, and just doing the straightforward and obvious
tests that are understandable _locally_.
So I claim that it's
(a) not an optimization at all
(b) completely locally unintuitive and unreadable
> Again, I think in all cases some more comment should be good indeed..
I really want more than a comment. I want that MM_CP_DIRTY_ACCT bit
testing gone.
The only point where it makes sense to check MM_CP_DIRTY_ACCT is
within the context of "is the page already dirty".
So I think the logic should be something along the lines of
- first:
if (!(vma->vm_flags & VM_WRITE))
return false;
because that logic is set in stone, and true regardless of anything
else. If the vma isn't writable, we're not going to set the write bit.
End of story.
- then, check the vma_is_anonumous() case:
if (vma_is_anonymous(vma))
return page_count(pte_page(pte)) == 1;
because if it's a writable mapping, and anonymous, then we can
mark it writable if we're the exclusive owners of that page.
- and THEN we can handle the "ok, shared mapping, now let's start
thinking about dirty accounting" cases.
Make it obvious and correct. This is not a sequence where you should
try to (incorrectly) optimize away individual instructions.
Linus
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: [patch 128/192] mm: improve mprotect(R|W) efficiency on pages referenced once
2021-06-30 2:25 ` Linus Torvalds
@ 2021-06-30 16:42 ` Peter Xu
2021-06-30 18:03 ` Linus Torvalds
0 siblings, 1 reply; 409+ messages in thread
From: Peter Xu @ 2021-06-30 16:42 UTC (permalink / raw)
To: Linus Torvalds
Cc: Andrew Morton, Andrea Arcangeli, Evgeniy Stepanov, kostyak,
Linux-MM, mm-commits, Peter Collingbourne
[-- Attachment #1: Type: text/plain, Size: 5239 bytes --]
On Tue, Jun 29, 2021 at 07:25:42PM -0700, Linus Torvalds wrote:
> On Tue, Jun 29, 2021 at 6:39 PM Peter Xu <peterx@redhat.com> wrote:
> >
> > And since MM_CP_DIRTY_ACCT implies "VM_WRITE|VM_SHARED" all set, above should
> > be a slightly faster version of below:
>
> That's way too subtle, particularly since the MM_CP_DIRTY_ACCT logic
> comes from another file entirely.
>
> I don't think it's even faster, considering that presumably the
> anonymous mapping case is the common one, and that's the one that
> needs all the extra tests, it's likely better to _not_ test that very
> subtle flag at all, and just doing the straightforward and obvious
> tests that are understandable _locally_.
>
> So I claim that it's
>
> (a) not an optimization at all
>
> (b) completely locally unintuitive and unreadable
>
> > Again, I think in all cases some more comment should be good indeed..
>
> I really want more than a comment. I want that MM_CP_DIRTY_ACCT bit
> testing gone.
My understanding is that MM_CP_DIRTY_ACCT contains all check results from
vma_wants_writenotify(), so if we drop it we'd need to have something like that
to be checked within change_pte_range(), which is again slower (I have totally
no idea how slow to check vma->vm_flags & VM_WRITE, but moving the whole
vma_wants_writenotify here is definitely even slower).
>
> The only point where it makes sense to check MM_CP_DIRTY_ACCT is
> within the context of "is the page already dirty".
>
> So I think the logic should be something along the lines of
>
> - first:
>
> if (!(vma->vm_flags & VM_WRITE))
> return false;
>
> because that logic is set in stone, and true regardless of anything
> else. If the vma isn't writable, we're not going to set the write bit.
> End of story.
>
> - then, check the vma_is_anonumous() case:
>
> if (vma_is_anonymous(vma))
> return page_count(pte_page(pte)) == 1;
>
> because if it's a writable mapping, and anonymous, then we can
> mark it writable if we're the exclusive owners of that page.
Shouldn't we still at least checks [soft-]dirty bits and uffd-wp bits to make
sure it's either not dirty tracked or uffd wr-protected? Say, IMHO it's
possible that soft-dirty tracking enabled on this anonymous vma range, then we
still depend on the write bit removed to set the soft-dirty later in the fault
handler.
>
> - and THEN we can handle the "ok, shared mapping, now let's start
> thinking about dirty accounting" cases.
>
> Make it obvious and correct. This is not a sequence where you should
> try to (incorrectly) optimize away individual instructions.
Yes I still fully agree it's very un-obvious. So far the best thing I can come
up with is something like below (patch attached too but not yet tested). I
moved VM_WRITE out so hopefully it'll be very clear; then I also rearranged the
checks so the final outcome looks like below:
static bool may_avoid_write_fault(pte_t pte, struct vm_area_struct *vma,
unsigned long cp_flags)
{
/*
* It is unclear whether this optimization can be done safely for NUMA
* pages.
*/
if (cp_flags & MM_CP_PROT_NUMA)
return false;
/*
* Never apply write bit if VM_WRITE not set. Note that this is
* actually checked for VM_SHARED when MM_CP_DIRTY_ACCT is set, so
* logically we only need to check it for !MM_CP_DIRTY_ACCT, but just
* make it even more obvious.
*/
if (!(vma->vm_flags & VM_WRITE))
return false;
/*
* Don't do this optimization for clean pages as we need to be notified
* of the transition from clean to dirty.
*/
if (!pte_dirty(pte))
return false;
/* Same for softdirty. */
if (!pte_soft_dirty(pte) && (vma->vm_flags & VM_SOFTDIRTY))
return false;
/*
* For userfaultfd the user program needs to monitor write faults so we
* can't do this optimization.
*/
if (pte_uffd_wp(pte))
return false;
/*
* MM_CP_DIRTY_ACCT indicates that we can always make the page writable
* regardless of the number of references. Time to set the write bit.
*/
if (cp_flags & MM_CP_DIRTY_ACCT)
return true;
/*
* Othewise it means !MM_CP_DIRTY_ACCT. We can only apply write bit
* early if it's anonymous page and we exclusively own it.
*/
if (vma_is_anonymous(vma) && (page_count(pte_page(pte)) == 1))
return true;
/* Don't play any trick */
return false;
}
The logic should be the same as before, it's just that we'll do an extra check
on VM_WRITE for MM_CP_DIRTY_ACCT but assuming it's ok.
Another side note is that I still think the VM_SOFTDIRTY check is wrong in
may_avoid_write_fault() and even in the old code (I mentioned it previously
when reviewing the patch), as !VM_SOFTDIRTY should mean soft dirty tracking
enabled while VM_SOFTDIRTY means disabled. So I wonder whether it should be:
- if (!pte_soft_dirty(pte) && (vma->vm_flags & VM_SOFTDIRTY))
+ if (!pte_soft_dirty(pte) && !(vma->vm_flags & VM_SOFTDIRTY))
However I didn't touch it up there as it may need more justifications (I feel
it's okay in the old code, as vma_wants_writenotify actually checks it too and
in the right way; however after the anonymous fast path it seems to prone to
error if it's anonymous; I'll check later).
Thanks,
--
Peter Xu
[-- Attachment #2: 0001-mm-mprotect-Optimize-layout-of-may_avoid_write_fault.patch --]
[-- Type: text/plain, Size: 2964 bytes --]
From 4fb32ad7c949d5ec6b6ea364d3388b50bf674c9c Mon Sep 17 00:00:00 2001
From: Peter Xu <peterx@redhat.com>
Date: Wed, 30 Jun 2021 12:20:12 -0400
Subject: [PATCH] mm/mprotect: Optimize layout of may_avoid_write_fault()
Firstly move VM_WRITE check to be outside of !MM_CP_DIRTY_ACCT chunk, so as to
make it clear that we won't accidentally set the write bit to !VM_WRITE vmas.
The old logic is hard to read in that it was written in reversed logic. Put
things backward by moving the soft-dirty and uffd-wp checks earlier. Make the
NUMA check even earlier than those as it's a cheap check and straightforward.
Make the only "return true" case to be either the MM_CP_DIRTY_ACCT (which
stands for the VM_SHARED cases when write bit can be applied), or the special
anonymous page when we exclusively own it.
Signed-off-by: Peter Xu <peterx@redhat.com>
---
mm/mprotect.c | 39 +++++++++++++++++++++++++--------------
1 file changed, 25 insertions(+), 14 deletions(-)
diff --git a/mm/mprotect.c b/mm/mprotect.c
index 4cb240fd9936..3977bfd55f62 100644
--- a/mm/mprotect.c
+++ b/mm/mprotect.c
@@ -40,17 +40,20 @@ static bool may_avoid_write_fault(pte_t pte, struct vm_area_struct *vma,
unsigned long cp_flags)
{
/*
- * The dirty accountable bit indicates that we can always make the page
- * writable regardless of the number of references.
+ * It is unclear whether this optimization can be done safely for NUMA
+ * pages.
*/
- if (!(cp_flags & MM_CP_DIRTY_ACCT)) {
- /* Otherwise, we must have exclusive access to the page. */
- if (!(vma_is_anonymous(vma) && (vma->vm_flags & VM_WRITE)))
- return false;
+ if (cp_flags & MM_CP_PROT_NUMA)
+ return false;
- if (page_count(pte_page(pte)) != 1)
- return false;
- }
+ /*
+ * Never apply write bit if VM_WRITE not set. Note that this is
+ * actually checked for VM_SHARED when MM_CP_DIRTY_ACCT is set, so
+ * logically we only need to check it for !MM_CP_DIRTY_ACCT, but just
+ * make it even more obvious.
+ */
+ if (!(vma->vm_flags & VM_WRITE))
+ return false;
/*
* Don't do this optimization for clean pages as we need to be notified
@@ -71,13 +74,21 @@ static bool may_avoid_write_fault(pte_t pte, struct vm_area_struct *vma,
return false;
/*
- * It is unclear whether this optimization can be done safely for NUMA
- * pages.
+ * MM_CP_DIRTY_ACCT indicates that we can always make the page writable
+ * regardless of the number of references. Time to set the write bit.
*/
- if (cp_flags & MM_CP_PROT_NUMA)
- return false;
+ if (cp_flags & MM_CP_DIRTY_ACCT)
+ return true;
+
+ /*
+ * Othewise it means !MM_CP_DIRTY_ACCT. We can only apply write bit
+ * early if it's anonymous page and we exclusively own it.
+ */
+ if (vma_is_anonymous(vma) && (page_count(pte_page(pte)) == 1))
+ return true;
- return true;
+ /* Don't play any trick */
+ return false;
}
static unsigned long change_pte_range(struct vm_area_struct *vma, pmd_t *pmd,
--
2.31.1
^ permalink raw reply related [flat|nested] 409+ messages in thread
* Re: [patch 128/192] mm: improve mprotect(R|W) efficiency on pages referenced once
2021-06-30 16:42 ` Peter Xu
@ 2021-06-30 18:03 ` Linus Torvalds
2021-07-01 1:27 ` Peter Xu
0 siblings, 1 reply; 409+ messages in thread
From: Linus Torvalds @ 2021-06-30 18:03 UTC (permalink / raw)
To: Peter Xu
Cc: Andrew Morton, Andrea Arcangeli, Evgeniy Stepanov, kostyak,
Linux-MM, mm-commits, Peter Collingbourne
On Wed, Jun 30, 2021 at 9:42 AM Peter Xu <peterx@redhat.com> wrote:
>
> Yes I still fully agree it's very un-obvious. So far the best thing I can come
> up with is something like below (patch attached too but not yet tested). I
> moved VM_WRITE out so hopefully it'll be very clear; then I also rearranged the
> checks so the final outcome looks like below:
>
> static bool may_avoid_write_fault(pte_t pte, struct vm_area_struct *vma,
> unsigned long cp_flags)
> {
> /*
> * It is unclear whether this optimization can be done safely for NUMA
> * pages.
> */
> if (cp_flags & MM_CP_PROT_NUMA)
> return false;
Please just put that VM_WRITE test first. It's the one that really
*matters*. There's no "it's unclear if" about that part. Just handle
the obvious and important check first.
Yeah, yeah, they both return false, so order doesn't matter from a
semantic standpoint, but from a clarity standpoint just do the clear
and unambiguous and security-relevant test first.
The rest of the tests are implementation details, the VM_WRITE test is
fundamental behavior. It's the one that made me worry about this patch
in the first place.
> /*
> * Don't do this optimization for clean pages as we need to be notified
> * of the transition from clean to dirty.
> */
> if (!pte_dirty(pte))
> return false;
>
> /* Same for softdirty. */
> if (!pte_soft_dirty(pte) && (vma->vm_flags & VM_SOFTDIRTY))
> return false;
>
> /*
> * For userfaultfd the user program needs to monitor write faults so we
> * can't do this optimization.
> */
> if (pte_uffd_wp(pte))
> return false;
So all of these are a bit special.
Why? Because if I look at the actual page fault path, these are not
the tests there.
I'd really like to have some obvious situation where we keep this
"make it writable" in sync with what would actually happen on a write
fault when it's not writable.
And it's not at all obvious to me for these cases.
The do_wp_page() code doesn't even use pte_uffd_wp(). It uses
userfaultfd_pte_wp(vma, pte), and I don't even know why. Yes, I can
see the code (it additionally tests the VM_UFFD_WP flag in the vma),
but a number of other paths then only do that pte_uffd_wp() test.
I get the feeling that we really should try to match what the
do_wp_page() path does, though.
Which brings up another issue: the do_wp_page() path treats PageKsm()
pages differently. And it locks the page before double-checking the
page count.
Why does mprotect() not need to do the same thing? I think this has
come up before, and "change_protection()" can get called with the
mmap_sem held just for reading - see userfaultfd - so it has all the
same issues as a page fault does, afaik.
> /*
> * MM_CP_DIRTY_ACCT indicates that we can always make the page writable
> * regardless of the number of references. Time to set the write bit.
> */
> if (cp_flags & MM_CP_DIRTY_ACCT)
> return true;
>
> /*
> * Othewise it means !MM_CP_DIRTY_ACCT. We can only apply write bit
> * early if it's anonymous page and we exclusively own it.
> */
> if (vma_is_anonymous(vma) && (page_count(pte_page(pte)) == 1))
> return true;
>
> /* Don't play any trick */
> return false;
> }
>
> The logic should be the same as before, it's just that we'll do an extra check
> on VM_WRITE for MM_CP_DIRTY_ACCT but assuming it's ok.
See above. I don't think the logic before was all that clear either.
The one case that is clear is that if it's a shared mapping, and
MM_CP_DIRTY_ACCT is set, and it was already dirty (and softdirty),
then it's ok.,
That's the old code. I don't like how the old code was written
(because I think that MM_CP_DIRTY_ACCT bit wasx too subtle), but I
think the old code was at least correct.
The new code, it just worries me. It adds all those new cases for when
we can make the page writable early - that's the whole point of the
patch, after all - but my point here is that it's not at all obvious
that those new cases are actually correct.
MAYBE it's all correct. I'm not saying it's wrong. I'm just saying
it's not _obvious_ that it's correct.
What about that page_count() test, for example: it has a comment, it
looks obvious, but it's very different from what do_wp_page() does. So
what happens if we have a page-out at the same time that turns that
page into a swap cache page, and increments the page count? What about
that race? Do we end up with a writable page that is shared with a
swap cache entry? Is that ok? Why isn't it ok in the page fault case?
See why this patch worries me so much?
Linus
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: [patch 128/192] mm: improve mprotect(R|W) efficiency on pages referenced once
2021-06-30 18:03 ` Linus Torvalds
@ 2021-07-01 1:27 ` Peter Xu
2021-07-01 18:29 ` Linus Torvalds
0 siblings, 1 reply; 409+ messages in thread
From: Peter Xu @ 2021-07-01 1:27 UTC (permalink / raw)
To: Linus Torvalds
Cc: Andrew Morton, Andrea Arcangeli, Evgeniy Stepanov, kostyak,
Linux-MM, mm-commits, Peter Collingbourne
On Wed, Jun 30, 2021 at 11:03:25AM -0700, Linus Torvalds wrote:
> On Wed, Jun 30, 2021 at 9:42 AM Peter Xu <peterx@redhat.com> wrote:
> >
> > Yes I still fully agree it's very un-obvious. So far the best thing I can come
> > up with is something like below (patch attached too but not yet tested). I
> > moved VM_WRITE out so hopefully it'll be very clear; then I also rearranged the
> > checks so the final outcome looks like below:
> >
> > static bool may_avoid_write_fault(pte_t pte, struct vm_area_struct *vma,
> > unsigned long cp_flags)
> > {
> > /*
> > * It is unclear whether this optimization can be done safely for NUMA
> > * pages.
> > */
> > if (cp_flags & MM_CP_PROT_NUMA)
> > return false;
>
> Please just put that VM_WRITE test first. It's the one that really
> *matters*. There's no "it's unclear if" about that part. Just handle
> the obvious and important check first.
>
> Yeah, yeah, they both return false, so order doesn't matter from a
> semantic standpoint, but from a clarity standpoint just do the clear
> and unambiguous and security-relevant test first.
>
> The rest of the tests are implementation details, the VM_WRITE test is
> fundamental behavior. It's the one that made me worry about this patch
> in the first place.
Sure.
>
> > /*
> > * Don't do this optimization for clean pages as we need to be notified
> > * of the transition from clean to dirty.
> > */
> > if (!pte_dirty(pte))
> > return false;
> >
> > /* Same for softdirty. */
> > if (!pte_soft_dirty(pte) && (vma->vm_flags & VM_SOFTDIRTY))
> > return false;
> >
> > /*
> > * For userfaultfd the user program needs to monitor write faults so we
> > * can't do this optimization.
> > */
> > if (pte_uffd_wp(pte))
> > return false;
>
> So all of these are a bit special.
>
> Why? Because if I look at the actual page fault path, these are not
> the tests there.
>
> I'd really like to have some obvious situation where we keep this
> "make it writable" in sync with what would actually happen on a write
> fault when it's not writable.
>
> And it's not at all obvious to me for these cases.
>
> The do_wp_page() code doesn't even use pte_uffd_wp(). It uses
> userfaultfd_pte_wp(vma, pte), and I don't even know why. Yes, I can
> see the code (it additionally tests the VM_UFFD_WP flag in the vma),
> but a number of other paths then only do that pte_uffd_wp() test.
The vma check is a safety net to make sure it's not the case e.g. when the vma
has already unregistered from uffd-wp while there's uffd-wp bit left over.
E.g., currently UFFDIO_UNREGISTER is lazy on removing uffd-wp bits that applied
to ptes, so even vma is unregistered there could still have pte_uffd_wp() being
true for some ptes. That vma check makes sure when it happens the uffd-wp bit
will be auto-removed.
>
> I get the feeling that we really should try to match what the
> do_wp_page() path does, though.
Makes sense.
>
> Which brings up another issue: the do_wp_page() path treats PageKsm()
> pages differently. And it locks the page before double-checking the
> page count.
>
> Why does mprotect() not need to do the same thing? I think this has
> come up before, and "change_protection()" can get called with the
> mmap_sem held just for reading - see userfaultfd - so it has all the
> same issues as a page fault does, afaik.
Good point.. I overlooked ksm when reviewing the patch, while I should really
have looked at do_wp_page() as you suggested (hmm.. the truth is I wasn't even
aware of this patch and never planned to try to review it, until it breaks the
uffd-wp anonymous tests in its initial versions when I was testing mmots...).
Maybe something like this (to be squashed into the previously attached patch):
---8<---
diff --git a/mm/mprotect.c b/mm/mprotect.c
index 3977bfd55f62..7aab30ac9c9f 100644
--- a/mm/mprotect.c
+++ b/mm/mprotect.c
@@ -39,12 +39,8 @@
static bool may_avoid_write_fault(pte_t pte, struct vm_area_struct *vma,
unsigned long cp_flags)
{
- /*
- * It is unclear whether this optimization can be done safely for NUMA
- * pages.
- */
- if (cp_flags & MM_CP_PROT_NUMA)
- return false;
+ struct page *page;
+ bool ret = false;
/*
* Never apply write bit if VM_WRITE not set. Note that this is
@@ -55,6 +51,13 @@ static bool may_avoid_write_fault(pte_t pte, struct vm_area_struct *vma,
if (!(vma->vm_flags & VM_WRITE))
return false;
+ /*
+ * It is unclear whether this optimization can be done safely for NUMA
+ * pages.
+ */
+ if (cp_flags & MM_CP_PROT_NUMA)
+ return false;
+
/*
* Don't do this optimization for clean pages as we need to be notified
* of the transition from clean to dirty.
@@ -80,15 +83,22 @@ static bool may_avoid_write_fault(pte_t pte, struct vm_area_struct *vma,
if (cp_flags & MM_CP_DIRTY_ACCT)
return true;
+ page = pte_page(pte);
+ /* Best effort to take page lock, don't play trick if failed */
+ if (!trylock_page(page))
+ return false;
+ /* KSM pages needs COW; leave them be */
+ if (PageKsm(page))
+ goto unlock_fail;
/*
- * Othewise it means !MM_CP_DIRTY_ACCT. We can only apply write bit
- * early if it's anonymous page and we exclusively own it.
+ * Othewise it means !MM_CP_DIRTY_ACCT and !KSM. We can only apply
+ * write bit early if it's anonymous page and we exclusively own it.
*/
- if (vma_is_anonymous(vma) && (page_count(pte_page(pte)) == 1))
- return true;
-
- /* Don't play any trick */
- return false;
+ if (vma_is_anonymous(vma) && (page_count(page) == 1))
+ ret = true;
+unlock_fail:
+ unlock_page(page);
+ return ret;
}
static unsigned long change_pte_range(struct vm_area_struct *vma, pmd_t *pmd,
---8<---
I hope I didn't overlook something else..
Today when I was looking at ksm code, I found that I got lost on why we say
"PageKsm() doesn't necessarily raise the page refcount", as in do_wp_page(). I
was looking at replace_page() where, afaict, we still do proper refcounting for
the stable nodes with "get_page(kpage)".
I know I must missed something but I can't quickly tell. In all cases with
above PageKsm check I think it'll be safe, and it's definitely clear that page
lock will stablize PageKsm()'s return value, like do_wp_page().
>
> > /*
> > * MM_CP_DIRTY_ACCT indicates that we can always make the page writable
> > * regardless of the number of references. Time to set the write bit.
> > */
> > if (cp_flags & MM_CP_DIRTY_ACCT)
> > return true;
> >
> > /*
> > * Othewise it means !MM_CP_DIRTY_ACCT. We can only apply write bit
> > * early if it's anonymous page and we exclusively own it.
> > */
> > if (vma_is_anonymous(vma) && (page_count(pte_page(pte)) == 1))
> > return true;
> >
> > /* Don't play any trick */
> > return false;
> > }
> >
> > The logic should be the same as before, it's just that we'll do an extra check
> > on VM_WRITE for MM_CP_DIRTY_ACCT but assuming it's ok.
>
> See above. I don't think the logic before was all that clear either.
>
> The one case that is clear is that if it's a shared mapping, and
> MM_CP_DIRTY_ACCT is set, and it was already dirty (and softdirty),
> then it's ok.,
>
> That's the old code. I don't like how the old code was written
> (because I think that MM_CP_DIRTY_ACCT bit wasx too subtle), but I
> think the old code was at least correct.
>
> The new code, it just worries me. It adds all those new cases for when
> we can make the page writable early - that's the whole point of the
> patch, after all - but my point here is that it's not at all obvious
> that those new cases are actually correct.
Yes agreed the MM_CP_DIRTY_ACCT bit is very subtle and not easy to get. It's
just that I don't have a good idea to make it better, yet..
>
> MAYBE it's all correct. I'm not saying it's wrong. I'm just saying
> it's not _obvious_ that it's correct.
>
> What about that page_count() test, for example: it has a comment, it
> looks obvious, but it's very different from what do_wp_page() does. So
> what happens if we have a page-out at the same time that turns that
> page into a swap cache page, and increments the page count? What about
> that race? Do we end up with a writable page that is shared with a
> swap cache entry? Is that ok? Why isn't it ok in the page fault case?
That looks fine to me: when the race happens we must have checked page_count==1
first and granted the write bit, then add_to_swap_cache() happens after the
page_count==1 check (as it takes another refcount, so >2 otherwise). Then it
also means unmap mappings should happen even after that point. If my above
understanding is correct, our newly installed pte will be zapped safely soon,
but of course after we release the pgtable lock in change_pte_range().
Thanks,
--
Peter Xu
^ permalink raw reply related [flat|nested] 409+ messages in thread
* Re: [patch 128/192] mm: improve mprotect(R|W) efficiency on pages referenced once
2021-07-01 1:27 ` Peter Xu
@ 2021-07-01 18:29 ` Linus Torvalds
2021-07-06 1:24 ` Peter Xu
0 siblings, 1 reply; 409+ messages in thread
From: Linus Torvalds @ 2021-07-01 18:29 UTC (permalink / raw)
To: Peter Xu
Cc: Andrew Morton, Andrea Arcangeli, Evgeniy Stepanov,
Kostya Kortchinsky, Linux-MM, mm-commits, Peter Collingbourne
On Wed, Jun 30, 2021 at 6:27 PM Peter Xu <peterx@redhat.com> wrote:
>
> On Wed, Jun 30, 2021 at 11:03:25AM -0700, Linus Torvalds wrote:
> >
> > What about that page_count() test, for example: it has a comment, it
> > looks obvious, but it's very different from what do_wp_page() does. So
> > what happens if we have a page-out at the same time that turns that
> > page into a swap cache page, and increments the page count? What about
> > that race? Do we end up with a writable page that is shared with a
> > swap cache entry? Is that ok? Why isn't it ok in the page fault case?
>
> That looks fine to me: when the race happens we must have checked page_count==1
> first and granted the write bit, then add_to_swap_cache() happens after the
> page_count==1 check (as it takes another refcount, so >2 otherwise). Then it
> also means unmap mappings should happen even after that point. If my above
> understanding is correct, our newly installed pte will be zapped safely soon,
> but of course after we release the pgtable lock in change_pte_range().
So if this is fine, then maybe we should just remove the page lock in
the do_wp_page() path (and remove the PageKSM check while at it)?
If it's not required by mprotect() to say "I can make the page
writable directly", then it really shouldn't be required by the page
fault path either.
Which I'd love to do, and was really itching to do (it's a nasty
lock), but I worried about it..
I'd hate to have mprotect do one thing, and page faulting do another
thing, and not have some logic to why they have to be different.
Linus
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: [patch 128/192] mm: improve mprotect(R|W) efficiency on pages referenced once
2021-07-01 18:29 ` Linus Torvalds
@ 2021-07-06 1:24 ` Peter Xu
0 siblings, 0 replies; 409+ messages in thread
From: Peter Xu @ 2021-07-06 1:24 UTC (permalink / raw)
To: Linus Torvalds
Cc: Andrew Morton, Andrea Arcangeli, Evgeniy Stepanov,
Kostya Kortchinsky, Linux-MM, mm-commits, Peter Collingbourne
(sorry for a very late reply)
On Thu, Jul 01, 2021 at 11:29:50AM -0700, Linus Torvalds wrote:
> On Wed, Jun 30, 2021 at 6:27 PM Peter Xu <peterx@redhat.com> wrote:
> >
> > On Wed, Jun 30, 2021 at 11:03:25AM -0700, Linus Torvalds wrote:
> > >
> > > What about that page_count() test, for example: it has a comment, it
> > > looks obvious, but it's very different from what do_wp_page() does. So
> > > what happens if we have a page-out at the same time that turns that
> > > page into a swap cache page, and increments the page count? What about
> > > that race? Do we end up with a writable page that is shared with a
> > > swap cache entry? Is that ok? Why isn't it ok in the page fault case?
> >
> > That looks fine to me: when the race happens we must have checked page_count==1
> > first and granted the write bit, then add_to_swap_cache() happens after the
> > page_count==1 check (as it takes another refcount, so >2 otherwise). Then it
> > also means unmap mappings should happen even after that point. If my above
> > understanding is correct, our newly installed pte will be zapped safely soon,
> > but of course after we release the pgtable lock in change_pte_range().
>
> So if this is fine, then maybe we should just remove the page lock in
> the do_wp_page() path (and remove the PageKSM check while at it)?
I could be wrong, but I thought the page lock in do_wp_page() is more for the
PageKsm() race (e.g., to make sure we don't grant write to a page that is
becoming a ksm page in parallel).
>
> If it's not required by mprotect() to say "I can make the page
> writable directly", then it really shouldn't be required by the page
> fault path either.
>
> Which I'd love to do, and was really itching to do (it's a nasty
> lock), but I worried about it..
>
> I'd hate to have mprotect do one thing, and page faulting do another
> thing, and not have some logic to why they have to be different.
Agreed; perhaps no need to be identical - I think the mprotect path can be even
stricter than the fault page, as it's a fast-path only. It should never apply
the write bit when the page fault path won't. So I think the original patch
does need a justification on why it didn't handle ksm page while do_wp_page
handled it.
Thanks,
--
Peter Xu
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2022-04-27 19:41 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2022-04-27 19:41 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits, patches
2 patches, based on d615b5416f8a1afeb82d13b238f8152c572d59c0.
Subsystems affected by this patch series:
mm/kasan
mm/debug
Subsystem: mm/kasan
Zqiang <qiang1.zhang@intel.com>:
kasan: prevent cpu_quarantine corruption when CPU offline and cache shrink occur at same time
Subsystem: mm/debug
Akira Yokosawa <akiyks@gmail.com>:
docs: vm/page_owner: use literal blocks for param description
Documentation/vm/page_owner.rst | 5 +++--
mm/kasan/quarantine.c | 7 +++++++
2 files changed, 10 insertions(+), 2 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2022-04-21 23:35 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2022-04-21 23:35 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm, patches
13 patches, based on b253435746d9a4a701b5f09211b9c14d3370d0da.
Subsystems affected by this patch series:
mm/memory-failure
mm/memcg
mm/userfaultfd
mm/hugetlbfs
mm/mremap
mm/oom-kill
mm/kasan
kcov
mm/hmm
Subsystem: mm/memory-failure
Naoya Horiguchi <naoya.horiguchi@nec.com>:
mm/hwpoison: fix race between hugetlb free/demotion and memory_failure_hugetlb()
Xu Yu <xuyu@linux.alibaba.com>:
mm/memory-failure.c: skip huge_zero_page in memory_failure()
Subsystem: mm/memcg
Shakeel Butt <shakeelb@google.com>:
memcg: sync flush only if periodic flush is delayed
Subsystem: mm/userfaultfd
Nadav Amit <namit@vmware.com>:
userfaultfd: mark uffd_wp regardless of VM_WRITE flag
Subsystem: mm/hugetlbfs
Christophe Leroy <christophe.leroy@csgroup.eu>:
mm, hugetlb: allow for "high" userspace addresses
Subsystem: mm/mremap
Sidhartha Kumar <sidhartha.kumar@oracle.com>:
selftest/vm: verify mmap addr in mremap_test
selftest/vm: verify remap destination address in mremap_test
selftest/vm: support xfail in mremap_test
selftest/vm: add skip support to mremap_test
Subsystem: mm/oom-kill
Nico Pache <npache@redhat.com>:
oom_kill.c: futex: delay the OOM reaper to allow time for proper futex cleanup
Subsystem: mm/kasan
Vincenzo Frascino <vincenzo.frascino@arm.com>:
MAINTAINERS: add Vincenzo Frascino to KASAN reviewers
Subsystem: kcov
Aleksandr Nogikh <nogikh@google.com>:
kcov: don't generate a warning on vm_insert_page()'s failure
Subsystem: mm/hmm
Alistair Popple <apopple@nvidia.com>:
mm/mmu_notifier.c: fix race in mmu_interval_notifier_remove()
MAINTAINERS | 1
fs/hugetlbfs/inode.c | 9 -
include/linux/hugetlb.h | 6 +
include/linux/memcontrol.h | 5
include/linux/mm.h | 8 +
include/linux/sched.h | 1
include/linux/sched/mm.h | 8 +
kernel/kcov.c | 7 -
mm/hugetlb.c | 10 +
mm/memcontrol.c | 12 ++
mm/memory-failure.c | 158 ++++++++++++++++++++++--------
mm/mmap.c | 8 -
mm/mmu_notifier.c | 14 ++
mm/oom_kill.c | 54 +++++++---
mm/userfaultfd.c | 15 +-
mm/workingset.c | 2
tools/testing/selftests/vm/mremap_test.c | 85 +++++++++++++++-
tools/testing/selftests/vm/run_vmtests.sh | 11 +-
18 files changed, 327 insertions(+), 87 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2022-04-15 2:12 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2022-04-15 2:12 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits, patches
14 patches, based on 115acbb56978941bb7537a97dfc303da286106c1.
Subsystems affected by this patch series:
MAINTAINERS
mm/tmpfs
m/secretmem
mm/kasan
mm/kfence
mm/pagealloc
mm/zram
mm/compaction
mm/hugetlb
binfmt
mm/vmalloc
mm/kmemleak
Subsystem: MAINTAINERS
Joe Perches <joe@perches.com>:
MAINTAINERS: Broadcom internal lists aren't maintainers
Subsystem: mm/tmpfs
Hugh Dickins <hughd@google.com>:
tmpfs: fix regressions from wider use of ZERO_PAGE
Subsystem: m/secretmem
Axel Rasmussen <axelrasmussen@google.com>:
mm/secretmem: fix panic when growing a memfd_secret
Subsystem: mm/kasan
Zqiang <qiang1.zhang@intel.com>:
irq_work: use kasan_record_aux_stack_noalloc() record callstack
Vincenzo Frascino <vincenzo.frascino@arm.com>:
kasan: fix hw tags enablement when KUNIT tests are disabled
Subsystem: mm/kfence
Marco Elver <elver@google.com>:
mm, kfence: support kmem_dump_obj() for KFENCE objects
Subsystem: mm/pagealloc
Juergen Gross <jgross@suse.com>:
mm, page_alloc: fix build_zonerefs_node()
Subsystem: mm/zram
Minchan Kim <minchan@kernel.org>:
mm: fix unexpected zeroed page mapping with zram swap
Subsystem: mm/compaction
Charan Teja Kalla <quic_charante@quicinc.com>:
mm: compaction: fix compiler warning when CONFIG_COMPACTION=n
Subsystem: mm/hugetlb
Mike Kravetz <mike.kravetz@oracle.com>:
hugetlb: do not demote poisoned hugetlb pages
Subsystem: binfmt
Andrew Morton <akpm@linux-foundation.org>:
revert "fs/binfmt_elf: fix PT_LOAD p_align values for loaders"
revert "fs/binfmt_elf: use PT_LOAD p_align values for static PIE"
Subsystem: mm/vmalloc
Omar Sandoval <osandov@fb.com>:
mm/vmalloc: fix spinning drain_vmap_work after reading from /proc/vmcore
Subsystem: mm/kmemleak
Patrick Wang <patrick.wang.shcn@gmail.com>:
mm: kmemleak: take a full lowmem check in kmemleak_*_phys()
MAINTAINERS | 64 ++++++++++++++++++++--------------------
arch/x86/include/asm/io.h | 2 -
arch/x86/kernel/crash_dump_64.c | 1
fs/binfmt_elf.c | 6 +--
include/linux/kfence.h | 24 +++++++++++++++
kernel/irq_work.c | 2 -
mm/compaction.c | 10 +++---
mm/filemap.c | 6 ---
mm/hugetlb.c | 17 ++++++----
mm/kasan/hw_tags.c | 5 +--
mm/kasan/kasan.h | 10 +++---
mm/kfence/core.c | 21 -------------
mm/kfence/kfence.h | 21 +++++++++++++
mm/kfence/report.c | 47 +++++++++++++++++++++++++++++
mm/kmemleak.c | 8 ++---
mm/page_alloc.c | 2 -
mm/page_io.c | 54 ---------------------------------
mm/secretmem.c | 17 ++++++++++
mm/shmem.c | 31 ++++++++++++-------
mm/slab.c | 2 -
mm/slab.h | 2 -
mm/slab_common.c | 9 +++++
mm/slob.c | 2 -
mm/slub.c | 2 -
mm/vmalloc.c | 11 ------
25 files changed, 207 insertions(+), 169 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2022-04-08 20:08 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2022-04-08 20:08 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits, patches
9 patches, based on d00c50b35101b862c3db270ffeba53a63a1063d9.
Subsystems affected by this patch series:
mm/migration
mm/highmem
lz4
mm/sparsemem
mm/mremap
mm/mempolicy
mailmap
mm/memcg
MAINTAINERS
Subsystem: mm/migration
Zi Yan <ziy@nvidia.com>:
mm: migrate: use thp_order instead of HPAGE_PMD_ORDER for new page allocation.
Subsystem: mm/highmem
Max Filippov <jcmvbkbc@gmail.com>:
highmem: fix checks in __kmap_local_sched_{in,out}
Subsystem: lz4
Guo Xuenan <guoxuenan@huawei.com>:
lz4: fix LZ4_decompress_safe_partial read out of bound
Subsystem: mm/sparsemem
Waiman Long <longman@redhat.com>:
mm/sparsemem: fix 'mem_section' will never be NULL gcc 12 warning
Subsystem: mm/mremap
Paolo Bonzini <pbonzini@redhat.com>:
mmmremap.c: avoid pointless invalidate_range_start/end on mremap(old_size=0)
Subsystem: mm/mempolicy
Miaohe Lin <linmiaohe@huawei.com>:
mm/mempolicy: fix mpol_new leak in shared_policy_replace
Subsystem: mailmap
Vasily Averin <vasily.averin@linux.dev>:
mailmap: update Vasily Averin's email address
Subsystem: mm/memcg
Andrew Morton <akpm@linux-foundation.org>:
mm/list_lru.c: revert "mm/list_lru: optimize memcg_reparent_list_lru_node()"
Subsystem: MAINTAINERS
Tom Rix <trix@redhat.com>:
MAINTAINERS: add Tom as clang reviewer
.mailmap | 4 ++++
MAINTAINERS | 1 +
include/linux/mmzone.h | 11 +++++++----
lib/lz4/lz4_decompress.c | 8 ++++++--
mm/highmem.c | 4 ++--
mm/list_lru.c | 6 ------
mm/mempolicy.c | 3 ++-
mm/migrate.c | 2 +-
mm/mremap.c | 3 +++
9 files changed, 26 insertions(+), 16 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2022-04-01 18:27 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2022-04-01 18:27 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits, patches
16 patches, based on e8b767f5e04097aaedcd6e06e2270f9fe5282696.
Subsystems affected by this patch series:
mm/madvise
ofs2
nilfs2
mm/mlock
mm/mfence
mailmap
mm/memory-failure
mm/kasan
mm/debug
mm/kmemleak
mm/damon
Subsystem: mm/madvise
Charan Teja Kalla <quic_charante@quicinc.com>:
Revert "mm: madvise: skip unmapped vma holes passed to process_madvise"
Subsystem: ofs2
Joseph Qi <joseph.qi@linux.alibaba.com>:
ocfs2: fix crash when mount with quota enabled
Subsystem: nilfs2
Ryusuke Konishi <konishi.ryusuke@gmail.com>:
Patch series "nilfs2 lockdep warning fixes":
nilfs2: fix lockdep warnings in page operations for btree nodes
nilfs2: fix lockdep warnings during disk space reclamation
nilfs2: get rid of nilfs_mapping_init()
Subsystem: mm/mlock
Hugh Dickins <hughd@google.com>:
mm/munlock: add lru_add_drain() to fix memcg_stat_test
mm/munlock: update Documentation/vm/unevictable-lru.rst
Sebastian Andrzej Siewior <bigeasy@linutronix.de>:
mm/munlock: protect the per-CPU pagevec by a local_lock_t
Subsystem: mm/kfence
Muchun Song <songmuchun@bytedance.com>:
mm: kfence: fix objcgs vector allocation
Subsystem: mailmap
Kirill Tkhai <kirill.tkhai@openvz.org>:
mailmap: update Kirill's email
Subsystem: mm/memory-failure
Rik van Riel <riel@surriel.com>:
mm,hwpoison: unmap poisoned page before invalidation
Subsystem: mm/kasan
Andrey Konovalov <andreyknvl@google.com>:
mm, kasan: fix __GFP_BITS_SHIFT definition breaking LOCKDEP
Subsystem: mm/debug
Yinan Zhang <zhangyinan2019@email.szu.edu.cn>:
tools/vm/page_owner_sort.c: remove -c option
doc/vm/page_owner.rst: remove content related to -c option
Subsystem: mm/kmemleak
Kuan-Ying Lee <Kuan-Ying.Lee@mediatek.com>:
mm/kmemleak: reset tag when compare object pointer
Subsystem: mm/damon
Jonghyeon Kim <tome01@ajou.ac.kr>:
mm/damon: prevent activated scheme from sleeping by deactivated schemes
.mailmap | 1
Documentation/vm/page_owner.rst | 1
Documentation/vm/unevictable-lru.rst | 473 +++++++++++++++--------------------
fs/nilfs2/btnode.c | 23 +
fs/nilfs2/btnode.h | 1
fs/nilfs2/btree.c | 27 +
fs/nilfs2/dat.c | 4
fs/nilfs2/gcinode.c | 7
fs/nilfs2/inode.c | 167 +++++++++++-
fs/nilfs2/mdt.c | 45 ++-
fs/nilfs2/mdt.h | 6
fs/nilfs2/nilfs.h | 16 -
fs/nilfs2/page.c | 16 -
fs/nilfs2/page.h | 1
fs/nilfs2/segment.c | 9
fs/nilfs2/super.c | 5
fs/ocfs2/quota_global.c | 23 -
fs/ocfs2/quota_local.c | 2
include/linux/gfp.h | 4
mm/damon/core.c | 5
mm/gup.c | 10
mm/internal.h | 6
mm/kfence/core.c | 11
mm/kfence/kfence.h | 3
mm/kmemleak.c | 9
mm/madvise.c | 9
mm/memory.c | 12
mm/migrate.c | 2
mm/mlock.c | 46 ++-
mm/page_alloc.c | 1
mm/rmap.c | 4
mm/swap.c | 4
tools/vm/page_owner_sort.c | 6
33 files changed, 560 insertions(+), 399 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2022-04-01 18:20 incoming Andrew Morton
@ 2022-04-01 18:27 ` Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2022-04-01 18:27 UTC (permalink / raw)
To: Linus Torvalds, linux-mm, mm-commits, patches
Argh, messed up in-reply-to. Let me redo...
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2022-04-01 18:20 Andrew Morton
2022-04-01 18:27 ` incoming Andrew Morton
0 siblings, 1 reply; 409+ messages in thread
From: Andrew Morton @ 2022-04-01 18:20 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits, patches
16 patches, based on e8b767f5e04097aaedcd6e06e2270f9fe5282696.
Subsystems affected by this patch series:
mm/madvise
ofs2
nilfs2
mm/mlock
mm/mfence
mailmap
mm/memory-failure
mm/kasan
mm/debug
mm/kmemleak
mm/damon
Subsystem: mm/madvise
Charan Teja Kalla <quic_charante@quicinc.com>:
Revert "mm: madvise: skip unmapped vma holes passed to process_madvise"
Subsystem: ofs2
Joseph Qi <joseph.qi@linux.alibaba.com>:
ocfs2: fix crash when mount with quota enabled
Subsystem: nilfs2
Ryusuke Konishi <konishi.ryusuke@gmail.com>:
Patch series "nilfs2 lockdep warning fixes":
nilfs2: fix lockdep warnings in page operations for btree nodes
nilfs2: fix lockdep warnings during disk space reclamation
nilfs2: get rid of nilfs_mapping_init()
Subsystem: mm/mlock
Hugh Dickins <hughd@google.com>:
mm/munlock: add lru_add_drain() to fix memcg_stat_test
mm/munlock: update Documentation/vm/unevictable-lru.rst
Sebastian Andrzej Siewior <bigeasy@linutronix.de>:
mm/munlock: protect the per-CPU pagevec by a local_lock_t
Subsystem: mm/kfence
Muchun Song <songmuchun@bytedance.com>:
mm: kfence: fix objcgs vector allocation
Subsystem: mailmap
Kirill Tkhai <kirill.tkhai@openvz.org>:
mailmap: update Kirill's email
Subsystem: mm/memory-failure
Rik van Riel <riel@surriel.com>:
mm,hwpoison: unmap poisoned page before invalidation
Subsystem: mm/kasan
Andrey Konovalov <andreyknvl@google.com>:
mm, kasan: fix __GFP_BITS_SHIFT definition breaking LOCKDEP
Subsystem: mm/debug
Yinan Zhang <zhangyinan2019@email.szu.edu.cn>:
tools/vm/page_owner_sort.c: remove -c option
doc/vm/page_owner.rst: remove content related to -c option
Subsystem: mm/kmemleak
Kuan-Ying Lee <Kuan-Ying.Lee@mediatek.com>:
mm/kmemleak: reset tag when compare object pointer
Subsystem: mm/damon
Jonghyeon Kim <tome01@ajou.ac.kr>:
mm/damon: prevent activated scheme from sleeping by deactivated schemes
.mailmap | 1
Documentation/vm/page_owner.rst | 1
Documentation/vm/unevictable-lru.rst | 473 +++++++++++++++--------------------
fs/nilfs2/btnode.c | 23 +
fs/nilfs2/btnode.h | 1
fs/nilfs2/btree.c | 27 +
fs/nilfs2/dat.c | 4
fs/nilfs2/gcinode.c | 7
fs/nilfs2/inode.c | 167 +++++++++++-
fs/nilfs2/mdt.c | 45 ++-
fs/nilfs2/mdt.h | 6
fs/nilfs2/nilfs.h | 16 -
fs/nilfs2/page.c | 16 -
fs/nilfs2/page.h | 1
fs/nilfs2/segment.c | 9
fs/nilfs2/super.c | 5
fs/ocfs2/quota_global.c | 23 -
fs/ocfs2/quota_local.c | 2
include/linux/gfp.h | 4
mm/damon/core.c | 5
mm/gup.c | 10
mm/internal.h | 6
mm/kfence/core.c | 11
mm/kfence/kfence.h | 3
mm/kmemleak.c | 9
mm/madvise.c | 9
mm/memory.c | 12
mm/migrate.c | 2
mm/mlock.c | 46 ++-
mm/page_alloc.c | 1
mm/rmap.c | 4
mm/swap.c | 4
tools/vm/page_owner_sort.c | 6
33 files changed, 560 insertions(+), 399 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2022-03-25 1:07 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2022-03-25 1:07 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm, patches
This is the material which was staged after willystuff in linux-next.
Everything applied seamlessly on your latest, all looks well.
114 patches, based on 52deda9551a01879b3562e7b41748e85c591f14c.
Subsystems affected by this patch series:
mm/debug
mm/selftests
mm/pagecache
mm/thp
mm/rmap
mm/migration
mm/kasan
mm/hugetlb
mm/pagemap
mm/madvise
selftests
Subsystem: mm/debug
Sean Anderson <seanga2@gmail.com>:
tools/vm/page_owner_sort.c: sort by stacktrace before culling
tools/vm/page_owner_sort.c: support sorting by stack trace
Yinan Zhang <zhangyinan2019@email.szu.edu.cn>:
tools/vm/page_owner_sort.c: add switch between culling by stacktrace and txt
Chongxi Zhao <zhaochongxi2019@email.szu.edu.cn>:
tools/vm/page_owner_sort.c: support sorting pid and time
Shenghong Han <hanshenghong2019@email.szu.edu.cn>:
tools/vm/page_owner_sort.c: two trivial fixes
Yixuan Cao <caoyixuan2019@email.szu.edu.cn>:
tools/vm/page_owner_sort.c: delete invalid duplicate code
Shenghong Han <hanshenghong2019@email.szu.edu.cn>:
Documentation/vm/page_owner.rst: update the documentation
Shuah Khan <skhan@linuxfoundation.org>:
Documentation/vm/page_owner.rst: fix unexpected indentation warns
Waiman Long <longman@redhat.com>:
Patch series "mm/page_owner: Extend page_owner to show memcg information", v4:
lib/vsprintf: avoid redundant work with 0 size
mm/page_owner: use scnprintf() to avoid excessive buffer overrun check
mm/page_owner: print memcg information
mm/page_owner: record task command name
Yixuan Cao <caoyixuan2019@email.szu.edu.cn>:
mm/page_owner.c: record tgid
tools/vm/page_owner_sort.c: fix the instructions for use
Jiajian Ye <yejiajian2018@email.szu.edu.cn>:
tools/vm/page_owner_sort.c: fix comments
tools/vm/page_owner_sort.c: add a security check
tools/vm/page_owner_sort.c: support sorting by tgid and update documentation
tools/vm/page_owner_sort: fix three trivival places
tools/vm/page_owner_sort: support for sorting by task command name
tools/vm/page_owner_sort.c: support for selecting by PID, TGID or task command name
tools/vm/page_owner_sort.c: support for user-defined culling rules
Christoph Hellwig <hch@lst.de>:
mm: unexport page_init_poison
Subsystem: mm/selftests
"Aneesh Kumar K.V" <aneesh.kumar@linux.ibm.com>:
selftest/vm: add util.h and and move helper functions there
Mike Rapoport <rppt@kernel.org>:
selftest/vm: add helpers to detect PAGE_SIZE and PAGE_SHIFT
Subsystem: mm/pagecache
Hugh Dickins <hughd@google.com>:
mm: delete __ClearPageWaiters()
mm: filemap_unaccount_folio() large skip mapcount fixup
Subsystem: mm/thp
Hugh Dickins <hughd@google.com>:
mm/thp: fix NR_FILE_MAPPED accounting in page_*_file_rmap()
Subsystem: mm/rmap
Subsystem: mm/migration
Anshuman Khandual <anshuman.khandual@arm.com>:
Patch series "mm/migration: Add trace events", v3:
mm/migration: add trace events for THP migrations
mm/migration: add trace events for base page and HugeTLB migrations
Subsystem: mm/kasan
Andrey Konovalov <andreyknvl@google.com>:
Patch series "kasan, vmalloc, arm64: add vmalloc tagging support for SW/HW_TAGS", v6:
kasan, page_alloc: deduplicate should_skip_kasan_poison
kasan, page_alloc: move tag_clear_highpage out of kernel_init_free_pages
kasan, page_alloc: merge kasan_free_pages into free_pages_prepare
kasan, page_alloc: simplify kasan_poison_pages call site
kasan, page_alloc: init memory of skipped pages on free
kasan: drop skip_kasan_poison variable in free_pages_prepare
mm: clarify __GFP_ZEROTAGS comment
kasan: only apply __GFP_ZEROTAGS when memory is zeroed
kasan, page_alloc: refactor init checks in post_alloc_hook
kasan, page_alloc: merge kasan_alloc_pages into post_alloc_hook
kasan, page_alloc: combine tag_clear_highpage calls in post_alloc_hook
kasan, page_alloc: move SetPageSkipKASanPoison in post_alloc_hook
kasan, page_alloc: move kernel_init_free_pages in post_alloc_hook
kasan, page_alloc: rework kasan_unpoison_pages call site
kasan: clean up metadata byte definitions
kasan: define KASAN_VMALLOC_INVALID for SW_TAGS
kasan, x86, arm64, s390: rename functions for modules shadow
kasan, vmalloc: drop outdated VM_KASAN comment
kasan: reorder vmalloc hooks
kasan: add wrappers for vmalloc hooks
kasan, vmalloc: reset tags in vmalloc functions
kasan, fork: reset pointer tags of vmapped stacks
kasan, arm64: reset pointer tags of vmapped stacks
kasan, vmalloc: add vmalloc tagging for SW_TAGS
kasan, vmalloc, arm64: mark vmalloc mappings as pgprot_tagged
kasan, vmalloc: unpoison VM_ALLOC pages after mapping
kasan, mm: only define ___GFP_SKIP_KASAN_POISON with HW_TAGS
kasan, page_alloc: allow skipping unpoisoning for HW_TAGS
kasan, page_alloc: allow skipping memory init for HW_TAGS
kasan, vmalloc: add vmalloc tagging for HW_TAGS
kasan, vmalloc: only tag normal vmalloc allocations
kasan, arm64: don't tag executable vmalloc allocations
kasan: mark kasan_arg_stacktrace as __initdata
kasan: clean up feature flags for HW_TAGS mode
kasan: add kasan.vmalloc command line flag
kasan: allow enabling KASAN_VMALLOC and SW/HW_TAGS
arm64: select KASAN_VMALLOC for SW/HW_TAGS modes
kasan: documentation updates
kasan: improve vmalloc tests
kasan: test: support async (again) and asymm modes for HW_TAGS
tangmeng <tangmeng@uniontech.com>:
mm/kasan: remove unnecessary CONFIG_KASAN option
Peter Collingbourne <pcc@google.com>:
kasan: update function name in comments
Andrey Konovalov <andreyknvl@google.com>:
kasan: print virtual mapping info in reports
Patch series "kasan: report clean-ups and improvements":
kasan: drop addr check from describe_object_addr
kasan: more line breaks in reports
kasan: rearrange stack frame info in reports
kasan: improve stack frame info in reports
kasan: print basic stack frame info for SW_TAGS
kasan: simplify async check in end_report()
kasan: simplify kasan_update_kunit_status() and call sites
kasan: check CONFIG_KASAN_KUNIT_TEST instead of CONFIG_KUNIT
kasan: move update_kunit_status to start_report
kasan: move disable_trace_on_warning to start_report
kasan: split out print_report from __kasan_report
kasan: simplify kasan_find_first_bad_addr call sites
kasan: restructure kasan_report
kasan: merge __kasan_report into kasan_report
kasan: call print_report from kasan_report_invalid_free
kasan: move and simplify kasan_report_async
kasan: rename kasan_access_info to kasan_report_info
kasan: add comment about UACCESS regions to kasan_report
kasan: respect KASAN_BIT_REPORTED in all reporting routines
kasan: reorder reporting functions
kasan: move and hide kasan_save_enable/restore_multi_shot
kasan: disable LOCKDEP when printing reports
Subsystem: mm/hugetlb
Mike Kravetz <mike.kravetz@oracle.com>:
Patch series "Add hugetlb MADV_DONTNEED support", v3:
mm: enable MADV_DONTNEED for hugetlb mappings
selftests/vm: add hugetlb madvise MADV_DONTNEED MADV_REMOVE test
userfaultfd/selftests: enable hugetlb remap and remove event testing
Miaohe Lin <linmiaohe@huawei.com>:
mm/huge_memory: make is_transparent_hugepage() static
Subsystem: mm/pagemap
David Hildenbrand <david@redhat.com>:
Patch series "mm: COW fixes part 1: fix the COW security issue for THP and swap", v3:
mm: optimize do_wp_page() for exclusive pages in the swapcache
mm: optimize do_wp_page() for fresh pages in local LRU pagevecs
mm: slightly clarify KSM logic in do_swap_page()
mm: streamline COW logic in do_swap_page()
mm/huge_memory: streamline COW logic in do_huge_pmd_wp_page()
mm/khugepaged: remove reuse_swap_page() usage
mm/swapfile: remove stale reuse_swap_page()
mm/huge_memory: remove stale page_trans_huge_mapcount()
mm/huge_memory: remove stale locking logic from __split_huge_pmd()
Hugh Dickins <hughd@google.com>:
mm: warn on deleting redirtied only if accounted
mm: unmap_mapping_range_tree() with i_mmap_rwsem shared
Anshuman Khandual <anshuman.khandual@arm.com>:
mm: generalize ARCH_HAS_FILTER_PGPROT
Subsystem: mm/madvise
Mauricio Faria de Oliveira <mfo@canonical.com>:
mm: fix race between MADV_FREE reclaim and blkdev direct IO read
Johannes Weiner <hannes@cmpxchg.org>:
mm: madvise: MADV_DONTNEED_LOCKED
Subsystem: selftests
Muhammad Usama Anjum <usama.anjum@collabora.com>:
selftests: vm: remove dependecy from internal kernel macros
Kees Cook <keescook@chromium.org>:
selftests: kselftest framework: provide "finished" helper
Documentation/dev-tools/kasan.rst | 17
Documentation/vm/page_owner.rst | 72 ++
arch/alpha/include/uapi/asm/mman.h | 2
arch/arm64/Kconfig | 2
arch/arm64/include/asm/vmalloc.h | 6
arch/arm64/include/asm/vmap_stack.h | 5
arch/arm64/kernel/module.c | 5
arch/arm64/mm/pageattr.c | 2
arch/arm64/net/bpf_jit_comp.c | 3
arch/mips/include/uapi/asm/mman.h | 2
arch/parisc/include/uapi/asm/mman.h | 2
arch/powerpc/mm/book3s64/trace.c | 1
arch/s390/kernel/module.c | 2
arch/x86/Kconfig | 3
arch/x86/kernel/module.c | 2
arch/x86/mm/init.c | 1
arch/xtensa/include/uapi/asm/mman.h | 2
include/linux/gfp.h | 53 +-
include/linux/huge_mm.h | 6
include/linux/kasan.h | 136 +++--
include/linux/mm.h | 5
include/linux/page-flags.h | 2
include/linux/pagemap.h | 3
include/linux/swap.h | 4
include/linux/vmalloc.h | 18
include/trace/events/huge_memory.h | 1
include/trace/events/migrate.h | 31 +
include/trace/events/mmflags.h | 18
include/trace/events/thp.h | 27 +
include/uapi/asm-generic/mman-common.h | 2
kernel/fork.c | 13
kernel/scs.c | 16
lib/Kconfig.kasan | 18
lib/test_kasan.c | 239 ++++++++-
lib/vsprintf.c | 8
mm/Kconfig | 3
mm/debug.c | 1
mm/filemap.c | 63 +-
mm/huge_memory.c | 109 ----
mm/kasan/Makefile | 2
mm/kasan/common.c | 4
mm/kasan/hw_tags.c | 243 +++++++---
mm/kasan/kasan.h | 76 ++-
mm/kasan/report.c | 516 +++++++++++----------
mm/kasan/report_generic.c | 34 -
mm/kasan/report_hw_tags.c | 1
mm/kasan/report_sw_tags.c | 16
mm/kasan/report_tags.c | 2
mm/kasan/shadow.c | 76 +--
mm/khugepaged.c | 11
mm/madvise.c | 57 +-
mm/memory.c | 129 +++--
mm/memremap.c | 2
mm/migrate.c | 4
mm/page-writeback.c | 18
mm/page_alloc.c | 270 ++++++-----
mm/page_owner.c | 86 ++-
mm/rmap.c | 62 +-
mm/swap.c | 4
mm/swapfile.c | 104 ----
mm/vmalloc.c | 167 ++++--
tools/testing/selftests/kselftest.h | 10
tools/testing/selftests/vm/.gitignore | 1
tools/testing/selftests/vm/Makefile | 1
tools/testing/selftests/vm/gup_test.c | 3
tools/testing/selftests/vm/hugetlb-madvise.c | 410 ++++++++++++++++
tools/testing/selftests/vm/ksm_tests.c | 38 -
tools/testing/selftests/vm/memfd_secret.c | 2
tools/testing/selftests/vm/run_vmtests.sh | 15
tools/testing/selftests/vm/transhuge-stress.c | 41 -
tools/testing/selftests/vm/userfaultfd.c | 72 +-
tools/testing/selftests/vm/util.h | 75 ++-
tools/vm/page_owner_sort.c | 628 +++++++++++++++++++++-----
73 files changed, 2797 insertions(+), 1288 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2022-03-23 23:04 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2022-03-23 23:04 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm, patches
Various misc subsystems, before getting into the post-linux-next material.
This is all based on v5.17. I tested applying and compiling against
today's 1bc191051dca28fa6. One patch required an extra whack, all
looks good.
41 patches, based on f443e374ae131c168a065ea1748feac6b2e76613.
Subsystems affected by this patch series:
procfs
misc
core-kernel
lib
checkpatch
init
pipe
minix
fat
cgroups
kexec
kdump
taskstats
panic
kcov
resource
ubsan
Subsystem: procfs
Hao Lee <haolee.swjtu@gmail.com>:
proc: alloc PATH_MAX bytes for /proc/${pid}/fd/ symlinks
David Hildenbrand <david@redhat.com>:
proc/vmcore: fix possible deadlock on concurrent mmap and read
Yang Li <yang.lee@linux.alibaba.com>:
proc/vmcore: fix vmcore_alloc_buf() kernel-doc comment
Subsystem: misc
Bjorn Helgaas <bhelgaas@google.com>:
linux/types.h: remove unnecessary __bitwise__
Documentation/sparse: add hints about __CHECKER__
Subsystem: core-kernel
Miaohe Lin <linmiaohe@huawei.com>:
kernel/ksysfs.c: use helper macro __ATTR_RW
Subsystem: lib
Kees Cook <keescook@chromium.org>:
Kconfig.debug: make DEBUG_INFO selectable from a choice
Rasmus Villemoes <linux@rasmusvillemoes.dk>:
include: drop pointless __compiler_offsetof indirection
Christophe Leroy <christophe.leroy@csgroup.eu>:
ilog2: force inlining of __ilog2_u32() and __ilog2_u64()
Andy Shevchenko <andriy.shevchenko@linux.intel.com>:
bitfield: add explicit inclusions to the example
Feng Tang <feng.tang@intel.com>:
lib/Kconfig.debug: add ARCH dependency for FUNCTION_ALIGN option
Randy Dunlap <rdunlap@infradead.org>:
lib: bitmap: fix many kernel-doc warnings
Subsystem: checkpatch
Joe Perches <joe@perches.com>:
checkpatch: prefer MODULE_LICENSE("GPL") over MODULE_LICENSE("GPL v2")
checkpatch: add --fix option for some TRAILING_STATEMENTS
checkpatch: add early_param exception to blank line after struct/function test
Sagar Patel <sagarmp@cs.unc.edu>:
checkpatch: use python3 to find codespell dictionary
Subsystem: init
Mark-PK Tsai <mark-pk.tsai@mediatek.com>:
init: use ktime_us_delta() to make initcall_debug log more precise
Randy Dunlap <rdunlap@infradead.org>:
init.h: improve __setup and early_param documentation
init/main.c: return 1 from handled __setup() functions
Subsystem: pipe
Andrei Vagin <avagin@gmail.com>:
fs/pipe: use kvcalloc to allocate a pipe_buffer array
fs/pipe.c: local vars have to match types of proper pipe_inode_info fields
Subsystem: minix
Qinghua Jin <qhjin.dev@gmail.com>:
minix: fix bug when opening a file with O_DIRECT
Subsystem: fat
Helge Deller <deller@gmx.de>:
fat: use pointer to simple type in put_user()
Subsystem: cgroups
Sebastian Andrzej Siewior <bigeasy@linutronix.de>:
cgroup: use irqsave in cgroup_rstat_flush_locked().
cgroup: add a comment to cgroup_rstat_flush_locked().
Subsystem: kexec
Jisheng Zhang <jszhang@kernel.org>:
Patch series "kexec: use IS_ENABLED(CONFIG_KEXEC_CORE) instead of #ifdef", v2:
kexec: make crashk_res, crashk_low_res and crash_notes symbols always visible
riscv: mm: init: use IS_ENABLED(CONFIG_KEXEC_CORE) instead of #ifdef
x86/setup: use IS_ENABLED(CONFIG_KEXEC_CORE) instead of #ifdef
arm64: mm: use IS_ENABLED(CONFIG_KEXEC_CORE) instead of #ifdef
Subsystem: kdump
Tiezhu Yang <yangtiezhu@loongson.cn>:
Patch series "Update doc and fix some issues about kdump", v2:
docs: kdump: update description about sysfs file system support
docs: kdump: add scp example to write out the dump file
panic: unset panic_on_warn inside panic()
ubsan: no need to unset panic_on_warn in ubsan_epilogue()
kasan: no need to unset panic_on_warn in end_report()
Subsystem: taskstats
Lukas Bulwahn <lukas.bulwahn@gmail.com>:
taskstats: remove unneeded dead assignment
Subsystem: panic
"Guilherme G. Piccoli" <gpiccoli@igalia.com>:
Patch series "Some improvements on panic_print":
docs: sysctl/kernel: add missing bit to panic_print
panic: add option to dump all CPUs backtraces in panic_print
panic: move panic_print before kmsg dumpers
Subsystem: kcov
Aleksandr Nogikh <nogikh@google.com>:
Patch series "kcov: improve mmap processing", v3:
kcov: split ioctl handling into locked and unlocked parts
kcov: properly handle subsequent mmap calls
Subsystem: resource
Miaohe Lin <linmiaohe@huawei.com>:
kernel/resource: fix kfree() of bootmem memory again
Subsystem: ubsan
Marco Elver <elver@google.com>:
Revert "ubsan, kcsan: Don't combine sanitizer with kcov on clang"
Documentation/admin-guide/kdump/kdump.rst | 10 +
Documentation/admin-guide/kernel-parameters.txt | 5
Documentation/admin-guide/sysctl/kernel.rst | 2
Documentation/dev-tools/sparse.rst | 2
arch/arm64/mm/init.c | 9 -
arch/riscv/mm/init.c | 6 -
arch/x86/kernel/setup.c | 10 -
fs/fat/dir.c | 2
fs/minix/inode.c | 3
fs/pipe.c | 13 +-
fs/proc/base.c | 8 -
fs/proc/vmcore.c | 43 +++----
include/linux/bitfield.h | 3
include/linux/compiler_types.h | 3
include/linux/init.h | 11 +
include/linux/kexec.h | 12 +-
include/linux/log2.h | 4
include/linux/stddef.h | 6 -
include/uapi/linux/types.h | 6 -
init/main.c | 14 +-
kernel/cgroup/rstat.c | 13 +-
kernel/kcov.c | 102 ++++++++---------
kernel/ksysfs.c | 3
kernel/panic.c | 37 ++++--
kernel/resource.c | 41 +-----
kernel/taskstats.c | 5
lib/Kconfig.debug | 142 ++++++++++++------------
lib/Kconfig.kcsan | 11 -
lib/Kconfig.ubsan | 12 --
lib/bitmap.c | 24 ++--
lib/ubsan.c | 10 -
mm/kasan/report.c | 10 -
scripts/checkpatch.pl | 31 ++++-
tools/include/linux/types.h | 5
34 files changed, 313 insertions(+), 305 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2022-03-22 21:38 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2022-03-22 21:38 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits, patches
- A few misc subsystems
- There is a lot of MM material in Willy's tree. Folio work and
non-folio patches which depended on that work.
Here I send almost all the MM patches which precede the patches in
Willy's tree. The remaining ~100 MM patches are staged on Willy's
tree and I'll send those along once Willy is merged up.
I tried this batch against your current tree (as of
51912904076680281) and a couple need some extra persuasion to apply,
but all looks OK otherwise.
227 patches, based on f443e374ae131c168a065ea1748feac6b2e76613
Subsystems affected by this patch series:
kthread
scripts
ntfs
ocfs2
block
vfs
mm/kasan
mm/pagecache
mm/gup
mm/swap
mm/shmem
mm/memcg
mm/selftests
mm/pagemap
mm/mremap
mm/sparsemem
mm/vmalloc
mm/pagealloc
mm/memory-failure
mm/mlock
mm/hugetlb
mm/userfaultfd
mm/vmscan
mm/compaction
mm/mempolicy
mm/oom-kill
mm/migration
mm/thp
mm/cma
mm/autonuma
mm/psi
mm/ksm
mm/page-poison
mm/madvise
mm/memory-hotplug
mm/rmap
mm/zswap
mm/uaccess
mm/ioremap
mm/highmem
mm/cleanups
mm/kfence
mm/hmm
mm/damon
Subsystem: kthread
Rasmus Villemoes <linux@rasmusvillemoes.dk>:
linux/kthread.h: remove unused macros
Subsystem: scripts
Colin Ian King <colin.i.king@gmail.com>:
scripts/spelling.txt: add more spellings to spelling.txt
Subsystem: ntfs
Dongliang Mu <mudongliangabcd@gmail.com>:
ntfs: add sanity check on allocation size
Subsystem: ocfs2
Joseph Qi <joseph.qi@linux.alibaba.com>:
ocfs2: cleanup some return variables
hongnanli <hongnan.li@linux.alibaba.com>:
fs/ocfs2: fix comments mentioning i_mutex
Subsystem: block
NeilBrown <neilb@suse.de>:
Patch series "Remove remaining parts of congestion tracking code", v2:
doc: convert 'subsection' to 'section' in gfp.h
mm: document and polish read-ahead code
mm: improve cleanup when ->readpages doesn't process all pages
fuse: remove reliance on bdi congestion
nfs: remove reliance on bdi congestion
ceph: remove reliance on bdi congestion
remove inode_congested()
remove bdi_congested() and wb_congested() and related functions
f2fs: replace congestion_wait() calls with io_schedule_timeout()
block/bfq-iosched.c: use "false" rather than "BLK_RW_ASYNC"
remove congestion tracking framework
Subsystem: vfs
Anthony Iliopoulos <ailiop@suse.com>:
mount: warn only once about timestamp range expiration
Subsystem: mm/kasan
Miaohe Lin <linmiaohe@huawei.com>:
mm/memremap: avoid calling kasan_remove_zero_shadow() for device private memory
Subsystem: mm/pagecache
Miaohe Lin <linmiaohe@huawei.com>:
filemap: remove find_get_pages()
mm/writeback: minor clean up for highmem_dirtyable_memory
Minchan Kim <minchan@kernel.org>:
mm: fs: fix lru_cache_disabled race in bh_lru
Subsystem: mm/gup
Peter Xu <peterx@redhat.com>:
Patch series "mm/gup: some cleanups", v5:
mm: fix invalid page pointer returned with FOLL_PIN gups
John Hubbard <jhubbard@nvidia.com>:
mm/gup: follow_pfn_pte(): -EEXIST cleanup
mm/gup: remove unused pin_user_pages_locked()
mm: change lookup_node() to use get_user_pages_fast()
mm/gup: remove unused get_user_pages_locked()
Subsystem: mm/swap
Bang Li <libang.linuxer@gmail.com>:
mm/swap: fix confusing comment in folio_mark_accessed
Subsystem: mm/shmem
Xavier Roche <xavier.roche@algolia.com>:
tmpfs: support for file creation time
Hugh Dickins <hughd@google.com>:
shmem: mapping_set_exiting() to help mapped resilience
tmpfs: do not allocate pages on read
Miaohe Lin <linmiaohe@huawei.com>:
mm: shmem: use helper macro __ATTR_RW
Subsystem: mm/memcg
Shakeel Butt <shakeelb@google.com>:
memcg: replace in_interrupt() with !in_task()
Yosry Ahmed <yosryahmed@google.com>:
memcg: add per-memcg total kernel memory stat
Wei Yang <richard.weiyang@gmail.com>:
mm/memcg: mem_cgroup_per_node is already set to 0 on allocation
mm/memcg: retrieve parent memcg from css.parent
Shakeel Butt <shakeelb@google.com>:
Patch series "memcg: robust enforcement of memory.high", v2:
memcg: refactor mem_cgroup_oom
memcg: unify force charging conditions
selftests: memcg: test high limit for single entry allocation
memcg: synchronously enforce memory.high for large overcharges
Randy Dunlap <rdunlap@infradead.org>:
mm/memcontrol: return 1 from cgroup.memory __setup() handler
Michal Hocko <mhocko@suse.com>:
Patch series "mm/memcg: Address PREEMPT_RT problems instead of disabling it", v5:
mm/memcg: revert ("mm/memcg: optimize user context object stock access")
Sebastian Andrzej Siewior <bigeasy@linutronix.de>:
mm/memcg: disable threshold event handlers on PREEMPT_RT
mm/memcg: protect per-CPU counter by disabling preemption on PREEMPT_RT where needed.
Johannes Weiner <hannes@cmpxchg.org>:
mm/memcg: opencode the inner part of obj_cgroup_uncharge_pages() in drain_obj_stock()
Sebastian Andrzej Siewior <bigeasy@linutronix.de>:
mm/memcg: protect memcg_stock with a local_lock_t
mm/memcg: disable migration instead of preemption in drain_all_stock().
Muchun Song <songmuchun@bytedance.com>:
Patch series "Optimize list lru memory consumption", v6:
mm: list_lru: transpose the array of per-node per-memcg lru lists
mm: introduce kmem_cache_alloc_lru
fs: introduce alloc_inode_sb() to allocate filesystems specific inode
fs: allocate inode by using alloc_inode_sb()
f2fs: allocate inode by using alloc_inode_sb()
mm: dcache: use kmem_cache_alloc_lru() to allocate dentry
xarray: use kmem_cache_alloc_lru to allocate xa_node
mm: memcontrol: move memcg_online_kmem() to mem_cgroup_css_online()
mm: list_lru: allocate list_lru_one only when needed
mm: list_lru: rename memcg_drain_all_list_lrus to memcg_reparent_list_lrus
mm: list_lru: replace linear array with xarray
mm: memcontrol: reuse memory cgroup ID for kmem ID
mm: memcontrol: fix cannot alloc the maximum memcg ID
mm: list_lru: rename list_lru_per_memcg to list_lru_memcg
mm: memcontrol: rename memcg_cache_id to memcg_kmem_id
Vasily Averin <vvs@virtuozzo.com>:
memcg: enable accounting for tty-related objects
Subsystem: mm/selftests
Guillaume Tucker <guillaume.tucker@collabora.com>:
selftests, x86: fix how check_cc.sh is being invoked
Subsystem: mm/pagemap
Anshuman Khandual <anshuman.khandual@arm.com>:
mm: merge pte_mkhuge() call into arch_make_huge_pte()
Stafford Horne <shorne@gmail.com>:
mm: remove mmu_gathers storage from remaining architectures
Muchun Song <songmuchun@bytedance.com>:
Patch series "Fix some cache flush bugs", v5:
mm: thp: fix wrong cache flush in remove_migration_pmd()
mm: fix missing cache flush for all tail pages of compound page
mm: hugetlb: fix missing cache flush in copy_huge_page_from_user()
mm: hugetlb: fix missing cache flush in hugetlb_mcopy_atomic_pte()
mm: shmem: fix missing cache flush in shmem_mfill_atomic_pte()
mm: userfaultfd: fix missing cache flush in mcopy_atomic_pte() and __mcopy_atomic()
mm: replace multiple dcache flush with flush_dcache_folio()
Peter Xu <peterx@redhat.com>:
Patch series "mm: Rework zap ptes on swap entries", v5:
mm: don't skip swap entry even if zap_details specified
mm: rename zap_skip_check_mapping() to should_zap_page()
mm: change zap_details.zap_mapping into even_cows
mm: rework swap handling of zap_pte_range
Randy Dunlap <rdunlap@infradead.org>:
mm/mmap: return 1 from stack_guard_gap __setup() handler
Miaohe Lin <linmiaohe@huawei.com>:
mm/memory.c: use helper function range_in_vma()
mm/memory.c: use helper macro min and max in unmap_mapping_range_tree()
Hugh Dickins <hughd@google.com>:
mm: _install_special_mapping() apply VM_LOCKED_CLEAR_MASK
Miaohe Lin <linmiaohe@huawei.com>:
mm/mmap: remove obsolete comment in ksys_mmap_pgoff
Subsystem: mm/mremap
Miaohe Lin <linmiaohe@huawei.com>:
mm/mremap:: use vma_lookup() instead of find_vma()
Subsystem: mm/sparsemem
Miaohe Lin <linmiaohe@huawei.com>:
mm/sparse: make mminit_validate_memmodel_limits() static
Subsystem: mm/vmalloc
Miaohe Lin <linmiaohe@huawei.com>:
mm/vmalloc: remove unneeded function forward declaration
"Uladzislau Rezki (Sony)" <urezki@gmail.com>:
mm/vmalloc: Move draining areas out of caller context
Uladzislau Rezki <uladzislau.rezki@sony.com>:
mm/vmalloc: add adjust_search_size parameter
"Uladzislau Rezki (Sony)" <urezki@gmail.com>:
mm/vmalloc: eliminate an extra orig_gfp_mask
Jiapeng Chong <jiapeng.chong@linux.alibaba.com>:
mm/vmalloc.c: fix "unused function" warning
Bang Li <libang.linuxer@gmail.com>:
mm/vmalloc: fix comments about vmap_area struct
Subsystem: mm/pagealloc
Zi Yan <ziy@nvidia.com>:
mm: page_alloc: avoid merging non-fallbackable pageblocks with others
Peter Collingbourne <pcc@google.com>:
mm/mmzone.c: use try_cmpxchg() in page_cpupid_xchg_last()
Miaohe Lin <linmiaohe@huawei.com>:
mm/mmzone.h: remove unused macros
Nicolas Saenz Julienne <nsaenzju@redhat.com>:
mm/page_alloc: don't pass pfn to free_unref_page_commit()
David Hildenbrand <david@redhat.com>:
Patch series "mm: enforce pageblock_order < MAX_ORDER":
cma: factor out minimum alignment requirement
mm: enforce pageblock_order < MAX_ORDER
Nathan Chancellor <nathan@kernel.org>:
mm/page_alloc: mark pagesets as __maybe_unused
Alistair Popple <apopple@nvidia.com>:
mm/pages_alloc.c: don't create ZONE_MOVABLE beyond the end of a node
Mel Gorman <mgorman@techsingularity.net>:
Patch series "Follow-up on high-order PCP caching", v2:
mm/page_alloc: fetch the correct pcp buddy during bulk free
mm/page_alloc: track range of active PCP lists during bulk free
mm/page_alloc: simplify how many pages are selected per pcp list during bulk free
mm/page_alloc: drain the requested list first during bulk free
mm/page_alloc: free pages in a single pass during bulk free
mm/page_alloc: limit number of high-order pages on PCP during bulk free
mm/page_alloc: do not prefetch buddies during bulk free
Oscar Salvador <osalvador@suse.de>:
arch/x86/mm/numa: Do not initialize nodes twice
Suren Baghdasaryan <surenb@google.com>:
mm: count time in drain_all_pages during direct reclaim as memory pressure
Eric Dumazet <edumazet@google.com>:
mm/page_alloc: call check_new_pages() while zone spinlock is not held
Mel Gorman <mgorman@techsingularity.net>:
mm/page_alloc: check high-order pages for corruption during PCP operations
Subsystem: mm/memory-failure
Naoya Horiguchi <naoya.horiguchi@nec.com>:
mm/memory-failure.c: remove obsolete comment
mm/hwpoison: fix error page recovered but reported "not recovered"
Rik van Riel <riel@surriel.com>:
mm: invalidate hwpoison page cache page in fault path
Miaohe Lin <linmiaohe@huawei.com>:
Patch series "A few cleanup and fixup patches for memory failure", v3:
mm/memory-failure.c: minor clean up for memory_failure_dev_pagemap
mm/memory-failure.c: catch unexpected -EFAULT from vma_address()
mm/memory-failure.c: rework the signaling logic in kill_proc
mm/memory-failure.c: fix race with changing page more robustly
mm/memory-failure.c: remove PageSlab check in hwpoison_filter_dev
mm/memory-failure.c: rework the try_to_unmap logic in hwpoison_user_mappings()
mm/memory-failure.c: remove obsolete comment in __soft_offline_page
mm/memory-failure.c: remove unnecessary PageTransTail check
mm/hwpoison-inject: support injecting hwpoison to free page
luofei <luofei@unicloud.com>:
mm/hwpoison: avoid the impact of hwpoison_filter() return value on mce handler
mm/hwpoison: add in-use hugepage hwpoison filter judgement
Miaohe Lin <linmiaohe@huawei.com>:
Patch series "A few fixup patches for memory failure", v2:
mm/memory-failure.c: fix race with changing page compound again
mm/memory-failure.c: avoid calling invalidate_inode_page() with unexpected pages
mm/memory-failure.c: make non-LRU movable pages unhandlable
Vlastimil Babka <vbabka@suse.cz>:
mm, fault-injection: declare should_fail_alloc_page()
Subsystem: mm/mlock
Miaohe Lin <linmiaohe@huawei.com>:
mm/mlock: fix potential imbalanced rlimit ucounts adjustment
Subsystem: mm/hugetlb
Muchun Song <songmuchun@bytedance.com>:
Patch series "Free the 2nd vmemmap page associated with each HugeTLB page", v7:
mm: hugetlb: free the 2nd vmemmap page associated with each HugeTLB page
mm: hugetlb: replace hugetlb_free_vmemmap_enabled with a static_key
mm: sparsemem: use page table lock to protect kernel pmd operations
selftests: vm: add a hugetlb test case
mm: sparsemem: move vmemmap related to HugeTLB to CONFIG_HUGETLB_PAGE_FREE_VMEMMAP
Anshuman Khandual <anshuman.khandual@arm.com>:
mm/hugetlb: generalize ARCH_WANT_GENERAL_HUGETLB
Mike Kravetz <mike.kravetz@oracle.com>:
hugetlb: clean up potential spectre issue warnings
Miaohe Lin <linmiaohe@huawei.com>:
mm/hugetlb: use helper macro __ATTR_RW
David Howells <dhowells@redhat.com>:
mm/hugetlb.c: export PageHeadHuge()
Miaohe Lin <linmiaohe@huawei.com>:
mm: remove unneeded local variable follflags
Subsystem: mm/userfaultfd
Nadav Amit <namit@vmware.com>:
userfaultfd: provide unmasked address on page-fault
Guo Zhengkui <guozhengkui@vivo.com>:
userfaultfd/selftests: fix uninitialized_var.cocci warning
Subsystem: mm/vmscan
Hugh Dickins <hughd@google.com>:
mm/fs: delete PF_SWAPWRITE
mm: __isolate_lru_page_prepare() in isolate_migratepages_block()
Waiman Long <longman@redhat.com>:
mm/list_lru: optimize memcg_reparent_list_lru_node()
Marcelo Tosatti <mtosatti@redhat.com>:
mm: lru_cache_disable: replace work queue synchronization with synchronize_rcu
Sebastian Andrzej Siewior <bigeasy@linutronix.de>:
mm: workingset: replace IRQ-off check with a lockdep assert.
Charan Teja Kalla <quic_charante@quicinc.com>:
mm: vmscan: fix documentation for page_check_references()
Subsystem: mm/compaction
Baolin Wang <baolin.wang@linux.alibaba.com>:
mm: compaction: cleanup the compaction trace events
Subsystem: mm/mempolicy
Hugh Dickins <hughd@google.com>:
mempolicy: mbind_range() set_policy() after vma_merge()
Subsystem: mm/oom-kill
Miaohe Lin <linmiaohe@huawei.com>:
mm/oom_kill: remove unneeded is_memcg_oom check
Subsystem: mm/migration
Huang Ying <ying.huang@intel.com>:
mm,migrate: fix establishing demotion target
"andrew.yang" <andrew.yang@mediatek.com>:
mm/migrate: fix race between lock page and clear PG_Isolated
Subsystem: mm/thp
Hugh Dickins <hughd@google.com>:
mm/thp: refix __split_huge_pmd_locked() for migration PMD
Subsystem: mm/cma
Hari Bathini <hbathini@linux.ibm.com>:
Patch series "powerpc/fadump: handle CMA activation failure appropriately", v3:
mm/cma: provide option to opt out from exposing pages on activation failure
powerpc/fadump: opt out from freeing pages on cma activation failure
Subsystem: mm/autonuma
Huang Ying <ying.huang@intel.com>:
Patch series "NUMA balancing: optimize memory placement for memory tiering system", v13:
NUMA Balancing: add page promotion counter
NUMA balancing: optimize page placement for memory tiering system
memory tiering: skip to scan fast memory
Subsystem: mm/psi
Johannes Weiner <hannes@cmpxchg.org>:
mm: page_io: fix psi memory pressure error on cold swapins
Subsystem: mm/ksm
Yang Yang <yang.yang29@zte.com.cn>:
mm/vmstat: add event for ksm swapping in copy
Miaohe Lin <linmiaohe@huawei.com>:
mm/ksm: use helper macro __ATTR_RW
Subsystem: mm/page-poison
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
mm/hwpoison: check the subpage, not the head page
Subsystem: mm/madvise
Miaohe Lin <linmiaohe@huawei.com>:
mm/madvise: use vma_lookup() instead of find_vma()
Charan Teja Kalla <quic_charante@quicinc.com>:
Patch series "mm: madvise: return correct bytes processed with:
mm: madvise: return correct bytes advised with process_madvise
mm: madvise: skip unmapped vma holes passed to process_madvise
Subsystem: mm/memory-hotplug
Michal Hocko <mhocko@suse.com>:
Patch series "mm, memory_hotplug: handle unitialized numa node gracefully":
mm, memory_hotplug: make arch_alloc_nodedata independent on CONFIG_MEMORY_HOTPLUG
mm: handle uninitialized numa nodes gracefully
mm, memory_hotplug: drop arch_free_nodedata
mm, memory_hotplug: reorganize new pgdat initialization
mm: make free_area_init_node aware of memory less nodes
Wei Yang <richard.weiyang@gmail.com>:
memcg: do not tweak node in alloc_mem_cgroup_per_node_info
David Hildenbrand <david@redhat.com>:
drivers/base/memory: add memory block to memory group after registration succeeded
drivers/base/node: consolidate node device subsystem initialization in node_dev_init()
Miaohe Lin <linmiaohe@huawei.com>:
Patch series "A few cleanup patches around memory_hotplug":
mm/memory_hotplug: remove obsolete comment of __add_pages
mm/memory_hotplug: avoid calling zone_intersects() for ZONE_NORMAL
mm/memory_hotplug: clean up try_offline_node
mm/memory_hotplug: fix misplaced comment in offline_pages
David Hildenbrand <david@redhat.com>:
Patch series "drivers/base/memory: determine and store zone for single-zone memory blocks", v2:
drivers/base/node: rename link_mem_sections() to register_memory_block_under_node()
drivers/base/memory: determine and store zone for single-zone memory blocks
drivers/base/memory: clarify adding and removing of memory blocks
Oscar Salvador <osalvador@suse.de>:
mm: only re-generate demotion targets when a numa node changes its N_CPU state
Subsystem: mm/rmap
Hugh Dickins <hughd@google.com>:
mm/thp: ClearPageDoubleMap in first page_add_file_rmap()
Subsystem: mm/zswap
"Maciej S. Szmigiero" <maciej.szmigiero@oracle.com>:
mm/zswap.c: allow handling just same-value filled pages
Subsystem: mm/uaccess
Christophe Leroy <christophe.leroy@csgroup.eu>:
mm: remove usercopy_warn()
mm: uninline copy_overflow()
Randy Dunlap <rdunlap@infradead.org>:
mm/usercopy: return 1 from hardened_usercopy __setup() handler
Subsystem: mm/ioremap
Vlastimil Babka <vbabka@suse.cz>:
mm/early_ioremap: declare early_memremap_pgprot_adjust()
Subsystem: mm/highmem
Ira Weiny <ira.weiny@intel.com>:
highmem: document kunmap_local()
Miaohe Lin <linmiaohe@huawei.com>:
mm/highmem: remove unnecessary done label
Subsystem: mm/cleanups
"Dr. David Alan Gilbert" <linux@treblig.org>:
mm/page_table_check.c: use strtobool for param parsing
Subsystem: mm/kfence
tangmeng <tangmeng@uniontech.com>:
mm/kfence: remove unnecessary CONFIG_KFENCE option
Tianchen Ding <dtcccc@linux.alibaba.com>:
Patch series "provide the flexibility to enable KFENCE", v3:
kfence: allow re-enabling KFENCE after system startup
kfence: alloc kfence_pool after system startup
Peng Liu <liupeng256@huawei.com>:
Patch series "kunit: fix a UAF bug and do some optimization", v2:
kunit: fix UAF when run kfence test case test_gfpzero
kunit: make kunit_test_timeout compatible with comment
kfence: test: try to avoid test_gfpzero trigger rcu_stall
Marco Elver <elver@google.com>:
kfence: allow use of a deferrable timer
Subsystem: mm/hmm
Miaohe Lin <linmiaohe@huawei.com>:
mm/hmm.c: remove unneeded local variable ret
Subsystem: mm/damon
SeongJae Park <sj@kernel.org>:
Patch series "Remove the type-unclear target id concept":
mm/damon/dbgfs/init_regions: use target index instead of target id
Docs/admin-guide/mm/damon/usage: update for changed initail_regions file input
mm/damon/core: move damon_set_targets() into dbgfs
mm/damon: remove the target id concept
Baolin Wang <baolin.wang@linux.alibaba.com>:
mm/damon: remove redundant page validation
SeongJae Park <sj@kernel.org>:
Patch series "Allow DAMON user code independent of monitoring primitives":
mm/damon: rename damon_primitives to damon_operations
mm/damon: let monitoring operations can be registered and selected
mm/damon/paddr,vaddr: register themselves to DAMON in subsys_initcall
mm/damon/reclaim: use damon_select_ops() instead of damon_{v,p}a_set_operations()
mm/damon/dbgfs: use damon_select_ops() instead of damon_{v,p}a_set_operations()
mm/damon/dbgfs: use operations id for knowing if the target has pid
mm/damon/dbgfs-test: fix is_target_id() change
mm/damon/paddr,vaddr: remove damon_{p,v}a_{target_valid,set_operations}()
tangmeng <tangmeng@uniontech.com>:
mm/damon: remove unnecessary CONFIG_DAMON option
SeongJae Park <sj@kernel.org>:
Patch series "Docs/damon: Update documents for better consistency":
Docs/vm/damon: call low level monitoring primitives the operations
Docs/vm/damon/design: update DAMON-Idle Page Tracking interference handling
Docs/damon: update outdated term 'regions update interval'
Patch series "Introduce DAMON sysfs interface", v3:
mm/damon/core: allow non-exclusive DAMON start/stop
mm/damon/core: add number of each enum type values
mm/damon: implement a minimal stub for sysfs-based DAMON interface
mm/damon/sysfs: link DAMON for virtual address spaces monitoring
mm/damon/sysfs: support the physical address space monitoring
mm/damon/sysfs: support DAMON-based Operation Schemes
mm/damon/sysfs: support DAMOS quotas
mm/damon/sysfs: support schemes prioritization
mm/damon/sysfs: support DAMOS watermarks
mm/damon/sysfs: support DAMOS stats
selftests/damon: add a test for DAMON sysfs interface
Docs/admin-guide/mm/damon/usage: document DAMON sysfs interface
Docs/ABI/testing: add DAMON sysfs interface ABI document
Xin Hao <xhao@linux.alibaba.com>:
mm/damon/sysfs: remove repeat container_of() in damon_sysfs_kdamond_release()
Documentation/ABI/testing/sysfs-kernel-mm-damon | 274 ++
Documentation/admin-guide/cgroup-v1/memory.rst | 2
Documentation/admin-guide/cgroup-v2.rst | 5
Documentation/admin-guide/kernel-parameters.txt | 2
Documentation/admin-guide/mm/damon/usage.rst | 380 +++
Documentation/admin-guide/mm/zswap.rst | 22
Documentation/admin-guide/sysctl/kernel.rst | 31
Documentation/core-api/mm-api.rst | 19
Documentation/dev-tools/kfence.rst | 12
Documentation/filesystems/porting.rst | 6
Documentation/filesystems/vfs.rst | 16
Documentation/vm/damon/design.rst | 43
Documentation/vm/damon/faq.rst | 2
MAINTAINERS | 1
arch/arm/Kconfig | 4
arch/arm64/kernel/setup.c | 3
arch/arm64/mm/hugetlbpage.c | 1
arch/hexagon/mm/init.c | 2
arch/ia64/kernel/topology.c | 10
arch/ia64/mm/discontig.c | 11
arch/mips/kernel/topology.c | 5
arch/nds32/mm/init.c | 1
arch/openrisc/mm/init.c | 2
arch/powerpc/include/asm/fadump-internal.h | 5
arch/powerpc/include/asm/nohash/32/hugetlb-8xx.h | 4
arch/powerpc/kernel/fadump.c | 8
arch/powerpc/kernel/sysfs.c | 17
arch/riscv/Kconfig | 4
arch/riscv/kernel/setup.c | 3
arch/s390/kernel/numa.c | 7
arch/sh/kernel/topology.c | 5
arch/sparc/kernel/sysfs.c | 12
arch/sparc/mm/hugetlbpage.c | 1
arch/x86/Kconfig | 4
arch/x86/kernel/cpu/mce/core.c | 8
arch/x86/kernel/topology.c | 5
arch/x86/mm/numa.c | 33
block/bdev.c | 2
block/bfq-iosched.c | 2
drivers/base/init.c | 1
drivers/base/memory.c | 149 +
drivers/base/node.c | 48
drivers/block/drbd/drbd_int.h | 3
drivers/block/drbd/drbd_req.c | 3
drivers/dax/super.c | 2
drivers/of/of_reserved_mem.c | 9
drivers/tty/tty_io.c | 2
drivers/virtio/virtio_mem.c | 9
fs/9p/vfs_inode.c | 2
fs/adfs/super.c | 2
fs/affs/super.c | 2
fs/afs/super.c | 2
fs/befs/linuxvfs.c | 2
fs/bfs/inode.c | 2
fs/btrfs/inode.c | 2
fs/buffer.c | 8
fs/ceph/addr.c | 22
fs/ceph/inode.c | 2
fs/ceph/super.c | 1
fs/ceph/super.h | 1
fs/cifs/cifsfs.c | 2
fs/coda/inode.c | 2
fs/dcache.c | 3
fs/ecryptfs/super.c | 2
fs/efs/super.c | 2
fs/erofs/super.c | 2
fs/exfat/super.c | 2
fs/ext2/ialloc.c | 5
fs/ext2/super.c | 2
fs/ext4/super.c | 2
fs/f2fs/compress.c | 4
fs/f2fs/data.c | 3
fs/f2fs/f2fs.h | 6
fs/f2fs/segment.c | 8
fs/f2fs/super.c | 14
fs/fat/inode.c | 2
fs/freevxfs/vxfs_super.c | 2
fs/fs-writeback.c | 40
fs/fuse/control.c | 17
fs/fuse/dev.c | 8
fs/fuse/file.c | 17
fs/fuse/inode.c | 2
fs/gfs2/super.c | 2
fs/hfs/super.c | 2
fs/hfsplus/super.c | 2
fs/hostfs/hostfs_kern.c | 2
fs/hpfs/super.c | 2
fs/hugetlbfs/inode.c | 2
fs/inode.c | 2
fs/isofs/inode.c | 2
fs/jffs2/super.c | 2
fs/jfs/super.c | 2
fs/minix/inode.c | 2
fs/namespace.c | 2
fs/nfs/inode.c | 2
fs/nfs/write.c | 14
fs/nilfs2/segbuf.c | 16
fs/nilfs2/super.c | 2
fs/ntfs/inode.c | 6
fs/ntfs3/super.c | 2
fs/ocfs2/alloc.c | 2
fs/ocfs2/aops.c | 2
fs/ocfs2/cluster/nodemanager.c | 2
fs/ocfs2/dir.c | 4
fs/ocfs2/dlmfs/dlmfs.c | 2
fs/ocfs2/file.c | 13
fs/ocfs2/inode.c | 2
fs/ocfs2/localalloc.c | 6
fs/ocfs2/namei.c | 2
fs/ocfs2/ocfs2.h | 4
fs/ocfs2/quota_global.c | 2
fs/ocfs2/stack_user.c | 18
fs/ocfs2/super.c | 2
fs/ocfs2/xattr.c | 2
fs/openpromfs/inode.c | 2
fs/orangefs/super.c | 2
fs/overlayfs/super.c | 2
fs/proc/inode.c | 2
fs/qnx4/inode.c | 2
fs/qnx6/inode.c | 2
fs/reiserfs/super.c | 2
fs/romfs/super.c | 2
fs/squashfs/super.c | 2
fs/sysv/inode.c | 2
fs/ubifs/super.c | 2
fs/udf/super.c | 2
fs/ufs/super.c | 2
fs/userfaultfd.c | 5
fs/vboxsf/super.c | 2
fs/xfs/libxfs/xfs_btree.c | 2
fs/xfs/xfs_buf.c | 3
fs/xfs/xfs_icache.c | 2
fs/zonefs/super.c | 2
include/linux/backing-dev-defs.h | 8
include/linux/backing-dev.h | 50
include/linux/cma.h | 14
include/linux/damon.h | 95
include/linux/fault-inject.h | 2
include/linux/fs.h | 21
include/linux/gfp.h | 10
include/linux/highmem-internal.h | 10
include/linux/hugetlb.h | 8
include/linux/kthread.h | 22
include/linux/list_lru.h | 45
include/linux/memcontrol.h | 46
include/linux/memory.h | 12
include/linux/memory_hotplug.h | 132 -
include/linux/migrate.h | 8
include/linux/mm.h | 11
include/linux/mmzone.h | 22
include/linux/nfs_fs_sb.h | 1
include/linux/node.h | 25
include/linux/page-flags.h | 96
include/linux/pageblock-flags.h | 7
include/linux/pagemap.h | 7
include/linux/sched.h | 1
include/linux/sched/sysctl.h | 10
include/linux/shmem_fs.h | 1
include/linux/slab.h | 3
include/linux/swap.h | 6
include/linux/thread_info.h | 5
include/linux/uaccess.h | 2
include/linux/vm_event_item.h | 3
include/linux/vmalloc.h | 4
include/linux/xarray.h | 9
include/ras/ras_event.h | 1
include/trace/events/compaction.h | 26
include/trace/events/writeback.h | 28
include/uapi/linux/userfaultfd.h | 8
ipc/mqueue.c | 2
kernel/dma/contiguous.c | 4
kernel/sched/core.c | 21
kernel/sysctl.c | 2
lib/Kconfig.kfence | 12
lib/kunit/try-catch.c | 3
lib/xarray.c | 10
mm/Kconfig | 6
mm/backing-dev.c | 57
mm/cma.c | 31
mm/cma.h | 1
mm/compaction.c | 60
mm/damon/Kconfig | 19
mm/damon/Makefile | 7
mm/damon/core-test.h | 23
mm/damon/core.c | 190 +
mm/damon/dbgfs-test.h | 103
mm/damon/dbgfs.c | 264 +-
mm/damon/ops-common.c | 133 +
mm/damon/ops-common.h | 16
mm/damon/paddr.c | 62
mm/damon/prmtv-common.c | 133 -
mm/damon/prmtv-common.h | 16
mm/damon/reclaim.c | 11
mm/damon/sysfs.c | 2632 ++++++++++++++++++++++-
mm/damon/vaddr-test.h | 8
mm/damon/vaddr.c | 67
mm/early_ioremap.c | 1
mm/fadvise.c | 5
mm/filemap.c | 17
mm/gup.c | 103
mm/highmem.c | 9
mm/hmm.c | 3
mm/huge_memory.c | 41
mm/hugetlb.c | 23
mm/hugetlb_vmemmap.c | 74
mm/hwpoison-inject.c | 7
mm/internal.h | 19
mm/kfence/Makefile | 2
mm/kfence/core.c | 147 +
mm/kfence/kfence_test.c | 3
mm/ksm.c | 6
mm/list_lru.c | 690 ++----
mm/maccess.c | 6
mm/madvise.c | 18
mm/memcontrol.c | 549 ++--
mm/memory-failure.c | 148 -
mm/memory.c | 116 -
mm/memory_hotplug.c | 136 -
mm/mempolicy.c | 29
mm/memremap.c | 3
mm/migrate.c | 128 -
mm/mlock.c | 1
mm/mmap.c | 5
mm/mmzone.c | 7
mm/mprotect.c | 13
mm/mremap.c | 4
mm/oom_kill.c | 3
mm/page-writeback.c | 12
mm/page_alloc.c | 429 +--
mm/page_io.c | 7
mm/page_table_check.c | 10
mm/ptdump.c | 16
mm/readahead.c | 124 +
mm/rmap.c | 15
mm/shmem.c | 46
mm/slab.c | 39
mm/slab.h | 25
mm/slob.c | 6
mm/slub.c | 42
mm/sparse-vmemmap.c | 70
mm/sparse.c | 2
mm/swap.c | 25
mm/swapfile.c | 1
mm/usercopy.c | 16
mm/userfaultfd.c | 3
mm/vmalloc.c | 102
mm/vmscan.c | 138 -
mm/vmstat.c | 19
mm/workingset.c | 7
mm/zswap.c | 15
net/socket.c | 2
net/sunrpc/rpc_pipe.c | 2
scripts/spelling.txt | 16
tools/testing/selftests/cgroup/cgroup_util.c | 15
tools/testing/selftests/cgroup/cgroup_util.h | 1
tools/testing/selftests/cgroup/test_memcontrol.c | 78
tools/testing/selftests/damon/Makefile | 1
tools/testing/selftests/damon/sysfs.sh | 306 ++
tools/testing/selftests/vm/.gitignore | 1
tools/testing/selftests/vm/Makefile | 7
tools/testing/selftests/vm/hugepage-vmemmap.c | 144 +
tools/testing/selftests/vm/run_vmtests.sh | 11
tools/testing/selftests/vm/userfaultfd.c | 2
tools/testing/selftests/x86/Makefile | 6
264 files changed, 7205 insertions(+), 3090 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2022-03-16 23:14 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2022-03-16 23:14 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm, patches
4 patches, based on 56e337f2cf1326323844927a04e9dbce9a244835.
Subsystems affected by this patch series:
mm/swap
kconfig
ocfs2
selftests
Subsystem: mm/swap
Guo Ziliang <guo.ziliang@zte.com.cn>:
mm: swap: get rid of deadloop in swapin readahead
Subsystem: kconfig
Qian Cai <quic_qiancai@quicinc.com>:
configs/debug: restore DEBUG_INFO=y for overriding
Subsystem: ocfs2
Joseph Qi <joseph.qi@linux.alibaba.com>:
ocfs2: fix crash when initialize filecheck kobj fails
Subsystem: selftests
Yosry Ahmed <yosryahmed@google.com>:
selftests: vm: fix clang build error multiple output files
fs/ocfs2/super.c | 22 +++++++++++-----------
kernel/configs/debug.config | 1 +
mm/swap_state.c | 2 +-
tools/testing/selftests/vm/Makefile | 6 ++----
4 files changed, 15 insertions(+), 16 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2022-03-05 4:28 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2022-03-05 4:28 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm, patches
8 patches, based on 07ebd38a0da24d2534da57b4841346379db9f354.
Subsystems affected by this patch series:
mm/hugetlb
mm/pagemap
memfd
selftests
mm/userfaultfd
kconfig
Subsystem: mm/hugetlb
Mike Kravetz <mike.kravetz@oracle.com>:
selftests/vm: cleanup hugetlb file after mremap test
Subsystem: mm/pagemap
Suren Baghdasaryan <surenb@google.com>:
mm: refactor vm_area_struct::anon_vma_name usage code
mm: prevent vm_area_struct::anon_name refcount saturation
mm: fix use-after-free when anon vma name is used after vma is freed
Subsystem: memfd
Hugh Dickins <hughd@google.com>:
memfd: fix F_SEAL_WRITE after shmem huge page allocated
Subsystem: selftests
Chengming Zhou <zhouchengming@bytedance.com>:
kselftest/vm: fix tests build with old libc
Subsystem: mm/userfaultfd
Yun Zhou <yun.zhou@windriver.com>:
proc: fix documentation and description of pagemap
Subsystem: kconfig
Qian Cai <quic_qiancai@quicinc.com>:
configs/debug: set CONFIG_DEBUG_INFO=y properly
Documentation/admin-guide/mm/pagemap.rst | 2
fs/proc/task_mmu.c | 9 +-
fs/userfaultfd.c | 6 -
include/linux/mm.h | 7 +
include/linux/mm_inline.h | 105 ++++++++++++++++++---------
include/linux/mm_types.h | 5 +
kernel/configs/debug.config | 2
kernel/fork.c | 4 -
kernel/sys.c | 19 +++-
mm/madvise.c | 98 +++++++++----------------
mm/memfd.c | 40 +++++++---
mm/mempolicy.c | 2
mm/mlock.c | 2
mm/mmap.c | 12 +--
mm/mprotect.c | 2
tools/testing/selftests/vm/hugepage-mremap.c | 26 ++++--
tools/testing/selftests/vm/run_vmtests.sh | 3
tools/testing/selftests/vm/userfaultfd.c | 1
18 files changed, 201 insertions(+), 144 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2022-02-26 3:10 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2022-02-26 3:10 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm, patches
12 patches, based on c47658311d60be064b839f329c0e4d34f5f0735b.
Subsystems affected by this patch series:
MAINTAINERS
mm/hugetlb
mm/kasan
mm/hugetlbfs
mm/pagemap
mm/selftests
mm/memcg
m/slab
mailmap
memfd
Subsystem: MAINTAINERS
Luis Chamberlain <mcgrof@kernel.org>:
MAINTAINERS: add sysctl-next git tree
Subsystem: mm/hugetlb
"Aneesh Kumar K.V" <aneesh.kumar@linux.ibm.com>:
mm/hugetlb: fix kernel crash with hugetlb mremap
Subsystem: mm/kasan
Andrey Konovalov <andreyknvl@google.com>:
kasan: test: prevent cache merging in kmem_cache_double_destroy
Subsystem: mm/hugetlbfs
Liu Yuntao <liuyuntao10@huawei.com>:
hugetlbfs: fix a truncation issue in hugepages parameter
Subsystem: mm/pagemap
Suren Baghdasaryan <surenb@google.com>:
mm: fix use-after-free bug when mm->mmap is reused after being freed
Subsystem: mm/selftests
"Aneesh Kumar K.V" <aneesh.kumar@linux.ibm.com>:
selftest/vm: fix map_fixed_noreplace test failure
Subsystem: mm/memcg
Roman Gushchin <roman.gushchin@linux.dev>:
MAINTAINERS: add Roman as a memcg co-maintainer
Vladimir Davydov <vdavydov.dev@gmail.com>:
MAINTAINERS: remove Vladimir from memcg maintainers
Shakeel Butt <shakeelb@google.com>:
MAINTAINERS: add Shakeel as a memcg co-maintainer
Subsystem: m/slab
Vlastimil Babka <vbabka@suse.cz>:
MAINTAINERS, SLAB: add Roman as reviewer, git tree
Subsystem: mailmap
Roman Gushchin <roman.gushchin@linux.dev>:
mailmap: update Roman Gushchin's email
Subsystem: memfd
Mike Kravetz <mike.kravetz@oracle.com>:
selftests/memfd: clean up mapping in mfd_fail_write
.mailmap | 3 +
MAINTAINERS | 6 ++
lib/test_kasan.c | 5 +-
mm/hugetlb.c | 11 ++---
mm/mmap.c | 1
tools/testing/selftests/memfd/memfd_test.c | 1
tools/testing/selftests/vm/map_fixed_noreplace.c | 49 +++++++++++++++++------
7 files changed, 56 insertions(+), 20 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2022-02-12 2:02 ` incoming Linus Torvalds
@ 2022-02-12 5:24 ` Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2022-02-12 5:24 UTC (permalink / raw)
To: Linus Torvalds; +Cc: Linux-MM, mm-commits, patches
On Fri, 11 Feb 2022 18:02:53 -0800 Linus Torvalds <torvalds@linux-foundation.org> wrote:
> On Fri, Feb 11, 2022 at 4:27 PM Andrew Morton <akpm@linux-foundation.org> wrote:
> >
> > 5 patches, based on f1baf68e1383f6ed93eb9cff2866d46562607a43.
>
> So this *completely* flummoxed 'b4', because you first sent the wrong
> series, and then sent the right one in the same thread.
>
> I fetched the emails manually, but honestly, this was confusing even
> then, with two "[PATCH x/5]" series where the only way to tell the
> right one was basically by date of email. They did arrive in the same
> order in my mailbox, but even that wouldn't have been guaranteed if
> there had been some mailer delays somewhere..
Yes, I wondered. Sorry bout that.
> So next time when you mess up, resend it all as a completely new
> series and completely new threading - so with a new header email too.
> Please?
Wilco.
> And since I'm here, let me just verify that yes, the series you
> actually want me to apply is this one (as described by the head
> email):
>
> Subject: [patch 1/5] fs/binfmt_elf: fix PT_LOAD p_align values ..
> Subject: [patch 2/5] fs/proc: task_mmu.c: don't read mapcount f..
> Subject: [patch 3/5] mm: vmscan: remove deadlock due to throttl..
> Subject: [patch 4/5] mm: memcg: synchronize objcg lists with a ..
> Subject: [patch 5/5] kfence: make test case compatible with run..
>
> and not the other one with GUP patches?
Those are the ones. Five fixes, three with cc:stable.
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2022-02-12 0:27 incoming Andrew Morton
@ 2022-02-12 2:02 ` Linus Torvalds
2022-02-12 5:24 ` incoming Andrew Morton
0 siblings, 1 reply; 409+ messages in thread
From: Linus Torvalds @ 2022-02-12 2:02 UTC (permalink / raw)
To: Andrew Morton; +Cc: Linux-MM, mm-commits, patches
On Fri, Feb 11, 2022 at 4:27 PM Andrew Morton <akpm@linux-foundation.org> wrote:
>
> 5 patches, based on f1baf68e1383f6ed93eb9cff2866d46562607a43.
So this *completely* flummoxed 'b4', because you first sent the wrong
series, and then sent the right one in the same thread.
I fetched the emails manually, but honestly, this was confusing even
then, with two "[PATCH x/5]" series where the only way to tell the
right one was basically by date of email. They did arrive in the same
order in my mailbox, but even that wouldn't have been guaranteed if
there had been some mailer delays somewhere..
So next time when you mess up, resend it all as a completely new
series and completely new threading - so with a new header email too.
Please?
And since I'm here, let me just verify that yes, the series you
actually want me to apply is this one (as described by the head
email):
Subject: [patch 1/5] fs/binfmt_elf: fix PT_LOAD p_align values ..
Subject: [patch 2/5] fs/proc: task_mmu.c: don't read mapcount f..
Subject: [patch 3/5] mm: vmscan: remove deadlock due to throttl..
Subject: [patch 4/5] mm: memcg: synchronize objcg lists with a ..
Subject: [patch 5/5] kfence: make test case compatible with run..
and not the other one with GUP patches?
Linus
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2022-02-12 0:27 Andrew Morton
2022-02-12 2:02 ` incoming Linus Torvalds
0 siblings, 1 reply; 409+ messages in thread
From: Andrew Morton @ 2022-02-12 0:27 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits, patches
5 patches, based on f1baf68e1383f6ed93eb9cff2866d46562607a43.
Subsystems affected by this patch series:
binfmt
procfs
mm/vmscan
mm/memcg
mm/kfence
Subsystem: binfmt
Mike Rapoport <rppt@linux.ibm.com>:
fs/binfmt_elf: fix PT_LOAD p_align values for loaders
Subsystem: procfs
Yang Shi <shy828301@gmail.com>:
fs/proc: task_mmu.c: don't read mapcount for migration entry
Subsystem: mm/vmscan
Mel Gorman <mgorman@suse.de>:
mm: vmscan: remove deadlock due to throttling failing to make progress
Subsystem: mm/memcg
Roman Gushchin <guro@fb.com>:
mm: memcg: synchronize objcg lists with a dedicated spinlock
Subsystem: mm/kfence
Peng Liu <liupeng256@huawei.com>:
kfence: make test case compatible with run time set sample interval
fs/binfmt_elf.c | 2 +-
fs/proc/task_mmu.c | 40 +++++++++++++++++++++++++++++++---------
include/linux/kfence.h | 2 ++
include/linux/memcontrol.h | 5 +++--
mm/kfence/core.c | 3 ++-
mm/kfence/kfence_test.c | 8 ++++----
mm/memcontrol.c | 10 +++++-----
mm/vmscan.c | 4 +++-
8 files changed, 51 insertions(+), 23 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2022-02-04 4:48 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2022-02-04 4:48 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
10 patches, based on 1f2cfdd349b7647f438c1e552dc1b983da86d830.
Subsystems affected by this patch series:
mm/vmscan
mm/debug
mm/pagemap
ipc
mm/kmemleak
MAINTAINERS
mm/selftests
Subsystem: mm/vmscan
Chen Wandun <chenwandun@huawei.com>:
Revert "mm/page_isolation: unset migratetype directly for non Buddy page"
Subsystem: mm/debug
Pasha Tatashin <pasha.tatashin@soleen.com>:
Patch series "page table check fixes and cleanups", v5:
mm/debug_vm_pgtable: remove pte entry from the page table
mm/page_table_check: use unsigned long for page counters and cleanup
mm/khugepaged: unify collapse pmd clear, flush and free
mm/page_table_check: check entries at pmd levels
Subsystem: mm/pagemap
Mike Rapoport <rppt@linux.ibm.com>:
mm/pgtable: define pte_index so that preprocessor could recognize it
Subsystem: ipc
Minghao Chi <chi.minghao@zte.com.cn>:
ipc/sem: do not sleep with a spin lock held
Subsystem: mm/kmemleak
Lang Yu <lang.yu@amd.com>:
mm/kmemleak: avoid scanning potential huge holes
Subsystem: MAINTAINERS
Mike Rapoport <rppt@linux.ibm.com>:
MAINTAINERS: update rppt's email
Subsystem: mm/selftests
Shuah Khan <skhan@linuxfoundation.org>:
kselftest/vm: revert "tools/testing/selftests/vm/userfaultfd.c: use swap() to make code cleaner"
MAINTAINERS | 2 -
include/linux/page_table_check.h | 19 ++++++++++
include/linux/pgtable.h | 1
ipc/sem.c | 4 +-
mm/debug_vm_pgtable.c | 2 +
mm/khugepaged.c | 37 +++++++++++---------
mm/kmemleak.c | 13 +++----
mm/page_isolation.c | 2 -
mm/page_table_check.c | 55 +++++++++++++++----------------
tools/testing/selftests/vm/userfaultfd.c | 11 ++++--
10 files changed, 89 insertions(+), 57 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2022-01-29 21:40 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2022-01-29 21:40 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
12 patches, based on f8c7e4ede46fe63ff10000669652648aab09d112.
Subsystems affected by this patch series:
sysctl
binfmt
ia64
mm/memory-failure
mm/folios
selftests
mm/kasan
mm/psi
ocfs2
Subsystem: sysctl
Andrew Morton <akpm@linux-foundation.org>:
include/linux/sysctl.h: fix register_sysctl_mount_point() return type
Subsystem: binfmt
Tong Zhang <ztong0001@gmail.com>:
binfmt_misc: fix crash when load/unload module
Subsystem: ia64
Randy Dunlap <rdunlap@infradead.org>:
ia64: make IA64_MCA_RECOVERY bool instead of tristate
Subsystem: mm/memory-failure
Joao Martins <joao.m.martins@oracle.com>:
memory-failure: fetch compound_head after pgmap_pfn_valid()
Subsystem: mm/folios
Wei Yang <richard.weiyang@gmail.com>:
mm: page->mapping folio->mapping should have the same offset
Subsystem: selftests
Maor Gottlieb <maorg@nvidia.com>:
tools/testing/scatterlist: add missing defines
Subsystem: mm/kasan
Marco Elver <elver@google.com>:
kasan: test: fix compatibility with FORTIFY_SOURCE
Peter Collingbourne <pcc@google.com>:
mm, kasan: use compare-exchange operation to set KASAN page tag
Subsystem: mm/psi
Suren Baghdasaryan <surenb@google.com>:
psi: fix "no previous prototype" warnings when CONFIG_CGROUPS=n
psi: fix "defined but not used" warnings when CONFIG_PROC_FS=n
Subsystem: ocfs2
Joseph Qi <joseph.qi@linux.alibaba.com>:
Patch series "ocfs2: fix a deadlock case":
jbd2: export jbd2_journal_[grab|put]_journal_head
ocfs2: fix a deadlock when commit trans
arch/ia64/Kconfig | 2
fs/binfmt_misc.c | 8 +--
fs/jbd2/journal.c | 2
fs/ocfs2/suballoc.c | 25 ++++-------
include/linux/mm.h | 17 +++++--
include/linux/mm_types.h | 1
include/linux/psi.h | 11 ++--
include/linux/sysctl.h | 2
kernel/sched/psi.c | 79 ++++++++++++++++++-----------------
lib/test_kasan.c | 5 ++
mm/memory-failure.c | 6 ++
tools/testing/scatterlist/linux/mm.h | 3 -
12 files changed, 91 insertions(+), 70 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2022-01-29 4:25 ` incoming Matthew Wilcox
@ 2022-01-29 6:23 ` Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2022-01-29 6:23 UTC (permalink / raw)
To: Matthew Wilcox; +Cc: Linus Torvalds, mm-commits, linux-mm
On Sat, 29 Jan 2022 04:25:33 +0000 Matthew Wilcox <willy@infradead.org> wrote:
> On Fri, Jan 28, 2022 at 06:13:41PM -0800, Andrew Morton wrote:
> > 12 patches, based on 169387e2aa291a4e3cb856053730fe99d6cec06f.
> ^^
>
> I see 7?
Crap, sorry, ignore all this, shall redo tomorrow.
(It wasn't a good day over here. The thing with disk drives is that
the bigger they are, the harder they fall).
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2022-01-29 2:13 incoming Andrew Morton
@ 2022-01-29 4:25 ` Matthew Wilcox
2022-01-29 6:23 ` incoming Andrew Morton
0 siblings, 1 reply; 409+ messages in thread
From: Matthew Wilcox @ 2022-01-29 4:25 UTC (permalink / raw)
To: Andrew Morton; +Cc: Linus Torvalds, mm-commits, linux-mm
On Fri, Jan 28, 2022 at 06:13:41PM -0800, Andrew Morton wrote:
> 12 patches, based on 169387e2aa291a4e3cb856053730fe99d6cec06f.
^^
I see 7?
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2022-01-29 2:13 Andrew Morton
2022-01-29 4:25 ` incoming Matthew Wilcox
0 siblings, 1 reply; 409+ messages in thread
From: Andrew Morton @ 2022-01-29 2:13 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
12 patches, based on 169387e2aa291a4e3cb856053730fe99d6cec06f.
Subsystems affected by this patch series:
sysctl
binfmt
ia64
mm/memory-failure
mm/folios
selftests
mm/kasan
mm/psi
ocfs2
Subsystem: sysctl
Andrew Morton <akpm@linux-foundation.org>:
include/linux/sysctl.h: fix register_sysctl_mount_point() return type
Subsystem: binfmt
Tong Zhang <ztong0001@gmail.com>:
binfmt_misc: fix crash when load/unload module
Subsystem: ia64
Randy Dunlap <rdunlap@infradead.org>:
ia64: make IA64_MCA_RECOVERY bool instead of tristate
Subsystem: mm/memory-failure
Joao Martins <joao.m.martins@oracle.com>:
memory-failure: fetch compound_head after pgmap_pfn_valid()
Subsystem: mm/folios
Wei Yang <richard.weiyang@gmail.com>:
mm: page->mapping folio->mapping should have the same offset
Subsystem: selftests
Maor Gottlieb <maorg@nvidia.com>:
tools/testing/scatterlist: add missing defines
Subsystem: mm/kasan
Marco Elver <elver@google.com>:
kasan: test: fix compatibility with FORTIFY_SOURCE
Peter Collingbourne <pcc@google.com>:
mm, kasan: use compare-exchange operation to set KASAN page tag
Subsystem: mm/psi
Suren Baghdasaryan <surenb@google.com>:
psi: fix "no previous prototype" warnings when CONFIG_CGROUPS=n
psi: fix "defined but not used" warnings when CONFIG_PROC_FS=n
Subsystem: ocfs2
Joseph Qi <joseph.qi@linux.alibaba.com>:
Patch series "ocfs2: fix a deadlock case":
jbd2: export jbd2_journal_[grab|put]_journal_head
ocfs2: fix a deadlock when commit trans
arch/ia64/Kconfig | 2
fs/binfmt_misc.c | 8 +--
fs/jbd2/journal.c | 2
fs/ocfs2/suballoc.c | 25 ++++-------
include/linux/mm.h | 17 +++++--
include/linux/mm_types.h | 1
include/linux/psi.h | 11 ++--
include/linux/sysctl.h | 2
kernel/sched/psi.c | 79 ++++++++++++++++++-----------------
lib/test_kasan.c | 5 ++
mm/memory-failure.c | 6 ++
tools/testing/scatterlist/linux/mm.h | 3 -
12 files changed, 91 insertions(+), 70 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2022-01-22 6:10 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2022-01-22 6:10 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
This is the post-linux-next queue. Material which was based on or
dependent upon material which was in -next.
69 patches, based on 9b57f458985742bd1c585f4c7f36d04634ce1143.
Subsystems affected by this patch series:
mm/migration
sysctl
mm/zsmalloc
proc
lib
Subsystem: mm/migration
Alistair Popple <apopple@nvidia.com>:
mm/migrate.c: rework migration_entry_wait() to not take a pageref
Subsystem: sysctl
Xiaoming Ni <nixiaoming@huawei.com>:
Patch series "sysctl: first set of kernel/sysctl cleanups", v2:
sysctl: add a new register_sysctl_init() interface
sysctl: move some boundary constants from sysctl.c to sysctl_vals
hung_task: move hung_task sysctl interface to hung_task.c
watchdog: move watchdog sysctl interface to watchdog.c
Stephen Kitt <steve@sk2.org>:
sysctl: make ngroups_max const
Xiaoming Ni <nixiaoming@huawei.com>:
sysctl: use const for typically used max/min proc sysctls
sysctl: use SYSCTL_ZERO to replace some static int zero uses
aio: move aio sysctl to aio.c
dnotify: move dnotify sysctl to dnotify.c
Luis Chamberlain <mcgrof@kernel.org>:
Patch series "sysctl: second set of kernel/sysctl cleanups", v2:
hpet: simplify subdirectory registration with register_sysctl()
i915: simplify subdirectory registration with register_sysctl()
macintosh/mac_hid.c: simplify subdirectory registration with register_sysctl()
ocfs2: simplify subdirectory registration with register_sysctl()
test_sysctl: simplify subdirectory registration with register_sysctl()
Xiaoming Ni <nixiaoming@huawei.com>:
inotify: simplify subdirectory registration with register_sysctl()
Luis Chamberlain <mcgrof@kernel.org>:
cdrom: simplify subdirectory registration with register_sysctl()
Xiaoming Ni <nixiaoming@huawei.com>:
eventpoll: simplify sysctl declaration with register_sysctl()
Patch series "sysctl: 3rd set of kernel/sysctl cleanups", v2:
firmware_loader: move firmware sysctl to its own files
random: move the random sysctl declarations to its own file
Luis Chamberlain <mcgrof@kernel.org>:
sysctl: add helper to register a sysctl mount point
fs: move binfmt_misc sysctl to its own file
Xiaoming Ni <nixiaoming@huawei.com>:
printk: move printk sysctl to printk/sysctl.c
scsi/sg: move sg-big-buff sysctl to scsi/sg.c
stackleak: move stack_erasing sysctl to stackleak.c
Luis Chamberlain <mcgrof@kernel.org>:
sysctl: share unsigned long const values
Patch series "sysctl: 4th set of kernel/sysctl cleanups":
fs: move inode sysctls to its own file
fs: move fs stat sysctls to file_table.c
fs: move dcache sysctls to its own file
sysctl: move maxolduid as a sysctl specific const
fs: move shared sysctls to fs/sysctls.c
fs: move locking sysctls where they are used
fs: move namei sysctls to its own file
fs: move fs/exec.c sysctls into its own file
fs: move pipe sysctls to is own file
Patch series "sysctl: add and use base directory declarer and registration helper":
sysctl: add and use base directory declarer and registration helper
fs: move namespace sysctls and declare fs base directory
kernel/sysctl.c: rename sysctl_init() to sysctl_init_bases()
Xiaoming Ni <nixiaoming@huawei.com>:
printk: fix build warning when CONFIG_PRINTK=n
fs/coredump: move coredump sysctls into its own file
kprobe: move sysctl_kprobes_optimization to kprobes.c
Colin Ian King <colin.i.king@gmail.com>:
kernel/sysctl.c: remove unused variable ten_thousand
Baokun Li <libaokun1@huawei.com>:
sysctl: returns -EINVAL when a negative value is passed to proc_doulongvec_minmax
Subsystem: mm/zsmalloc
Minchan Kim <minchan@kernel.org>:
Patch series "zsmalloc: remove bit_spin_lock", v2:
zsmalloc: introduce some helper functions
zsmalloc: rename zs_stat_type to class_stat_type
zsmalloc: decouple class actions from zspage works
zsmalloc: introduce obj_allocated
zsmalloc: move huge compressed obj from page to zspage
zsmalloc: remove zspage isolation for migration
locking/rwlocks: introduce write_lock_nested
zsmalloc: replace per zpage lock with pool->migrate_lock
Mike Galbraith <umgwanakikbuti@gmail.com>:
zsmalloc: replace get_cpu_var with local_lock
Subsystem: proc
Muchun Song <songmuchun@bytedance.com>:
fs: proc: store PDE()->data into inode->i_private
proc: remove PDE_DATA() completely
Subsystem: lib
Vlastimil Babka <vbabka@suse.cz>:
lib/stackdepot: allow optional init and stack_table allocation by kvmalloc()
lib/stackdepot: fix spelling mistake and grammar in pr_err message
lib/stackdepot: allow optional init and stack_table allocation by kvmalloc() - fixup
lib/stackdepot: allow optional init and stack_table allocation by kvmalloc() - fixup3
lib/stackdepot: allow optional init and stack_table allocation by kvmalloc() - fixup4
Marco Elver <elver@google.com>:
lib/stackdepot: always do filter_irq_stacks() in stack_depot_save()
Christoph Hellwig <hch@lst.de>:
Patch series "remove Xen tmem leftovers":
mm: remove cleancache
frontswap: remove frontswap_writethrough
frontswap: remove frontswap_tmem_exclusive_gets
frontswap: remove frontswap_shrink
frontswap: remove frontswap_curr_pages
frontswap: simplify frontswap_init
frontswap: remove the frontswap exports
mm: simplify try_to_unuse
frontswap: remove frontswap_test
frontswap: simplify frontswap_register_ops
mm: mark swap_lock and swap_active_head static
frontswap: remove support for multiple ops
mm: hide the FRONTSWAP Kconfig symbol
Documentation/vm/cleancache.rst | 296 ------
Documentation/vm/frontswap.rst | 31
Documentation/vm/index.rst | 1
MAINTAINERS | 7
arch/alpha/kernel/srm_env.c | 4
arch/arm/configs/bcm2835_defconfig | 1
arch/arm/configs/qcom_defconfig | 1
arch/arm/kernel/atags_proc.c | 2
arch/arm/mm/alignment.c | 2
arch/ia64/kernel/salinfo.c | 10
arch/m68k/configs/amiga_defconfig | 1
arch/m68k/configs/apollo_defconfig | 1
arch/m68k/configs/atari_defconfig | 1
arch/m68k/configs/bvme6000_defconfig | 1
arch/m68k/configs/hp300_defconfig | 1
arch/m68k/configs/mac_defconfig | 1
arch/m68k/configs/multi_defconfig | 1
arch/m68k/configs/mvme147_defconfig | 1
arch/m68k/configs/mvme16x_defconfig | 1
arch/m68k/configs/q40_defconfig | 1
arch/m68k/configs/sun3_defconfig | 1
arch/m68k/configs/sun3x_defconfig | 1
arch/powerpc/kernel/proc_powerpc.c | 4
arch/s390/configs/debug_defconfig | 1
arch/s390/configs/defconfig | 1
arch/sh/mm/alignment.c | 4
arch/xtensa/platforms/iss/simdisk.c | 4
block/bdev.c | 5
drivers/acpi/proc.c | 2
drivers/base/firmware_loader/fallback.c | 7
drivers/base/firmware_loader/fallback.h | 11
drivers/base/firmware_loader/fallback_table.c | 25
drivers/cdrom/cdrom.c | 23
drivers/char/hpet.c | 22
drivers/char/random.c | 14
drivers/gpu/drm/drm_dp_mst_topology.c | 1
drivers/gpu/drm/drm_mm.c | 4
drivers/gpu/drm/drm_modeset_lock.c | 9
drivers/gpu/drm/i915/i915_perf.c | 22
drivers/gpu/drm/i915/intel_runtime_pm.c | 3
drivers/hwmon/dell-smm-hwmon.c | 4
drivers/macintosh/mac_hid.c | 24
drivers/net/bonding/bond_procfs.c | 8
drivers/net/wireless/cisco/airo.c | 22
drivers/net/wireless/intersil/hostap/hostap_ap.c | 16
drivers/net/wireless/intersil/hostap/hostap_download.c | 2
drivers/net/wireless/intersil/hostap/hostap_proc.c | 24
drivers/net/wireless/ray_cs.c | 2
drivers/nubus/proc.c | 36
drivers/parisc/led.c | 4
drivers/pci/proc.c | 10
drivers/platform/x86/thinkpad_acpi.c | 4
drivers/platform/x86/toshiba_acpi.c | 16
drivers/pnp/isapnp/proc.c | 2
drivers/pnp/pnpbios/proc.c | 4
drivers/scsi/scsi_proc.c | 4
drivers/scsi/sg.c | 35
drivers/usb/gadget/function/rndis.c | 4
drivers/zorro/proc.c | 2
fs/Makefile | 4
fs/afs/proc.c | 6
fs/aio.c | 31
fs/binfmt_misc.c | 6
fs/btrfs/extent_io.c | 10
fs/btrfs/super.c | 2
fs/coredump.c | 66 +
fs/dcache.c | 37
fs/eventpoll.c | 10
fs/exec.c | 145 +--
fs/ext4/mballoc.c | 14
fs/ext4/readpage.c | 6
fs/ext4/super.c | 3
fs/f2fs/data.c | 13
fs/file_table.c | 47 -
fs/inode.c | 39
fs/jbd2/journal.c | 2
fs/locks.c | 34
fs/mpage.c | 7
fs/namei.c | 58 +
fs/namespace.c | 24
fs/notify/dnotify/dnotify.c | 21
fs/notify/fanotify/fanotify_user.c | 10
fs/notify/inotify/inotify_user.c | 11
fs/ntfs3/ntfs_fs.h | 1
fs/ocfs2/stackglue.c | 25
fs/ocfs2/super.c | 2
fs/pipe.c | 64 +
fs/proc/generic.c | 6
fs/proc/inode.c | 1
fs/proc/internal.h | 5
fs/proc/proc_net.c | 8
fs/proc/proc_sysctl.c | 67 +
fs/super.c | 3
fs/sysctls.c | 47 -
include/linux/aio.h | 4
include/linux/cleancache.h | 124 --
include/linux/coredump.h | 10
include/linux/dcache.h | 10
include/linux/dnotify.h | 1
include/linux/fanotify.h | 2
include/linux/frontswap.h | 35
include/linux/fs.h | 18
include/linux/inotify.h | 3
include/linux/kprobes.h | 6
include/linux/migrate.h | 2
include/linux/mount.h | 3
include/linux/pipe_fs_i.h | 4
include/linux/poll.h | 2
include/linux/printk.h | 4
include/linux/proc_fs.h | 17
include/linux/ref_tracker.h | 2
include/linux/rwlock.h | 6
include/linux/rwlock_api_smp.h | 8
include/linux/rwlock_rt.h | 10
include/linux/sched/sysctl.h | 14
include/linux/seq_file.h | 2
include/linux/shmem_fs.h | 3
include/linux/spinlock_api_up.h | 1
include/linux/stackdepot.h | 25
include/linux/stackleak.h | 5
include/linux/swapfile.h | 3
include/linux/sysctl.h | 67 +
include/scsi/sg.h | 4
init/main.c | 9
ipc/util.c | 2
kernel/hung_task.c | 81 +
kernel/irq/proc.c | 8
kernel/kprobes.c | 30
kernel/locking/spinlock.c | 10
kernel/locking/spinlock_rt.c | 12
kernel/printk/Makefile | 5
kernel/printk/internal.h | 8
kernel/printk/printk.c | 4
kernel/printk/sysctl.c | 85 +
kernel/resource.c | 4
kernel/stackleak.c | 26
kernel/sysctl.c | 790 +----------------
kernel/watchdog.c | 101 ++
lib/Kconfig | 4
lib/Kconfig.kasan | 2
lib/stackdepot.c | 46
lib/test_sysctl.c | 22
mm/Kconfig | 40
mm/Makefile | 1
mm/cleancache.c | 315 ------
mm/filemap.c | 102 +-
mm/frontswap.c | 259 -----
mm/kasan/common.c | 1
mm/migrate.c | 38
mm/page_owner.c | 2
mm/shmem.c | 33
mm/swapfile.c | 90 -
mm/truncate.c | 15
mm/zsmalloc.c | 557 ++++-------
mm/zswap.c | 8
net/atm/proc.c | 4
net/bluetooth/af_bluetooth.c | 8
net/can/bcm.c | 2
net/can/proc.c | 2
net/core/neighbour.c | 6
net/core/pktgen.c | 6
net/ipv4/netfilter/ipt_CLUSTERIP.c | 6
net/ipv4/raw.c | 8
net/ipv4/tcp_ipv4.c | 2
net/ipv4/udp.c | 6
net/netfilter/x_tables.c | 10
net/netfilter/xt_hashlimit.c | 18
net/netfilter/xt_recent.c | 4
net/sunrpc/auth_gss/svcauth_gss.c | 4
net/sunrpc/cache.c | 24
net/sunrpc/stats.c | 2
sound/core/info.c | 4
172 files changed, 1877 insertions(+), 2931 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2022-01-20 2:07 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2022-01-20 2:07 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
55 patches, based on df0cc57e057f18e44dac8e6c18aba47ab53202f9 ("Linux 5.16")
Subsystems affected by this patch series:
percpu
procfs
sysctl
misc
core-kernel
get_maintainer
lib
checkpatch
binfmt
nilfs2
hfs
fat
adfs
panic
delayacct
kconfig
kcov
ubsan
Subsystem: percpu
Kefeng Wang <wangkefeng.wang@huawei.com>:
Patch series "mm: percpu: Cleanup percpu first chunk function":
mm: percpu: generalize percpu related config
mm: percpu: add pcpu_fc_cpu_to_node_fn_t typedef
mm: percpu: add generic pcpu_fc_alloc/free funciton
mm: percpu: add generic pcpu_populate_pte() function
Subsystem: procfs
David Hildenbrand <david@redhat.com>:
proc/vmcore: don't fake reading zeroes on surprise vmcore_cb unregistration
Hans de Goede <hdegoede@redhat.com>:
proc: make the proc_create[_data]() stubs static inlines
Qi Zheng <zhengqi.arch@bytedance.com>:
proc: convert the return type of proc_fd_access_allowed() to be boolean
Subsystem: sysctl
Geert Uytterhoeven <geert+renesas@glider.be>:
sysctl: fix duplicate path separator in printed entries
luo penghao <luo.penghao@zte.com.cn>:
sysctl: remove redundant ret assignment
Subsystem: misc
Andy Shevchenko <andriy.shevchenko@linux.intel.com>:
include/linux/unaligned: replace kernel.h with the necessary inclusions
kernel.h: include a note to discourage people from including it in headers
Subsystem: core-kernel
Yafang Shao <laoar.shao@gmail.com>:
Patch series "task comm cleanups", v2:
fs/exec: replace strlcpy with strscpy_pad in __set_task_comm
fs/exec: replace strncpy with strscpy_pad in __get_task_comm
drivers/infiniband: replace open-coded string copy with get_task_comm
fs/binfmt_elf: replace open-coded string copy with get_task_comm
samples/bpf/test_overhead_kprobe_kern: replace bpf_probe_read_kernel with bpf_probe_read_kernel_str to get task comm
tools/bpf/bpftool/skeleton: replace bpf_probe_read_kernel with bpf_probe_read_kernel_str to get task comm
tools/testing/selftests/bpf: replace open-coded 16 with TASK_COMM_LEN
kthread: dynamically allocate memory to store kthread's full name
Davidlohr Bueso <dave@stgolabs.net>:
kernel/sys.c: only take tasklist_lock for get/setpriority(PRIO_PGRP)
Subsystem: get_maintainer
Randy Dunlap <rdunlap@infradead.org>:
get_maintainer: don't remind about no git repo when --nogit is used
Subsystem: lib
Alexey Dobriyan <adobriyan@gmail.com>:
kstrtox: uninline everything
Andy Shevchenko <andriy.shevchenko@linux.intel.com>:
list: introduce list_is_head() helper and re-use it in list.h
Zhen Lei <thunder.leizhen@huawei.com>:
lib/list_debug.c: print more list debugging context in __list_del_entry_valid()
Isabella Basso <isabbasso@riseup.net>:
Patch series "test_hash.c: refactor into KUnit", v3:
hash.h: remove unused define directive
test_hash.c: split test_int_hash into arch-specific functions
test_hash.c: split test_hash_init
lib/Kconfig.debug: properly split hash test kernel entries
test_hash.c: refactor into kunit
Andy Shevchenko <andriy.shevchenko@linux.intel.com>:
kunit: replace kernel.h with the necessary inclusions
uuid: discourage people from using UAPI header in new code
uuid: remove licence boilerplate text from the header
Andrey Konovalov <andreyknvl@google.com>:
lib/test_meminit: destroy cache in kmem_cache_alloc_bulk() test
Subsystem: checkpatch
Jerome Forissier <jerome@forissier.org>:
checkpatch: relax regexp for COMMIT_LOG_LONG_LINE
Joe Perches <joe@perches.com>:
checkpatch: improve Kconfig help test
Rikard Falkeborn <rikard.falkeborn@gmail.com>:
const_structs.checkpatch: add frequently used ops structs
Subsystem: binfmt
"H.J. Lu" <hjl.tools@gmail.com>:
fs/binfmt_elf: use PT_LOAD p_align values for static PIE
Subsystem: nilfs2
Colin Ian King <colin.i.king@gmail.com>:
nilfs2: remove redundant pointer sbufs
Subsystem: hfs
Kees Cook <keescook@chromium.org>:
hfsplus: use struct_group_attr() for memcpy() region
Subsystem: fat
"NeilBrown" <neilb@suse.de>:
FAT: use io_schedule_timeout() instead of congestion_wait()
Subsystem: adfs
Minghao Chi <chi.minghao@zte.com.cn>:
fs/adfs: remove unneeded variable make code cleaner
Subsystem: panic
Marco Elver <elver@google.com>:
panic: use error_report_end tracepoint on warnings
Sebastian Andrzej Siewior <bigeasy@linutronix.de>:
panic: remove oops_id
Subsystem: delayacct
Yang Yang <yang.yang29@zte.com.cn>:
delayacct: support swapin delay accounting for swapping without blkio
delayacct: fix incomplete disable operation when switch enable to disable
delayacct: cleanup flags in struct task_delay_info and functions use it
wangyong <wang.yong12@zte.com.cn>:
Documentation/accounting/delay-accounting.rst: add thrashing page cache and direct compact
delayacct: track delays from memory compact
Subsystem: kconfig
Qian Cai <quic_qiancai@quicinc.com>:
configs: introduce debug.config for CI-like setup
Nathan Chancellor <nathan@kernel.org>:
Patch series "Fix CONFIG_TEST_KMOD with 256kB page size":
arch/Kconfig: split PAGE_SIZE_LESS_THAN_256KB from PAGE_SIZE_LESS_THAN_64KB
btrfs: use generic Kconfig option for 256kB page size limit
lib/Kconfig.debug: make TEST_KMOD depend on PAGE_SIZE_LESS_THAN_256KB
Subsystem: kcov
Marco Elver <elver@google.com>:
kcov: fix generic Kconfig dependencies if ARCH_WANTS_NO_INSTR
Subsystem: ubsan
Kees Cook <keescook@chromium.org>:
ubsan: remove CONFIG_UBSAN_OBJECT_SIZE
Colin Ian King <colin.i.king@gmail.com>:
lib: remove redundant assignment to variable ret
Documentation/accounting/delay-accounting.rst | 63 +-
arch/Kconfig | 4
arch/arm64/Kconfig | 20
arch/ia64/Kconfig | 9
arch/mips/Kconfig | 10
arch/mips/mm/init.c | 28 -
arch/powerpc/Kconfig | 17
arch/powerpc/kernel/setup_64.c | 113 ----
arch/riscv/Kconfig | 10
arch/sparc/Kconfig | 12
arch/sparc/kernel/led.c | 8
arch/sparc/kernel/smp_64.c | 119 -----
arch/x86/Kconfig | 19
arch/x86/kernel/setup_percpu.c | 82 ---
drivers/base/arch_numa.c | 78 ---
drivers/infiniband/hw/qib/qib.h | 2
drivers/infiniband/hw/qib/qib_file_ops.c | 2
drivers/infiniband/sw/rxe/rxe_qp.c | 3
drivers/net/wireless/broadcom/brcm80211/brcmfmac/xtlv.c | 2
fs/adfs/inode.c | 4
fs/binfmt_elf.c | 6
fs/btrfs/Kconfig | 3
fs/exec.c | 5
fs/fat/file.c | 5
fs/hfsplus/hfsplus_raw.h | 12
fs/hfsplus/xattr.c | 4
fs/nilfs2/page.c | 4
fs/proc/array.c | 3
fs/proc/base.c | 4
fs/proc/proc_sysctl.c | 9
fs/proc/vmcore.c | 10
include/kunit/assert.h | 2
include/linux/delayacct.h | 107 ++--
include/linux/elfcore-compat.h | 5
include/linux/elfcore.h | 5
include/linux/hash.h | 5
include/linux/kernel.h | 9
include/linux/kthread.h | 1
include/linux/list.h | 36 -
include/linux/percpu.h | 21
include/linux/proc_fs.h | 12
include/linux/sched.h | 9
include/linux/unaligned/packed_struct.h | 2
include/trace/events/error_report.h | 8
include/uapi/linux/taskstats.h | 6
include/uapi/linux/uuid.h | 10
kernel/configs/debug.config | 105 ++++
kernel/delayacct.c | 49 +-
kernel/kthread.c | 32 +
kernel/panic.c | 21
kernel/sys.c | 16
lib/Kconfig.debug | 45 +
lib/Kconfig.ubsan | 13
lib/Makefile | 5
lib/asn1_encoder.c | 2
lib/kstrtox.c | 12
lib/list_debug.c | 8
lib/lz4/lz4defs.h | 2
lib/test_hash.c | 375 +++++++---------
lib/test_meminit.c | 1
lib/test_ubsan.c | 22
mm/Kconfig | 12
mm/memory.c | 4
mm/page_alloc.c | 3
mm/page_io.c | 3
mm/percpu.c | 168 +++++--
samples/bpf/offwaketime_kern.c | 4
samples/bpf/test_overhead_kprobe_kern.c | 11
samples/bpf/test_overhead_tp_kern.c | 5
scripts/Makefile.ubsan | 1
scripts/checkpatch.pl | 54 +-
scripts/const_structs.checkpatch | 23
scripts/get_maintainer.pl | 2
tools/accounting/getdelays.c | 8
tools/bpf/bpftool/skeleton/pid_iter.bpf.c | 4
tools/include/linux/hash.h | 5
tools/testing/selftests/bpf/progs/test_stacktrace_map.c | 6
tools/testing/selftests/bpf/progs/test_tracepoint.c | 6
78 files changed, 943 insertions(+), 992 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2022-01-14 22:02 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2022-01-14 22:02 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
146 patches, based on df0cc57e057f18e44dac8e6c18aba47ab53202f9 ("Linux 5.16")
Subsystems affected by this patch series:
kthread
ia64
scripts
ntfs
squashfs
ocfs2
vfs
mm/slab-generic
mm/slab
mm/kmemleak
mm/dax
mm/kasan
mm/debug
mm/pagecache
mm/gup
mm/shmem
mm/frontswap
mm/memremap
mm/memcg
mm/selftests
mm/pagemap
mm/dma
mm/vmalloc
mm/memory-failure
mm/hugetlb
mm/userfaultfd
mm/vmscan
mm/mempolicy
mm/oom-kill
mm/hugetlbfs
mm/migration
mm/thp
mm/ksm
mm/page-poison
mm/percpu
mm/rmap
mm/zswap
mm/zram
mm/cleanups
mm/hmm
mm/damon
Subsystem: kthread
Cai Huoqing <caihuoqing@baidu.com>:
kthread: add the helper function kthread_run_on_cpu()
RDMA/siw: make use of the helper function kthread_run_on_cpu()
ring-buffer: make use of the helper function kthread_run_on_cpu()
rcutorture: make use of the helper function kthread_run_on_cpu()
trace/osnoise: make use of the helper function kthread_run_on_cpu()
trace/hwlat: make use of the helper function kthread_run_on_cpu()
Subsystem: ia64
Yang Guang <yang.guang5@zte.com.cn>:
ia64: module: use swap() to make code cleaner
arch/ia64/kernel/setup.c: use swap() to make code cleaner
Jason Wang <wangborong@cdjrlc.com>:
ia64: fix typo in a comment
Greg Kroah-Hartman <gregkh@linuxfoundation.org>:
ia64: topology: use default_groups in kobj_type
Subsystem: scripts
Drew Fustini <dfustini@baylibre.com>:
scripts/spelling.txt: add "oveflow"
Subsystem: ntfs
Yang Li <yang.lee@linux.alibaba.com>:
fs/ntfs/attrib.c: fix one kernel-doc comment
Subsystem: squashfs
Zheng Liang <zhengliang6@huawei.com>:
squashfs: provide backing_dev_info in order to disable read-ahead
Subsystem: ocfs2
Zhang Mingyu <zhang.mingyu@zte.com.cn>:
ocfs2: use BUG_ON instead of if condition followed by BUG.
Joseph Qi <joseph.qi@linux.alibaba.com>:
ocfs2: clearly handle ocfs2_grab_pages_for_write() return value
Greg Kroah-Hartman <gregkh@linuxfoundation.org>:
ocfs2: use default_groups in kobj_type
Colin Ian King <colin.i.king@gmail.com>:
ocfs2: remove redundant assignment to pointer root_bh
Greg Kroah-Hartman <gregkh@linuxfoundation.org>:
ocfs2: cluster: use default_groups in kobj_type
Colin Ian King <colin.i.king@gmail.com>:
ocfs2: remove redundant assignment to variable free_space
Subsystem: vfs
Amit Daniel Kachhap <amit.kachhap@arm.com>:
fs/ioctl: remove unnecessary __user annotation
Subsystem: mm/slab-generic
Marco Elver <elver@google.com>:
mm/slab_common: use WARN() if cache still has objects on destroy
Subsystem: mm/slab
Muchun Song <songmuchun@bytedance.com>:
mm: slab: make slab iterator functions static
Subsystem: mm/kmemleak
Kuan-Ying Lee <Kuan-Ying.Lee@mediatek.com>:
kmemleak: fix kmemleak false positive report with HW tag-based kasan enable
Calvin Zhang <calvinzhang.cool@gmail.com>:
mm: kmemleak: alloc gray object for reserved region with direct map
Kefeng Wang <wangkefeng.wang@huawei.com>:
mm: defer kmemleak object creation of module_alloc()
Subsystem: mm/dax
Joao Martins <joao.m.martins@oracle.com>:
Patch series "mm, device-dax: Introduce compound pages in devmap", v7:
mm/page_alloc: split prep_compound_page into head and tail subparts
mm/page_alloc: refactor memmap_init_zone_device() page init
mm/memremap: add ZONE_DEVICE support for compound pages
device-dax: use ALIGN() for determining pgoff
device-dax: use struct_size()
device-dax: ensure dev_dax->pgmap is valid for dynamic devices
device-dax: factor out page mapping initialization
device-dax: set mapping prior to vmf_insert_pfn{,_pmd,pud}()
device-dax: remove pfn from __dev_dax_{pte,pmd,pud}_fault()
device-dax: compound devmap support
Subsystem: mm/kasan
Marco Elver <elver@google.com>:
kasan: test: add globals left-out-of-bounds test
kasan: add ability to detect double-kmem_cache_destroy()
kasan: test: add test case for double-kmem_cache_destroy()
Andrey Konovalov <andreyknvl@google.com>:
kasan: fix quarantine conflicting with init_on_free
Subsystem: mm/debug
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
mm,fs: split dump_mapping() out from dump_page()
Anshuman Khandual <anshuman.khandual@arm.com>:
mm/debug_vm_pgtable: update comments regarding migration swap entries
Subsystem: mm/pagecache
chiminghao <chi.minghao@zte.com.cn>:
mm/truncate.c: remove unneeded variable
Subsystem: mm/gup
Christophe Leroy <christophe.leroy@csgroup.eu>:
gup: avoid multiple user access locking/unlocking in fault_in_{read/write}able
Li Xinhai <lixinhai.lxh@gmail.com>:
mm/gup.c: stricter check on THP migration entry during follow_pmd_mask
Subsystem: mm/shmem
Yang Shi <shy828301@gmail.com>:
mm: shmem: don't truncate page if memory failure happens
Gang Li <ligang.bdlg@bytedance.com>:
shmem: fix a race between shmem_unused_huge_shrink and shmem_evict_inode
Subsystem: mm/frontswap
Christophe JAILLET <christophe.jaillet@wanadoo.fr>:
mm/frontswap.c: use non-atomic '__set_bit()' when possible
Subsystem: mm/memremap
Subsystem: mm/memcg
Muchun Song <songmuchun@bytedance.com>:
mm: memcontrol: make cgroup_memory_nokmem static
Donghai Qiao <dqiao@redhat.com>:
mm/page_counter: remove an incorrect call to propagate_protected_usage()
Dan Schatzberg <schatzberg.dan@gmail.com>:
mm/memcg: add oom_group_kill memory event
Shakeel Butt <shakeelb@google.com>:
memcg: better bounds on the memcg stats updates
Wang Weiyang <wangweiyang2@huawei.com>:
mm/memcg: use struct_size() helper in kzalloc()
Shakeel Butt <shakeelb@google.com>:
memcg: add per-memcg vmalloc stat
Subsystem: mm/selftests
chiminghao <chi.minghao@zte.com.cn>:
tools/testing/selftests/vm/userfaultfd.c: use swap() to make code cleaner
Subsystem: mm/pagemap
Qi Zheng <zhengqi.arch@bytedance.com>:
mm: remove redundant check about FAULT_FLAG_ALLOW_RETRY bit
Colin Cross <ccross@google.com>:
Patch series "mm: rearrange madvise code to allow for reuse", v11:
mm: rearrange madvise code to allow for reuse
mm: add a field to store names for private anonymous memory
Suren Baghdasaryan <surenb@google.com>:
mm: add anonymous vma name refcounting
Arnd Bergmann <arnd@arndb.de>:
mm: move anon_vma declarations to linux/mm_inline.h
mm: move tlb_flush_pending inline helpers to mm_inline.h
Suren Baghdasaryan <surenb@google.com>:
mm: protect free_pgtables with mmap_lock write lock in exit_mmap
mm: document locking restrictions for vm_operations_struct::close
mm/oom_kill: allow process_mrelease to run under mmap_lock protection
Shuah Khan <skhan@linuxfoundation.org>:
docs/vm: add vmalloced-kernel-stacks document
Pasha Tatashin <pasha.tatashin@soleen.com>:
Patch series "page table check", v3:
mm: change page type prior to adding page table entry
mm: ptep_clear() page table helper
mm: page table check
x86: mm: add x86_64 support for page table check
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
mm: remove last argument of reuse_swap_page()
mm: remove the total_mapcount argument from page_trans_huge_map_swapcount()
mm: remove the total_mapcount argument from page_trans_huge_mapcount()
Subsystem: mm/dma
Christian König <christian.koenig@amd.com>:
mm/dmapool.c: revert "make dma pool to use kmalloc_node"
Subsystem: mm/vmalloc
Michal Hocko <mhocko@suse.com>:
Patch series "extend vmalloc support for constrained allocations", v2:
mm/vmalloc: alloc GFP_NO{FS,IO} for vmalloc
mm/vmalloc: add support for __GFP_NOFAIL
mm/vmalloc: be more explicit about supported gfp flags.
mm: allow !GFP_KERNEL allocations for kvmalloc
mm: make slab and vmalloc allocators __GFP_NOLOCKDEP aware
"NeilBrown" <neilb@suse.de>:
mm: introduce memalloc_retry_wait()
Suren Baghdasaryan <surenb@google.com>:
mm/pagealloc: sysctl: change watermark_scale_factor max limit to 30%
Changcheng Deng <deng.changcheng@zte.com.cn>:
mm: fix boolreturn.cocci warning
Xiongwei Song <sxwjean@gmail.com>:
mm: page_alloc: fix building error on -Werror=array-compare
Michal Hocko <mhocko@suse.com>:
mm: drop node from alloc_pages_vma
Miles Chen <miles.chen@mediatek.com>:
include/linux/gfp.h: further document GFP_DMA32
Anshuman Khandual <anshuman.khandual@arm.com>:
mm/page_alloc.c: modify the comment section for alloc_contig_pages()
Baoquan He <bhe@redhat.com>:
Patch series "Handle warning of allocation failure on DMA zone w/o managed pages", v4:
mm_zone: add function to check if managed dma zone exists
dma/pool: create dma atomic pool only if dma zone has managed pages
mm/page_alloc.c: do not warn allocation failure on zone DMA if no managed pages
Subsystem: mm/memory-failure
Subsystem: mm/hugetlb
Mina Almasry <almasrymina@google.com>:
hugetlb: add hugetlb.*.numa_stat file
Yosry Ahmed <yosryahmed@google.com>:
mm, hugepages: make memory size variable in hugepage-mremap selftest
Yang Yang <yang.yang29@zte.com.cn>:
mm/vmstat: add events for THP max_ptes_* exceeds
Waiman Long <longman@redhat.com>:
selftests/vm: make charge_reserved_hugetlb.sh work with existing cgroup setting
Subsystem: mm/userfaultfd
Peter Xu <peterx@redhat.com>:
selftests/uffd: allow EINTR/EAGAIN
Mike Kravetz <mike.kravetz@oracle.com>:
userfaultfd/selftests: clean up hugetlb allocation code
Subsystem: mm/vmscan
Gang Li <ligang.bdlg@bytedance.com>:
vmscan: make drop_slab_node static
Chen Wandun <chenwandun@huawei.com>:
mm/page_isolation: unset migratetype directly for non Buddy page
Subsystem: mm/mempolicy
"Aneesh Kumar K.V" <aneesh.kumar@linux.ibm.com>:
Patch series "mm: add new syscall set_mempolicy_home_node", v6:
mm/mempolicy: use policy_node helper with MPOL_PREFERRED_MANY
mm/mempolicy: add set_mempolicy_home_node syscall
mm/mempolicy: wire up syscall set_mempolicy_home_node
Randy Dunlap <rdunlap@infradead.org>:
mm/mempolicy: fix all kernel-doc warnings
Subsystem: mm/oom-kill
Jann Horn <jannh@google.com>:
mm, oom: OOM sysrq should always kill a process
Subsystem: mm/hugetlbfs
Sean Christopherson <seanjc@google.com>:
hugetlbfs: fix off-by-one error in hugetlb_vmdelete_list()
Subsystem: mm/migration
Baolin Wang <baolin.wang@linux.alibaba.com>:
Patch series "Improve the migration stats":
mm: migrate: fix the return value of migrate_pages()
mm: migrate: correct the hugetlb migration stats
mm: compaction: fix the migration stats in trace_mm_compaction_migratepages()
mm: migrate: support multiple target nodes demotion
mm: migrate: add more comments for selecting target node randomly
Huang Ying <ying.huang@intel.com>:
mm/migrate: move node demotion code to near its user
Colin Ian King <colin.i.king@gmail.com>:
mm/migrate: remove redundant variables used in a for-loop
Subsystem: mm/thp
Anshuman Khandual <anshuman.khandual@arm.com>:
mm/thp: drop unused trace events hugepage_[invalidate|splitting]
Subsystem: mm/ksm
Nanyong Sun <sunnanyong@huawei.com>:
mm: ksm: fix use-after-free kasan report in ksm_might_need_to_copy
Subsystem: mm/page-poison
Naoya Horiguchi <naoya.horiguchi@nec.com>:
Patch series "mm/hwpoison: fix unpoison_memory()", v4:
mm/hwpoison: mf_mutex for soft offline and unpoison
mm/hwpoison: remove MF_MSG_BUDDY_2ND and MF_MSG_POISONED_HUGE
mm/hwpoison: fix unpoison_memory()
Subsystem: mm/percpu
Qi Zheng <zhengqi.arch@bytedance.com>:
mm: memcg/percpu: account extra objcg space to memory cgroups
Subsystem: mm/rmap
Huang Ying <ying.huang@intel.com>:
mm/rmap: fix potential batched TLB flush race
Subsystem: mm/zswap
Zhaoyu Liu <zackary.liu.pro@gmail.com>:
zpool: remove the list of pools_head
Subsystem: mm/zram
Luis Chamberlain <mcgrof@kernel.org>:
zram: use ATTRIBUTE_GROUPS
Subsystem: mm/cleanups
Quanfa Fu <fuqf0919@gmail.com>:
mm: fix some comment errors
Ting Liu <liuting.0x7c00@bytedance.com>:
mm: make some vars and functions static or __init
Subsystem: mm/hmm
Alistair Popple <apopple@nvidia.com>:
mm/hmm.c: allow VM_MIXEDMAP to work with hmm_range_fault
Subsystem: mm/damon
Xin Hao <xhao@linux.alibaba.com>:
Patch series "mm/damon: Do some small changes", v4:
mm/damon: unified access_check function naming rules
mm/damon: add 'age' of region tracepoint support
mm/damon/core: use abs() instead of diff_of()
mm/damon: remove some unneeded function definitions in damon.h
Yihao Han <hanyihao@vivo.com>:
mm/damon/vaddr: remove swap_ranges() and replace it with swap()
Xin Hao <xhao@linux.alibaba.com>:
mm/damon/schemes: add the validity judgment of thresholds
mm/damon: move damon_rand() definition into damon.h
mm/damon: modify damon_rand() macro to static inline function
SeongJae Park <sj@kernel.org>:
Patch series "mm/damon: Misc cleanups":
mm/damon: convert macro functions to static inline functions
Docs/admin-guide/mm/damon/usage: update for scheme quotas and watermarks
Docs/admin-guide/mm/damon/usage: remove redundant information
Docs/admin-guide/mm/damon/usage: mention tracepoint at the beginning
Docs/admin-guide/mm/damon/usage: update for kdamond_pid and (mk|rm)_contexts
mm/damon: remove a mistakenly added comment for a future feature
Patch series "mm/damon/schemes: Extend stats for better online analysis and tuning":
mm/damon/schemes: account scheme actions that successfully applied
mm/damon/schemes: account how many times quota limit has exceeded
mm/damon/reclaim: provide reclamation statistics
Docs/admin-guide/mm/damon/reclaim: document statistics parameters
mm/damon/dbgfs: support all DAMOS stats
Docs/admin-guide/mm/damon/usage: update for schemes statistics
Baolin Wang <baolin.wang@linux.alibaba.com>:
mm/damon: add access checking for hugetlb pages
Guoqing Jiang <guoqing.jiang@linux.dev>:
mm/damon: move the implementation of damon_insert_region to damon.h
SeongJae Park <sj@kernel.org>:
Patch series "mm/damon: Hide unnecessary information disclosures":
mm/damon/dbgfs: remove an unnecessary variable
mm/damon/vaddr: use pr_debug() for damon_va_three_regions() failure logging
mm/damon/vaddr: hide kernel pointer from damon_va_three_regions() failure log
mm/damon: hide kernel pointer from tracepoint event
Documentation/admin-guide/cgroup-v1/hugetlb.rst | 4
Documentation/admin-guide/cgroup-v2.rst | 11
Documentation/admin-guide/mm/damon/reclaim.rst | 25
Documentation/admin-guide/mm/damon/usage.rst | 235 +++++--
Documentation/admin-guide/mm/numa_memory_policy.rst | 16
Documentation/admin-guide/sysctl/vm.rst | 2
Documentation/filesystems/proc.rst | 6
Documentation/vm/arch_pgtable_helpers.rst | 20
Documentation/vm/index.rst | 2
Documentation/vm/page_migration.rst | 12
Documentation/vm/page_table_check.rst | 56 +
Documentation/vm/vmalloced-kernel-stacks.rst | 153 ++++
MAINTAINERS | 9
arch/Kconfig | 3
arch/alpha/kernel/syscalls/syscall.tbl | 1
arch/alpha/mm/fault.c | 16
arch/arc/mm/fault.c | 3
arch/arm/mm/fault.c | 2
arch/arm/tools/syscall.tbl | 1
arch/arm64/include/asm/unistd.h | 2
arch/arm64/include/asm/unistd32.h | 2
arch/arm64/kernel/module.c | 4
arch/arm64/mm/fault.c | 6
arch/hexagon/mm/vm_fault.c | 8
arch/ia64/kernel/module.c | 6
arch/ia64/kernel/setup.c | 5
arch/ia64/kernel/syscalls/syscall.tbl | 1
arch/ia64/kernel/topology.c | 3
arch/ia64/kernel/uncached.c | 2
arch/ia64/mm/fault.c | 16
arch/m68k/kernel/syscalls/syscall.tbl | 1
arch/m68k/mm/fault.c | 18
arch/microblaze/kernel/syscalls/syscall.tbl | 1
arch/microblaze/mm/fault.c | 18
arch/mips/kernel/syscalls/syscall_n32.tbl | 1
arch/mips/kernel/syscalls/syscall_n64.tbl | 1
arch/mips/kernel/syscalls/syscall_o32.tbl | 1
arch/mips/mm/fault.c | 19
arch/nds32/mm/fault.c | 16
arch/nios2/mm/fault.c | 18
arch/openrisc/mm/fault.c | 18
arch/parisc/kernel/syscalls/syscall.tbl | 1
arch/parisc/mm/fault.c | 18
arch/powerpc/kernel/syscalls/syscall.tbl | 1
arch/powerpc/mm/fault.c | 6
arch/riscv/mm/fault.c | 2
arch/s390/kernel/module.c | 5
arch/s390/kernel/syscalls/syscall.tbl | 1
arch/s390/mm/fault.c | 28
arch/sh/kernel/syscalls/syscall.tbl | 1
arch/sh/mm/fault.c | 18
arch/sparc/kernel/syscalls/syscall.tbl | 1
arch/sparc/mm/fault_32.c | 16
arch/sparc/mm/fault_64.c | 16
arch/um/kernel/trap.c | 8
arch/x86/Kconfig | 1
arch/x86/entry/syscalls/syscall_32.tbl | 1
arch/x86/entry/syscalls/syscall_64.tbl | 1
arch/x86/include/asm/pgtable.h | 31 -
arch/x86/kernel/module.c | 7
arch/x86/mm/fault.c | 3
arch/xtensa/kernel/syscalls/syscall.tbl | 1
arch/xtensa/mm/fault.c | 17
drivers/block/zram/zram_drv.c | 11
drivers/dax/bus.c | 32 +
drivers/dax/bus.h | 1
drivers/dax/device.c | 140 ++--
drivers/infiniband/sw/siw/siw_main.c | 7
drivers/of/fdt.c | 6
fs/ext4/extents.c | 8
fs/ext4/inline.c | 5
fs/ext4/page-io.c | 9
fs/f2fs/data.c | 4
fs/f2fs/gc.c | 5
fs/f2fs/inode.c | 4
fs/f2fs/node.c | 4
fs/f2fs/recovery.c | 6
fs/f2fs/segment.c | 9
fs/f2fs/super.c | 5
fs/hugetlbfs/inode.c | 7
fs/inode.c | 49 +
fs/ioctl.c | 2
fs/ntfs/attrib.c | 2
fs/ocfs2/alloc.c | 2
fs/ocfs2/aops.c | 26
fs/ocfs2/cluster/masklog.c | 11
fs/ocfs2/dir.c | 2
fs/ocfs2/filecheck.c | 3
fs/ocfs2/journal.c | 6
fs/proc/task_mmu.c | 13
fs/squashfs/super.c | 33 +
fs/userfaultfd.c | 8
fs/xfs/kmem.c | 3
fs/xfs/xfs_buf.c | 2
include/linux/ceph/libceph.h | 1
include/linux/damon.h | 93 +--
include/linux/fs.h | 1
include/linux/gfp.h | 12
include/linux/hugetlb.h | 4
include/linux/hugetlb_cgroup.h | 7
include/linux/kasan.h | 4
include/linux/kthread.h | 25
include/linux/memcontrol.h | 22
include/linux/mempolicy.h | 1
include/linux/memremap.h | 11
include/linux/mm.h | 76 --
include/linux/mm_inline.h | 136 ++++
include/linux/mm_types.h | 252 +++-----
include/linux/mmzone.h | 9
include/linux/page-flags.h | 6
include/linux/page_idle.h | 1
include/linux/page_table_check.h | 147 ++++
include/linux/pgtable.h | 8
include/linux/sched/mm.h | 26
include/linux/swap.h | 8
include/linux/syscalls.h | 3
include/linux/vm_event_item.h | 3
include/linux/vmalloc.h | 7
include/ras/ras_event.h | 2
include/trace/events/compaction.h | 24
include/trace/events/damon.h | 15
include/trace/events/thp.h | 35 -
include/uapi/asm-generic/unistd.h | 5
include/uapi/linux/prctl.h | 3
kernel/dma/pool.c | 4
kernel/fork.c | 3
kernel/kthread.c | 1
kernel/rcu/rcutorture.c | 7
kernel/sys.c | 63 ++
kernel/sys_ni.c | 1
kernel/sysctl.c | 3
kernel/trace/ring_buffer.c | 7
kernel/trace/trace_hwlat.c | 6
kernel/trace/trace_osnoise.c | 3
lib/test_hmm.c | 24
lib/test_kasan.c | 30
mm/Kconfig | 14
mm/Kconfig.debug | 24
mm/Makefile | 1
mm/compaction.c | 7
mm/damon/core.c | 45 -
mm/damon/dbgfs.c | 20
mm/damon/paddr.c | 24
mm/damon/prmtv-common.h | 4
mm/damon/reclaim.c | 46 +
mm/damon/vaddr.c | 186 ++++--
mm/debug.c | 52 -
mm/debug_vm_pgtable.c | 6
mm/dmapool.c | 2
mm/frontswap.c | 4
mm/gup.c | 31 -
mm/hmm.c | 5
mm/huge_memory.c | 32 -
mm/hugetlb.c | 6
mm/hugetlb_cgroup.c | 133 +++-
mm/internal.h | 7
mm/kasan/quarantine.c | 11
mm/kasan/shadow.c | 9
mm/khugepaged.c | 23
mm/kmemleak.c | 21
mm/ksm.c | 5
mm/madvise.c | 510 ++++++++++------
mm/mapping_dirty_helpers.c | 1
mm/memcontrol.c | 44 -
mm/memory-failure.c | 189 +++---
mm/memory.c | 12
mm/mempolicy.c | 95 ++-
mm/memremap.c | 18
mm/migrate.c | 527 ++++++++++-------
mm/mlock.c | 2
mm/mmap.c | 55 +
mm/mmu_gather.c | 1
mm/mprotect.c | 2
mm/oom_kill.c | 30
mm/page_alloc.c | 198 ++++--
mm/page_counter.c | 1
mm/page_ext.c | 8
mm/page_isolation.c | 2
mm/page_owner.c | 4
mm/page_table_check.c | 270 ++++++++
mm/percpu-internal.h | 18
mm/percpu.c | 10
mm/pgtable-generic.c | 1
mm/rmap.c | 43 +
mm/shmem.c | 91 ++
mm/slab.h | 5
mm/slab_common.c | 34 -
mm/swap.c | 2
mm/swapfile.c | 46 -
mm/truncate.c | 5
mm/userfaultfd.c | 5
mm/util.c | 15
mm/vmalloc.c | 75 +-
mm/vmscan.c | 2
mm/vmstat.c | 3
mm/zpool.c | 12
net/ceph/buffer.c | 4
net/ceph/ceph_common.c | 27
net/ceph/crypto.c | 2
net/ceph/messenger.c | 2
net/ceph/messenger_v2.c | 2
net/ceph/osdmap.c | 12
net/sunrpc/svc_xprt.c | 3
scripts/spelling.txt | 1
tools/testing/selftests/vm/charge_reserved_hugetlb.sh | 34 -
tools/testing/selftests/vm/hmm-tests.c | 42 +
tools/testing/selftests/vm/hugepage-mremap.c | 46 -
tools/testing/selftests/vm/hugetlb_reparenting_test.sh | 21
tools/testing/selftests/vm/run_vmtests.sh | 2
tools/testing/selftests/vm/userfaultfd.c | 33 -
tools/testing/selftests/vm/write_hugetlb_memory.sh | 2
211 files changed, 3980 insertions(+), 1759 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-12-31 4:12 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-12-31 4:12 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
2 patches, based on 4f3d93c6eaff6b84e43b63e0d7a119c5920e1020.
Subsystems affected by this patch series:
mm/userfaultfd
mm/damon
Subsystem: mm/userfaultfd
Mike Kravetz <mike.kravetz@oracle.com>:
userfaultfd/selftests: fix hugetlb area allocations
Subsystem: mm/damon
SeongJae Park <sj@kernel.org>:
mm/damon/dbgfs: fix 'struct pid' leaks in 'dbgfs_target_ids_write()'
mm/damon/dbgfs.c | 9 +++++++--
tools/testing/selftests/vm/userfaultfd.c | 16 ++++++++++------
2 files changed, 17 insertions(+), 8 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-12-25 5:11 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-12-25 5:11 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
9 patches, based on bc491fb12513e79702c6f936c838f792b5389129.
Subsystems affected by this patch series:
mm/kfence
mm/mempolicy
core-kernel
MAINTAINERS
mm/memory-failure
mm/pagemap
mm/pagealloc
mm/damon
mm/memory-failure
Subsystem: mm/kfence
Baokun Li <libaokun1@huawei.com>:
kfence: fix memory leak when cat kfence objects
Subsystem: mm/mempolicy
Andrey Ryabinin <arbn@yandex-team.com>:
mm: mempolicy: fix THP allocations escaping mempolicy restrictions
Subsystem: core-kernel
Philipp Rudo <prudo@redhat.com>:
kernel/crash_core: suppress unknown crashkernel parameter warning
Subsystem: MAINTAINERS
Randy Dunlap <rdunlap@infradead.org>:
MAINTAINERS: mark more list instances as moderated
Subsystem: mm/memory-failure
Naoya Horiguchi <naoya.horiguchi@nec.com>:
mm, hwpoison: fix condition in free hugetlb page path
Subsystem: mm/pagemap
Hugh Dickins <hughd@google.com>:
mm: delete unsafe BUG from page_cache_add_speculative()
Subsystem: mm/pagealloc
Thibaut Sautereau <thibaut.sautereau@ssi.gouv.fr>:
mm/page_alloc: fix __alloc_size attribute for alloc_pages_exact_nid
Subsystem: mm/damon
SeongJae Park <sj@kernel.org>:
mm/damon/dbgfs: protect targets destructions with kdamond_lock
Subsystem: mm/memory-failure
Liu Shixin <liushixin2@huawei.com>:
mm/hwpoison: clear MF_COUNT_INCREASED before retrying get_any_page()
MAINTAINERS | 4 ++--
include/linux/gfp.h | 2 +-
include/linux/pagemap.h | 1 -
kernel/crash_core.c | 11 +++++++++++
mm/damon/dbgfs.c | 2 ++
mm/kfence/core.c | 1 +
mm/memory-failure.c | 14 +++++---------
mm/mempolicy.c | 3 +--
8 files changed, 23 insertions(+), 15 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-12-10 22:45 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-12-10 22:45 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
21 patches, based on c741e49150dbb0c0aebe234389f4aa8b47958fa8.
Subsystems affected by this patch series:
mm/mlock
MAINTAINERS
mailmap
mm/pagecache
mm/damon
mm/slub
mm/memcg
mm/hugetlb
mm/pagecache
Subsystem: mm/mlock
Drew DeVault <sir@cmpwn.com>:
Increase default MLOCK_LIMIT to 8 MiB
Subsystem: MAINTAINERS
Dave Young <dyoung@redhat.com>:
MAINTAINERS: update kdump maintainers
Subsystem: mailmap
Guo Ren <guoren@linux.alibaba.com>:
mailmap: update email address for Guo Ren
Subsystem: mm/pagecache
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
filemap: remove PageHWPoison check from next_uptodate_page()
Subsystem: mm/damon
SeongJae Park <sj@kernel.org>:
Patch series "mm/damon: Fix fake /proc/loadavg reports", v3:
timers: implement usleep_idle_range()
mm/damon/core: fix fake load reports due to uninterruptible sleeps
Patch series "mm/damon: Trivial fixups and improvements":
mm/damon/core: use better timer mechanisms selection threshold
mm/damon/dbgfs: remove an unnecessary error message
mm/damon/core: remove unnecessary error messages
mm/damon/vaddr: remove an unnecessary warning message
mm/damon/vaddr-test: split a test function having >1024 bytes frame size
mm/damon/vaddr-test: remove unnecessary variables
selftests/damon: skip test if DAMON is running
selftests/damon: test DAMON enabling with empty target_ids case
selftests/damon: test wrong DAMOS condition ranges input
selftests/damon: test debugfs file reads/writes with huge count
selftests/damon: split test cases
Subsystem: mm/slub
Gerald Schaefer <gerald.schaefer@linux.ibm.com>:
mm/slub: fix endianness bug for alloc/free_traces attributes
Subsystem: mm/memcg
Waiman Long <longman@redhat.com>:
mm/memcg: relocate mod_objcg_mlstate(), get_obj_stock() and put_obj_stock()
Subsystem: mm/hugetlb
Zhenguo Yao <yaozhenguo1@gmail.com>:
hugetlbfs: fix issue of preallocation of gigantic pages can't work
Subsystem: mm/pagecache
Manjong Lee <mj0123.lee@samsung.com>:
mm: bdi: initialize bdi_min_ratio when bdi is unregistered
.mailmap | 2
MAINTAINERS | 2
include/linux/delay.h | 14
include/uapi/linux/resource.h | 13
kernel/time/timer.c | 16 -
mm/backing-dev.c | 7
mm/damon/core.c | 20 -
mm/damon/dbgfs.c | 4
mm/damon/vaddr-test.h | 85 ++---
mm/damon/vaddr.c | 1
mm/filemap.c | 2
mm/hugetlb.c | 2
mm/memcontrol.c | 106 +++----
mm/slub.c | 15 -
tools/testing/selftests/damon/.gitignore | 2
tools/testing/selftests/damon/Makefile | 7
tools/testing/selftests/damon/_debugfs_common.sh | 52 +++
tools/testing/selftests/damon/debugfs_attrs.sh | 149 ++--------
tools/testing/selftests/damon/debugfs_empty_targets.sh | 13
tools/testing/selftests/damon/debugfs_huge_count_read_write.sh | 22 +
tools/testing/selftests/damon/debugfs_schemes.sh | 19 +
tools/testing/selftests/damon/debugfs_target_ids.sh | 19 +
tools/testing/selftests/damon/huge_count_read_write.c | 39 ++
23 files changed, 363 insertions(+), 248 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-11-20 0:42 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-11-20 0:42 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
15 patches, based on a90af8f15bdc9449ee2d24e1d73fa3f7e8633f81.
Subsystems affected by this patch series:
mm/swap
ipc
mm/slab-generic
hexagon
mm/kmemleak
mm/hugetlb
mm/kasan
mm/damon
mm/highmem
proc
Subsystem: mm/swap
Matthew Wilcox <willy@infradead.org>:
mm/swap.c:put_pages_list(): reinitialise the page list
Subsystem: ipc
Alexander Mikhalitsyn <alexander.mikhalitsyn@virtuozzo.com>:
Patch series "shm: shm_rmid_forced feature fixes":
ipc: WARN if trying to remove ipc object which is absent
shm: extend forced shm destroy to support objects from several IPC nses
Subsystem: mm/slab-generic
Yunfeng Ye <yeyunfeng@huawei.com>:
mm: emit the "free" trace report before freeing memory in kmem_cache_free()
Subsystem: hexagon
Nathan Chancellor <nathan@kernel.org>:
Patch series "Fixes for ARCH=hexagon allmodconfig", v2:
hexagon: export raw I/O routines for modules
hexagon: clean up timer-regs.h
hexagon: ignore vmlinux.lds
Subsystem: mm/kmemleak
Rustam Kovhaev <rkovhaev@gmail.com>:
mm: kmemleak: slob: respect SLAB_NOLEAKTRACE flag
Subsystem: mm/hugetlb
Bui Quang Minh <minhquangbui99@gmail.com>:
hugetlb: fix hugetlb cgroup refcounting during mremap
Mina Almasry <almasrymina@google.com>:
hugetlb, userfaultfd: fix reservation restore on userfaultfd error
Subsystem: mm/kasan
Kees Cook <keescook@chromium.org>:
kasan: test: silence intentional read overflow warnings
Subsystem: mm/damon
SeongJae Park <sj@kernel.org>:
Patch series "DAMON fixes":
mm/damon/dbgfs: use '__GFP_NOWARN' for user-specified size buffer allocation
mm/damon/dbgfs: fix missed use of damon_dbgfs_lock
Subsystem: mm/highmem
Ard Biesheuvel <ardb@kernel.org>:
kmap_local: don't assume kmap PTEs are linear arrays in memory
Subsystem: proc
David Hildenbrand <david@redhat.com>:
proc/vmcore: fix clearing user buffer by properly using clear_user()
arch/arm/Kconfig | 1
arch/hexagon/include/asm/timer-regs.h | 26 ----
arch/hexagon/include/asm/timex.h | 3
arch/hexagon/kernel/.gitignore | 1
arch/hexagon/kernel/time.c | 12 +-
arch/hexagon/lib/io.c | 4
fs/proc/vmcore.c | 20 ++-
include/linux/hugetlb_cgroup.h | 12 ++
include/linux/ipc_namespace.h | 15 ++
include/linux/sched/task.h | 2
ipc/shm.c | 189 +++++++++++++++++++++++++---------
ipc/util.c | 6 -
lib/test_kasan.c | 2
mm/Kconfig | 3
mm/damon/dbgfs.c | 20 ++-
mm/highmem.c | 32 +++--
mm/hugetlb.c | 11 +
mm/slab.c | 3
mm/slab.h | 2
mm/slob.c | 3
mm/slub.c | 2
mm/swap.c | 1
22 files changed, 254 insertions(+), 116 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-11-11 4:32 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-11-11 4:32 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
The post-linux-next material.
7 patches, based on debe436e77c72fcee804fb867f275e6d31aa999c.
Subsystems affected by this patch series:
mm/debug
mm/slab-generic
mm/migration
mm/memcg
mm/kasan
Subsystem: mm/debug
Yixuan Cao <caoyixuan2019@email.szu.edu.cn>:
mm/page_owner.c: modify the type of argument "order" in some functions
Subsystem: mm/slab-generic
Ingo Molnar <mingo@kernel.org>:
mm: allow only SLUB on PREEMPT_RT
Subsystem: mm/migration
Baolin Wang <baolin.wang@linux.alibaba.com>:
mm: migrate: simplify the file-backed pages validation when migrating its mapping
Alistair Popple <apopple@nvidia.com>:
mm/migrate.c: remove MIGRATE_PFN_LOCKED
Subsystem: mm/memcg
Christoph Hellwig <hch@lst.de>:
Patch series "unexport memcg locking helpers":
mm: unexport folio_memcg_{,un}lock
mm: unexport {,un}lock_page_memcg
Subsystem: mm/kasan
Kuan-Ying Lee <Kuan-Ying.Lee@mediatek.com>:
kasan: add kasan mode messages when kasan init
Documentation/vm/hmm.rst | 2
arch/arm64/mm/kasan_init.c | 2
arch/powerpc/kvm/book3s_hv_uvmem.c | 4
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c | 2
drivers/gpu/drm/nouveau/nouveau_dmem.c | 4
include/linux/migrate.h | 1
include/linux/page_owner.h | 12 +-
init/Kconfig | 2
lib/test_hmm.c | 5 -
mm/kasan/hw_tags.c | 14 ++
mm/kasan/sw_tags.c | 2
mm/memcontrol.c | 4
mm/migrate.c | 151 +++++--------------------------
mm/page_owner.c | 6 -
14 files changed, 61 insertions(+), 150 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-11-09 2:30 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-11-09 2:30 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
87 patches, based on 8bb7eca972ad531c9b149c0a51ab43a417385813, plus
previously sent material.
Subsystems affected by this patch series:
mm/pagecache
mm/hugetlb
procfs
misc
MAINTAINERS
lib
checkpatch
binfmt
kallsyms
ramfs
init
codafs
nilfs2
hfs
crash_dump
signals
seq_file
fork
sysvfs
kcov
gdb
resource
selftests
ipc
Subsystem: mm/pagecache
Johannes Weiner <hannes@cmpxchg.org>:
vfs: keep inodes with page cache off the inode shrinker LRU
Subsystem: mm/hugetlb
zhangyiru <zhangyiru3@huawei.com>:
mm,hugetlb: remove mlock ulimit for SHM_HUGETLB
Subsystem: procfs
Florian Weimer <fweimer@redhat.com>:
procfs: do not list TID 0 in /proc/<pid>/task
David Hildenbrand <david@redhat.com>:
x86/xen: update xen_oldmem_pfn_is_ram() documentation
x86/xen: simplify xen_oldmem_pfn_is_ram()
x86/xen: print a warning when HVMOP_get_mem_type fails
proc/vmcore: let pfn_is_ram() return a bool
proc/vmcore: convert oldmem_pfn_is_ram callback to more generic vmcore callbacks
virtio-mem: factor out hotplug specifics from virtio_mem_init() into virtio_mem_init_hotplug()
virtio-mem: factor out hotplug specifics from virtio_mem_probe() into virtio_mem_init_hotplug()
virtio-mem: factor out hotplug specifics from virtio_mem_remove() into virtio_mem_deinit_hotplug()
virtio-mem: kdump mode to sanitize /proc/vmcore access
Stephen Brennan <stephen.s.brennan@oracle.com>:
proc: allow pid_revalidate() during LOOKUP_RCU
Subsystem: misc
Andy Shevchenko <andriy.shevchenko@linux.intel.com>:
Patch series "kernel.h further split", v5:
kernel.h: drop unneeded <linux/kernel.h> inclusion from other headers
kernel.h: split out container_of() and typeof_member() macros
include/kunit/test.h: replace kernel.h with the necessary inclusions
include/linux/list.h: replace kernel.h with the necessary inclusions
include/linux/llist.h: replace kernel.h with the necessary inclusions
include/linux/plist.h: replace kernel.h with the necessary inclusions
include/media/media-entity.h: replace kernel.h with the necessary inclusions
include/linux/delay.h: replace kernel.h with the necessary inclusions
include/linux/sbitmap.h: replace kernel.h with the necessary inclusions
include/linux/radix-tree.h: replace kernel.h with the necessary inclusions
include/linux/generic-radix-tree.h: replace kernel.h with the necessary inclusions
Stephen Rothwell <sfr@canb.auug.org.au>:
kernel.h: split out instruction pointer accessors
Rasmus Villemoes <linux@rasmusvillemoes.dk>:
linux/container_of.h: switch to static_assert
Colin Ian King <colin.i.king@googlemail.com>:
mailmap: update email address for Colin King
Subsystem: MAINTAINERS
Kees Cook <keescook@chromium.org>:
MAINTAINERS: add "exec & binfmt" section with myself and Eric
Lukas Bulwahn <lukas.bulwahn@gmail.com>:
Patch series "Rectify file references for dt-bindings in MAINTAINERS", v5:
MAINTAINERS: rectify entry for ARM/TOSHIBA VISCONTI ARCHITECTURE
MAINTAINERS: rectify entry for HIKEY960 ONBOARD USB GPIO HUB DRIVER
MAINTAINERS: rectify entry for INTEL KEEM BAY DRM DRIVER
MAINTAINERS: rectify entry for ALLWINNER HARDWARE SPINLOCK SUPPORT
Subsystem: lib
Imran Khan <imran.f.khan@oracle.com>:
Patch series "lib, stackdepot: check stackdepot handle before accessing slabs", v2:
lib, stackdepot: check stackdepot handle before accessing slabs
lib, stackdepot: add helper to print stack entries
lib, stackdepot: add helper to print stack entries into buffer
Lucas De Marchi <lucas.demarchi@intel.com>:
include/linux/string_helpers.h: add linux/string.h for strlen()
Alexey Dobriyan <adobriyan@gmail.com>:
lib: uninline simple_strntoull() as well
Thomas Gleixner <tglx@linutronix.de>:
mm/scatterlist: replace the !preemptible warning in sg_miter_stop()
Subsystem: checkpatch
Rikard Falkeborn <rikard.falkeborn@gmail.com>:
const_structs.checkpatch: add a few sound ops structs
Joe Perches <joe@perches.com>:
checkpatch: improve EXPORT_SYMBOL test for EXPORT_SYMBOL_NS uses
Peter Ujfalusi <peter.ujfalusi@linux.intel.com>:
checkpatch: get default codespell dictionary path from package location
Subsystem: binfmt
Kees Cook <keescook@chromium.org>:
binfmt_elf: reintroduce using MAP_FIXED_NOREPLACE
Alexey Dobriyan <adobriyan@gmail.com>:
ELF: simplify STACK_ALLOC macro
Subsystem: kallsyms
Kefeng Wang <wangkefeng.wang@huawei.com>:
Patch series "sections: Unify kernel sections range check and use", v4:
kallsyms: remove arch specific text and data check
kallsyms: fix address-checks for kernel related range
sections: move and rename core_kernel_data() to is_kernel_core_data()
sections: move is_kernel_inittext() into sections.h
x86: mm: rename __is_kernel_text() to is_x86_32_kernel_text()
sections: provide internal __is_kernel() and __is_kernel_text() helper
mm: kasan: use is_kernel() helper
extable: use is_kernel_text() helper
powerpc/mm: use core_kernel_text() helper
microblaze: use is_kernel_text() helper
alpha: use is_kernel_text() helper
Subsystem: ramfs
yangerkun <yangerkun@huawei.com>:
ramfs: fix mount source show for ramfs
Subsystem: init
Andrew Halaney <ahalaney@redhat.com>:
init: make unknown command line param message clearer
Subsystem: codafs
Jan Harkes <jaharkes@cs.cmu.edu>:
Patch series "Coda updates for -next":
coda: avoid NULL pointer dereference from a bad inode
coda: check for async upcall request using local state
Alex Shi <alex.shi@linux.alibaba.com>:
coda: remove err which no one care
Jan Harkes <jaharkes@cs.cmu.edu>:
coda: avoid flagging NULL inodes
coda: avoid hidden code duplication in rename
coda: avoid doing bad things on inode type changes during revalidation
Xiyu Yang <xiyuyang19@fudan.edu.cn>:
coda: convert from atomic_t to refcount_t on coda_vm_ops->refcnt
Jing Yangyang <jing.yangyang@zte.com.cn>:
coda: use vmemdup_user to replace the open code
Jan Harkes <jaharkes@cs.cmu.edu>:
coda: bump module version to 7.2
Subsystem: nilfs2
Qing Wang <wangqing@vivo.com>:
Patch series "nilfs2 updates":
nilfs2: replace snprintf in show functions with sysfs_emit
Ryusuke Konishi <konishi.ryusuke@gmail.com>:
nilfs2: remove filenames from file comments
Subsystem: hfs
Arnd Bergmann <arnd@arndb.de>:
hfs/hfsplus: use WARN_ON for sanity check
Subsystem: crash_dump
Changcheng Deng <deng.changcheng@zte.com.cn>:
crash_dump: fix boolreturn.cocci warning
Ye Guojin <ye.guojin@zte.com.cn>:
crash_dump: remove duplicate include in crash_dump.h
Subsystem: signals
Ye Guojin <ye.guojin@zte.com.cn>:
signal: remove duplicate include in signal.h
Subsystem: seq_file
Andy Shevchenko <andriy.shevchenko@linux.intel.com>:
seq_file: move seq_escape() to a header
Muchun Song <songmuchun@bytedance.com>:
seq_file: fix passing wrong private data
Subsystem: fork
Ran Xiaokai <ran.xiaokai@zte.com.cn>:
kernel/fork.c: unshare(): use swap() to make code cleaner
Subsystem: sysvfs
Pavel Skripkin <paskripkin@gmail.com>:
sysv: use BUILD_BUG_ON instead of runtime check
Subsystem: kcov
Sebastian Andrzej Siewior <bigeasy@linutronix.de>:
Patch series "kcov: PREEMPT_RT fixup + misc", v2:
Documentation/kcov: include types.h in the example
Documentation/kcov: define `ip' in the example
kcov: allocate per-CPU memory on the relevant node
kcov: avoid enable+disable interrupts if !in_task()
kcov: replace local_irq_save() with a local_lock_t
Subsystem: gdb
Douglas Anderson <dianders@chromium.org>:
scripts/gdb: handle split debug for vmlinux
Subsystem: resource
David Hildenbrand <david@redhat.com>:
Patch series "virtio-mem: disallow mapping virtio-mem memory via /dev/mem", v5:
kernel/resource: clean up and optimize iomem_is_exclusive()
kernel/resource: disallow access to exclusive system RAM regions
virtio-mem: disallow mapping virtio-mem memory via /dev/mem
Subsystem: selftests
SeongJae Park <sjpark@amazon.de>:
selftests/kselftest/runner/run_one(): allow running non-executable files
Subsystem: ipc
Michal Clapinski <mclapinski@google.com>:
ipc: check checkpoint_restore_ns_capable() to modify C/R proc files
Manfred Spraul <manfred@colorfullife.com>:
ipc/ipc_sysctl.c: remove fallback for !CONFIG_PROC_SYSCTL
.mailmap | 2
Documentation/dev-tools/kcov.rst | 5
MAINTAINERS | 21 +
arch/alpha/kernel/traps.c | 4
arch/microblaze/mm/pgtable.c | 3
arch/powerpc/mm/pgtable_32.c | 7
arch/riscv/lib/delay.c | 4
arch/s390/include/asm/facility.h | 4
arch/x86/kernel/aperture_64.c | 13
arch/x86/kernel/unwind_orc.c | 2
arch/x86/mm/init_32.c | 14
arch/x86/xen/mmu_hvm.c | 39 --
drivers/gpu/drm/drm_dp_mst_topology.c | 5
drivers/gpu/drm/drm_mm.c | 5
drivers/gpu/drm/i915/i915_vma.c | 5
drivers/gpu/drm/i915/intel_runtime_pm.c | 20 -
drivers/media/dvb-frontends/cxd2880/cxd2880_common.h | 1
drivers/virtio/Kconfig | 1
drivers/virtio/virtio_mem.c | 321 +++++++++++++------
fs/binfmt_elf.c | 33 +
fs/coda/cnode.c | 13
fs/coda/coda_linux.c | 39 +-
fs/coda/coda_linux.h | 6
fs/coda/dir.c | 20 -
fs/coda/file.c | 12
fs/coda/psdev.c | 14
fs/coda/upcall.c | 3
fs/hfs/inode.c | 6
fs/hfsplus/inode.c | 12
fs/hugetlbfs/inode.c | 23 -
fs/inode.c | 46 +-
fs/internal.h | 1
fs/nilfs2/alloc.c | 2
fs/nilfs2/alloc.h | 2
fs/nilfs2/bmap.c | 2
fs/nilfs2/bmap.h | 2
fs/nilfs2/btnode.c | 2
fs/nilfs2/btnode.h | 2
fs/nilfs2/btree.c | 2
fs/nilfs2/btree.h | 2
fs/nilfs2/cpfile.c | 2
fs/nilfs2/cpfile.h | 2
fs/nilfs2/dat.c | 2
fs/nilfs2/dat.h | 2
fs/nilfs2/dir.c | 2
fs/nilfs2/direct.c | 2
fs/nilfs2/direct.h | 2
fs/nilfs2/file.c | 2
fs/nilfs2/gcinode.c | 2
fs/nilfs2/ifile.c | 2
fs/nilfs2/ifile.h | 2
fs/nilfs2/inode.c | 2
fs/nilfs2/ioctl.c | 2
fs/nilfs2/mdt.c | 2
fs/nilfs2/mdt.h | 2
fs/nilfs2/namei.c | 2
fs/nilfs2/nilfs.h | 2
fs/nilfs2/page.c | 2
fs/nilfs2/page.h | 2
fs/nilfs2/recovery.c | 2
fs/nilfs2/segbuf.c | 2
fs/nilfs2/segbuf.h | 2
fs/nilfs2/segment.c | 2
fs/nilfs2/segment.h | 2
fs/nilfs2/sufile.c | 2
fs/nilfs2/sufile.h | 2
fs/nilfs2/super.c | 2
fs/nilfs2/sysfs.c | 78 ++--
fs/nilfs2/sysfs.h | 2
fs/nilfs2/the_nilfs.c | 2
fs/nilfs2/the_nilfs.h | 2
fs/proc/base.c | 21 -
fs/proc/vmcore.c | 109 ++++--
fs/ramfs/inode.c | 11
fs/seq_file.c | 16
fs/sysv/super.c | 6
include/asm-generic/sections.h | 75 +++-
include/kunit/test.h | 13
include/linux/bottom_half.h | 3
include/linux/container_of.h | 52 ++-
include/linux/crash_dump.h | 30 +
include/linux/delay.h | 2
include/linux/fs.h | 1
include/linux/fwnode.h | 1
include/linux/generic-radix-tree.h | 3
include/linux/hugetlb.h | 6
include/linux/instruction_pointer.h | 8
include/linux/kallsyms.h | 21 -
include/linux/kernel.h | 39 --
include/linux/list.h | 4
include/linux/llist.h | 4
include/linux/pagemap.h | 50 ++
include/linux/plist.h | 5
include/linux/radix-tree.h | 4
include/linux/rwsem.h | 1
include/linux/sbitmap.h | 11
include/linux/seq_file.h | 19 +
include/linux/signal.h | 1
include/linux/smp.h | 1
include/linux/spinlock.h | 1
include/linux/stackdepot.h | 5
include/linux/string_helpers.h | 1
include/media/media-entity.h | 3
init/main.c | 4
ipc/ipc_sysctl.c | 42 +-
ipc/shm.c | 8
kernel/extable.c | 33 -
kernel/fork.c | 9
kernel/kcov.c | 40 +-
kernel/locking/lockdep.c | 3
kernel/resource.c | 54 ++-
kernel/trace/ftrace.c | 2
lib/scatterlist.c | 11
lib/stackdepot.c | 46 ++
lib/vsprintf.c | 3
mm/Kconfig | 7
mm/filemap.c | 8
mm/kasan/report.c | 17 -
mm/memfd.c | 4
mm/mmap.c | 3
mm/page_owner.c | 18 -
mm/truncate.c | 19 +
mm/vmscan.c | 7
mm/workingset.c | 10
net/sysctl_net.c | 2
scripts/checkpatch.pl | 33 +
scripts/const_structs.checkpatch | 4
scripts/gdb/linux/symbols.py | 3
tools/testing/selftests/kselftest/runner.sh | 28 +
tools/testing/selftests/proc/.gitignore | 1
tools/testing/selftests/proc/Makefile | 2
tools/testing/selftests/proc/proc-tid0.c | 81 ++++
132 files changed, 1206 insertions(+), 681 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-11-05 20:34 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-11-05 20:34 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
262 patches, based on 8bb7eca972ad531c9b149c0a51ab43a417385813
Subsystems affected by this patch series:
scripts
ocfs2
vfs
mm/slab-generic
mm/slab
mm/slub
mm/kconfig
mm/dax
mm/kasan
mm/debug
mm/pagecache
mm/gup
mm/swap
mm/memcg
mm/pagemap
mm/mprotect
mm/mremap
mm/iomap
mm/tracing
mm/vmalloc
mm/pagealloc
mm/memory-failure
mm/hugetlb
mm/userfaultfd
mm/vmscan
mm/tools
mm/memblock
mm/oom-kill
mm/hugetlbfs
mm/migration
mm/thp
mm/readahead
mm/nommu
mm/ksm
mm/vmstat
mm/madvise
mm/memory-hotplug
mm/rmap
mm/zsmalloc
mm/highmem
mm/zram
mm/cleanups
mm/kfence
mm/damon
Subsystem: scripts
Colin Ian King <colin.king@canonical.com>:
scripts/spelling.txt: add more spellings to spelling.txt
Sven Eckelmann <sven@narfation.org>:
scripts/spelling.txt: fix "mistake" version of "synchronization"
weidonghui <weidonghui@allwinnertech.com>:
scripts/decodecode: fix faulting instruction no print when opps.file is DOS format
Subsystem: ocfs2
Chenyuan Mi <cymi20@fudan.edu.cn>:
ocfs2: fix handle refcount leak in two exception handling paths
Valentin Vidic <vvidic@valentin-vidic.from.hr>:
ocfs2: cleanup journal init and shutdown
Colin Ian King <colin.king@canonical.com>:
ocfs2/dlm: remove redundant assignment of variable ret
Jan Kara <jack@suse.cz>:
Patch series "ocfs2: Truncate data corruption fix":
ocfs2: fix data corruption on truncate
ocfs2: do not zero pages beyond i_size
Subsystem: vfs
Arnd Bergmann <arnd@arndb.de>:
fs/posix_acl.c: avoid -Wempty-body warning
Jia He <justin.he@arm.com>:
d_path: fix Kernel doc validator complaining
Subsystem: mm/slab-generic
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
mm: move kvmalloc-related functions to slab.h
Subsystem: mm/slab
Shi Lei <shi_lei@massclouds.com>:
mm/slab.c: remove useless lines in enable_cpucache()
Subsystem: mm/slub
Kefeng Wang <wangkefeng.wang@huawei.com>:
slub: add back check for free nonslab objects
Vlastimil Babka <vbabka@suse.cz>:
mm, slub: change percpu partial accounting from objects to pages
mm/slub: increase default cpu partial list sizes
Hyeonggon Yoo <42.hyeyoo@gmail.com>:
mm, slub: use prefetchw instead of prefetch
Subsystem: mm/kconfig
Sebastian Andrzej Siewior <bigeasy@linutronix.de>:
mm: disable NUMA_BALANCING_DEFAULT_ENABLED and TRANSPARENT_HUGEPAGE on PREEMPT_RT
Subsystem: mm/dax
Christoph Hellwig <hch@lst.de>:
mm: don't include <linux/dax.h> in <linux/mempolicy.h>
Subsystem: mm/kasan
Marco Elver <elver@google.com>:
Patch series "stackdepot, kasan, workqueue: Avoid expanding stackdepot slabs when holding raw_spin_lock", v2:
lib/stackdepot: include gfp.h
lib/stackdepot: remove unused function argument
lib/stackdepot: introduce __stack_depot_save()
kasan: common: provide can_alloc in kasan_save_stack()
kasan: generic: introduce kasan_record_aux_stack_noalloc()
workqueue, kasan: avoid alloc_pages() when recording stack
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
kasan: fix tag for large allocations when using CONFIG_SLAB
Peter Collingbourne <pcc@google.com>:
kasan: test: add memcpy test that avoids out-of-bounds write
Subsystem: mm/debug
Peter Xu <peterx@redhat.com>:
Patch series "mm/smaps: Fixes and optimizations on shmem swap handling":
mm/smaps: fix shmem pte hole swap calculation
mm/smaps: use vma->vm_pgoff directly when counting partial swap
mm/smaps: simplify shmem handling of pte holes
Guo Ren <guoren@linux.alibaba.com>:
mm: debug_vm_pgtable: don't use __P000 directly
Kees Cook <keescook@chromium.org>:
kasan: test: bypass __alloc_size checks
Patch series "Add __alloc_size()", v3:
rapidio: avoid bogus __alloc_size warning
Compiler Attributes: add __alloc_size() for better bounds checking
slab: clean up function prototypes
slab: add __alloc_size attributes for better bounds checking
mm/kvmalloc: add __alloc_size attributes for better bounds checking
mm/vmalloc: add __alloc_size attributes for better bounds checking
mm/page_alloc: add __alloc_size attributes for better bounds checking
percpu: add __alloc_size attributes for better bounds checking
Yinan Zhang <zhangyinan2019@email.szu.edu.cn>:
mm/page_ext.c: fix a comment
Subsystem: mm/pagecache
David Howells <dhowells@redhat.com>:
mm: stop filemap_read() from grabbing a superfluous page
Christoph Hellwig <hch@lst.de>:
Patch series "simplify bdi unregistation":
mm: export bdi_unregister
mtd: call bdi_unregister explicitly
fs: explicitly unregister per-superblock BDIs
mm: don't automatically unregister bdis
mm: simplify bdi refcounting
Jens Axboe <axboe@kernel.dk>:
mm: don't read i_size of inode unless we need it
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
mm/filemap.c: remove bogus VM_BUG_ON
Jens Axboe <axboe@kernel.dk>:
mm: move more expensive part of XA setup out of mapping check
Subsystem: mm/gup
John Hubbard <jhubbard@nvidia.com>:
mm/gup: further simplify __gup_device_huge()
Subsystem: mm/swap
Xu Wang <vulab@iscas.ac.cn>:
mm/swapfile: remove needless request_queue NULL pointer check
Rafael Aquini <aquini@redhat.com>:
mm/swapfile: fix an integer overflow in swap_show()
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
mm: optimise put_pages_list()
Subsystem: mm/memcg
Peter Xu <peterx@redhat.com>:
mm/memcg: drop swp_entry_t* in mc_handle_file_pte()
Shakeel Butt <shakeelb@google.com>:
memcg: flush stats only if updated
memcg: unify memcg stat flushing
Waiman Long <longman@redhat.com>:
mm/memcg: remove obsolete memcg_free_kmem()
Len Baker <len.baker@gmx.com>:
mm/list_lru.c: prefer struct_size over open coded arithmetic
Shakeel Butt <shakeelb@google.com>:
memcg, kmem: further deprecate kmem.limit_in_bytes
Muchun Song <songmuchun@bytedance.com>:
mm: list_lru: remove holding lru lock
mm: list_lru: fix the return value of list_lru_count_one()
mm: memcontrol: remove kmemcg_id reparenting
mm: memcontrol: remove the kmem states
mm: list_lru: only add memcg-aware lrus to the global lru list
Vasily Averin <vvs@virtuozzo.com>:
Patch series "memcg: prohibit unconditional exceeding the limit of dying tasks", v3:
mm, oom: pagefault_out_of_memory: don't force global OOM for dying tasks
Michal Hocko <mhocko@suse.com>:
mm, oom: do not trigger out_of_memory from the #PF
Vasily Averin <vvs@virtuozzo.com>:
memcg: prohibit unconditional exceeding the limit of dying tasks
Subsystem: mm/pagemap
Peng Liu <liupeng256@huawei.com>:
mm/mmap.c: fix a data race of mm->total_vm
Rolf Eike Beer <eb@emlix.com>:
mm: use __pfn_to_section() instead of open coding it
Amit Daniel Kachhap <amit.kachhap@arm.com>:
mm/memory.c: avoid unnecessary kernel/user pointer conversion
Nadav Amit <namit@vmware.com>:
mm/memory.c: use correct VMA flags when freeing page-tables
Peter Xu <peterx@redhat.com>:
Patch series "mm: A few cleanup patches around zap, shmem and uffd", v4:
mm/shmem: unconditionally set pte dirty in mfill_atomic_install_pte
mm: clear vmf->pte after pte_unmap_same() returns
mm: drop first_index/last_index in zap_details
mm: add zap_skip_check_mapping() helper
Qi Zheng <zhengqi.arch@bytedance.com>:
Patch series "Do some code cleanups related to mm", v3:
mm: introduce pmd_install() helper
mm: remove redundant smp_wmb()
Tiberiu A Georgescu <tiberiu.georgescu@nutanix.com>:
Documentation: update pagemap with shmem exceptions
Nicholas Piggin <npiggin@gmail.com>:
Patch series "shoot lazy tlbs", v4:
lazy tlb: introduce lazy mm refcount helper functions
lazy tlb: allow lazy tlb mm refcounting to be configurable
lazy tlb: shoot lazies, a non-refcounting lazy tlb option
powerpc/64s: enable MMU_LAZY_TLB_SHOOTDOWN
Lukas Bulwahn <lukas.bulwahn@gmail.com>:
memory: remove unused CONFIG_MEM_BLOCK_SIZE
Subsystem: mm/mprotect
Liu Song <liu.song11@zte.com.cn>:
mm/mprotect.c: avoid repeated assignment in do_mprotect_pkey()
Subsystem: mm/mremap
Dmitry Safonov <dima@arista.com>:
mm/mremap: don't account pages in vma_to_resize()
Subsystem: mm/iomap
Lucas De Marchi <lucas.demarchi@intel.com>:
include/linux/io-mapping.h: remove fallback for writecombine
Subsystem: mm/tracing
Gang Li <ligang.bdlg@bytedance.com>:
mm: mmap_lock: remove redundant newline in TP_printk
mm: mmap_lock: use DECLARE_EVENT_CLASS and DEFINE_EVENT_FN
Subsystem: mm/vmalloc
Vasily Averin <vvs@virtuozzo.com>:
mm/vmalloc: repair warn_alloc()s in __vmalloc_area_node()
Peter Zijlstra <peterz@infradead.org>:
mm/vmalloc: don't allow VM_NO_GUARD on vmap()
Eric Dumazet <edumazet@google.com>:
mm/vmalloc: make show_numa_info() aware of hugepage mappings
mm/vmalloc: make sure to dump unpurged areas in /proc/vmallocinfo
"Uladzislau Rezki (Sony)" <urezki@gmail.com>:
mm/vmalloc: do not adjust the search size for alignment overhead
mm/vmalloc: check various alignments when debugging
Vasily Averin <vvs@virtuozzo.com>:
vmalloc: back off when the current task is OOM-killed
Kefeng Wang <wangkefeng.wang@huawei.com>:
vmalloc: choose a better start address in vm_area_register_early()
arm64: support page mapping percpu first chunk allocator
kasan: arm64: fix pcpu_page_first_chunk crash with KASAN_VMALLOC
Michal Hocko <mhocko@suse.com>:
mm/vmalloc: be more explicit about supported gfp flags
Chen Wandun <chenwandun@huawei.com>:
mm/vmalloc: introduce alloc_pages_bulk_array_mempolicy to accelerate memory allocation
Changcheng Deng <deng.changcheng@zte.com.cn>:
lib/test_vmalloc.c: use swap() to make code cleaner
Subsystem: mm/pagealloc
Eric Dumazet <edumazet@google.com>:
mm/large system hash: avoid possible NULL deref in alloc_large_system_hash
Miaohe Lin <linmiaohe@huawei.com>:
Patch series "Cleanups and fixup for page_alloc", v2:
mm/page_alloc.c: remove meaningless VM_BUG_ON() in pindex_to_order()
mm/page_alloc.c: simplify the code by using macro K()
mm/page_alloc.c: fix obsolete comment in free_pcppages_bulk()
mm/page_alloc.c: use helper function zone_spans_pfn()
mm/page_alloc.c: avoid allocating highmem pages via alloc_pages_exact[_nid]
Bharata B Rao <bharata@amd.com>:
Patch series "Fix NUMA nodes fallback list ordering":
mm/page_alloc: print node fallback order
Krupa Ramakrishnan <krupa.ramakrishnan@amd.com>:
mm/page_alloc: use accumulated load when building node fallback list
Geert Uytterhoeven <geert+renesas@glider.be>:
Patch series "Fix NUMA without SMP":
mm: move node_reclaim_distance to fix NUMA without SMP
mm: move fold_vm_numa_events() to fix NUMA without SMP
Eric Dumazet <edumazet@google.com>:
mm/page_alloc.c: do not acquire zone lock in is_free_buddy_page()
Feng Tang <feng.tang@intel.com>:
mm/page_alloc: detect allocation forbidden by cpuset and bail out early
Liangcai Fan <liangcaifan19@gmail.com>:
mm/page_alloc.c: show watermark_boost of zone in zoneinfo
Christophe Leroy <christophe.leroy@csgroup.eu>:
mm: create a new system state and fix core_kernel_text()
mm: make generic arch_is_kernel_initmem_freed() do what it says
powerpc: use generic version of arch_is_kernel_initmem_freed()
s390: use generic version of arch_is_kernel_initmem_freed()
Sebastian Andrzej Siewior <bigeasy@linutronix.de>:
mm: page_alloc: use migrate_disable() in drain_local_pages_wq()
Wang ShaoBo <bobo.shaobowang@huawei.com>:
mm/page_alloc: use clamp() to simplify code
Subsystem: mm/memory-failure
Marco Elver <elver@google.com>:
mm: fix data race in PagePoisoned()
Rikard Falkeborn <rikard.falkeborn@gmail.com>:
mm/memory_failure: constify static mm_walk_ops
Yang Shi <shy828301@gmail.com>:
Patch series "Solve silent data loss caused by poisoned page cache (shmem/tmpfs)", v5:
mm: filemap: coding style cleanup for filemap_map_pmd()
mm: hwpoison: refactor refcount check handling
mm: shmem: don't truncate page if memory failure happens
mm: hwpoison: handle non-anonymous THP correctly
Subsystem: mm/hugetlb
Peter Xu <peterx@redhat.com>:
mm/hugetlb: drop __unmap_hugepage_range definition from hugetlb.h
Mike Kravetz <mike.kravetz@oracle.com>:
Patch series "hugetlb: add demote/split page functionality", v4:
hugetlb: add demote hugetlb page sysfs interfaces
mm/cma: add cma_pages_valid to determine if pages are in CMA
hugetlb: be sure to free demoted CMA pages to CMA
hugetlb: add demote bool to gigantic page routines
hugetlb: add hugetlb demote page support
Liangcai Fan <liangcaifan19@gmail.com>:
mm: khugepaged: recalculate min_free_kbytes after stopping khugepaged
Mina Almasry <almasrymina@google.com>:
mm, hugepages: add mremap() support for hugepage backed vma
mm, hugepages: add hugetlb vma mremap() test
Baolin Wang <baolin.wang@linux.alibaba.com>:
hugetlb: support node specified when using cma for gigantic hugepages
Ran Jianping <ran.jianping@zte.com.cn>:
mm: remove duplicate include in hugepage-mremap.c
Baolin Wang <baolin.wang@linux.alibaba.com>:
Patch series "Some cleanups and improvements for hugetlb":
hugetlb_cgroup: remove unused hugetlb_cgroup_from_counter macro
hugetlb: replace the obsolete hugetlb_instantiation_mutex in the comments
hugetlb: remove redundant validation in has_same_uncharge_info()
hugetlb: remove redundant VM_BUG_ON() in add_reservation_in_range()
Mike Kravetz <mike.kravetz@oracle.com>:
hugetlb: remove unnecessary set_page_count in prep_compound_gigantic_page
Subsystem: mm/userfaultfd
Axel Rasmussen <axelrasmussen@google.com>:
Patch series "Small userfaultfd selftest fixups", v2:
userfaultfd/selftests: don't rely on GNU extensions for random numbers
userfaultfd/selftests: fix feature support detection
userfaultfd/selftests: fix calculation of expected ioctls
Subsystem: mm/vmscan
Miaohe Lin <linmiaohe@huawei.com>:
mm/page_isolation: fix potential missing call to unset_migratetype_isolate()
mm/page_isolation: guard against possible putback unisolated page
Kai Song <songkai01@inspur.com>:
mm/vmscan.c: fix -Wunused-but-set-variable warning
Mel Gorman <mgorman@techsingularity.net>:
Patch series "Remove dependency on congestion_wait in mm/", v5. Patch series:
mm/vmscan: throttle reclaim until some writeback completes if congested
mm/vmscan: throttle reclaim and compaction when too may pages are isolated
mm/vmscan: throttle reclaim when no progress is being made
mm/writeback: throttle based on page writeback instead of congestion
mm/page_alloc: remove the throttling logic from the page allocator
mm/vmscan: centralise timeout values for reclaim_throttle
mm/vmscan: increase the timeout if page reclaim is not making progress
mm/vmscan: delay waking of tasks throttled on NOPROGRESS
Yuanzheng Song <songyuanzheng@huawei.com>:
mm/vmpressure: fix data-race with memcg->socket_pressure
Subsystem: mm/tools
Zhenliang Wei <weizhenliang@huawei.com>:
tools/vm/page_owner_sort.c: count and sort by mem
Naoya Horiguchi <naoya.horiguchi@nec.com>:
Patch series "tools/vm/page-types.c: a few improvements":
tools/vm/page-types.c: make walk_file() aware of address range option
tools/vm/page-types.c: move show_file() to summary output
tools/vm/page-types.c: print file offset in hexadecimal
Subsystem: mm/memblock
Mike Rapoport <rppt@linux.ibm.com>:
Patch series "memblock: cleanup memblock_free interface", v2:
arch_numa: simplify numa_distance allocation
xen/x86: free_p2m_page: use memblock_free_ptr() to free a virtual pointer
memblock: drop memblock_free_early_nid() and memblock_free_early()
memblock: stop aliasing __memblock_free_late with memblock_free_late
memblock: rename memblock_free to memblock_phys_free
memblock: use memblock_free for freeing virtual pointers
Subsystem: mm/oom-kill
Sultan Alsawaf <sultan@kerneltoast.com>:
mm: mark the OOM reaper thread as freezable
Subsystem: mm/hugetlbfs
Zhenguo Yao <yaozhenguo1@gmail.com>:
hugetlbfs: extend the definition of hugepages parameter to support node allocation
Subsystem: mm/migration
John Hubbard <jhubbard@nvidia.com>:
mm/migrate: de-duplicate migrate_reason strings
Yang Shi <shy828301@gmail.com>:
mm: migrate: make demotion knob depend on migration
Subsystem: mm/thp
"George G. Davis" <davis.george@siemens.com>:
selftests/vm/transhuge-stress: fix ram size thinko
Rongwei Wang <rongwei.wang@linux.alibaba.com>:
Patch series "fix two bugs for file THP":
mm, thp: lock filemap when truncating page cache
mm, thp: fix incorrect unmap behavior for private pages
Subsystem: mm/readahead
Lin Feng <linf@wangsu.com>:
mm/readahead.c: fix incorrect comments for get_init_ra_size
Subsystem: mm/nommu
Kefeng Wang <wangkefeng.wang@huawei.com>:
mm: nommu: kill arch_get_unmapped_area()
Subsystem: mm/ksm
"Aneesh Kumar K.V" <aneesh.kumar@linux.ibm.com>:
selftest/vm: fix ksm selftest to run with different NUMA topologies
Pedro Demarchi Gomes <pedrodemargomes@gmail.com>:
selftests: vm: add KSM huge pages merging time test
Subsystem: mm/vmstat
Liu Shixin <liushixin2@huawei.com>:
mm/vmstat: annotate data race for zone->free_area[order].nr_free
Lin Feng <linf@wangsu.com>:
mm: vmstat.c: make extfrag_index show more pretty
Subsystem: mm/madvise
David Hildenbrand <david@redhat.com>:
selftests/vm: make MADV_POPULATE_(READ|WRITE) use in-tree headers
Subsystem: mm/memory-hotplug
Tang Yizhou <tangyizhou@huawei.com>:
mm/memory_hotplug: add static qualifier for online_policy_to_str()
David Hildenbrand <david@redhat.com>:
Patch series "memory-hotplug.rst: document the "auto-movable" online policy":
memory-hotplug.rst: fix two instances of "movablecore" that should be "movable_node"
memory-hotplug.rst: fix wrong /sys/module/memory_hotplug/parameters/ path
memory-hotplug.rst: document the "auto-movable" online policy
Patch series "mm/memory_hotplug: Kconfig and 32 bit cleanups":
mm/memory_hotplug: remove CONFIG_X86_64_ACPI_NUMA dependency from CONFIG_MEMORY_HOTPLUG
mm/memory_hotplug: remove CONFIG_MEMORY_HOTPLUG_SPARSE
mm/memory_hotplug: restrict CONFIG_MEMORY_HOTPLUG to 64 bit
mm/memory_hotplug: remove HIGHMEM leftovers
mm/memory_hotplug: remove stale function declarations
x86: remove memory hotplug support on X86_32
Patch series "mm/memory_hotplug: full support for add_memory_driver_managed() with CONFIG_ARCH_KEEP_MEMBLOCK", v2:
mm/memory_hotplug: handle memblock_add_node() failures in add_memory_resource()
memblock: improve MEMBLOCK_HOTPLUG documentation
memblock: allow to specify flags with memblock_add_node()
memblock: add MEMBLOCK_DRIVER_MANAGED to mimic IORESOURCE_SYSRAM_DRIVER_MANAGED
mm/memory_hotplug: indicate MEMBLOCK_DRIVER_MANAGED with IORESOURCE_SYSRAM_DRIVER_MANAGED
Subsystem: mm/rmap
Alistair Popple <apopple@nvidia.com>:
mm/rmap.c: avoid double faults migrating device private pages
Subsystem: mm/zsmalloc
Miaohe Lin <linmiaohe@huawei.com>:
mm/zsmalloc.c: close race window between zs_pool_dec_isolated() and zs_unregister_migration()
Subsystem: mm/highmem
Ira Weiny <ira.weiny@intel.com>:
mm/highmem: remove deprecated kmap_atomic
Subsystem: mm/zram
Jaewon Kim <jaewon31.kim@samsung.com>:
zram_drv: allow reclaim on bio_alloc
Dan Carpenter <dan.carpenter@oracle.com>:
zram: off by one in read_block_state()
Brian Geffon <bgeffon@google.com>:
zram: introduce an aged idle interface
Subsystem: mm/cleanups
Stephen Kitt <steve@sk2.org>:
mm: remove HARDENED_USERCOPY_FALLBACK
Mianhan Liu <liumh1@shanghaitech.edu.cn>:
include/linux/mm.h: move nr_free_buffer_pages from swap.h to mm.h
Subsystem: mm/kfence
Marco Elver <elver@google.com>:
stacktrace: move filter_irq_stacks() to kernel/stacktrace.c
kfence: count unexpectedly skipped allocations
kfence: move saving stack trace of allocations into __kfence_alloc()
kfence: limit currently covered allocations when pool nearly full
kfence: add note to documentation about skipping covered allocations
kfence: test: use kunit_skip() to skip tests
kfence: shorten critical sections of alloc/free
kfence: always use static branches to guard kfence_alloc()
kfence: default to dynamic branch instead of static keys mode
Subsystem: mm/damon
Geert Uytterhoeven <geert@linux-m68k.org>:
mm/damon: grammar s/works/work/
SeongJae Park <sjpark@amazon.de>:
Documentation/vm: move user guides to admin-guide/mm/
SeongJae Park <sj@kernel.org>:
MAINTAINERS: update SeongJae's email address
SeongJae Park <sjpark@amazon.de>:
docs/vm/damon: remove broken reference
include/linux/damon.h: fix kernel-doc comments for 'damon_callback'
SeongJae Park <sj@kernel.org>:
mm/damon/core: print kdamond start log in debug mode only
Changbin Du <changbin.du@gmail.com>:
mm/damon: remove unnecessary do_exit() from kdamond
mm/damon: needn't hold kdamond_lock to print pid of kdamond
Colin Ian King <colin.king@canonical.com>:
mm/damon/core: nullify pointer ctx->kdamond with a NULL
SeongJae Park <sj@kernel.org>:
Patch series "Implement Data Access Monitoring-based Memory Operation Schemes":
mm/damon/core: account age of target regions
mm/damon/core: implement DAMON-based Operation Schemes (DAMOS)
mm/damon/vaddr: support DAMON-based Operation Schemes
mm/damon/dbgfs: support DAMON-based Operation Schemes
mm/damon/schemes: implement statistics feature
selftests/damon: add 'schemes' debugfs tests
Docs/admin-guide/mm/damon: document DAMON-based Operation Schemes
Patch series "DAMON: Support Physical Memory Address Space Monitoring::
mm/damon/dbgfs: allow users to set initial monitoring target regions
mm/damon/dbgfs-test: add a unit test case for 'init_regions'
Docs/admin-guide/mm/damon: document 'init_regions' feature
mm/damon/vaddr: separate commonly usable functions
mm/damon: implement primitives for physical address space monitoring
mm/damon/dbgfs: support physical memory monitoring
Docs/DAMON: document physical memory monitoring support
Rikard Falkeborn <rikard.falkeborn@gmail.com>:
mm/damon/vaddr: constify static mm_walk_ops
Rongwei Wang <rongwei.wang@linux.alibaba.com>:
mm/damon/dbgfs: remove unnecessary variables
SeongJae Park <sj@kernel.org>:
mm/damon/paddr: support the pageout scheme
mm/damon/schemes: implement size quota for schemes application speed control
mm/damon/schemes: skip already charged targets and regions
mm/damon/schemes: implement time quota
mm/damon/dbgfs: support quotas of schemes
mm/damon/selftests: support schemes quotas
mm/damon/schemes: prioritize regions within the quotas
mm/damon/vaddr,paddr: support pageout prioritization
mm/damon/dbgfs: support prioritization weights
tools/selftests/damon: update for regions prioritization of schemes
mm/damon/schemes: activate schemes based on a watermarks mechanism
mm/damon/dbgfs: support watermarks
selftests/damon: support watermarks
mm/damon: introduce DAMON-based Reclamation (DAMON_RECLAIM)
Documentation/admin-guide/mm/damon: add a document for DAMON_RECLAIM
Xin Hao <xhao@linux.alibaba.com>:
Patch series "mm/damon: Fix some small bugs", v4:
mm/damon: remove unnecessary variable initialization
mm/damon/dbgfs: add adaptive_targets list check before enable monitor_on
SeongJae Park <sj@kernel.org>:
Patch series "Fix trivial nits in Documentation/admin-guide/mm":
Docs/admin-guide/mm/damon/start: fix wrong example commands
Docs/admin-guide/mm/damon/start: fix a wrong link
Docs/admin-guide/mm/damon/start: simplify the content
Docs/admin-guide/mm/pagemap: wordsmith page flags descriptions
Changbin Du <changbin.du@gmail.com>:
mm/damon: simplify stop mechanism
Colin Ian King <colin.i.king@googlemail.com>:
mm/damon: fix a few spelling mistakes in comments and a pr_debug message
Changbin Du <changbin.du@gmail.com>:
mm/damon: remove return value from before_terminate callback
a/Documentation/admin-guide/blockdev/zram.rst | 8
a/Documentation/admin-guide/cgroup-v1/memory.rst | 11
a/Documentation/admin-guide/kernel-parameters.txt | 14
a/Documentation/admin-guide/mm/damon/index.rst | 1
a/Documentation/admin-guide/mm/damon/reclaim.rst | 235 +++
a/Documentation/admin-guide/mm/damon/start.rst | 140 +
a/Documentation/admin-guide/mm/damon/usage.rst | 117 +
a/Documentation/admin-guide/mm/hugetlbpage.rst | 42
a/Documentation/admin-guide/mm/memory-hotplug.rst | 147 +-
a/Documentation/admin-guide/mm/pagemap.rst | 75 -
a/Documentation/core-api/memory-hotplug.rst | 3
a/Documentation/dev-tools/kfence.rst | 23
a/Documentation/translations/zh_CN/core-api/memory-hotplug.rst | 4
a/Documentation/vm/damon/design.rst | 29
a/Documentation/vm/damon/faq.rst | 5
a/Documentation/vm/damon/index.rst | 1
a/Documentation/vm/page_owner.rst | 23
a/MAINTAINERS | 2
a/Makefile | 15
a/arch/Kconfig | 28
a/arch/alpha/kernel/core_irongate.c | 6
a/arch/arc/mm/init.c | 6
a/arch/arm/mach-hisi/platmcpm.c | 2
a/arch/arm/mach-rpc/ecard.c | 2
a/arch/arm/mm/init.c | 2
a/arch/arm64/Kconfig | 4
a/arch/arm64/mm/kasan_init.c | 16
a/arch/arm64/mm/mmu.c | 4
a/arch/ia64/mm/contig.c | 2
a/arch/ia64/mm/init.c | 2
a/arch/m68k/mm/mcfmmu.c | 3
a/arch/m68k/mm/motorola.c | 6
a/arch/mips/loongson64/init.c | 4
a/arch/mips/mm/init.c | 6
a/arch/mips/sgi-ip27/ip27-memory.c | 3
a/arch/mips/sgi-ip30/ip30-setup.c | 6
a/arch/powerpc/Kconfig | 1
a/arch/powerpc/configs/skiroot_defconfig | 1
a/arch/powerpc/include/asm/machdep.h | 2
a/arch/powerpc/include/asm/sections.h | 13
a/arch/powerpc/kernel/dt_cpu_ftrs.c | 8
a/arch/powerpc/kernel/paca.c | 8
a/arch/powerpc/kernel/setup-common.c | 4
a/arch/powerpc/kernel/setup_64.c | 6
a/arch/powerpc/kernel/smp.c | 2
a/arch/powerpc/mm/book3s64/radix_tlb.c | 4
a/arch/powerpc/mm/hugetlbpage.c | 9
a/arch/powerpc/platforms/powernv/pci-ioda.c | 4
a/arch/powerpc/platforms/powernv/setup.c | 4
a/arch/powerpc/platforms/pseries/setup.c | 2
a/arch/powerpc/platforms/pseries/svm.c | 9
a/arch/riscv/kernel/setup.c | 10
a/arch/s390/include/asm/sections.h | 12
a/arch/s390/kernel/setup.c | 11
a/arch/s390/kernel/smp.c | 6
a/arch/s390/kernel/uv.c | 2
a/arch/s390/mm/init.c | 3
a/arch/s390/mm/kasan_init.c | 2
a/arch/sh/boards/mach-ap325rxa/setup.c | 2
a/arch/sh/boards/mach-ecovec24/setup.c | 4
a/arch/sh/boards/mach-kfr2r09/setup.c | 2
a/arch/sh/boards/mach-migor/setup.c | 2
a/arch/sh/boards/mach-se/7724/setup.c | 4
a/arch/sparc/kernel/smp_64.c | 4
a/arch/um/kernel/mem.c | 4
a/arch/x86/Kconfig | 6
a/arch/x86/kernel/setup.c | 4
a/arch/x86/kernel/setup_percpu.c | 2
a/arch/x86/mm/init.c | 2
a/arch/x86/mm/init_32.c | 31
a/arch/x86/mm/kasan_init_64.c | 4
a/arch/x86/mm/numa.c | 2
a/arch/x86/mm/numa_emulation.c | 2
a/arch/x86/xen/mmu_pv.c | 8
a/arch/x86/xen/p2m.c | 4
a/arch/x86/xen/setup.c | 6
a/drivers/base/Makefile | 2
a/drivers/base/arch_numa.c | 96 +
a/drivers/base/node.c | 9
a/drivers/block/zram/zram_drv.c | 66
a/drivers/firmware/efi/memmap.c | 2
a/drivers/hwmon/occ/p9_sbe.c | 1
a/drivers/macintosh/smu.c | 2
a/drivers/mmc/core/mmc_test.c | 1
a/drivers/mtd/mtdcore.c | 1
a/drivers/of/kexec.c | 4
a/drivers/of/of_reserved_mem.c | 5
a/drivers/rapidio/devices/rio_mport_cdev.c | 9
a/drivers/s390/char/sclp_early.c | 4
a/drivers/usb/early/xhci-dbc.c | 10
a/drivers/virtio/Kconfig | 2
a/drivers/xen/swiotlb-xen.c | 4
a/fs/d_path.c | 8
a/fs/exec.c | 4
a/fs/ocfs2/alloc.c | 21
a/fs/ocfs2/dlm/dlmrecovery.c | 1
a/fs/ocfs2/file.c | 8
a/fs/ocfs2/inode.c | 4
a/fs/ocfs2/journal.c | 28
a/fs/ocfs2/journal.h | 3
a/fs/ocfs2/super.c | 40
a/fs/open.c | 16
a/fs/posix_acl.c | 3
a/fs/proc/task_mmu.c | 28
a/fs/super.c | 3
a/include/asm-generic/sections.h | 14
a/include/linux/backing-dev-defs.h | 3
a/include/linux/backing-dev.h | 1
a/include/linux/cma.h | 1
a/include/linux/compiler-gcc.h | 8
a/include/linux/compiler_attributes.h | 10
a/include/linux/compiler_types.h | 12
a/include/linux/cpuset.h | 17
a/include/linux/damon.h | 258 +++
a/include/linux/fs.h | 1
a/include/linux/gfp.h | 8
a/include/linux/highmem.h | 28
a/include/linux/hugetlb.h | 36
a/include/linux/io-mapping.h | 6
a/include/linux/kasan.h | 8
a/include/linux/kernel.h | 1
a/include/linux/kfence.h | 21
a/include/linux/memblock.h | 48
a/include/linux/memcontrol.h | 9
a/include/linux/memory.h | 26
a/include/linux/memory_hotplug.h | 3
a/include/linux/mempolicy.h | 5
a/include/linux/migrate.h | 23
a/include/linux/migrate_mode.h | 13
a/include/linux/mm.h | 57
a/include/linux/mm_types.h | 2
a/include/linux/mmzone.h | 41
a/include/linux/node.h | 4
a/include/linux/page-flags.h | 2
a/include/linux/percpu.h | 6
a/include/linux/sched/mm.h | 25
a/include/linux/slab.h | 181 +-
a/include/linux/slub_def.h | 13
a/include/linux/stackdepot.h | 8
a/include/linux/stacktrace.h | 1
a/include/linux/swap.h | 1
a/include/linux/vmalloc.h | 24
a/include/trace/events/mmap_lock.h | 50
a/include/trace/events/vmscan.h | 42
a/include/trace/events/writeback.h | 7
a/init/Kconfig | 2
a/init/initramfs.c | 4
a/init/main.c | 6
a/kernel/cgroup/cpuset.c | 23
a/kernel/cpu.c | 2
a/kernel/dma/swiotlb.c | 6
a/kernel/exit.c | 2
a/kernel/extable.c | 2
a/kernel/fork.c | 51
a/kernel/kexec_file.c | 5
a/kernel/kthread.c | 21
a/kernel/locking/lockdep.c | 15
a/kernel/printk/printk.c | 4
a/kernel/sched/core.c | 37
a/kernel/sched/sched.h | 4
a/kernel/sched/topology.c | 1
a/kernel/stacktrace.c | 30
a/kernel/tsacct.c | 2
a/kernel/workqueue.c | 2
a/lib/Kconfig.debug | 2
a/lib/Kconfig.kfence | 26
a/lib/bootconfig.c | 2
a/lib/cpumask.c | 6
a/lib/stackdepot.c | 76 -
a/lib/test_kasan.c | 26
a/lib/test_kasan_module.c | 2
a/lib/test_vmalloc.c | 6
a/mm/Kconfig | 10
a/mm/backing-dev.c | 65
a/mm/cma.c | 26
a/mm/compaction.c | 12
a/mm/damon/Kconfig | 24
a/mm/damon/Makefile | 4
a/mm/damon/core.c | 500 ++++++-
a/mm/damon/dbgfs-test.h | 56
a/mm/damon/dbgfs.c | 486 +++++-
a/mm/damon/paddr.c | 275 +++
a/mm/damon/prmtv-common.c | 133 +
a/mm/damon/prmtv-common.h | 20
a/mm/damon/reclaim.c | 356 ++++
a/mm/damon/vaddr-test.h | 2
a/mm/damon/vaddr.c | 167 +-
a/mm/debug.c | 20
a/mm/debug_vm_pgtable.c | 7
a/mm/filemap.c | 78 -
a/mm/gup.c | 5
a/mm/highmem.c | 6
a/mm/hugetlb.c | 713 +++++++++-
a/mm/hugetlb_cgroup.c | 3
a/mm/internal.h | 26
a/mm/kasan/common.c | 8
a/mm/kasan/generic.c | 16
a/mm/kasan/kasan.h | 2
a/mm/kasan/shadow.c | 5
a/mm/kfence/core.c | 214 ++-
a/mm/kfence/kfence.h | 2
a/mm/kfence/kfence_test.c | 14
a/mm/khugepaged.c | 10
a/mm/list_lru.c | 58
a/mm/memblock.c | 35
a/mm/memcontrol.c | 217 +--
a/mm/memory-failure.c | 117 +
a/mm/memory.c | 166 +-
a/mm/memory_hotplug.c | 57
a/mm/mempolicy.c | 143 +-
a/mm/migrate.c | 61
a/mm/mmap.c | 2
a/mm/mprotect.c | 5
a/mm/mremap.c | 86 -
a/mm/nommu.c | 6
a/mm/oom_kill.c | 27
a/mm/page-writeback.c | 13
a/mm/page_alloc.c | 119 -
a/mm/page_ext.c | 2
a/mm/page_isolation.c | 29
a/mm/percpu.c | 24
a/mm/readahead.c | 2
a/mm/rmap.c | 8
a/mm/shmem.c | 44
a/mm/slab.c | 16
a/mm/slab_common.c | 8
a/mm/slub.c | 117 -
a/mm/sparse-vmemmap.c | 2
a/mm/sparse.c | 6
a/mm/swap.c | 23
a/mm/swapfile.c | 6
a/mm/userfaultfd.c | 8
a/mm/vmalloc.c | 107 +
a/mm/vmpressure.c | 2
a/mm/vmscan.c | 194 ++
a/mm/vmstat.c | 76 -
a/mm/zsmalloc.c | 7
a/net/ipv4/tcp.c | 1
a/net/ipv4/udp.c | 1
a/net/netfilter/ipvs/ip_vs_ctl.c | 1
a/net/openvswitch/meter.c | 1
a/net/sctp/protocol.c | 1
a/scripts/checkpatch.pl | 3
a/scripts/decodecode | 2
a/scripts/spelling.txt | 18
a/security/Kconfig | 14
a/tools/testing/selftests/damon/debugfs_attrs.sh | 25
a/tools/testing/selftests/memory-hotplug/config | 1
a/tools/testing/selftests/vm/.gitignore | 1
a/tools/testing/selftests/vm/Makefile | 1
a/tools/testing/selftests/vm/hugepage-mremap.c | 161 ++
a/tools/testing/selftests/vm/ksm_tests.c | 154 ++
a/tools/testing/selftests/vm/madv_populate.c | 15
a/tools/testing/selftests/vm/run_vmtests.sh | 11
a/tools/testing/selftests/vm/transhuge-stress.c | 2
a/tools/testing/selftests/vm/userfaultfd.c | 157 +-
a/tools/vm/page-types.c | 38
a/tools/vm/page_owner_sort.c | 94 +
b/Documentation/admin-guide/mm/index.rst | 2
b/Documentation/vm/index.rst | 26
260 files changed, 6448 insertions(+), 2327 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-10-28 21:35 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-10-28 21:35 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
11 patches, based on 411a44c24a561e449b592ff631b7ae321f1eb559.
Subsystems affected by this patch series:
mm/memcg
mm/memory-failure
mm/oom-kill
ocfs2
mm/secretmem
mm/vmalloc
mm/hugetlb
mm/damon
mm/tools
Subsystem: mm/memcg
Shakeel Butt <shakeelb@google.com>:
memcg: page_alloc: skip bulk allocator for __GFP_ACCOUNT
Subsystem: mm/memory-failure
Yang Shi <shy828301@gmail.com>:
mm: hwpoison: remove the unnecessary THP check
mm: filemap: check if THP has hwpoisoned subpage for PMD page fault
Subsystem: mm/oom-kill
Suren Baghdasaryan <surenb@google.com>:
mm/oom_kill.c: prevent a race between process_mrelease and exit_mmap
Subsystem: ocfs2
Gautham Ananthakrishna <gautham.ananthakrishna@oracle.com>:
ocfs2: fix race between searching chunks and release journal_head from buffer_head
Subsystem: mm/secretmem
Kees Cook <keescook@chromium.org>:
mm/secretmem: avoid letting secretmem_users drop to zero
Subsystem: mm/vmalloc
Chen Wandun <chenwandun@huawei.com>:
mm/vmalloc: fix numa spreading for large hash tables
Subsystem: mm/hugetlb
Rongwei Wang <rongwei.wang@linux.alibaba.com>:
mm, thp: bail out early in collapse_file for writeback page
Yang Shi <shy828301@gmail.com>:
mm: khugepaged: skip huge page collapse for special files
Subsystem: mm/damon
SeongJae Park <sj@kernel.org>:
mm/damon/core-test: fix wrong expectations for 'damon_split_regions_of()'
Subsystem: mm/tools
David Yang <davidcomponentone@gmail.com>:
tools/testing/selftests/vm/split_huge_page_test.c: fix application of sizeof to pointer
fs/ocfs2/suballoc.c | 22 ++++++++++-------
include/linux/page-flags.h | 23 ++++++++++++++++++
mm/damon/core-test.h | 4 +--
mm/huge_memory.c | 2 +
mm/khugepaged.c | 26 +++++++++++++-------
mm/memory-failure.c | 28 +++++++++++-----------
mm/memory.c | 9 +++++++
mm/oom_kill.c | 23 +++++++++---------
mm/page_alloc.c | 8 +++++-
mm/secretmem.c | 2 -
mm/vmalloc.c | 15 +++++++----
tools/testing/selftests/vm/split_huge_page_test.c | 2 -
12 files changed, 110 insertions(+), 54 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-10-18 22:14 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-10-18 22:14 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
19 patches, based on 519d81956ee277b4419c723adfb154603c2565ba.
Subsystems affected by this patch series:
mm/userfaultfd
mm/migration
ocfs2
mm/memblock
mm/mempolicy
mm/slub
binfmt
vfs
mm/secretmem
mm/thp
misc
Subsystem: mm/userfaultfd
Peter Xu <peterx@redhat.com>:
mm/userfaultfd: selftests: fix memory corruption with thp enabled
Nadav Amit <namit@vmware.com>:
userfaultfd: fix a race between writeprotect and exit_mmap()
Subsystem: mm/migration
Dave Hansen <dave.hansen@linux.intel.com>:
Patch series "mm/migrate: 5.15 fixes for automatic demotion", v2:
mm/migrate: optimize hotplug-time demotion order updates
mm/migrate: add CPU hotplug to demotion #ifdef
Huang Ying <ying.huang@intel.com>:
mm/migrate: fix CPUHP state to update node demotion order
Subsystem: ocfs2
Jan Kara <jack@suse.cz>:
ocfs2: fix data corruption after conversion from inline format
Valentin Vidic <vvidic@valentin-vidic.from.hr>:
ocfs2: mount fails with buffer overflow in strlen
Subsystem: mm/memblock
Peng Fan <peng.fan@nxp.com>:
memblock: check memory total_size
Subsystem: mm/mempolicy
Eric Dumazet <edumazet@google.com>:
mm/mempolicy: do not allow illegal MPOL_F_NUMA_BALANCING | MPOL_LOCAL in mbind()
Subsystem: mm/slub
Miaohe Lin <linmiaohe@huawei.com>:
Patch series "Fixups for slub":
mm, slub: fix two bugs in slab_debug_trace_open()
mm, slub: fix mismatch between reconstructed freelist depth and cnt
mm, slub: fix potential memoryleak in kmem_cache_open()
mm, slub: fix potential use-after-free in slab_debugfs_fops
mm, slub: fix incorrect memcg slab count for bulk free
Subsystem: binfmt
Lukas Bulwahn <lukas.bulwahn@gmail.com>:
elfcore: correct reference to CONFIG_UML
Subsystem: vfs
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
vfs: check fd has read access in kernel_read_file_from_fd()
Subsystem: mm/secretmem
Sean Christopherson <seanjc@google.com>:
mm/secretmem: fix NULL page->mapping dereference in page_is_secretmem()
Subsystem: mm/thp
Marek Szyprowski <m.szyprowski@samsung.com>:
mm/thp: decrease nr_thps in file's mapping on THP split
Subsystem: misc
Andrej Shadura <andrew.shadura@collabora.co.uk>:
mailmap: add Andrej Shadura
.mailmap | 2 +
fs/kernel_read_file.c | 2 -
fs/ocfs2/alloc.c | 46 ++++++-----------------
fs/ocfs2/super.c | 14 +++++--
fs/userfaultfd.c | 12 ++++--
include/linux/cpuhotplug.h | 4 ++
include/linux/elfcore.h | 2 -
include/linux/memory.h | 5 ++
include/linux/secretmem.h | 2 -
mm/huge_memory.c | 6 ++-
mm/memblock.c | 2 -
mm/mempolicy.c | 16 ++------
mm/migrate.c | 62 ++++++++++++++++++-------------
mm/page_ext.c | 4 --
mm/slab.c | 4 +-
mm/slub.c | 31 ++++++++++++---
tools/testing/selftests/vm/userfaultfd.c | 23 ++++++++++-
17 files changed, 138 insertions(+), 99 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-09-24 22:42 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-09-24 22:42 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
16 patches, based on 7d42e98182586f57f376406d033f05fe135edb75.
Subsystems affected by this patch series:
mm/memory-failure
mm/kasan
mm/damon
xtensa
mm/shmem
ocfs2
scripts
mm/tools
lib
mm/pagecache
mm/debug
sh
mm/kasan
mm/memory-failure
mm/pagemap
Subsystem: mm/memory-failure
Naoya Horiguchi <naoya.horiguchi@nec.com>:
mm, hwpoison: add is_free_buddy_page() in HWPoisonHandlable()
Subsystem: mm/kasan
Marco Elver <elver@google.com>:
kasan: fix Kconfig check of CC_HAS_WORKING_NOSANITIZE_ADDRESS
Subsystem: mm/damon
Adam Borowski <kilobyte@angband.pl>:
mm/damon: don't use strnlen() with known-bogus source length
Subsystem: xtensa
Guenter Roeck <linux@roeck-us.net>:
xtensa: increase size of gcc stack frame check
Subsystem: mm/shmem
Liu Yuntao <liuyuntao10@huawei.com>:
mm/shmem.c: fix judgment error in shmem_is_huge()
Subsystem: ocfs2
Wengang Wang <wen.gang.wang@oracle.com>:
ocfs2: drop acl cache for directories too
Subsystem: scripts
Miles Chen <miles.chen@mediatek.com>:
scripts/sorttable: riscv: fix undeclared identifier 'EM_RISCV' error
Subsystem: mm/tools
Changbin Du <changbin.du@gmail.com>:
tools/vm/page-types: remove dependency on opt_file for idle page tracking
Subsystem: lib
Paul Menzel <pmenzel@molgen.mpg.de>:
lib/zlib_inflate/inffast: check config in C to avoid unused function warning
Subsystem: mm/pagecache
Minchan Kim <minchan@kernel.org>:
mm: fs: invalidate bh_lrus for only cold path
Subsystem: mm/debug
Weizhao Ouyang <o451686892@gmail.com>:
mm/debug: sync up MR_CONTIG_RANGE and MR_LONGTERM_PIN
mm/debug: sync up latest migrate_reason to migrate_reason_names
Subsystem: sh
Geert Uytterhoeven <geert+renesas@glider.be>:
sh: pgtable-3level: fix cast to pointer from integer of different size
Subsystem: mm/kasan
Nathan Chancellor <nathan@kernel.org>:
kasan: always respect CONFIG_KASAN_STACK
Subsystem: mm/memory-failure
Qi Zheng <zhengqi.arch@bytedance.com>:
mm/memory_failure: fix the missing pte_unmap() call
Subsystem: mm/pagemap
Chen Jun <chenjun102@huawei.com>:
mm: fix uninitialized use in overcommit_policy_handler
arch/sh/include/asm/pgtable-3level.h | 2 +-
fs/buffer.c | 8 ++++++--
fs/ocfs2/dlmglue.c | 3 ++-
include/linux/buffer_head.h | 4 ++--
include/linux/migrate.h | 6 +++++-
lib/Kconfig.debug | 2 +-
lib/Kconfig.kasan | 2 ++
lib/zlib_inflate/inffast.c | 13 ++++++-------
mm/damon/dbgfs-test.h | 16 ++++++++--------
mm/debug.c | 4 +++-
mm/memory-failure.c | 12 ++++++------
mm/shmem.c | 4 ++--
mm/swap.c | 19 ++++++++++++++++---
mm/util.c | 4 ++--
scripts/Makefile.kasan | 3 ++-
scripts/sorttable.c | 4 ++++
tools/vm/page-types.c | 2 +-
17 files changed, 69 insertions(+), 39 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2021-09-10 17:11 ` incoming Kees Cook
@ 2021-09-10 20:13 ` Kees Cook
0 siblings, 0 replies; 409+ messages in thread
From: Kees Cook @ 2021-09-10 20:13 UTC (permalink / raw)
To: linux-kernel; +Cc: Linus Torvalds, Andrew Morton, linux-mm, mm-commits
On Fri, Sep 10, 2021 at 10:11:53AM -0700, Kees Cook wrote:
> On Thu, Sep 09, 2021 at 08:09:48PM -0700, Andrew Morton wrote:
> >
> > More post linux-next material.
> >
> > 9 patches, based on f154c806676ad7153c6e161f30c53a44855329d6.
> >
> > Subsystems affected by this patch series:
> >
> > mm/slab-generic
> > rapidio
> > mm/debug
> >
> > Subsystem: mm/slab-generic
> >
> > "Matthew Wilcox (Oracle)" <willy@infradead.org>:
> > mm: move kvmalloc-related functions to slab.h
> >
> > Subsystem: rapidio
> >
> > Kees Cook <keescook@chromium.org>:
> > rapidio: avoid bogus __alloc_size warning
> >
> > Subsystem: mm/debug
> >
> > Kees Cook <keescook@chromium.org>:
> > Patch series "Add __alloc_size() for better bounds checking", v2:
> > Compiler Attributes: add __alloc_size() for better bounds checking
> > checkpatch: add __alloc_size() to known $Attribute
> > slab: clean up function declarations
> > slab: add __alloc_size attributes for better bounds checking
> > mm/page_alloc: add __alloc_size attributes for better bounds checking
> > percpu: add __alloc_size attributes for better bounds checking
> > mm/vmalloc: add __alloc_size attributes for better bounds checking
>
> Hi,
>
> FYI, in overnight build testing I found yet another corner case in
> GCC's handling of the __alloc_size attribute. It's the gift that keeps
> on giving. The fix is here:
>
> https://lore.kernel.org/lkml/20210910165851.3296624-1-keescook@chromium.org/
I'm so glad it's Friday. Here's the v2 fix... *sigh*
https://lore.kernel.org/lkml/20210910201132.3809437-1-keescook@chromium.org/
-Kees
>
> >
> > Makefile | 15 +++
> > drivers/of/kexec.c | 1
> > drivers/rapidio/devices/rio_mport_cdev.c | 9 +-
> > include/linux/compiler_attributes.h | 6 +
> > include/linux/gfp.h | 2
> > include/linux/mm.h | 34 --------
> > include/linux/percpu.h | 3
> > include/linux/slab.h | 122 ++++++++++++++++++++++---------
> > include/linux/vmalloc.h | 11 ++
> > scripts/checkpatch.pl | 3
> > 10 files changed, 132 insertions(+), 74 deletions(-)
> >
>
> --
> Kees Cook
--
Kees Cook
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2021-09-10 3:09 incoming Andrew Morton
@ 2021-09-10 17:11 ` Kees Cook
2021-09-10 20:13 ` incoming Kees Cook
0 siblings, 1 reply; 409+ messages in thread
From: Kees Cook @ 2021-09-10 17:11 UTC (permalink / raw)
To: Linus Torvalds, Andrew Morton; +Cc: linux-mm, mm-commits
On Thu, Sep 09, 2021 at 08:09:48PM -0700, Andrew Morton wrote:
>
> More post linux-next material.
>
> 9 patches, based on f154c806676ad7153c6e161f30c53a44855329d6.
>
> Subsystems affected by this patch series:
>
> mm/slab-generic
> rapidio
> mm/debug
>
> Subsystem: mm/slab-generic
>
> "Matthew Wilcox (Oracle)" <willy@infradead.org>:
> mm: move kvmalloc-related functions to slab.h
>
> Subsystem: rapidio
>
> Kees Cook <keescook@chromium.org>:
> rapidio: avoid bogus __alloc_size warning
>
> Subsystem: mm/debug
>
> Kees Cook <keescook@chromium.org>:
> Patch series "Add __alloc_size() for better bounds checking", v2:
> Compiler Attributes: add __alloc_size() for better bounds checking
> checkpatch: add __alloc_size() to known $Attribute
> slab: clean up function declarations
> slab: add __alloc_size attributes for better bounds checking
> mm/page_alloc: add __alloc_size attributes for better bounds checking
> percpu: add __alloc_size attributes for better bounds checking
> mm/vmalloc: add __alloc_size attributes for better bounds checking
Hi,
FYI, in overnight build testing I found yet another corner case in
GCC's handling of the __alloc_size attribute. It's the gift that keeps
on giving. The fix is here:
https://lore.kernel.org/lkml/20210910165851.3296624-1-keescook@chromium.org/
>
> Makefile | 15 +++
> drivers/of/kexec.c | 1
> drivers/rapidio/devices/rio_mport_cdev.c | 9 +-
> include/linux/compiler_attributes.h | 6 +
> include/linux/gfp.h | 2
> include/linux/mm.h | 34 --------
> include/linux/percpu.h | 3
> include/linux/slab.h | 122 ++++++++++++++++++++++---------
> include/linux/vmalloc.h | 11 ++
> scripts/checkpatch.pl | 3
> 10 files changed, 132 insertions(+), 74 deletions(-)
>
--
Kees Cook
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-09-10 3:09 Andrew Morton
2021-09-10 17:11 ` incoming Kees Cook
0 siblings, 1 reply; 409+ messages in thread
From: Andrew Morton @ 2021-09-10 3:09 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
More post linux-next material.
9 patches, based on f154c806676ad7153c6e161f30c53a44855329d6.
Subsystems affected by this patch series:
mm/slab-generic
rapidio
mm/debug
Subsystem: mm/slab-generic
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
mm: move kvmalloc-related functions to slab.h
Subsystem: rapidio
Kees Cook <keescook@chromium.org>:
rapidio: avoid bogus __alloc_size warning
Subsystem: mm/debug
Kees Cook <keescook@chromium.org>:
Patch series "Add __alloc_size() for better bounds checking", v2:
Compiler Attributes: add __alloc_size() for better bounds checking
checkpatch: add __alloc_size() to known $Attribute
slab: clean up function declarations
slab: add __alloc_size attributes for better bounds checking
mm/page_alloc: add __alloc_size attributes for better bounds checking
percpu: add __alloc_size attributes for better bounds checking
mm/vmalloc: add __alloc_size attributes for better bounds checking
Makefile | 15 +++
drivers/of/kexec.c | 1
drivers/rapidio/devices/rio_mport_cdev.c | 9 +-
include/linux/compiler_attributes.h | 6 +
include/linux/gfp.h | 2
include/linux/mm.h | 34 --------
include/linux/percpu.h | 3
include/linux/slab.h | 122 ++++++++++++++++++++++---------
include/linux/vmalloc.h | 11 ++
scripts/checkpatch.pl | 3
10 files changed, 132 insertions(+), 74 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-09-09 1:08 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-09-09 1:08 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
A bunch of hotfixes, mostly cc:stable.
8 patches, based on 2d338201d5311bcd79d42f66df4cecbcbc5f4f2c.
Subsystems affected by this patch series:
mm/hmm
mm/hugetlb
mm/vmscan
mm/pagealloc
mm/pagemap
mm/kmemleak
mm/mempolicy
mm/memblock
Subsystem: mm/hmm
Li Zhijian <lizhijian@cn.fujitsu.com>:
mm/hmm: bypass devmap pte when all pfn requested flags are fulfilled
Subsystem: mm/hugetlb
Liu Zixian <liuzixian4@huawei.com>:
mm/hugetlb: initialize hugetlb_usage in mm_init
Subsystem: mm/vmscan
Rik van Riel <riel@surriel.com>:
mm,vmscan: fix divide by zero in get_scan_count
Subsystem: mm/pagealloc
Miaohe Lin <linmiaohe@huawei.com>:
mm/page_alloc.c: avoid accessing uninitialized pcp page migratetype
Subsystem: mm/pagemap
Liam Howlett <liam.howlett@oracle.com>:
mmap_lock: change trace and locking order
Subsystem: mm/kmemleak
Naohiro Aota <naohiro.aota@wdc.com>:
mm/kmemleak: allow __GFP_NOLOCKDEP passed to kmemleak's gfp
Subsystem: mm/mempolicy
yanghui <yanghui.def@bytedance.com>:
mm/mempolicy: fix a race between offset_il_node and mpol_rebind_task
Subsystem: mm/memblock
Mike Rapoport <rppt@linux.ibm.com>:
nds32/setup: remove unused memblock_region variable in setup_memory()
arch/nds32/kernel/setup.c | 1 -
include/linux/hugetlb.h | 9 +++++++++
include/linux/mmap_lock.h | 8 ++++----
kernel/fork.c | 1 +
mm/hmm.c | 5 ++++-
mm/kmemleak.c | 3 ++-
mm/mempolicy.c | 17 +++++++++++++----
mm/page_alloc.c | 4 +++-
mm/vmscan.c | 2 +-
9 files changed, 37 insertions(+), 13 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-09-08 22:17 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-09-08 22:17 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
This is the post-linux-next material, so it is based upon latest
upstream to catch the now-merged dependencies.
10 patches, based on 2d338201d5311bcd79d42f66df4cecbcbc5f4f2c.
Subsystems affected by this patch series:
mm/vmstat
mm/migration
compat
Subsystem: mm/vmstat
Ingo Molnar <mingo@elte.hu>:
mm/vmstat: protect per cpu variables with preempt disable on RT
Subsystem: mm/migration
Baolin Wang <baolin.wang@linux.alibaba.com>:
mm: migrate: introduce a local variable to get the number of pages
mm: migrate: fix the incorrect function name in comments
mm: migrate: change to use bool type for 'page_was_mapped'
Subsystem: compat
Arnd Bergmann <arnd@arndb.de>:
Patch series "compat: remove compat_alloc_user_space", v5:
kexec: move locking into do_kexec_load
kexec: avoid compat_alloc_user_space
mm: simplify compat_sys_move_pages
mm: simplify compat numa syscalls
compat: remove some compat entry points
arch: remove compat_alloc_user_space
arch/arm64/include/asm/compat.h | 5
arch/arm64/include/asm/uaccess.h | 11 -
arch/arm64/include/asm/unistd32.h | 10 -
arch/arm64/lib/Makefile | 2
arch/arm64/lib/copy_in_user.S | 77 ----------
arch/mips/cavium-octeon/octeon-memcpy.S | 2
arch/mips/include/asm/compat.h | 8 -
arch/mips/include/asm/uaccess.h | 26 ---
arch/mips/kernel/syscalls/syscall_n32.tbl | 10 -
arch/mips/kernel/syscalls/syscall_o32.tbl | 10 -
arch/mips/lib/memcpy.S | 11 -
arch/parisc/include/asm/compat.h | 6
arch/parisc/include/asm/uaccess.h | 2
arch/parisc/kernel/syscalls/syscall.tbl | 8 -
arch/parisc/lib/memcpy.c | 9 -
arch/powerpc/include/asm/compat.h | 16 --
arch/powerpc/kernel/syscalls/syscall.tbl | 10 -
arch/s390/include/asm/compat.h | 10 -
arch/s390/include/asm/uaccess.h | 3
arch/s390/kernel/syscalls/syscall.tbl | 10 -
arch/s390/lib/uaccess.c | 63 --------
arch/sparc/include/asm/compat.h | 19 --
arch/sparc/kernel/process_64.c | 2
arch/sparc/kernel/signal32.c | 12 -
arch/sparc/kernel/signal_64.c | 8 -
arch/sparc/kernel/syscalls/syscall.tbl | 10 -
arch/x86/entry/syscalls/syscall_32.tbl | 4
arch/x86/entry/syscalls/syscall_64.tbl | 2
arch/x86/include/asm/compat.h | 13 -
arch/x86/include/asm/uaccess_64.h | 7
include/linux/compat.h | 39 +----
include/linux/uaccess.h | 10 -
include/uapi/asm-generic/unistd.h | 10 -
kernel/compat.c | 21 --
kernel/kexec.c | 105 +++++---------
kernel/sys_ni.c | 5
mm/mempolicy.c | 213 +++++++-----------------------
mm/migrate.c | 69 +++++----
mm/vmstat.c | 48 ++++++
39 files changed, 243 insertions(+), 663 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2021-09-08 2:52 incoming Andrew Morton
@ 2021-09-08 8:57 ` Vlastimil Babka
0 siblings, 0 replies; 409+ messages in thread
From: Vlastimil Babka @ 2021-09-08 8:57 UTC (permalink / raw)
To: Andrew Morton, Linus Torvalds
Cc: linux-mm, mm-commits, Mike Galbraith, Mel Gorman
On 9/8/21 04:52, Andrew Morton wrote:
> Subsystem: mm/slub
>
> Vlastimil Babka <vbabka@suse.cz>:
> Patch series "SLUB: reduce irq disabled scope and make it RT compatible", v6:
> mm, slub: don't call flush_all() from slab_debug_trace_open()
> mm, slub: allocate private object map for debugfs listings
> mm, slub: allocate private object map for validate_slab_cache()
> mm, slub: don't disable irq for debug_check_no_locks_freed()
> mm, slub: remove redundant unfreeze_partials() from put_cpu_partial()
> mm, slub: extract get_partial() from new_slab_objects()
> mm, slub: dissolve new_slab_objects() into ___slab_alloc()
> mm, slub: return slab page from get_partial() and set c->page afterwards
> mm, slub: restructure new page checks in ___slab_alloc()
> mm, slub: simplify kmem_cache_cpu and tid setup
> mm, slub: move disabling/enabling irqs to ___slab_alloc()
> mm, slub: do initial checks in ___slab_alloc() with irqs enabled
> mm, slub: move disabling irqs closer to get_partial() in ___slab_alloc()
> mm, slub: restore irqs around calling new_slab()
> mm, slub: validate slab from partial list or page allocator before making it cpu slab
> mm, slub: check new pages with restored irqs
> mm, slub: stop disabling irqs around get_partial()
> mm, slub: move reset of c->page and freelist out of deactivate_slab()
> mm, slub: make locking in deactivate_slab() irq-safe
> mm, slub: call deactivate_slab() without disabling irqs
> mm, slub: move irq control into unfreeze_partials()
> mm, slub: discard slabs in unfreeze_partials() without irqs disabled
> mm, slub: detach whole partial list at once in unfreeze_partials()
> mm, slub: separate detaching of partial list in unfreeze_partials() from unfreezing
> mm, slub: only disable irq with spin_lock in __unfreeze_partials()
> mm, slub: don't disable irqs in slub_cpu_dead()
> mm, slab: split out the cpu offline variant of flush_slab()
>
> Sebastian Andrzej Siewior <bigeasy@linutronix.de>:
> mm: slub: move flush_cpu_slab() invocations __free_slab() invocations out of IRQ context
> mm: slub: make object_map_lock a raw_spinlock_t
>
> Vlastimil Babka <vbabka@suse.cz>:
> mm, slub: make slab_lock() disable irqs with PREEMPT_RT
> mm, slub: protect put_cpu_partial() with disabled irqs instead of cmpxchg
> mm, slub: use migrate_disable() on PREEMPT_RT
> mm, slub: convert kmem_cpu_slab protection to local_lock
For my own piece of mind, I've checked that this part (patches 1 to 33)
are identical to the v6 posting [1] and git version [2] that Mel and
Mike tested (replies to [1]).
[1] https://lore.kernel.org/all/20210904105003.11688-1-vbabka@suse.cz/
[2] git://git.kernel.org/pub/scm/linux/kernel/git/vbabka/linux.git
tags/mm-slub-5.15-rc1
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-09-08 2:52 Andrew Morton
2021-09-08 8:57 ` incoming Vlastimil Babka
0 siblings, 1 reply; 409+ messages in thread
From: Andrew Morton @ 2021-09-08 2:52 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
147 patches, based on 7d2a07b769330c34b4deabeed939325c77a7ec2f.
Subsystems affected by this patch series:
mm/slub
mm/memory-hotplug
mm/rmap
mm/ioremap
mm/highmem
mm/cleanups
mm/secretmem
mm/kfence
mm/damon
alpha
percpu
procfs
misc
core-kernel
MAINTAINERS
lib
bitops
checkpatch
epoll
init
nilfs2
coredump
fork
pids
criu
kconfig
selftests
ipc
mm/vmscan
scripts
Subsystem: mm/slub
Vlastimil Babka <vbabka@suse.cz>:
Patch series "SLUB: reduce irq disabled scope and make it RT compatible", v6:
mm, slub: don't call flush_all() from slab_debug_trace_open()
mm, slub: allocate private object map for debugfs listings
mm, slub: allocate private object map for validate_slab_cache()
mm, slub: don't disable irq for debug_check_no_locks_freed()
mm, slub: remove redundant unfreeze_partials() from put_cpu_partial()
mm, slub: extract get_partial() from new_slab_objects()
mm, slub: dissolve new_slab_objects() into ___slab_alloc()
mm, slub: return slab page from get_partial() and set c->page afterwards
mm, slub: restructure new page checks in ___slab_alloc()
mm, slub: simplify kmem_cache_cpu and tid setup
mm, slub: move disabling/enabling irqs to ___slab_alloc()
mm, slub: do initial checks in ___slab_alloc() with irqs enabled
mm, slub: move disabling irqs closer to get_partial() in ___slab_alloc()
mm, slub: restore irqs around calling new_slab()
mm, slub: validate slab from partial list or page allocator before making it cpu slab
mm, slub: check new pages with restored irqs
mm, slub: stop disabling irqs around get_partial()
mm, slub: move reset of c->page and freelist out of deactivate_slab()
mm, slub: make locking in deactivate_slab() irq-safe
mm, slub: call deactivate_slab() without disabling irqs
mm, slub: move irq control into unfreeze_partials()
mm, slub: discard slabs in unfreeze_partials() without irqs disabled
mm, slub: detach whole partial list at once in unfreeze_partials()
mm, slub: separate detaching of partial list in unfreeze_partials() from unfreezing
mm, slub: only disable irq with spin_lock in __unfreeze_partials()
mm, slub: don't disable irqs in slub_cpu_dead()
mm, slab: split out the cpu offline variant of flush_slab()
Sebastian Andrzej Siewior <bigeasy@linutronix.de>:
mm: slub: move flush_cpu_slab() invocations __free_slab() invocations out of IRQ context
mm: slub: make object_map_lock a raw_spinlock_t
Vlastimil Babka <vbabka@suse.cz>:
mm, slub: make slab_lock() disable irqs with PREEMPT_RT
mm, slub: protect put_cpu_partial() with disabled irqs instead of cmpxchg
mm, slub: use migrate_disable() on PREEMPT_RT
mm, slub: convert kmem_cpu_slab protection to local_lock
Subsystem: mm/memory-hotplug
David Hildenbrand <david@redhat.com>:
Patch series "memory-hotplug.rst: complete admin-guide overhaul", v3:
memory-hotplug.rst: remove locking details from admin-guide
memory-hotplug.rst: complete admin-guide overhaul
Mike Rapoport <rppt@linux.ibm.com>:
Patch series "mm: remove pfn_valid_within() and CONFIG_HOLES_IN_ZONE":
mm: remove pfn_valid_within() and CONFIG_HOLES_IN_ZONE
mm: memory_hotplug: cleanup after removal of pfn_valid_within()
David Hildenbrand <david@redhat.com>:
Patch series "mm/memory_hotplug: preparatory patches for new online policy and memory":
mm/memory_hotplug: use "unsigned long" for PFN in zone_for_pfn_range()
mm/memory_hotplug: remove nid parameter from arch_remove_memory()
mm/memory_hotplug: remove nid parameter from remove_memory() and friends
ACPI: memhotplug: memory resources cannot be enabled yet
Patch series "mm/memory_hotplug: "auto-movable" online policy and memory groups", v3:
mm: track present early pages per zone
mm/memory_hotplug: introduce "auto-movable" online policy
drivers/base/memory: introduce "memory groups" to logically group memory blocks
mm/memory_hotplug: track present pages in memory groups
ACPI: memhotplug: use a single static memory group for a single memory device
dax/kmem: use a single static memory group for a single probed unit
virtio-mem: use a single dynamic memory group for a single virtio-mem device
mm/memory_hotplug: memory group aware "auto-movable" online policy
mm/memory_hotplug: improved dynamic memory group aware "auto-movable" online policy
Miaohe Lin <linmiaohe@huawei.com>:
Patch series "Cleanup and fixups for memory hotplug":
mm/memory_hotplug: use helper zone_is_zone_device() to simplify the code
Subsystem: mm/rmap
Muchun Song <songmuchun@bytedance.com>:
mm: remove redundant compound_head() calling
Subsystem: mm/ioremap
Christoph Hellwig <hch@lst.de>:
riscv: only select GENERIC_IOREMAP if MMU support is enabled
Patch series "small ioremap cleanups":
mm: move ioremap_page_range to vmalloc.c
mm: don't allow executable ioremap mappings
Weizhao Ouyang <o451686892@gmail.com>:
mm/early_ioremap.c: remove redundant early_ioremap_shutdown()
Subsystem: mm/highmem
Sebastian Andrzej Siewior <bigeasy@linutronix.de>:
highmem: don't disable preemption on RT in kmap_atomic()
Subsystem: mm/cleanups
Changbin Du <changbin.du@gmail.com>:
mm: in_irq() cleanup
Muchun Song <songmuchun@bytedance.com>:
mm: introduce PAGEFLAGS_MASK to replace ((1UL << NR_PAGEFLAGS) - 1)
Subsystem: mm/secretmem
Jordy Zomer <jordy@jordyzomer.github.io>:
mm/secretmem: use refcount_t instead of atomic_t
Subsystem: mm/kfence
Marco Elver <elver@google.com>:
kfence: show cpu and timestamp in alloc/free info
kfence: test: fail fast if disabled at boot
Subsystem: mm/damon
SeongJae Park <sjpark@amazon.de>:
Patch series "Introduce Data Access MONitor (DAMON)", v34:
mm: introduce Data Access MONitor (DAMON)
mm/damon/core: implement region-based sampling
mm/damon: adaptively adjust regions
mm/idle_page_tracking: make PG_idle reusable
mm/damon: implement primitives for the virtual memory address spaces
mm/damon: add a tracepoint
mm/damon: implement a debugfs-based user space interface
mm/damon/dbgfs: export kdamond pid to the user space
mm/damon/dbgfs: support multiple contexts
Documentation: add documents for DAMON
mm/damon: add kunit tests
mm/damon: add user space selftests
MAINTAINERS: update for DAMON
Subsystem: alpha
Randy Dunlap <rdunlap@infradead.org>:
alpha: agp: make empty macros use do-while-0 style
alpha: pci-sysfs: fix all kernel-doc warnings
Subsystem: percpu
Greg Kroah-Hartman <gregkh@linuxfoundation.org>:
percpu: remove export of pcpu_base_addr
Subsystem: procfs
Feng Zhou <zhoufeng.zf@bytedance.com>:
fs/proc/kcore.c: add mmap interface
Christoph Hellwig <hch@lst.de>:
proc: stop using seq_get_buf in proc_task_name
Ohhoon Kwon <ohoono.kwon@samsung.com>:
connector: send event on write to /proc/[pid]/comm
Subsystem: misc
Colin Ian King <colin.king@canonical.com>:
arch: Kconfig: fix spelling mistake "seperate" -> "separate"
Andy Shevchenko <andriy.shevchenko@linux.intel.com>:
include/linux/once.h: fix trivia typo Not -> Note
Daniel Lezcano <daniel.lezcano@linaro.org>:
Patch series "Add Hz macros", v3:
units: change from 'L' to 'UL'
units: add the HZ macros
thermal/drivers/devfreq_cooling: use HZ macros
devfreq: use HZ macros
iio/drivers/as73211: use HZ macros
hwmon/drivers/mr75203: use HZ macros
iio/drivers/hid-sensor: use HZ macros
i2c/drivers/ov02q10: use HZ macros
mtd/drivers/nand: use HZ macros
phy/drivers/stm32: use HZ macros
Subsystem: core-kernel
Yang Yang <yang.yang29@zte.com.cn>:
kernel/acct.c: use dedicated helper to access rlimit values
Pavel Skripkin <paskripkin@gmail.com>:
profiling: fix shift-out-of-bounds bugs
Subsystem: MAINTAINERS
Nathan Chancellor <nathan@kernel.org>:
MAINTAINERS: update ClangBuiltLinux mailing list
Documentation/llvm: update mailing list
Documentation/llvm: update IRC location
Subsystem: lib
Geert Uytterhoeven <geert@linux-m68k.org>:
Patch series "math: RATIONAL and RATIONAL_KUNIT_TEST improvements":
math: make RATIONAL tristate
math: RATIONAL_KUNIT_TEST should depend on RATIONAL instead of selecting it
Matteo Croce <mcroce@microsoft.com>:
Patch series "lib/string: optimized mem* functions", v2:
lib/string: optimized memcpy
lib/string: optimized memmove
lib/string: optimized memset
Daniel Latypov <dlatypov@google.com>:
lib/test: convert test_sort.c to use KUnit
Randy Dunlap <rdunlap@infradead.org>:
lib/dump_stack: correct kernel-doc notation
lib/iov_iter.c: fix kernel-doc warnings
Subsystem: bitops
Yury Norov <yury.norov@gmail.com>:
Patch series "Resend bitmap patches":
bitops: protect find_first_{,zero}_bit properly
bitops: move find_bit_*_le functions from le.h to find.h
include: move find.h from asm_generic to linux
arch: remove GENERIC_FIND_FIRST_BIT entirely
lib: add find_first_and_bit()
cpumask: use find_first_and_bit()
all: replace find_next{,_zero}_bit with find_first{,_zero}_bit where appropriate
tools: sync tools/bitmap with mother linux
cpumask: replace cpumask_next_* with cpumask_first_* where appropriate
include/linux: move for_each_bit() macros from bitops.h to find.h
find: micro-optimize for_each_{set,clear}_bit()
bitops: replace for_each_*_bit_from() with for_each_*_bit() where appropriate
Andy Shevchenko <andriy.shevchenko@linux.intel.com>:
tools: rename bitmap_alloc() to bitmap_zalloc()
Yury Norov <yury.norov@gmail.com>:
mm/percpu: micro-optimize pcpu_is_populated()
bitmap: unify find_bit operations
lib: bitmap: add performance test for bitmap_print_to_pagebuf
vsprintf: rework bitmap_list_string
Subsystem: checkpatch
Joe Perches <joe@perches.com>:
checkpatch: support wide strings
Mimi Zohar <zohar@linux.ibm.com>:
checkpatch: make email address check case insensitive
Joe Perches <joe@perches.com>:
checkpatch: improve GIT_COMMIT_ID test
Subsystem: epoll
Nicholas Piggin <npiggin@gmail.com>:
fs/epoll: use a per-cpu counter for user's watches count
Subsystem: init
Rasmus Villemoes <linux@rasmusvillemoes.dk>:
init: move usermodehelper_enable() to populate_rootfs()
Kefeng Wang <wangkefeng.wang@huawei.com>:
trap: cleanup trap_init()
Subsystem: nilfs2
Nanyong Sun <sunnanyong@huawei.com>:
Patch series "nilfs2: fix incorrect usage of kobject":
nilfs2: fix memory leak in nilfs_sysfs_create_device_group
nilfs2: fix NULL pointer in nilfs_##name##_attr_release
nilfs2: fix memory leak in nilfs_sysfs_create_##name##_group
nilfs2: fix memory leak in nilfs_sysfs_delete_##name##_group
nilfs2: fix memory leak in nilfs_sysfs_create_snapshot_group
nilfs2: fix memory leak in nilfs_sysfs_delete_snapshot_group
Zhen Lei <thunder.leizhen@huawei.com>:
nilfs2: use refcount_dec_and_lock() to fix potential UAF
Subsystem: coredump
David Oberhollenzer <david.oberhollenzer@sigma-star.at>:
fs/coredump.c: log if a core dump is aborted due to changed file permissions
QiuXi <qiuxi1@huawei.com>:
coredump: fix memleak in dump_vma_snapshot()
Subsystem: fork
Christoph Hellwig <hch@lst.de>:
kernel/fork.c: unexport get_{mm,task}_exe_file
Subsystem: pids
Takahiro Itazuri <itazur@amazon.com>:
pid: cleanup the stale comment mentioning pidmap_init().
Subsystem: criu
Cyrill Gorcunov <gorcunov@gmail.com>:
prctl: allow to setup brk for et_dyn executables
Subsystem: kconfig
Zenghui Yu <yuzenghui@huawei.com>:
configs: remove the obsolete CONFIG_INPUT_POLLDEV
Lukas Bulwahn <lukas.bulwahn@gmail.com>:
Kconfig.debug: drop selecting non-existing HARDLOCKUP_DETECTOR_ARCH
Subsystem: selftests
Greg Thelen <gthelen@google.com>:
selftests/memfd: remove unused variable
Subsystem: ipc
Rafael Aquini <aquini@redhat.com>:
ipc: replace costly bailout check in sysvipc_find_ipc()
Subsystem: mm/vmscan
Randy Dunlap <rdunlap@infradead.org>:
mm/workingset: correct kernel-doc notations
Subsystem: scripts
Randy Dunlap <rdunlap@infradead.org>:
scripts: check_extable: fix typo in user error message
a/Documentation/admin-guide/mm/damon/index.rst | 15
a/Documentation/admin-guide/mm/damon/start.rst | 114 +
a/Documentation/admin-guide/mm/damon/usage.rst | 112 +
a/Documentation/admin-guide/mm/index.rst | 1
a/Documentation/admin-guide/mm/memory-hotplug.rst | 842 ++++++-----
a/Documentation/dev-tools/kfence.rst | 98 -
a/Documentation/kbuild/llvm.rst | 5
a/Documentation/vm/damon/api.rst | 20
a/Documentation/vm/damon/design.rst | 166 ++
a/Documentation/vm/damon/faq.rst | 51
a/Documentation/vm/damon/index.rst | 30
a/Documentation/vm/index.rst | 1
a/MAINTAINERS | 17
a/arch/Kconfig | 2
a/arch/alpha/include/asm/agp.h | 4
a/arch/alpha/include/asm/bitops.h | 2
a/arch/alpha/kernel/pci-sysfs.c | 12
a/arch/arc/Kconfig | 1
a/arch/arc/include/asm/bitops.h | 1
a/arch/arc/kernel/traps.c | 5
a/arch/arm/configs/dove_defconfig | 1
a/arch/arm/configs/pxa_defconfig | 1
a/arch/arm/include/asm/bitops.h | 1
a/arch/arm/kernel/traps.c | 5
a/arch/arm64/Kconfig | 1
a/arch/arm64/include/asm/bitops.h | 1
a/arch/arm64/mm/mmu.c | 3
a/arch/csky/include/asm/bitops.h | 1
a/arch/h8300/include/asm/bitops.h | 1
a/arch/h8300/kernel/traps.c | 4
a/arch/hexagon/include/asm/bitops.h | 1
a/arch/hexagon/kernel/traps.c | 4
a/arch/ia64/include/asm/bitops.h | 2
a/arch/ia64/mm/init.c | 3
a/arch/m68k/include/asm/bitops.h | 2
a/arch/mips/Kconfig | 1
a/arch/mips/configs/lemote2f_defconfig | 1
a/arch/mips/configs/pic32mzda_defconfig | 1
a/arch/mips/configs/rt305x_defconfig | 1
a/arch/mips/configs/xway_defconfig | 1
a/arch/mips/include/asm/bitops.h | 1
a/arch/nds32/kernel/traps.c | 5
a/arch/nios2/kernel/traps.c | 5
a/arch/openrisc/include/asm/bitops.h | 1
a/arch/openrisc/kernel/traps.c | 5
a/arch/parisc/configs/generic-32bit_defconfig | 1
a/arch/parisc/include/asm/bitops.h | 2
a/arch/parisc/kernel/traps.c | 4
a/arch/powerpc/include/asm/bitops.h | 2
a/arch/powerpc/include/asm/cputhreads.h | 2
a/arch/powerpc/kernel/traps.c | 5
a/arch/powerpc/mm/mem.c | 3
a/arch/powerpc/platforms/pasemi/dma_lib.c | 4
a/arch/powerpc/platforms/pseries/hotplug-memory.c | 9
a/arch/riscv/Kconfig | 2
a/arch/riscv/include/asm/bitops.h | 1
a/arch/riscv/kernel/traps.c | 5
a/arch/s390/Kconfig | 1
a/arch/s390/include/asm/bitops.h | 1
a/arch/s390/kvm/kvm-s390.c | 2
a/arch/s390/mm/init.c | 3
a/arch/sh/include/asm/bitops.h | 1
a/arch/sh/mm/init.c | 3
a/arch/sparc/include/asm/bitops_32.h | 1
a/arch/sparc/include/asm/bitops_64.h | 2
a/arch/um/kernel/trap.c | 4
a/arch/x86/Kconfig | 1
a/arch/x86/configs/i386_defconfig | 1
a/arch/x86/configs/x86_64_defconfig | 1
a/arch/x86/include/asm/bitops.h | 2
a/arch/x86/kernel/apic/vector.c | 4
a/arch/x86/mm/init_32.c | 3
a/arch/x86/mm/init_64.c | 3
a/arch/x86/um/Kconfig | 1
a/arch/xtensa/include/asm/bitops.h | 1
a/block/blk-mq.c | 2
a/drivers/acpi/acpi_memhotplug.c | 46
a/drivers/base/memory.c | 231 ++-
a/drivers/base/node.c | 2
a/drivers/block/rnbd/rnbd-clt.c | 2
a/drivers/dax/kmem.c | 43
a/drivers/devfreq/devfreq.c | 2
a/drivers/dma/ti/edma.c | 2
a/drivers/gpu/drm/etnaviv/etnaviv_gpu.c | 4
a/drivers/hwmon/ltc2992.c | 3
a/drivers/hwmon/mr75203.c | 2
a/drivers/iio/adc/ad7124.c | 2
a/drivers/iio/common/hid-sensors/hid-sensor-attributes.c | 3
a/drivers/iio/light/as73211.c | 3
a/drivers/infiniband/hw/irdma/hw.c | 16
a/drivers/media/cec/core/cec-core.c | 2
a/drivers/media/i2c/ov02a10.c | 2
a/drivers/media/mc/mc-devnode.c | 2
a/drivers/mmc/host/renesas_sdhi_core.c | 2
a/drivers/mtd/nand/raw/intel-nand-controller.c | 2
a/drivers/net/virtio_net.c | 2
a/drivers/pci/controller/dwc/pci-dra7xx.c | 2
a/drivers/phy/st/phy-stm32-usbphyc.c | 2
a/drivers/scsi/lpfc/lpfc_sli.c | 10
a/drivers/soc/fsl/qbman/bman_portal.c | 2
a/drivers/soc/fsl/qbman/qman_portal.c | 2
a/drivers/soc/ti/k3-ringacc.c | 4
a/drivers/thermal/devfreq_cooling.c | 2
a/drivers/tty/n_tty.c | 2
a/drivers/virt/acrn/ioreq.c | 3
a/drivers/virtio/virtio_mem.c | 26
a/fs/coredump.c | 15
a/fs/eventpoll.c | 18
a/fs/f2fs/segment.c | 8
a/fs/nilfs2/sysfs.c | 26
a/fs/nilfs2/the_nilfs.c | 9
a/fs/ocfs2/cluster/heartbeat.c | 2
a/fs/ocfs2/dlm/dlmdomain.c | 4
a/fs/ocfs2/dlm/dlmmaster.c | 18
a/fs/ocfs2/dlm/dlmrecovery.c | 2
a/fs/ocfs2/dlm/dlmthread.c | 2
a/fs/proc/array.c | 18
a/fs/proc/base.c | 5
a/fs/proc/kcore.c | 73
a/include/asm-generic/bitops.h | 1
a/include/asm-generic/bitops/find.h | 198 --
a/include/asm-generic/bitops/le.h | 64
a/include/asm-generic/early_ioremap.h | 6
a/include/linux/bitmap.h | 34
a/include/linux/bitops.h | 34
a/include/linux/cpumask.h | 46
a/include/linux/damon.h | 290 +++
a/include/linux/find.h | 134 +
a/include/linux/highmem-internal.h | 27
a/include/linux/memory.h | 55
a/include/linux/memory_hotplug.h | 40
a/include/linux/mmzone.h | 19
a/include/linux/once.h | 2
a/include/linux/page-flags.h | 17
a/include/linux/page_ext.h | 2
a/include/linux/page_idle.h | 6
a/include/linux/pagemap.h | 7
a/include/linux/sched/user.h | 3
a/include/linux/slub_def.h | 6
a/include/linux/threads.h | 2
a/include/linux/units.h | 10
a/include/linux/vmalloc.h | 3
a/include/trace/events/damon.h | 43
a/include/trace/events/mmflags.h | 2
a/include/trace/events/page_ref.h | 4
a/init/initramfs.c | 2
a/init/main.c | 3
a/init/noinitramfs.c | 2
a/ipc/util.c | 16
a/kernel/acct.c | 2
a/kernel/fork.c | 2
a/kernel/profile.c | 21
a/kernel/sys.c | 7
a/kernel/time/clocksource.c | 4
a/kernel/user.c | 25
a/lib/Kconfig | 3
a/lib/Kconfig.debug | 9
a/lib/dump_stack.c | 3
a/lib/find_bit.c | 21
a/lib/find_bit_benchmark.c | 21
a/lib/genalloc.c | 2
a/lib/iov_iter.c | 8
a/lib/math/Kconfig | 2
a/lib/math/rational.c | 3
a/lib/string.c | 130 +
a/lib/test_bitmap.c | 37
a/lib/test_printf.c | 2
a/lib/test_sort.c | 40
a/lib/vsprintf.c | 26
a/mm/Kconfig | 15
a/mm/Makefile | 4
a/mm/compaction.c | 20
a/mm/damon/Kconfig | 68
a/mm/damon/Makefile | 5
a/mm/damon/core-test.h | 253 +++
a/mm/damon/core.c | 748 ++++++++++
a/mm/damon/dbgfs-test.h | 126 +
a/mm/damon/dbgfs.c | 631 ++++++++
a/mm/damon/vaddr-test.h | 329 ++++
a/mm/damon/vaddr.c | 672 +++++++++
a/mm/early_ioremap.c | 5
a/mm/highmem.c | 2
a/mm/ioremap.c | 25
a/mm/kfence/core.c | 3
a/mm/kfence/kfence.h | 2
a/mm/kfence/kfence_test.c | 3
a/mm/kfence/report.c | 19
a/mm/kmemleak.c | 2
a/mm/memory_hotplug.c | 396 ++++-
a/mm/memremap.c | 5
a/mm/page_alloc.c | 27
a/mm/page_ext.c | 12
a/mm/page_idle.c | 10
a/mm/page_isolation.c | 7
a/mm/page_owner.c | 14
a/mm/percpu.c | 36
a/mm/rmap.c | 6
a/mm/secretmem.c | 9
a/mm/slab_common.c | 2
a/mm/slub.c | 1023 +++++++++-----
a/mm/vmalloc.c | 24
a/mm/workingset.c | 2
a/net/ncsi/ncsi-manage.c | 4
a/scripts/check_extable.sh | 2
a/scripts/checkpatch.pl | 93 -
a/tools/include/linux/bitmap.h | 4
a/tools/perf/bench/find-bit-bench.c | 2
a/tools/perf/builtin-c2c.c | 6
a/tools/perf/builtin-record.c | 2
a/tools/perf/tests/bitmap.c | 2
a/tools/perf/tests/mem2node.c | 2
a/tools/perf/util/affinity.c | 4
a/tools/perf/util/header.c | 4
a/tools/perf/util/metricgroup.c | 2
a/tools/perf/util/mmap.c | 4
a/tools/testing/selftests/damon/Makefile | 7
a/tools/testing/selftests/damon/_chk_dependency.sh | 28
a/tools/testing/selftests/damon/debugfs_attrs.sh | 75 +
a/tools/testing/selftests/kvm/dirty_log_perf_test.c | 2
a/tools/testing/selftests/kvm/dirty_log_test.c | 4
a/tools/testing/selftests/kvm/x86_64/vmx_dirty_log_test.c | 2
a/tools/testing/selftests/memfd/memfd_test.c | 2
b/MAINTAINERS | 2
b/tools/include/asm-generic/bitops.h | 1
b/tools/include/linux/bitmap.h | 7
b/tools/include/linux/find.h | 81 +
b/tools/lib/find_bit.c | 20
227 files changed, 6695 insertions(+), 1875 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2021-09-02 21:48 incoming Andrew Morton
@ 2021-09-02 21:49 ` Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-09-02 21:49 UTC (permalink / raw)
To: Linus Torvalds, linux-mm, mm-commits
On Thu, 2 Sep 2021 14:48:20 -0700 Andrew Morton <akpm@linux-foundation.org> wrote:
> 212 patches, based on 4a3bb4200a5958d76cc26ebe4db4257efa56812b.
Make that "based on 7d2a07b769330c34b4deabeed939325c77a7ec2f".
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-09-02 21:48 Andrew Morton
2021-09-02 21:49 ` incoming Andrew Morton
0 siblings, 1 reply; 409+ messages in thread
From: Andrew Morton @ 2021-09-02 21:48 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
212 patches, based on 4a3bb4200a5958d76cc26ebe4db4257efa56812b.
Subsystems affected by this patch series:
ia64
ocfs2
block
mm/slub
mm/debug
mm/pagecache
mm/gup
mm/swap
mm/shmem
mm/memcg
mm/selftests
mm/pagemap
mm/mremap
mm/bootmem
mm/sparsemem
mm/vmalloc
mm/kasan
mm/pagealloc
mm/memory-failure
mm/hugetlb
mm/userfaultfd
mm/vmscan
mm/compaction
mm/mempolicy
mm/memblock
mm/oom-kill
mm/migration
mm/ksm
mm/percpu
mm/vmstat
mm/madvise
Subsystem: ia64
Jason Wang <wangborong@cdjrlc.com>:
ia64: fix typo in a comment
Geert Uytterhoeven <geert+renesas@glider.be>:
Patch series "ia64: Miscellaneous fixes and cleanups":
ia64: fix #endif comment for reserve_elfcorehdr()
ia64: make reserve_elfcorehdr() static
ia64: make num_rsvd_regions static
Subsystem: ocfs2
Dan Carpenter <dan.carpenter@oracle.com>:
ocfs2: remove an unnecessary condition
Tuo Li <islituo@gmail.com>:
ocfs2: quota_local: fix possible uninitialized-variable access in ocfs2_local_read_info()
Gang He <ghe@suse.com>:
ocfs2: ocfs2_downconvert_lock failure results in deadlock
Subsystem: block
kernel test robot <lkp@intel.com>:
arch/csky/kernel/probes/kprobes.c: fix bugon.cocci warnings
Subsystem: mm/slub
Vlastimil Babka <vbabka@suse.cz>:
Patch series "SLUB: reduce irq disabled scope and make it RT compatible", v4:
mm, slub: don't call flush_all() from slab_debug_trace_open()
mm, slub: allocate private object map for debugfs listings
mm, slub: allocate private object map for validate_slab_cache()
mm, slub: don't disable irq for debug_check_no_locks_freed()
mm, slub: remove redundant unfreeze_partials() from put_cpu_partial()
mm, slub: unify cmpxchg_double_slab() and __cmpxchg_double_slab()
mm, slub: extract get_partial() from new_slab_objects()
mm, slub: dissolve new_slab_objects() into ___slab_alloc()
mm, slub: return slab page from get_partial() and set c->page afterwards
mm, slub: restructure new page checks in ___slab_alloc()
mm, slub: simplify kmem_cache_cpu and tid setup
mm, slub: move disabling/enabling irqs to ___slab_alloc()
mm, slub: do initial checks in ___slab_alloc() with irqs enabled
mm, slub: move disabling irqs closer to get_partial() in ___slab_alloc()
mm, slub: restore irqs around calling new_slab()
mm, slub: validate slab from partial list or page allocator before making it cpu slab
mm, slub: check new pages with restored irqs
mm, slub: stop disabling irqs around get_partial()
mm, slub: move reset of c->page and freelist out of deactivate_slab()
mm, slub: make locking in deactivate_slab() irq-safe
mm, slub: call deactivate_slab() without disabling irqs
mm, slub: move irq control into unfreeze_partials()
mm, slub: discard slabs in unfreeze_partials() without irqs disabled
mm, slub: detach whole partial list at once in unfreeze_partials()
mm, slub: separate detaching of partial list in unfreeze_partials() from unfreezing
mm, slub: only disable irq with spin_lock in __unfreeze_partials()
mm, slub: don't disable irqs in slub_cpu_dead()
mm, slab: make flush_slab() possible to call with irqs enabled
Sebastian Andrzej Siewior <bigeasy@linutronix.de>:
mm: slub: move flush_cpu_slab() invocations __free_slab() invocations out of IRQ context
mm: slub: make object_map_lock a raw_spinlock_t
Vlastimil Babka <vbabka@suse.cz>:
mm, slub: optionally save/restore irqs in slab_[un]lock()/
mm, slub: make slab_lock() disable irqs with PREEMPT_RT
mm, slub: protect put_cpu_partial() with disabled irqs instead of cmpxchg
mm, slub: use migrate_disable() on PREEMPT_RT
mm, slub: convert kmem_cpu_slab protection to local_lock
Subsystem: mm/debug
Gavin Shan <gshan@redhat.com>:
Patch series "mm/debug_vm_pgtable: Enhancements", v6:
mm/debug_vm_pgtable: introduce struct pgtable_debug_args
mm/debug_vm_pgtable: use struct pgtable_debug_args in basic tests
mm/debug_vm_pgtable: use struct pgtable_debug_args in leaf and savewrite tests
mm/debug_vm_pgtable: use struct pgtable_debug_args in protnone and devmap tests
mm/debug_vm_pgtable: use struct pgtable_debug_args in soft_dirty and swap tests
mm/debug_vm_pgtable: use struct pgtable_debug_args in migration and thp tests
mm/debug_vm_pgtable: use struct pgtable_debug_args in PTE modifying tests
mm/debug_vm_pgtable: use struct pgtable_debug_args in PMD modifying tests
mm/debug_vm_pgtable: use struct pgtable_debug_args in PUD modifying tests
mm/debug_vm_pgtable: use struct pgtable_debug_args in PGD and P4D modifying tests
mm/debug_vm_pgtable: remove unused code
mm/debug_vm_pgtable: fix corrupted page flag
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
mm: report a more useful address for reclaim acquisition
liuhailong <liuhailong@oppo.com>:
mm: add kernel_misc_reclaimable in show_free_areas
Subsystem: mm/pagecache
Jan Kara <jack@suse.cz>:
Patch series "writeback: Fix bandwidth estimates", v4:
writeback: track number of inodes under writeback
writeback: reliably update bandwidth estimation
writeback: fix bandwidth estimate for spiky workload
writeback: rename domain_update_bandwidth()
writeback: use READ_ONCE for unlocked reads of writeback stats
Johannes Weiner <hannes@cmpxchg.org>:
mm: remove irqsave/restore locking from contexts with irqs enabled
fs: drop_caches: fix skipping over shadow cache inodes
fs: inode: count invalidated shadow pages in pginodesteal
Shakeel Butt <shakeelb@google.com>:
writeback: memcg: simplify cgroup_writeback_by_id
Jing Yangyang <jing.yangyang@zte.com.cn>:
include/linux/buffer_head.h: fix boolreturn.cocci warnings
Subsystem: mm/gup
Miaohe Lin <linmiaohe@huawei.com>:
Patch series "Cleanups and fixup for gup":
mm: gup: remove set but unused local variable major
mm: gup: remove unneed local variable orig_refs
mm: gup: remove useless BUG_ON in __get_user_pages()
mm: gup: fix potential pgmap refcnt leak in __gup_device_huge()
mm: gup: use helper PAGE_ALIGNED in populate_vma_page_range()
John Hubbard <jhubbard@nvidia.com>:
Patch series "A few gup refactorings and documentation updates", v3:
mm/gup: documentation corrections for gup/pup
mm/gup: small refactoring: simplify try_grab_page()
mm/gup: remove try_get_page(), call try_get_compound_head() directly
Subsystem: mm/swap
Hugh Dickins <hughd@google.com>:
fs, mm: fix race in unlinking swapfile
John Hubbard <jhubbard@nvidia.com>:
mm: delete unused get_kernel_page()
Subsystem: mm/shmem
Sebastian Andrzej Siewior <bigeasy@linutronix.de>:
shmem: use raw_spinlock_t for ->stat_lock
Miaohe Lin <linmiaohe@huawei.com>:
Patch series "Cleanups for shmem":
shmem: remove unneeded variable ret
shmem: remove unneeded header file
shmem: remove unneeded function forward declaration
shmem: include header file to declare swap_info
Hugh Dickins <hughd@google.com>:
Patch series "huge tmpfs: shmem_is_huge() fixes and cleanups":
huge tmpfs: fix fallocate(vanilla) advance over huge pages
huge tmpfs: fix split_huge_page() after FALLOC_FL_KEEP_SIZE
huge tmpfs: remove shrinklist addition from shmem_setattr()
huge tmpfs: revert shmem's use of transhuge_vma_enabled()
huge tmpfs: move shmem_huge_enabled() upwards
huge tmpfs: SGP_NOALLOC to stop collapse_file() on race
huge tmpfs: shmem_is_huge(vma, inode, index)
huge tmpfs: decide stat.st_blksize by shmem_is_huge()
shmem: shmem_writepage() split unlikely i915 THP
Subsystem: mm/memcg
Suren Baghdasaryan <surenb@google.com>:
mm, memcg: add mem_cgroup_disabled checks in vmpressure and swap-related functions
mm, memcg: inline mem_cgroup_{charge/uncharge} to improve disabled memcg config
mm, memcg: inline swap-related functions to improve disabled memcg config
Vasily Averin <vvs@virtuozzo.com>:
memcg: enable accounting for pids in nested pid namespaces
Shakeel Butt <shakeelb@google.com>:
memcg: switch lruvec stats to rstat
memcg: infrastructure to flush memcg stats
Yutian Yang <nglaive@gmail.com>:
memcg: charge fs_context and legacy_fs_context
Vasily Averin <vvs@virtuozzo.com>:
Patch series "memcg accounting from OpenVZ", v7:
memcg: enable accounting for mnt_cache entries
memcg: enable accounting for pollfd and select bits arrays
memcg: enable accounting for file lock caches
memcg: enable accounting for fasync_cache
memcg: enable accounting for new namesapces and struct nsproxy
memcg: enable accounting of ipc resources
memcg: enable accounting for signals
memcg: enable accounting for posix_timers_cache slab
memcg: enable accounting for ldt_struct objects
Shakeel Butt <shakeelb@google.com>:
memcg: cleanup racy sum avoidance code
Vasily Averin <vvs@virtuozzo.com>:
memcg: replace in_interrupt() by !in_task() in active_memcg()
Baolin Wang <baolin.wang@linux.alibaba.com>:
mm: memcontrol: set the correct memcg swappiness restriction
Miaohe Lin <linmiaohe@huawei.com>:
mm, memcg: remove unused functions
mm, memcg: save some atomic ops when flush is already true
Michal Hocko <mhocko@suse.com>:
memcg: fix up drain_local_stock comment
Shakeel Butt <shakeelb@google.com>:
memcg: make memcg->event_list_lock irqsafe
Subsystem: mm/selftests
Po-Hsu Lin <po-hsu.lin@canonical.com>:
selftests/vm: use kselftest skip code for skipped tests
Colin Ian King <colin.king@canonical.com>:
selftests: Fix spelling mistake "cann't" -> "cannot"
Subsystem: mm/pagemap
Nicholas Piggin <npiggin@gmail.com>:
Patch series "shoot lazy tlbs", v4:
lazy tlb: introduce lazy mm refcount helper functions
lazy tlb: allow lazy tlb mm refcounting to be configurable
lazy tlb: shoot lazies, a non-refcounting lazy tlb option
powerpc/64s: enable MMU_LAZY_TLB_SHOOTDOWN
Christoph Hellwig <hch@lst.de>:
Patch series "_kernel_dcache_page fixes and removal":
mmc: JZ4740: remove the flush_kernel_dcache_page call in jz4740_mmc_read_data
mmc: mmc_spi: replace flush_kernel_dcache_page with flush_dcache_page
scatterlist: replace flush_kernel_dcache_page with flush_dcache_page
mm: remove flush_kernel_dcache_page
Huang Ying <ying.huang@intel.com>:
mm,do_huge_pmd_numa_page: remove unnecessary TLB flushing code
Greg Kroah-Hartman <gregkh@linuxfoundation.org>:
mm: change fault_in_pages_* to have an unsigned size parameter
Luigi Rizzo <lrizzo@google.com>:
mm/pagemap: add mmap_assert_locked() annotations to find_vma*()
"Liam R. Howlett" <Liam.Howlett@Oracle.com>:
remap_file_pages: Use vma_lookup() instead of find_vma()
Subsystem: mm/mremap
Chen Wandun <chenwandun@huawei.com>:
mm/mremap: fix memory account on do_munmap() failure
Subsystem: mm/bootmem
Muchun Song <songmuchun@bytedance.com>:
mm/bootmem_info.c: mark __init on register_page_bootmem_info_section
Subsystem: mm/sparsemem
Ohhoon Kwon <ohoono.kwon@samsung.com>:
Patch series "mm: sparse: remove __section_nr() function", v4:
mm: sparse: pass section_nr to section_mark_present
mm: sparse: pass section_nr to find_memory_block
mm: sparse: remove __section_nr() function
Naoya Horiguchi <naoya.horiguchi@nec.com>:
mm/sparse: set SECTION_NID_SHIFT to 6
Matthew Wilcox <willy@infradead.org>:
include/linux/mmzone.h: avoid a warning in sparse memory support
Miles Chen <miles.chen@mediatek.com>:
mm/sparse: clarify pgdat_to_phys
Subsystem: mm/vmalloc
"Uladzislau Rezki (Sony)" <urezki@gmail.com>:
mm/vmalloc: use batched page requests in bulk-allocator
mm/vmalloc: remove gfpflags_allow_blocking() check
lib/test_vmalloc.c: add a new 'nr_pages' parameter
Chen Wandun <chenwandun@huawei.com>:
mm/vmalloc: fix wrong behavior in vread
Subsystem: mm/kasan
Woody Lin <woodylin@google.com>:
mm/kasan: move kasan.fault to mm/kasan/report.c
Andrey Konovalov <andreyknvl@gmail.com>:
Patch series "kasan: test: avoid crashing the kernel with HW_TAGS", v2:
kasan: test: rework kmalloc_oob_right
kasan: test: avoid writing invalid memory
kasan: test: avoid corrupting memory via memset
kasan: test: disable kmalloc_memmove_invalid_size for HW_TAGS
kasan: test: only do kmalloc_uaf_memset for generic mode
kasan: test: clean up ksize_uaf
kasan: test: avoid corrupting memory in copy_user_test
kasan: test: avoid corrupting memory in kasan_rcu_uaf
Subsystem: mm/pagealloc
Mike Rapoport <rppt@linux.ibm.com>:
Patch series "mm: ensure consistency of memory map poisoning":
mm/page_alloc: always initialize memory map for the holes
microblaze: simplify pte_alloc_one_kernel()
mm: introduce memmap_alloc() to unify memory map allocation
memblock: stop poisoning raw allocations
Nico Pache <npache@redhat.com>:
mm/page_alloc.c: fix 'zone_id' may be used uninitialized in this function warning
Mike Rapoport <rppt@linux.ibm.com>:
mm/page_alloc: make alloc_node_mem_map() __init rather than __ref
Vasily Averin <vvs@virtuozzo.com>:
mm/page_alloc.c: use in_task()
"George G. Davis" <davis.george@siemens.com>:
mm/page_isolation: tracing: trace all test_pages_isolated failures
Subsystem: mm/memory-failure
Miaohe Lin <linmiaohe@huawei.com>:
Patch series "Cleanups and fixup for hwpoison":
mm/hwpoison: remove unneeded variable unmap_success
mm/hwpoison: fix potential pte_unmap_unlock pte error
mm/hwpoison: change argument struct page **hpagep to *hpage
mm/hwpoison: fix some obsolete comments
Yang Shi <shy828301@gmail.com>:
mm: hwpoison: don't drop slab caches for offlining non-LRU page
doc: hwpoison: correct the support for hugepage
mm: hwpoison: dump page for unhandlable page
Michael Wang <yun.wang@linux.alibaba.com>:
mm: fix panic caused by __page_handle_poison()
Subsystem: mm/hugetlb
Mike Kravetz <mike.kravetz@oracle.com>:
hugetlb: simplify prep_compound_gigantic_page ref count racing code
hugetlb: drop ref count earlier after page allocation
hugetlb: before freeing hugetlb page set dtor to appropriate value
hugetlb: fix hugetlb cgroup refcounting during vma split
Subsystem: mm/userfaultfd
Nadav Amit <namit@vmware.com>:
Patch series "userfaultfd: minor bug fixes":
userfaultfd: change mmap_changing to atomic
userfaultfd: prevent concurrent API initialization
selftests/vm/userfaultfd: wake after copy failure
Subsystem: mm/vmscan
Dave Hansen <dave.hansen@linux.intel.com>:
Patch series "Migrate Pages in lieu of discard", v11:
mm/numa: automatically generate node migration order
mm/migrate: update node demotion order on hotplug events
Yang Shi <yang.shi@linux.alibaba.com>:
mm/migrate: enable returning precise migrate_pages() success count
Dave Hansen <dave.hansen@linux.intel.com>:
mm/migrate: demote pages during reclaim
Yang Shi <yang.shi@linux.alibaba.com>:
mm/vmscan: add page demotion counter
Dave Hansen <dave.hansen@linux.intel.com>:
mm/vmscan: add helper for querying ability to age anonymous pages
Keith Busch <kbusch@kernel.org>:
mm/vmscan: Consider anonymous pages without swap
Dave Hansen <dave.hansen@linux.intel.com>:
mm/vmscan: never demote for memcg reclaim
Huang Ying <ying.huang@intel.com>:
mm/migrate: add sysfs interface to enable reclaim migration
Hui Su <suhui@zeku.com>:
mm/vmpressure: replace vmpressure_to_css() with vmpressure_to_memcg()
Miaohe Lin <linmiaohe@huawei.com>:
Patch series "Cleanups for vmscan", v2:
mm/vmscan: remove the PageDirty check after MADV_FREE pages are page_ref_freezed
mm/vmscan: remove misleading setting to sc->priority
mm/vmscan: remove unneeded return value of kswapd_run()
mm/vmscan: add 'else' to remove check_pending label
Vlastimil Babka <vbabka@suse.cz>:
mm, vmscan: guarantee drop_slab_node() termination
Subsystem: mm/compaction
Charan Teja Reddy <charante@codeaurora.org>:
mm: compaction: optimize proactive compaction deferrals
mm: compaction: support triggering of proactive compaction by user
Subsystem: mm/mempolicy
Baolin Wang <baolin.wang@linux.alibaba.com>:
mm/mempolicy: use readable NUMA_NO_NODE macro instead of magic number
Dave Hansen <dave.hansen@linux.intel.com>:
Patch series "Introduce multi-preference mempolicy", v7:
mm/mempolicy: add MPOL_PREFERRED_MANY for multiple preferred nodes
Feng Tang <feng.tang@intel.com>:
mm/memplicy: add page allocation function for MPOL_PREFERRED_MANY policy
Ben Widawsky <ben.widawsky@intel.com>:
mm/hugetlb: add support for mempolicy MPOL_PREFERRED_MANY
mm/mempolicy: advertise new MPOL_PREFERRED_MANY
Feng Tang <feng.tang@intel.com>:
mm/mempolicy: unify the create() func for bind/interleave/prefer-many policies
Vasily Averin <vvs@virtuozzo.com>:
mm/mempolicy.c: use in_task() in mempolicy_slab_node()
Subsystem: mm/memblock
Mike Rapoport <rppt@linux.ibm.com>:
memblock: make memblock_find_in_range method private
Subsystem: mm/oom-kill
Suren Baghdasaryan <surenb@google.com>:
mm: introduce process_mrelease system call
mm: wire up syscall process_mrelease
Subsystem: mm/migration
Randy Dunlap <rdunlap@infradead.org>:
mm/migrate: correct kernel-doc notation
Subsystem: mm/ksm
Zhansaya Bagdauletkyzy <zhansayabagdaulet@gmail.com>:
Patch series "add KSM selftests":
selftests: vm: add KSM merge test
selftests: vm: add KSM unmerge test
selftests: vm: add KSM zero page merging test
selftests: vm: add KSM merging across nodes test
mm: KSM: fix data type
Patch series "add KSM performance tests", v3:
selftests: vm: add KSM merging time test
selftests: vm: add COW time test for KSM pages
Subsystem: mm/percpu
Jing Xiangfeng <jingxiangfeng@huawei.com>:
mm/percpu,c: remove obsolete comments of pcpu_chunk_populated()
Subsystem: mm/vmstat
Miaohe Lin <linmiaohe@huawei.com>:
Patch series "Cleanup for vmstat":
mm/vmstat: correct some wrong comments
mm/vmstat: simplify the array size calculation
mm/vmstat: remove unneeded return value
Subsystem: mm/madvise
zhangkui <zhangkui@oppo.com>:
mm/madvise: add MADV_WILLNEED to process_madvise()
Documentation/ABI/testing/sysfs-kernel-mm-numa | 24
Documentation/admin-guide/mm/numa_memory_policy.rst | 15
Documentation/admin-guide/sysctl/vm.rst | 3
Documentation/core-api/cachetlb.rst | 86 -
Documentation/dev-tools/kasan.rst | 13
Documentation/translations/zh_CN/core-api/cachetlb.rst | 9
Documentation/vm/hwpoison.rst | 1
arch/Kconfig | 28
arch/alpha/kernel/syscalls/syscall.tbl | 2
arch/arm/include/asm/cacheflush.h | 4
arch/arm/kernel/setup.c | 20
arch/arm/mach-rpc/ecard.c | 2
arch/arm/mm/flush.c | 33
arch/arm/mm/nommu.c | 6
arch/arm/tools/syscall.tbl | 2
arch/arm64/include/asm/unistd.h | 2
arch/arm64/include/asm/unistd32.h | 2
arch/arm64/kvm/hyp/reserved_mem.c | 9
arch/arm64/mm/init.c | 38
arch/csky/abiv1/cacheflush.c | 11
arch/csky/abiv1/inc/abi/cacheflush.h | 4
arch/csky/kernel/probes/kprobes.c | 3
arch/ia64/include/asm/meminit.h | 2
arch/ia64/kernel/acpi.c | 2
arch/ia64/kernel/setup.c | 55
arch/ia64/kernel/syscalls/syscall.tbl | 2
arch/m68k/kernel/syscalls/syscall.tbl | 2
arch/microblaze/include/asm/page.h | 3
arch/microblaze/include/asm/pgtable.h | 2
arch/microblaze/kernel/syscalls/syscall.tbl | 2
arch/microblaze/mm/init.c | 12
arch/microblaze/mm/pgtable.c | 17
arch/mips/include/asm/cacheflush.h | 8
arch/mips/kernel/setup.c | 14
arch/mips/kernel/syscalls/syscall_n32.tbl | 2
arch/mips/kernel/syscalls/syscall_n64.tbl | 2
arch/mips/kernel/syscalls/syscall_o32.tbl | 2
arch/nds32/include/asm/cacheflush.h | 3
arch/nds32/mm/cacheflush.c | 9
arch/parisc/include/asm/cacheflush.h | 8
arch/parisc/kernel/cache.c | 3
arch/parisc/kernel/syscalls/syscall.tbl | 2
arch/powerpc/Kconfig | 1
arch/powerpc/kernel/smp.c | 2
arch/powerpc/kernel/syscalls/syscall.tbl | 2
arch/powerpc/mm/book3s64/radix_tlb.c | 4
arch/powerpc/platforms/pseries/hotplug-memory.c | 4
arch/riscv/mm/init.c | 44
arch/s390/kernel/setup.c | 9
arch/s390/kernel/syscalls/syscall.tbl | 2
arch/s390/mm/fault.c | 2
arch/sh/include/asm/cacheflush.h | 8
arch/sh/kernel/syscalls/syscall.tbl | 2
arch/sparc/kernel/syscalls/syscall.tbl | 2
arch/x86/entry/syscalls/syscall_32.tbl | 1
arch/x86/entry/syscalls/syscall_64.tbl | 1
arch/x86/kernel/aperture_64.c | 5
arch/x86/kernel/ldt.c | 6
arch/x86/mm/init.c | 23
arch/x86/mm/numa.c | 5
arch/x86/mm/numa_emulation.c | 5
arch/x86/realmode/init.c | 2
arch/xtensa/kernel/syscalls/syscall.tbl | 2
block/blk-map.c | 2
drivers/acpi/tables.c | 5
drivers/base/arch_numa.c | 5
drivers/base/memory.c | 4
drivers/mmc/host/jz4740_mmc.c | 4
drivers/mmc/host/mmc_spi.c | 2
drivers/of/of_reserved_mem.c | 12
fs/drop_caches.c | 3
fs/exec.c | 12
fs/fcntl.c | 3
fs/fs-writeback.c | 28
fs/fs_context.c | 4
fs/inode.c | 2
fs/locks.c | 6
fs/namei.c | 8
fs/namespace.c | 7
fs/ocfs2/dlmglue.c | 14
fs/ocfs2/quota_global.c | 1
fs/ocfs2/quota_local.c | 2
fs/pipe.c | 2
fs/select.c | 4
fs/userfaultfd.c | 116 -
include/linux/backing-dev-defs.h | 2
include/linux/backing-dev.h | 19
include/linux/buffer_head.h | 2
include/linux/compaction.h | 2
include/linux/highmem.h | 5
include/linux/hugetlb_cgroup.h | 12
include/linux/memblock.h | 2
include/linux/memcontrol.h | 118 +
include/linux/memory.h | 2
include/linux/mempolicy.h | 16
include/linux/migrate.h | 14
include/linux/mm.h | 17
include/linux/mmzone.h | 4
include/linux/page-flags.h | 9
include/linux/pagemap.h | 4
include/linux/sched/mm.h | 35
include/linux/shmem_fs.h | 25
include/linux/slub_def.h | 6
include/linux/swap.h | 28
include/linux/syscalls.h | 1
include/linux/userfaultfd_k.h | 8
include/linux/vm_event_item.h | 2
include/linux/vmpressure.h | 2
include/linux/writeback.h | 4
include/trace/events/migrate.h | 3
include/uapi/asm-generic/unistd.h | 4
include/uapi/linux/mempolicy.h | 1
ipc/msg.c | 2
ipc/namespace.c | 2
ipc/sem.c | 9
ipc/shm.c | 2
kernel/cgroup/namespace.c | 2
kernel/cpu.c | 2
kernel/exit.c | 2
kernel/fork.c | 51
kernel/kthread.c | 21
kernel/nsproxy.c | 2
kernel/pid_namespace.c | 5
kernel/sched/core.c | 37
kernel/sched/sched.h | 4
kernel/signal.c | 2
kernel/sys_ni.c | 1
kernel/sysctl.c | 2
kernel/time/namespace.c | 4
kernel/time/posix-timers.c | 4
kernel/user_namespace.c | 2
lib/scatterlist.c | 5
lib/test_kasan.c | 80 -
lib/test_kasan_module.c | 20
lib/test_vmalloc.c | 5
mm/backing-dev.c | 11
mm/bootmem_info.c | 4
mm/compaction.c | 69 -
mm/debug_vm_pgtable.c | 982 +++++++++------
mm/filemap.c | 15
mm/gup.c | 109 -
mm/huge_memory.c | 32
mm/hugetlb.c | 173 ++
mm/hwpoison-inject.c | 2
mm/internal.h | 9
mm/kasan/hw_tags.c | 43
mm/kasan/kasan.h | 1
mm/kasan/report.c | 29
mm/khugepaged.c | 2
mm/ksm.c | 8
mm/madvise.c | 1
mm/memblock.c | 22
mm/memcontrol.c | 234 +--
mm/memory-failure.c | 53
mm/memory_hotplug.c | 2
mm/mempolicy.c | 207 ++-
mm/migrate.c | 319 ++++
mm/mmap.c | 7
mm/mremap.c | 2
mm/oom_kill.c | 70 +
mm/page-writeback.c | 133 +-
mm/page_alloc.c | 62
mm/page_isolation.c | 13
mm/percpu.c | 3
mm/shmem.c | 309 ++--
mm/slab_common.c | 2
mm/slub.c | 1085 ++++++++++-------
mm/sparse.c | 46
mm/swap.c | 22
mm/swapfile.c | 14
mm/truncate.c | 28
mm/userfaultfd.c | 15
mm/vmalloc.c | 79 -
mm/vmpressure.c | 10
mm/vmscan.c | 220 ++-
mm/vmstat.c | 25
security/tomoyo/domain.c | 13
tools/testing/scatterlist/linux/mm.h | 1
tools/testing/selftests/vm/.gitignore | 1
tools/testing/selftests/vm/Makefile | 3
tools/testing/selftests/vm/charge_reserved_hugetlb.sh | 5
tools/testing/selftests/vm/hugetlb_reparenting_test.sh | 5
tools/testing/selftests/vm/ksm_tests.c | 696 ++++++++++
tools/testing/selftests/vm/mlock-random-test.c | 2
tools/testing/selftests/vm/run_vmtests.sh | 98 +
tools/testing/selftests/vm/userfaultfd.c | 13
186 files changed, 4488 insertions(+), 2281 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-08-25 19:17 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-08-25 19:17 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
2 patches, based on 6e764bcd1cf72a2846c0e53d3975a09b242c04c9.
Subsystems affected by this patch series:
mm/memory-hotplug
MAINTAINERS
Subsystem: mm/memory-hotplug
Miaohe Lin <linmiaohe@huawei.com>:
mm/memory_hotplug: fix potential permanent lru cache disable
Subsystem: MAINTAINERS
Namjae Jeon <namjae.jeon@samsung.com>:
MAINTAINERS: exfat: update my email address
MAINTAINERS | 2 +-
mm/memory_hotplug.c | 1 +
2 files changed, 2 insertions(+), 1 deletion(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-08-20 2:03 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-08-20 2:03 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
10 patches, based on 614cb2751d3150850d459bee596c397f344a7936.
Subsystems affected by this patch series:
mm/shmem
mm/pagealloc
mm/tracing
MAINTAINERS
mm/memcg
mm/memory-failure
mm/vmscan
mm/kfence
mm/hugetlb
Subsystem: mm/shmem
Yang Shi <shy828301@gmail.com>:
Revert "mm/shmem: fix shmem_swapin() race with swapoff"
Revert "mm: swap: check if swap backing device is congested or not"
Subsystem: mm/pagealloc
Doug Berger <opendmb@gmail.com>:
mm/page_alloc: don't corrupt pcppage_migratetype
Subsystem: mm/tracing
Mike Rapoport <rppt@linux.ibm.com>:
mmflags.h: add missing __GFP_ZEROTAGS and __GFP_SKIP_KASAN_POISON names
Subsystem: MAINTAINERS
Nathan Chancellor <nathan@kernel.org>:
MAINTAINERS: update ClangBuiltLinux IRC chat
Subsystem: mm/memcg
Johannes Weiner <hannes@cmpxchg.org>:
mm: memcontrol: fix occasional OOMs due to proportional memory.low reclaim
Subsystem: mm/memory-failure
Naoya Horiguchi <naoya.horiguchi@nec.com>:
mm/hwpoison: retry with shake_page() for unhandlable pages
Subsystem: mm/vmscan
Johannes Weiner <hannes@cmpxchg.org>:
mm: vmscan: fix missing psi annotation for node_reclaim()
Subsystem: mm/kfence
Marco Elver <elver@google.com>:
kfence: fix is_kfence_address() for addresses below KFENCE_POOL_SIZE
Subsystem: mm/hugetlb
Mike Kravetz <mike.kravetz@oracle.com>:
hugetlb: don't pass page cache pages to restore_reserve_on_error
MAINTAINERS | 2 +-
include/linux/kfence.h | 7 ++++---
include/linux/memcontrol.h | 29 +++++++++++++++--------------
include/trace/events/mmflags.h | 4 +++-
mm/hugetlb.c | 19 ++++++++++++++-----
mm/memory-failure.c | 12 +++++++++---
mm/page_alloc.c | 25 ++++++++++++-------------
mm/shmem.c | 14 +-------------
mm/swap_state.c | 7 -------
mm/vmscan.c | 30 ++++++++++++++++++++++--------
10 files changed, 81 insertions(+), 68 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-08-13 23:53 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-08-13 23:53 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
7 patches, based on f8e6dfc64f6135d1b6c5215c14cd30b9b60a0008.
Subsystems affected by this patch series:
mm/kasan
mm/slub
mm/madvise
mm/memcg
lib
Subsystem: mm/kasan
Kuan-Ying Lee <Kuan-Ying.Lee@mediatek.com>:
Patch series "kasan, slub: reset tag when printing address", v3:
kasan, kmemleak: reset tags when scanning block
kasan, slub: reset tag when printing address
Subsystem: mm/slub
Shakeel Butt <shakeelb@google.com>:
slub: fix kmalloc_pagealloc_invalid_free unit test
Vlastimil Babka <vbabka@suse.cz>:
mm: slub: fix slub_debug disabling for list of slabs
Subsystem: mm/madvise
David Hildenbrand <david@redhat.com>:
mm/madvise: report SIGBUS as -EFAULT for MADV_POPULATE_(READ|WRITE)
Subsystem: mm/memcg
Waiman Long <longman@redhat.com>:
mm/memcg: fix incorrect flushing of lruvec data in obj_stock
Subsystem: lib
Liang Wang <wangliang101@huawei.com>:
lib: use PFN_PHYS() in devmem_is_allowed()
lib/devmem_is_allowed.c | 2 +-
mm/gup.c | 7 +++++--
mm/kmemleak.c | 6 +++---
mm/madvise.c | 4 +++-
mm/memcontrol.c | 6 ++++--
mm/slub.c | 25 ++++++++++++++-----------
6 files changed, 30 insertions(+), 20 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-07-29 21:52 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-07-29 21:52 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
7 patches, based on 7e96bf476270aecea66740a083e51b38c1371cd2.
Subsystems affected by this patch series:
lib
ocfs2
mm/memcg
mm/migration
mm/slub
mm/memcg
Subsystem: lib
Matteo Croce <mcroce@microsoft.com>:
lib/test_string.c: move string selftest in the Runtime Testing menu
Subsystem: ocfs2
Junxiao Bi <junxiao.bi@oracle.com>:
ocfs2: fix zero out valid data
ocfs2: issue zeroout to EOF blocks
Subsystem: mm/memcg
Johannes Weiner <hannes@cmpxchg.org>:
mm: memcontrol: fix blocking rstat function called from atomic cgroup1 thresholding code
Subsystem: mm/migration
"Aneesh Kumar K.V" <aneesh.kumar@linux.ibm.com>:
mm/migrate: fix NR_ISOLATED corruption on 64-bit
Subsystem: mm/slub
Shakeel Butt <shakeelb@google.com>:
slub: fix unreclaimable slab stat for bulk free
Subsystem: mm/memcg
Wang Hai <wanghai38@huawei.com>:
mm/memcg: fix NULL pointer dereference in memcg_slab_free_hook()
fs/ocfs2/file.c | 103 ++++++++++++++++++++++++++++++++----------------------
lib/Kconfig | 3 -
lib/Kconfig.debug | 3 +
mm/memcontrol.c | 3 +
mm/migrate.c | 2 -
mm/slab.h | 2 -
mm/slub.c | 22 ++++++-----
7 files changed, 81 insertions(+), 57 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-07-23 22:49 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-07-23 22:49 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
15 patches, based on 704f4cba43d4ed31ef4beb422313f1263d87bc55.
Subsystems affected by this patch series:
mm/userfaultfd
mm/kfence
mm/highmem
mm/pagealloc
mm/memblock
mm/pagecache
mm/secretmem
mm/pagemap
mm/hugetlbfs
Subsystem: mm/userfaultfd
Peter Collingbourne <pcc@google.com>:
Patch series "userfaultfd: do not untag user pointers", v5:
userfaultfd: do not untag user pointers
selftest: use mmap instead of posix_memalign to allocate memory
Subsystem: mm/kfence
Weizhao Ouyang <o451686892@gmail.com>:
kfence: defer kfence_test_init to ensure that kunit debugfs is created
Alexander Potapenko <glider@google.com>:
kfence: move the size check to the beginning of __kfence_alloc()
kfence: skip all GFP_ZONEMASK allocations
Subsystem: mm/highmem
Christoph Hellwig <hch@lst.de>:
mm: call flush_dcache_page() in memcpy_to_page() and memzero_page()
mm: use kmap_local_page in memzero_page
Subsystem: mm/pagealloc
Sergei Trofimovich <slyfox@gentoo.org>:
mm: page_alloc: fix page_poison=1 / INIT_ON_ALLOC_DEFAULT_ON interaction
Subsystem: mm/memblock
Mike Rapoport <rppt@linux.ibm.com>:
memblock: make for_each_mem_range() traverse MEMBLOCK_HOTPLUG regions
Subsystem: mm/pagecache
Roman Gushchin <guro@fb.com>:
writeback, cgroup: remove wb from offline list before releasing refcnt
writeback, cgroup: do not reparent dax inodes
Subsystem: mm/secretmem
Mike Rapoport <rppt@linux.ibm.com>:
mm/secretmem: wire up ->set_page_dirty
Subsystem: mm/pagemap
Muchun Song <songmuchun@bytedance.com>:
mm: mmap_lock: fix disabling preemption directly
Qi Zheng <zhengqi.arch@bytedance.com>:
mm: fix the deadlock in finish_fault()
Subsystem: mm/hugetlbfs
Mike Kravetz <mike.kravetz@oracle.com>:
hugetlbfs: fix mount mode command line processing
Documentation/arm64/tagged-address-abi.rst | 26 ++++++++++++++++++--------
fs/fs-writeback.c | 3 +++
fs/hugetlbfs/inode.c | 2 +-
fs/userfaultfd.c | 26 ++++++++++++--------------
include/linux/highmem.h | 6 ++++--
include/linux/memblock.h | 4 ++--
mm/backing-dev.c | 2 +-
mm/kfence/core.c | 19 ++++++++++++++++---
mm/kfence/kfence_test.c | 2 +-
mm/memblock.c | 3 ++-
mm/memory.c | 11 ++++++++++-
mm/mmap_lock.c | 4 ++--
mm/page_alloc.c | 29 ++++++++++++++++-------------
mm/secretmem.c | 1 +
tools/testing/selftests/vm/userfaultfd.c | 6 ++++--
15 files changed, 93 insertions(+), 51 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-07-15 4:26 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-07-15 4:26 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
13 patches, based on 40226a3d96ef8ab8980f032681c8bfd46d63874e.
Subsystems affected by this patch series:
mm/kasan
mm/pagealloc
mm/rmap
mm/hmm
hfs
mm/hugetlb
Subsystem: mm/kasan
Marco Elver <elver@google.com>:
mm: move helper to check slub_debug_enabled
Yee Lee <yee.lee@mediatek.com>:
kasan: add memzero init for unaligned size at DEBUG
Marco Elver <elver@google.com>:
kasan: fix build by including kernel.h
Subsystem: mm/pagealloc
Matteo Croce <mcroce@microsoft.com>:
Revert "mm/page_alloc: make should_fail_alloc_page() static"
Mel Gorman <mgorman@techsingularity.net>:
mm/page_alloc: avoid page allocator recursion with pagesets.lock held
Yanfei Xu <yanfei.xu@windriver.com>:
mm/page_alloc: correct return value when failing at preparing
Chuck Lever <chuck.lever@oracle.com>:
mm/page_alloc: further fix __alloc_pages_bulk() return value
Subsystem: mm/rmap
Christoph Hellwig <hch@lst.de>:
mm: fix the try_to_unmap prototype for !CONFIG_MMU
Subsystem: mm/hmm
Alistair Popple <apopple@nvidia.com>:
lib/test_hmm: remove set but unused page variable
Subsystem: hfs
Desmond Cheong Zhi Xi <desmondcheongzx@gmail.com>:
Patch series "hfs: fix various errors", v2:
hfs: add missing clean-up in hfs_fill_super
hfs: fix high memory mapping in hfs_bnode_read
hfs: add lock nesting notation to hfs_find_init
Subsystem: mm/hugetlb
Joao Martins <joao.m.martins@oracle.com>:
mm/hugetlb: fix refs calculation from unaligned @vaddr
fs/hfs/bfind.c | 14 +++++++++++++-
fs/hfs/bnode.c | 25 ++++++++++++++++++++-----
fs/hfs/btree.h | 7 +++++++
fs/hfs/super.c | 10 +++++-----
include/linux/kasan.h | 1 +
include/linux/rmap.h | 4 +++-
lib/test_hmm.c | 2 --
mm/hugetlb.c | 5 +++--
mm/kasan/kasan.h | 12 ++++++++++++
mm/page_alloc.c | 30 ++++++++++++++++++++++--------
mm/slab.h | 15 +++++++++++----
mm/slub.c | 14 --------------
12 files changed, 97 insertions(+), 42 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-07-08 0:59 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-07-08 0:59 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
54 patches, based on a931dd33d370896a683236bba67c0d6f3d01144d.
Subsystems affected by this patch series:
lib
mm/slub
mm/secretmem
mm/cleanups
mm/init
debug
mm/pagemap
mm/mremap
Subsystem: lib
Zhen Lei <thunder.leizhen@huawei.com>:
lib/test: fix spelling mistakes
lib: fix spelling mistakes
lib: fix spelling mistakes in header files
Subsystem: mm/slub
Nathan Chancellor <nathan@kernel.org>:
Patch series "hexagon: Fix build error with CONFIG_STACKDEPOT and select CONFIG_ARCH_WANT_LD_ORPHAN_WARN":
hexagon: handle {,SOFT}IRQENTRY_TEXT in linker script
hexagon: use common DISCARDS macro
hexagon: select ARCH_WANT_LD_ORPHAN_WARN
Oliver Glitta <glittao@gmail.com>:
mm/slub: use stackdepot to save stack trace in objects
Subsystem: mm/secretmem
Mike Rapoport <rppt@linux.ibm.com>:
Patch series "mm: introduce memfd_secret system call to create "secret" memory areas", v20:
mmap: make mlock_future_check() global
riscv/Kconfig: make direct map manipulation options depend on MMU
set_memory: allow querying whether set_direct_map_*() is actually enabled
mm: introduce memfd_secret system call to create "secret" memory areas
PM: hibernate: disable when there are active secretmem users
arch, mm: wire up memfd_secret system call where relevant
secretmem: test: add basic selftest for memfd_secret(2)
Subsystem: mm/cleanups
Zhen Lei <thunder.leizhen@huawei.com>:
mm: fix spelling mistakes in header files
Subsystem: mm/init
Kefeng Wang <wangkefeng.wang@huawei.com>:
Patch series "init_mm: cleanup ARCH's text/data/brk setup code", v3:
mm: add setup_initial_init_mm() helper
arc: convert to setup_initial_init_mm()
arm: convert to setup_initial_init_mm()
arm64: convert to setup_initial_init_mm()
csky: convert to setup_initial_init_mm()
h8300: convert to setup_initial_init_mm()
m68k: convert to setup_initial_init_mm()
nds32: convert to setup_initial_init_mm()
nios2: convert to setup_initial_init_mm()
openrisc: convert to setup_initial_init_mm()
powerpc: convert to setup_initial_init_mm()
riscv: convert to setup_initial_init_mm()
s390: convert to setup_initial_init_mm()
sh: convert to setup_initial_init_mm()
x86: convert to setup_initial_init_mm()
Subsystem: debug
Stephen Boyd <swboyd@chromium.org>:
Patch series "Add build ID to stacktraces", v6:
buildid: only consider GNU notes for build ID parsing
buildid: add API to parse build ID out of buffer
buildid: stash away kernels build ID on init
dump_stack: add vmlinux build ID to stack traces
module: add printk formats to add module build ID to stacktraces
arm64: stacktrace: use %pSb for backtrace printing
x86/dumpstack: use %pSb/%pBb for backtrace printing
scripts/decode_stacktrace.sh: support debuginfod
scripts/decode_stacktrace.sh: silence stderr messages from addr2line/nm
scripts/decode_stacktrace.sh: indicate 'auto' can be used for base path
buildid: mark some arguments const
buildid: fix kernel-doc notation
kdump: use vmlinux_build_id to simplify
Subsystem: mm/pagemap
"Aneesh Kumar K.V" <aneesh.kumar@linux.ibm.com>:
mm: rename pud_page_vaddr to pud_pgtable and make it return pmd_t *
mm: rename p4d_page_vaddr to p4d_pgtable and make it return pud_t *
Subsystem: mm/mremap
"Aneesh Kumar K.V" <aneesh.kumar@linux.ibm.com>:
Patch series "mrermap fixes", v2:
selftest/mremap_test: update the test to handle pagesize other than 4K
selftest/mremap_test: avoid crash with static build
mm/mremap: convert huge PUD move to separate helper
mm/mremap: don't enable optimized PUD move if page table levels is 2
mm/mremap: use pmd/pud_poplulate to update page table entries
mm/mremap: hold the rmap lock in write mode when moving page table entries.
Patch series "Speedup mremap on ppc64", v8:
mm/mremap: allow arch runtime override
powerpc/book3s64/mm: update flush_tlb_range to flush page walk cache
powerpc/mm: enable HAVE_MOVE_PMD support
Documentation/core-api/printk-formats.rst | 11
arch/alpha/include/asm/pgtable.h | 8
arch/arc/mm/init.c | 5
arch/arm/include/asm/pgtable-3level.h | 2
arch/arm/kernel/setup.c | 5
arch/arm64/include/asm/Kbuild | 1
arch/arm64/include/asm/cacheflush.h | 6
arch/arm64/include/asm/kfence.h | 2
arch/arm64/include/asm/pgtable.h | 8
arch/arm64/include/asm/set_memory.h | 17 +
arch/arm64/include/uapi/asm/unistd.h | 1
arch/arm64/kernel/machine_kexec.c | 1
arch/arm64/kernel/setup.c | 5
arch/arm64/kernel/stacktrace.c | 2
arch/arm64/mm/mmu.c | 7
arch/arm64/mm/pageattr.c | 13
arch/csky/kernel/setup.c | 5
arch/h8300/kernel/setup.c | 5
arch/hexagon/Kconfig | 1
arch/hexagon/kernel/vmlinux.lds.S | 9
arch/ia64/include/asm/pgtable.h | 4
arch/m68k/include/asm/motorola_pgtable.h | 2
arch/m68k/kernel/setup_mm.c | 5
arch/m68k/kernel/setup_no.c | 5
arch/mips/include/asm/pgtable-64.h | 8
arch/nds32/kernel/setup.c | 5
arch/nios2/kernel/setup.c | 5
arch/openrisc/kernel/setup.c | 5
arch/parisc/include/asm/pgtable.h | 4
arch/powerpc/include/asm/book3s/64/pgtable.h | 11
arch/powerpc/include/asm/book3s/64/tlbflush-radix.h | 2
arch/powerpc/include/asm/nohash/64/pgtable-4k.h | 6
arch/powerpc/include/asm/nohash/64/pgtable.h | 6
arch/powerpc/include/asm/tlb.h | 6
arch/powerpc/kernel/setup-common.c | 5
arch/powerpc/mm/book3s64/radix_hugetlbpage.c | 8
arch/powerpc/mm/book3s64/radix_pgtable.c | 6
arch/powerpc/mm/book3s64/radix_tlb.c | 44 +-
arch/powerpc/mm/pgtable_64.c | 4
arch/powerpc/platforms/Kconfig.cputype | 2
arch/riscv/Kconfig | 4
arch/riscv/include/asm/pgtable-64.h | 4
arch/riscv/include/asm/unistd.h | 1
arch/riscv/kernel/setup.c | 5
arch/s390/kernel/setup.c | 5
arch/sh/include/asm/pgtable-3level.h | 4
arch/sh/kernel/setup.c | 5
arch/sparc/include/asm/pgtable_32.h | 6
arch/sparc/include/asm/pgtable_64.h | 10
arch/um/include/asm/pgtable-3level.h | 2
arch/x86/entry/syscalls/syscall_32.tbl | 1
arch/x86/entry/syscalls/syscall_64.tbl | 1
arch/x86/include/asm/pgtable.h | 8
arch/x86/kernel/dumpstack.c | 2
arch/x86/kernel/setup.c | 5
arch/x86/mm/init_64.c | 4
arch/x86/mm/pat/set_memory.c | 4
arch/x86/mm/pgtable.c | 2
include/asm-generic/pgtable-nop4d.h | 2
include/asm-generic/pgtable-nopmd.h | 2
include/asm-generic/pgtable-nopud.h | 4
include/linux/bootconfig.h | 4
include/linux/buildid.h | 10
include/linux/compaction.h | 4
include/linux/cpumask.h | 2
include/linux/crash_core.h | 12
include/linux/debugobjects.h | 2
include/linux/hmm.h | 2
include/linux/hugetlb.h | 6
include/linux/kallsyms.h | 21 +
include/linux/list_lru.h | 4
include/linux/lru_cache.h | 8
include/linux/mm.h | 3
include/linux/mmu_notifier.h | 8
include/linux/module.h | 9
include/linux/nodemask.h | 6
include/linux/percpu-defs.h | 2
include/linux/percpu-refcount.h | 2
include/linux/pgtable.h | 4
include/linux/scatterlist.h | 2
include/linux/secretmem.h | 54 +++
include/linux/set_memory.h | 12
include/linux/shrinker.h | 2
include/linux/syscalls.h | 1
include/linux/vmalloc.h | 4
include/uapi/asm-generic/unistd.h | 7
include/uapi/linux/magic.h | 1
init/Kconfig | 1
init/main.c | 2
kernel/crash_core.c | 50 ---
kernel/kallsyms.c | 104 +++++--
kernel/module.c | 42 ++
kernel/power/hibernate.c | 5
kernel/sys_ni.c | 2
lib/Kconfig.debug | 17 -
lib/asn1_encoder.c | 2
lib/buildid.c | 80 ++++-
lib/devres.c | 2
lib/dump_stack.c | 13
lib/dynamic_debug.c | 2
lib/fonts/font_pearl_8x8.c | 2
lib/kfifo.c | 2
lib/list_sort.c | 2
lib/nlattr.c | 4
lib/oid_registry.c | 2
lib/pldmfw/pldmfw.c | 2
lib/reed_solomon/test_rslib.c | 2
lib/refcount.c | 2
lib/rhashtable.c | 2
lib/sbitmap.c | 2
lib/scatterlist.c | 4
lib/seq_buf.c | 2
lib/sort.c | 2
lib/stackdepot.c | 2
lib/test_bitops.c | 2
lib/test_bpf.c | 2
lib/test_kasan.c | 2
lib/test_kmod.c | 6
lib/test_scanf.c | 2
lib/vsprintf.c | 10
mm/Kconfig | 4
mm/Makefile | 1
mm/gup.c | 12
mm/init-mm.c | 9
mm/internal.h | 3
mm/mlock.c | 3
mm/mmap.c | 5
mm/mremap.c | 108 ++++++-
mm/secretmem.c | 254 +++++++++++++++++
mm/slub.c | 79 +++--
scripts/checksyscalls.sh | 4
scripts/decode_stacktrace.sh | 89 +++++-
tools/testing/selftests/vm/.gitignore | 1
tools/testing/selftests/vm/Makefile | 3
tools/testing/selftests/vm/memfd_secret.c | 296 ++++++++++++++++++++
tools/testing/selftests/vm/mremap_test.c | 116 ++++---
tools/testing/selftests/vm/run_vmtests.sh | 17 +
137 files changed, 1470 insertions(+), 442 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2021-07-03 0:28 ` incoming Linus Torvalds
@ 2021-07-03 1:06 ` Linus Torvalds
0 siblings, 0 replies; 409+ messages in thread
From: Linus Torvalds @ 2021-07-03 1:06 UTC (permalink / raw)
To: Andrew Morton; +Cc: Linux-MM, mm-commits
On Fri, Jul 2, 2021 at 5:28 PM Linus Torvalds
<torvalds@linux-foundation.org> wrote:
>
> Commit e058a84bfddc42ba356a2316f2cf1141974625c9 is good, and looking
> at the pulls and merges I've done since, this -mm series looks like
> the obvious culprit.
No, unless my bisection is wrong, the -mm branch is innocent, and was
discarded from the suspects on the very first bisection trial.
So never mind.
Linus
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2021-07-01 1:46 incoming Andrew Morton
@ 2021-07-03 0:28 ` Linus Torvalds
2021-07-03 1:06 ` incoming Linus Torvalds
0 siblings, 1 reply; 409+ messages in thread
From: Linus Torvalds @ 2021-07-03 0:28 UTC (permalink / raw)
To: Andrew Morton; +Cc: Linux-MM, mm-commits
On Wed, Jun 30, 2021 at 6:46 PM Andrew Morton <akpm@linux-foundation.org> wrote:
>
> This is the rest of the -mm tree, less 66 patches which are dependent on
> things which are (or were recently) in linux-next. I'll trickle that
> material over next week.
I haven't bisected this yet, but with the current -git I'm getting
watchdog: BUG: soft lockup - CPU#41 stuck for 49s!
and the common call chain seems to be in flush_tlb_mm_range ->
on_each_cpu_cond_mask.
Commit e058a84bfddc42ba356a2316f2cf1141974625c9 is good, and looking
at the pulls and merges I've done since, this -mm series looks like
the obvious culprit.
I'll go start bisection, but I thought I'd give a heads-up in case
somebody else has seen TLB-flush-related lockups and already figured
out the guilty party..
Linus
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-07-01 1:46 Andrew Morton
2021-07-03 0:28 ` incoming Linus Torvalds
0 siblings, 1 reply; 409+ messages in thread
From: Andrew Morton @ 2021-07-01 1:46 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
This is the rest of the -mm tree, less 66 patches which are dependent on
things which are (or were recently) in linux-next. I'll trickle that
material over next week.
192 patches, based on 7cf3dead1ad70c72edb03e2d98e1f3dcd332cdb2 plus the
June 28 sendings.
Subsystems affected by this patch series:
mm/hugetlb
mm/userfaultfd
mm/vmscan
mm/kconfig
mm/proc
mm/z3fold
mm/zbud
mm/ras
mm/mempolicy
mm/memblock
mm/migration
mm/thp
mm/nommu
mm/kconfig
mm/madvise
mm/memory-hotplug
mm/zswap
mm/zsmalloc
mm/zram
mm/cleanups
mm/kfence
mm/hmm
procfs
sysctl
misc
core-kernel
lib
lz4
checkpatch
init
kprobes
nilfs2
hfs
signals
exec
kcov
selftests
compress/decompress
ipc
Subsystem: mm/hugetlb
Muchun Song <songmuchun@bytedance.com>:
Patch series "Free some vmemmap pages of HugeTLB page", v23:
mm: memory_hotplug: factor out bootmem core functions to bootmem_info.c
mm: hugetlb: introduce a new config HUGETLB_PAGE_FREE_VMEMMAP
mm: hugetlb: gather discrete indexes of tail page
mm: hugetlb: free the vmemmap pages associated with each HugeTLB page
mm: hugetlb: defer freeing of HugeTLB pages
mm: hugetlb: alloc the vmemmap pages associated with each HugeTLB page
mm: hugetlb: add a kernel parameter hugetlb_free_vmemmap
mm: memory_hotplug: disable memmap_on_memory when hugetlb_free_vmemmap enabled
mm: hugetlb: introduce nr_free_vmemmap_pages in the struct hstate
Shixin Liu <liushixin2@huawei.com>:
mm/debug_vm_pgtable: move {pmd/pud}_huge_tests out of CONFIG_TRANSPARENT_HUGEPAGE
mm/debug_vm_pgtable: remove redundant pfn_{pmd/pte}() and fix one comment mistake
Miaohe Lin <linmiaohe@huawei.com>:
Patch series "Cleanup and fixup for huge_memory:, v3:
mm/huge_memory.c: remove dedicated macro HPAGE_CACHE_INDEX_MASK
mm/huge_memory.c: use page->deferred_list
mm/huge_memory.c: add missing read-only THP checking in transparent_hugepage_enabled()
mm/huge_memory.c: remove unnecessary tlb_remove_page_size() for huge zero pmd
mm/huge_memory.c: don't discard hugepage if other processes are mapping it
Christophe Leroy <christophe.leroy@csgroup.eu>:
Patch series "Subject: [PATCH v2 0/5] Implement huge VMAP and VMALLOC on powerpc 8xx", v2:
mm/hugetlb: change parameters of arch_make_huge_pte()
mm/pgtable: add stubs for {pmd/pub}_{set/clear}_huge
mm/vmalloc: enable mapping of huge pages at pte level in vmap
mm/vmalloc: enable mapping of huge pages at pte level in vmalloc
powerpc/8xx: add support for huge pages on VMAP and VMALLOC
Nanyong Sun <sunnanyong@huawei.com>:
khugepaged: selftests: remove debug_cow
Mina Almasry <almasrymina@google.com>:
mm, hugetlb: fix racy resv_huge_pages underflow on UFFDIO_COPY
Muchun Song <songmuchun@bytedance.com>:
Patch series "Split huge PMD mapping of vmemmap pages", v4:
mm: sparsemem: split the huge PMD mapping of vmemmap pages
mm: sparsemem: use huge PMD mapping for vmemmap pages
mm: hugetlb: introduce CONFIG_HUGETLB_PAGE_FREE_VMEMMAP_DEFAULT_ON
Mike Kravetz <mike.kravetz@oracle.com>:
Patch series "Fix prep_compound_gigantic_page ref count adjustment":
hugetlb: remove prep_compound_huge_page cleanup
hugetlb: address ref count racing in prep_compound_gigantic_page
Naoya Horiguchi <naoya.horiguchi@nec.com>:
mm/hwpoison: disable pcp for page_handle_poison()
Subsystem: mm/userfaultfd
Peter Xu <peterx@redhat.com>:
Patch series "userfaultfd/selftests: A few cleanups", v2:
userfaultfd/selftests: use user mode only
userfaultfd/selftests: remove the time() check on delayed uffd
userfaultfd/selftests: dropping VERIFY check in locking_thread
userfaultfd/selftests: only dump counts if mode enabled
userfaultfd/selftests: unify error handling
Patch series "mm/uffd: Misc fix for uffd-wp and one more test":
mm/thp: simplify copying of huge zero page pmd when fork
mm/userfaultfd: fix uffd-wp special cases for fork()
mm/userfaultfd: fail uffd-wp registration if not supported
mm/pagemap: export uffd-wp protection information
userfaultfd/selftests: add pagemap uffd-wp test
Axel Rasmussen <axelrasmussen@google.com>:
Patch series "userfaultfd: add minor fault handling for shmem", v6:
userfaultfd/shmem: combine shmem_{mcopy_atomic,mfill_zeropage}_pte
userfaultfd/shmem: support minor fault registration for shmem
userfaultfd/shmem: support UFFDIO_CONTINUE for shmem
userfaultfd/shmem: advertise shmem minor fault support
userfaultfd/shmem: modify shmem_mfill_atomic_pte to use install_pte()
userfaultfd/selftests: use memfd_create for shmem test type
userfaultfd/selftests: create alias mappings in the shmem test
userfaultfd/selftests: reinitialize test context in each test
userfaultfd/selftests: exercise minor fault handling shmem support
Subsystem: mm/vmscan
Yu Zhao <yuzhao@google.com>:
mm/vmscan.c: fix potential deadlock in reclaim_pages()
include/trace/events/vmscan.h: remove mm_vmscan_inactive_list_is_low
Miaohe Lin <linmiaohe@huawei.com>:
mm: workingset: define macro WORKINGSET_SHIFT
Subsystem: mm/kconfig
Kefeng Wang <wangkefeng.wang@huawei.com>:
mm/kconfig: move HOLES_IN_ZONE into mm
Subsystem: mm/proc
Mike Rapoport <rppt@linux.ibm.com>:
docs: proc.rst: meminfo: briefly describe gaps in memory accounting
David Hildenbrand <david@redhat.com>:
Patch series "fs/proc/kcore: don't read offline sections, logically offline pages and hwpoisoned pages", v3:
fs/proc/kcore: drop KCORE_REMAP and KCORE_OTHER
fs/proc/kcore: pfn_is_ram check only applies to KCORE_RAM
fs/proc/kcore: don't read offline sections, logically offline pages and hwpoisoned pages
mm: introduce page_offline_(begin|end|freeze|thaw) to synchronize setting PageOffline()
virtio-mem: use page_offline_(start|end) when setting PageOffline()
fs/proc/kcore: use page_offline_(freeze|thaw)
Subsystem: mm/z3fold
Miaohe Lin <linmiaohe@huawei.com>:
Patch series "Cleanup and fixup for z3fold":
mm/z3fold: define macro NCHUNKS as TOTAL_CHUNKS - ZHDR_CHUNKS
mm/z3fold: avoid possible underflow in z3fold_alloc()
mm/z3fold: remove magic number in z3fold_create_pool()
mm/z3fold: remove unused function handle_to_z3fold_header()
mm/z3fold: fix potential memory leak in z3fold_destroy_pool()
mm/z3fold: use release_z3fold_page_locked() to release locked z3fold page
Subsystem: mm/zbud
Miaohe Lin <linmiaohe@huawei.com>:
Patch series "Cleanups for zbud", v2:
mm/zbud: reuse unbuddied[0] as buddied in zbud_pool
mm/zbud: don't export any zbud API
Subsystem: mm/ras
YueHaibing <yuehaibing@huawei.com>:
mm/compaction: use DEVICE_ATTR_WO macro
Liu Xiang <liu.xiang@zlingsmart.com>:
mm: compaction: remove duplicate !list_empty(&sublist) check
Wonhyuk Yang <vvghjk1234@gmail.com>:
mm/compaction: fix 'limit' in fast_isolate_freepages
Subsystem: mm/mempolicy
Feng Tang <feng.tang@intel.com>:
Patch series "mm/mempolicy: some fix and semantics cleanup", v4:
mm/mempolicy: cleanup nodemask intersection check for oom
mm/mempolicy: don't handle MPOL_LOCAL like a fake MPOL_PREFERRED policy
mm/mempolicy: unify the parameter sanity check for mbind and set_mempolicy
Yang Shi <shy828301@gmail.com>:
mm: mempolicy: don't have to split pmd for huge zero page
Ben Widawsky <ben.widawsky@intel.com>:
mm/mempolicy: use unified 'nodes' for bind/interleave/prefer policies
Subsystem: mm/memblock
Mike Rapoport <rppt@linux.ibm.com>:
Patch series "arm64: drop pfn_valid_within() and simplify pfn_valid()", v4:
include/linux/mmzone.h: add documentation for pfn_valid()
memblock: update initialization of reserved pages
arm64: decouple check whether pfn is in linear map from pfn_valid()
arm64: drop pfn_valid_within() and simplify pfn_valid()
Anshuman Khandual <anshuman.khandual@arm.com>:
arm64/mm: drop HAVE_ARCH_PFN_VALID
Subsystem: mm/migration
Muchun Song <songmuchun@bytedance.com>:
mm: migrate: fix missing update page_private to hugetlb_page_subpool
Subsystem: mm/thp
Collin Fijalkovich <cfijalkovich@google.com>:
mm, thp: relax the VM_DENYWRITE constraint on file-backed THPs
Yang Shi <shy828301@gmail.com>:
mm: memory: add orig_pmd to struct vm_fault
mm: memory: make numa_migrate_prep() non-static
mm: thp: refactor NUMA fault handling
mm: migrate: account THP NUMA migration counters correctly
mm: migrate: don't split THP for misplaced NUMA page
mm: migrate: check mapcount for THP instead of refcount
mm: thp: skip make PMD PROT_NONE if THP migration is not supported
Anshuman Khandual <anshuman.khandual@arm.com>:
mm/thp: make ARCH_ENABLE_SPLIT_PMD_PTLOCK dependent on PGTABLE_LEVELS > 2
Yang Shi <shy828301@gmail.com>:
mm: rmap: make try_to_unmap() void function
Hugh Dickins <hughd@google.com>:
mm/thp: remap_page() is only needed on anonymous THP
mm: hwpoison_user_mappings() try_to_unmap() with TTU_SYNC
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
mm/thp: fix strncpy warning
Subsystem: mm/nommu
Chen Li <chenli@uniontech.com>:
nommu: remove __GFP_HIGHMEM in vmalloc/vzalloc
Liam Howlett <liam.howlett@oracle.com>:
mm/nommu: unexport do_munmap()
Subsystem: mm/kconfig
Kefeng Wang <wangkefeng.wang@huawei.com>:
mm: generalize ZONE_[DMA|DMA32]
Subsystem: mm/madvise
David Hildenbrand <david@redhat.com>:
Patch series "mm/madvise: introduce MADV_POPULATE_(READ|WRITE) to prefault page tables", v2:
mm: make variable names for populate_vma_page_range() consistent
mm/madvise: introduce MADV_POPULATE_(READ|WRITE) to prefault page tables
MAINTAINERS: add tools/testing/selftests/vm/ to MEMORY MANAGEMENT
selftests/vm: add protection_keys_32 / protection_keys_64 to gitignore
selftests/vm: add test for MADV_POPULATE_(READ|WRITE)
Subsystem: mm/memory-hotplug
Liam Mark <lmark@codeaurora.org>:
mm/memory_hotplug: rate limit page migration warnings
Oscar Salvador <osalvador@suse.de>:
mm,memory_hotplug: drop unneeded locking
Subsystem: mm/zswap
Miaohe Lin <linmiaohe@huawei.com>:
Patch series "Cleanup and fixup for zswap":
mm/zswap.c: remove unused function zswap_debugfs_exit()
mm/zswap.c: avoid unnecessary copy-in at map time
mm/zswap.c: fix two bugs in zswap_writeback_entry()
Subsystem: mm/zsmalloc
Zhaoyang Huang <zhaoyang.huang@unisoc.com>:
mm: zram: amend SLAB_RECLAIM_ACCOUNT on zspage_cachep
Miaohe Lin <linmiaohe@huawei.com>:
Patch series "Cleanup for zsmalloc":
mm/zsmalloc.c: remove confusing code in obj_free()
mm/zsmalloc.c: improve readability for async_free_zspage()
Subsystem: mm/zram
Yue Hu <huyue2@yulong.com>:
zram: move backing_dev under macro CONFIG_ZRAM_WRITEBACK
Subsystem: mm/cleanups
Hyeonggon Yoo <42.hyeyoo@gmail.com>:
mm: fix typos and grammar error in comments
Anshuman Khandual <anshuman.khandual@arm.com>:
mm: define default value for FIRST_USER_ADDRESS
Zhen Lei <thunder.leizhen@huawei.com>:
mm: fix spelling mistakes
Mel Gorman <mgorman@techsingularity.net>:
Patch series "Clean W=1 build warnings for mm/":
mm/vmscan: remove kerneldoc-like comment from isolate_lru_pages
mm/vmalloc: include header for prototype of set_iounmap_nonlazy
mm/page_alloc: make should_fail_alloc_page() static
mm/mapping_dirty_helpers: remove double Note in kerneldoc
mm/memcontrol.c: fix kerneldoc comment for mem_cgroup_calculate_protection
mm/memory_hotplug: fix kerneldoc comment for __try_online_node
mm/memory_hotplug: fix kerneldoc comment for __remove_memory
mm/zbud: add kerneldoc fields for zbud_pool
mm/z3fold: add kerneldoc fields for z3fold_pool
mm/swap: make swap_address_space an inline function
mm/mmap_lock: remove dead code for !CONFIG_TRACING configurations
mm/page_alloc: move prototype for find_suitable_fallback
mm/swap: make NODE_DATA an inline function on CONFIG_FLATMEM
Anshuman Khandual <anshuman.khandual@arm.com>:
mm/thp: define default pmd_pgtable()
Subsystem: mm/kfence
Marco Elver <elver@google.com>:
kfence: unconditionally use unbound work queue
Subsystem: mm/hmm
Alistair Popple <apopple@nvidia.com>:
Patch series "Add support for SVM atomics in Nouveau", v11:
mm: remove special swap entry functions
mm/swapops: rework swap entry manipulation code
mm/rmap: split try_to_munlock from try_to_unmap
mm/rmap: split migration into its own function
mm: rename migrate_pgmap_owner
mm/memory.c: allow different return codes for copy_nonpresent_pte()
mm: device exclusive memory access
mm: selftests for exclusive device memory
nouveau/svm: refactor nouveau_range_fault
nouveau/svm: implement atomic SVM access
Subsystem: procfs
Marcelo Henrique Cerri <marcelo.cerri@canonical.com>:
proc: Avoid mixing integer types in mem_rw()
ZHOUFENG <zhoufeng.zf@bytedance.com>:
fs/proc/kcore.c: add mmap interface
Kalesh Singh <kaleshsingh@google.com>:
procfs: allow reading fdinfo with PTRACE_MODE_READ
procfs/dmabuf: add inode number to /proc/*/fdinfo
Subsystem: sysctl
Jiapeng Chong <jiapeng.chong@linux.alibaba.com>:
sysctl: remove redundant assignment to first
Subsystem: misc
Andy Shevchenko <andriy.shevchenko@linux.intel.com>:
drm: include only needed headers in ascii85.h
Subsystem: core-kernel
Andy Shevchenko <andriy.shevchenko@linux.intel.com>:
kernel.h: split out panic and oops helpers
Subsystem: lib
Zhen Lei <thunder.leizhen@huawei.com>:
lib: decompress_bunzip2: remove an unneeded semicolon
Andy Shevchenko <andriy.shevchenko@linux.intel.com>:
Patch series "lib/string_helpers: get rid of ugly *_escape_mem_ascii()", v3:
lib/string_helpers: switch to use BIT() macro
lib/string_helpers: move ESCAPE_NP check inside 'else' branch in a loop
lib/string_helpers: drop indentation level in string_escape_mem()
lib/string_helpers: introduce ESCAPE_NA for escaping non-ASCII
lib/string_helpers: introduce ESCAPE_NAP to escape non-ASCII and non-printable
lib/string_helpers: allow to append additional characters to be escaped
lib/test-string_helpers: print flags in hexadecimal format
lib/test-string_helpers: get rid of trailing comma in terminators
lib/test-string_helpers: add test cases for new features
MAINTAINERS: add myself as designated reviewer for generic string library
seq_file: introduce seq_escape_mem()
seq_file: add seq_escape_str() as replica of string_escape_str()
seq_file: convert seq_escape() to use seq_escape_str()
nfsd: avoid non-flexible API in seq_quote_mem()
seq_file: drop unused *_escape_mem_ascii()
Trent Piepho <tpiepho@gmail.com>:
lib/math/rational.c: fix divide by zero
lib/math/rational: add Kunit test cases
Zhen Lei <thunder.leizhen@huawei.com>:
lib/decompressors: fix spelling mistakes
lib/mpi: fix spelling mistakes
Alexey Dobriyan <adobriyan@gmail.com>:
lib: memscan() fixlet
lib: uninline simple_strtoull()
Matteo Croce <mcroce@microsoft.com>:
lib/test_string.c: allow module removal
Andy Shevchenko <andriy.shevchenko@linux.intel.com>:
kernel.h: split out kstrtox() and simple_strtox() to a separate header
Subsystem: lz4
Rajat Asthana <thisisrast7@gmail.com>:
lz4_decompress: declare LZ4_decompress_safe_withPrefix64k static
Dimitri John Ledkov <dimitri.ledkov@canonical.com>:
lib/decompress_unlz4.c: correctly handle zero-padding around initrds.
Subsystem: checkpatch
Guenter Roeck <linux@roeck-us.net>:
checkpatch: scripts/spdxcheck.py now requires python3
Joe Perches <joe@perches.com>:
checkpatch: improve the indented label test
Guenter Roeck <linux@roeck-us.net>:
checkpatch: do not complain about positive return values starting with EPOLL
Subsystem: init
Andrew Halaney <ahalaney@redhat.com>:
init: print out unknown kernel parameters
Subsystem: kprobes
Barry Song <song.bao.hua@hisilicon.com>:
kprobes: remove duplicated strong free_insn_page in x86 and s390
Subsystem: nilfs2
Colin Ian King <colin.king@canonical.com>:
nilfs2: remove redundant continue statement in a while-loop
Subsystem: hfs
Zhen Lei <thunder.leizhen@huawei.com>:
hfsplus: remove unnecessary oom message
Chung-Chiang Cheng <shepjeng@gmail.com>:
hfsplus: report create_date to kstat.btime
Subsystem: signals
Al Viro <viro@zeniv.linux.org.uk>:
x86: signal: don't do sas_ss_reset() until we are certain that sigframe won't be abandoned
Subsystem: exec
Alexey Dobriyan <adobriyan@gmail.com>:
exec: remove checks in __register_bimfmt()
Subsystem: kcov
Marco Elver <elver@google.com>:
kcov: add __no_sanitize_coverage to fix noinstr for all architectures
Subsystem: selftests
Dave Hansen <dave.hansen@linux.intel.com>:
Patch series "selftests/vm/pkeys: Bug fixes and a new test":
selftests/vm/pkeys: fix alloc_random_pkey() to make it really, really random
selftests/vm/pkeys: handle negative sys_pkey_alloc() return code
selftests/vm/pkeys: refill shadow register after implicit kernel write
selftests/vm/pkeys: exercise x86 XSAVE init state
Subsystem: compress/decompress
Yu Kuai <yukuai3@huawei.com>:
lib/decompressors: remove set but not used variabled 'level'
Subsystem: ipc
Vasily Averin <vvs@virtuozzo.com>:
Patch series "ipc: allocations cleanup", v2:
ipc sem: use kvmalloc for sem_undo allocation
ipc: use kmalloc for msg_queue and shmid_kernel
Manfred Spraul <manfred@colorfullife.com>:
ipc/sem.c: use READ_ONCE()/WRITE_ONCE() for use_global_lock
ipc/util.c: use binary search for max_idx
Documentation/admin-guide/kernel-parameters.txt | 35
Documentation/admin-guide/mm/hugetlbpage.rst | 11
Documentation/admin-guide/mm/memory-hotplug.rst | 13
Documentation/admin-guide/mm/pagemap.rst | 2
Documentation/admin-guide/mm/userfaultfd.rst | 3
Documentation/core-api/kernel-api.rst | 7
Documentation/filesystems/proc.rst | 48
Documentation/vm/hmm.rst | 19
Documentation/vm/unevictable-lru.rst | 33
MAINTAINERS | 10
arch/alpha/Kconfig | 5
arch/alpha/include/asm/pgalloc.h | 1
arch/alpha/include/asm/pgtable.h | 1
arch/alpha/include/uapi/asm/mman.h | 3
arch/alpha/kernel/setup.c | 2
arch/arc/include/asm/pgalloc.h | 2
arch/arc/include/asm/pgtable.h | 8
arch/arm/Kconfig | 3
arch/arm/include/asm/pgalloc.h | 1
arch/arm64/Kconfig | 15
arch/arm64/include/asm/hugetlb.h | 3
arch/arm64/include/asm/memory.h | 2
arch/arm64/include/asm/page.h | 4
arch/arm64/include/asm/pgalloc.h | 1
arch/arm64/include/asm/pgtable.h | 2
arch/arm64/kernel/setup.c | 1
arch/arm64/kvm/mmu.c | 2
arch/arm64/mm/hugetlbpage.c | 5
arch/arm64/mm/init.c | 51
arch/arm64/mm/ioremap.c | 4
arch/arm64/mm/mmu.c | 22
arch/csky/include/asm/pgalloc.h | 2
arch/csky/include/asm/pgtable.h | 1
arch/hexagon/include/asm/pgtable.h | 4
arch/ia64/Kconfig | 7
arch/ia64/include/asm/pal.h | 1
arch/ia64/include/asm/pgalloc.h | 1
arch/ia64/include/asm/pgtable.h | 1
arch/m68k/Kconfig | 5
arch/m68k/include/asm/mcf_pgalloc.h | 2
arch/m68k/include/asm/mcf_pgtable.h | 2
arch/m68k/include/asm/motorola_pgalloc.h | 1
arch/m68k/include/asm/motorola_pgtable.h | 2
arch/m68k/include/asm/pgtable_mm.h | 1
arch/m68k/include/asm/sun3_pgalloc.h | 1
arch/microblaze/Kconfig | 4
arch/microblaze/include/asm/pgalloc.h | 2
arch/microblaze/include/asm/pgtable.h | 2
arch/mips/Kconfig | 10
arch/mips/include/asm/pgalloc.h | 1
arch/mips/include/asm/pgtable-32.h | 1
arch/mips/include/asm/pgtable-64.h | 1
arch/mips/include/uapi/asm/mman.h | 3
arch/mips/kernel/relocate.c | 1
arch/mips/sgi-ip22/ip22-reset.c | 1
arch/mips/sgi-ip32/ip32-reset.c | 1
arch/nds32/include/asm/pgalloc.h | 5
arch/nios2/include/asm/pgalloc.h | 1
arch/nios2/include/asm/pgtable.h | 2
arch/openrisc/include/asm/pgalloc.h | 2
arch/openrisc/include/asm/pgtable.h | 1
arch/parisc/include/asm/pgalloc.h | 1
arch/parisc/include/asm/pgtable.h | 2
arch/parisc/include/uapi/asm/mman.h | 3
arch/parisc/kernel/pdc_chassis.c | 1
arch/powerpc/Kconfig | 6
arch/powerpc/include/asm/book3s/pgtable.h | 1
arch/powerpc/include/asm/nohash/32/hugetlb-8xx.h | 5
arch/powerpc/include/asm/nohash/32/mmu-8xx.h | 43
arch/powerpc/include/asm/nohash/32/pgtable.h | 1
arch/powerpc/include/asm/nohash/64/pgtable.h | 2
arch/powerpc/include/asm/pgalloc.h | 5
arch/powerpc/include/asm/pgtable.h | 6
arch/powerpc/kernel/setup-common.c | 1
arch/powerpc/platforms/Kconfig.cputype | 1
arch/riscv/Kconfig | 5
arch/riscv/include/asm/pgalloc.h | 2
arch/riscv/include/asm/pgtable.h | 2
arch/s390/Kconfig | 6
arch/s390/include/asm/pgalloc.h | 3
arch/s390/include/asm/pgtable.h | 5
arch/s390/kernel/ipl.c | 1
arch/s390/kernel/kprobes.c | 5
arch/s390/mm/pgtable.c | 2
arch/sh/include/asm/pgalloc.h | 1
arch/sh/include/asm/pgtable.h | 2
arch/sparc/Kconfig | 5
arch/sparc/include/asm/pgalloc_32.h | 1
arch/sparc/include/asm/pgalloc_64.h | 1
arch/sparc/include/asm/pgtable_32.h | 3
arch/sparc/include/asm/pgtable_64.h | 8
arch/sparc/kernel/sstate.c | 1
arch/sparc/mm/hugetlbpage.c | 6
arch/sparc/mm/init_64.c | 1
arch/um/drivers/mconsole_kern.c | 1
arch/um/include/asm/pgalloc.h | 1
arch/um/include/asm/pgtable-2level.h | 1
arch/um/include/asm/pgtable-3level.h | 1
arch/um/kernel/um_arch.c | 1
arch/x86/Kconfig | 17
arch/x86/include/asm/desc.h | 1
arch/x86/include/asm/pgalloc.h | 2
arch/x86/include/asm/pgtable_types.h | 2
arch/x86/kernel/cpu/mshyperv.c | 1
arch/x86/kernel/kprobes/core.c | 6
arch/x86/kernel/setup.c | 1
arch/x86/mm/init_64.c | 21
arch/x86/mm/pgtable.c | 34
arch/x86/purgatory/purgatory.c | 2
arch/x86/xen/enlighten.c | 1
arch/xtensa/include/asm/pgalloc.h | 2
arch/xtensa/include/asm/pgtable.h | 1
arch/xtensa/include/uapi/asm/mman.h | 3
arch/xtensa/platforms/iss/setup.c | 1
drivers/block/zram/zram_drv.h | 2
drivers/bus/brcmstb_gisb.c | 1
drivers/char/ipmi/ipmi_msghandler.c | 1
drivers/clk/analogbits/wrpll-cln28hpc.c | 4
drivers/edac/altera_edac.c | 1
drivers/firmware/google/gsmi.c | 1
drivers/gpu/drm/nouveau/include/nvif/if000c.h | 1
drivers/gpu/drm/nouveau/nouveau_svm.c | 162 ++-
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h | 1
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c | 6
drivers/hv/vmbus_drv.c | 1
drivers/hwtracing/coresight/coresight-cpu-debug.c | 1
drivers/leds/trigger/ledtrig-activity.c | 1
drivers/leds/trigger/ledtrig-heartbeat.c | 1
drivers/leds/trigger/ledtrig-panic.c | 1
drivers/misc/bcm-vk/bcm_vk_dev.c | 1
drivers/misc/ibmasm/heartbeat.c | 1
drivers/misc/pvpanic/pvpanic.c | 1
drivers/net/ipa/ipa_smp2p.c | 1
drivers/parisc/power.c | 1
drivers/power/reset/ltc2952-poweroff.c | 1
drivers/remoteproc/remoteproc_core.c | 1
drivers/s390/char/con3215.c | 1
drivers/s390/char/con3270.c | 1
drivers/s390/char/sclp.c | 1
drivers/s390/char/sclp_con.c | 1
drivers/s390/char/sclp_vt220.c | 1
drivers/s390/char/zcore.c | 1
drivers/soc/bcm/brcmstb/pm/pm-arm.c | 1
drivers/staging/olpc_dcon/olpc_dcon.c | 1
drivers/video/fbdev/hyperv_fb.c | 1
drivers/virtio/virtio_mem.c | 2
fs/Kconfig | 15
fs/exec.c | 3
fs/hfsplus/inode.c | 5
fs/hfsplus/xattr.c | 1
fs/nfsd/nfs4state.c | 2
fs/nilfs2/btree.c | 1
fs/open.c | 13
fs/proc/base.c | 6
fs/proc/fd.c | 20
fs/proc/kcore.c | 136 ++
fs/proc/task_mmu.c | 34
fs/seq_file.c | 43
fs/userfaultfd.c | 15
include/asm-generic/bug.h | 3
include/linux/ascii85.h | 3
include/linux/bootmem_info.h | 68 +
include/linux/compat.h | 2
include/linux/compiler-clang.h | 17
include/linux/compiler-gcc.h | 6
include/linux/compiler_types.h | 2
include/linux/huge_mm.h | 74 -
include/linux/hugetlb.h | 80 +
include/linux/hugetlb_cgroup.h | 19
include/linux/kcore.h | 3
include/linux/kernel.h | 227 ----
include/linux/kprobes.h | 1
include/linux/kstrtox.h | 155 ++
include/linux/memblock.h | 4
include/linux/memory_hotplug.h | 27
include/linux/mempolicy.h | 9
include/linux/memremap.h | 2
include/linux/migrate.h | 27
include/linux/mm.h | 18
include/linux/mm_types.h | 2
include/linux/mmu_notifier.h | 26
include/linux/mmzone.h | 27
include/linux/mpi.h | 4
include/linux/page-flags.h | 22
include/linux/panic.h | 98 +
include/linux/panic_notifier.h | 12
include/linux/pgtable.h | 44
include/linux/rmap.h | 13
include/linux/seq_file.h | 10
include/linux/shmem_fs.h | 19
include/linux/signal.h | 2
include/linux/string.h | 7
include/linux/string_helpers.h | 31
include/linux/sunrpc/cache.h | 1
include/linux/swap.h | 19
include/linux/swapops.h | 171 +--
include/linux/thread_info.h | 1
include/linux/userfaultfd_k.h | 5
include/linux/vmalloc.h | 15
include/linux/zbud.h | 23
include/trace/events/vmscan.h | 41
include/uapi/asm-generic/mman-common.h | 3
include/uapi/linux/mempolicy.h | 1
include/uapi/linux/userfaultfd.h | 7
init/main.c | 42
ipc/msg.c | 6
ipc/sem.c | 25
ipc/shm.c | 6
ipc/util.c | 44
ipc/util.h | 3
kernel/hung_task.c | 1
kernel/kexec_core.c | 1
kernel/kprobes.c | 2
kernel/panic.c | 1
kernel/rcu/tree.c | 2
kernel/signal.c | 14
kernel/sysctl.c | 4
kernel/trace/trace.c | 1
lib/Kconfig.debug | 12
lib/decompress_bunzip2.c | 6
lib/decompress_unlz4.c | 8
lib/decompress_unlzo.c | 3
lib/decompress_unxz.c | 2
lib/decompress_unzstd.c | 4
lib/kstrtox.c | 5
lib/lz4/lz4_decompress.c | 2
lib/math/Makefile | 1
lib/math/rational-test.c | 56 +
lib/math/rational.c | 16
lib/mpi/longlong.h | 4
lib/mpi/mpicoder.c | 6
lib/mpi/mpiutil.c | 2
lib/parser.c | 1
lib/string.c | 2
lib/string_helpers.c | 142 +-
lib/test-string_helpers.c | 157 ++-
lib/test_hmm.c | 127 ++
lib/test_hmm_uapi.h | 2
lib/test_string.c | 5
lib/vsprintf.c | 1
lib/xz/xz_dec_bcj.c | 2
lib/xz/xz_dec_lzma2.c | 8
lib/zlib_inflate/inffast.c | 2
lib/zstd/huf.h | 2
mm/Kconfig | 16
mm/Makefile | 2
mm/bootmem_info.c | 127 ++
mm/compaction.c | 20
mm/debug_vm_pgtable.c | 109 --
mm/gup.c | 58 +
mm/hmm.c | 12
mm/huge_memory.c | 269 ++---
mm/hugetlb.c | 369 +++++--
mm/hugetlb_vmemmap.c | 332 ++++++
mm/hugetlb_vmemmap.h | 53 -
mm/internal.h | 29
mm/kfence/core.c | 4
mm/khugepaged.c | 20
mm/madvise.c | 66 +
mm/mapping_dirty_helpers.c | 2
mm/memblock.c | 28
mm/memcontrol.c | 4
mm/memory-failure.c | 38
mm/memory.c | 239 +++-
mm/memory_hotplug.c | 161 ---
mm/mempolicy.c | 323 ++----
mm/migrate.c | 268 +----
mm/mlock.c | 12
mm/mmap_lock.c | 59 -
mm/mprotect.c | 18
mm/nommu.c | 5
mm/oom_kill.c | 2
mm/page_alloc.c | 5
mm/page_vma_mapped.c | 15
mm/rmap.c | 644 +++++++++---
mm/shmem.c | 125 --
mm/sparse-vmemmap.c | 432 +++++++-
mm/sparse.c | 1
mm/swap.c | 2
mm/swapfile.c | 2
mm/userfaultfd.c | 249 ++--
mm/util.c | 40
mm/vmalloc.c | 37
mm/vmscan.c | 20
mm/workingset.c | 10
mm/z3fold.c | 39
mm/zbud.c | 235 ++--
mm/zsmalloc.c | 5
mm/zswap.c | 26
scripts/checkpatch.pl | 16
tools/testing/selftests/vm/.gitignore | 3
tools/testing/selftests/vm/Makefile | 5
tools/testing/selftests/vm/hmm-tests.c | 158 +++
tools/testing/selftests/vm/khugepaged.c | 4
tools/testing/selftests/vm/madv_populate.c | 342 ++++++
tools/testing/selftests/vm/pkey-x86.h | 1
tools/testing/selftests/vm/protection_keys.c | 85 +
tools/testing/selftests/vm/run_vmtests.sh | 16
tools/testing/selftests/vm/userfaultfd.c | 1094 ++++++++++-----------
299 files changed, 6277 insertions(+), 3183 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-06-25 1:38 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-25 1:38 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
24 patches, based on 4a09d388f2ab382f217a764e6a152b3f614246f6.
Subsystems affected by this patch series:
mm/thp
nilfs2
mm/vmalloc
kthread
mm/hugetlb
mm/memory-failure
mm/pagealloc
MAINTAINERS
mailmap
Subsystem: mm/thp
Hugh Dickins <hughd@google.com>:
Patch series "mm: page_vma_mapped_walk() cleanup and THP fixes":
mm: page_vma_mapped_walk(): use page for pvmw->page
mm: page_vma_mapped_walk(): settle PageHuge on entry
mm: page_vma_mapped_walk(): use pmde for *pvmw->pmd
mm: page_vma_mapped_walk(): prettify PVMW_MIGRATION block
mm: page_vma_mapped_walk(): crossing page table boundary
mm: page_vma_mapped_walk(): add a level of indentation
mm: page_vma_mapped_walk(): use goto instead of while (1)
mm: page_vma_mapped_walk(): get vma_address_end() earlier
mm/thp: fix page_vma_mapped_walk() if THP mapped by ptes
mm/thp: another PVMW_SYNC fix in page_vma_mapped_walk()
Subsystem: nilfs2
Pavel Skripkin <paskripkin@gmail.com>:
nilfs2: fix memory leak in nilfs_sysfs_delete_device_group
Subsystem: mm/vmalloc
Claudio Imbrenda <imbrenda@linux.ibm.com>:
Patch series "mm: add vmalloc_no_huge and use it", v4:
mm/vmalloc: add vmalloc_no_huge
KVM: s390: prepare for hugepage vmalloc
Daniel Axtens <dja@axtens.net>:
mm/vmalloc: unbreak kasan vmalloc support
Subsystem: kthread
Petr Mladek <pmladek@suse.com>:
Patch series "kthread_worker: Fix race between kthread_mod_delayed_work():
kthread_worker: split code for canceling the delayed work timer
kthread: prevent deadlock when kthread_mod_delayed_work() races with kthread_cancel_delayed_work_sync()
Subsystem: mm/hugetlb
Hugh Dickins <hughd@google.com>:
mm, futex: fix shared futex pgoff on shmem huge page
Subsystem: mm/memory-failure
Tony Luck <tony.luck@intel.com>:
Patch series "mm,hwpoison: fix sending SIGBUS for Action Required MCE", v5:
mm/memory-failure: use a mutex to avoid memory_failure() races
Aili Yao <yaoaili@kingsoft.com>:
mm,hwpoison: return -EHWPOISON to denote that the page has already been poisoned
Naoya Horiguchi <naoya.horiguchi@nec.com>:
mm/hwpoison: do not lock page again when me_huge_page() successfully recovers
Subsystem: mm/pagealloc
Rasmus Villemoes <linux@rasmusvillemoes.dk>:
mm/page_alloc: __alloc_pages_bulk(): do bounds check before accessing array
Mel Gorman <mgorman@techsingularity.net>:
mm/page_alloc: do bulk array bounds check after checking populated elements
Subsystem: MAINTAINERS
Marek Behún <kabel@kernel.org>:
MAINTAINERS: fix Marek's identity again
Subsystem: mailmap
Marek Behún <kabel@kernel.org>:
mailmap: add Marek's other e-mail address and identity without diacritics
.mailmap | 2
MAINTAINERS | 4
arch/s390/kvm/pv.c | 7 +
fs/nilfs2/sysfs.c | 1
include/linux/hugetlb.h | 16 ---
include/linux/pagemap.h | 13 +-
include/linux/vmalloc.h | 1
kernel/futex.c | 3
kernel/kthread.c | 81 ++++++++++------
mm/hugetlb.c | 5 -
mm/memory-failure.c | 83 +++++++++++------
mm/page_alloc.c | 6 +
mm/page_vma_mapped.c | 233 +++++++++++++++++++++++++++---------------------
mm/vmalloc.c | 41 ++++++--
14 files changed, 297 insertions(+), 199 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-06-16 1:22 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-16 1:22 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
18 patches, based on 94f0b2d4a1d0c52035aef425da5e022bd2cb1c71.
Subsystems affected by this patch series:
mm/memory-failure
mm/swap
mm/slub
mm/hugetlb
mm/memory-failure
coredump
mm/slub
mm/thp
mm/sparsemem
Subsystem: mm/memory-failure
Naoya Horiguchi <naoya.horiguchi@nec.com>:
mm,hwpoison: fix race with hugetlb page allocation
Subsystem: mm/swap
Peter Xu <peterx@redhat.com>:
mm/swap: fix pte_same_as_swp() not removing uffd-wp bit when compare
Subsystem: mm/slub
Kees Cook <keescook@chromium.org>:
Patch series "Actually fix freelist pointer vs redzoning", v4:
mm/slub: clarify verification reporting
mm/slub: fix redzoning for small allocations
mm/slub: actually fix freelist pointer vs redzoning
Subsystem: mm/hugetlb
Mike Kravetz <mike.kravetz@oracle.com>:
mm/hugetlb: expand restore_reserve_on_error functionality
Subsystem: mm/memory-failure
yangerkun <yangerkun@huawei.com>:
mm/memory-failure: make sure wait for page writeback in memory_failure
Subsystem: coredump
Pingfan Liu <kernelfans@gmail.com>:
crash_core, vmcoreinfo: append 'SECTION_SIZE_BITS' to vmcoreinfo
Subsystem: mm/slub
Andrew Morton <akpm@linux-foundation.org>:
mm/slub.c: include swab.h
Subsystem: mm/thp
Xu Yu <xuyu@linux.alibaba.com>:
mm, thp: use head page in __migration_entry_wait()
Hugh Dickins <hughd@google.com>:
Patch series "mm/thp: fix THP splitting unmap BUGs and related", v10:
mm/thp: fix __split_huge_pmd_locked() on shmem migration entry
mm/thp: make is_huge_zero_pmd() safe and quicker
mm/thp: try_to_unmap() use TTU_SYNC for safe splitting
mm/thp: fix vma_address() if virtual address below file offset
Jue Wang <juew@google.com>:
mm/thp: fix page_address_in_vma() on file THP tails
Hugh Dickins <hughd@google.com>:
mm/thp: unmap_mapping_page() to fix THP truncate_cleanup_page()
Yang Shi <shy828301@gmail.com>:
mm: thp: replace DEBUG_VM BUG with VM_WARN when unmap fails for split
Subsystem: mm/sparsemem
Miles Chen <miles.chen@mediatek.com>:
mm/sparse: fix check_usemap_section_nr warnings
Documentation/vm/slub.rst | 10 +--
fs/hugetlbfs/inode.c | 1
include/linux/huge_mm.h | 8 ++
include/linux/hugetlb.h | 8 ++
include/linux/mm.h | 3 +
include/linux/rmap.h | 1
include/linux/swapops.h | 15 +++--
kernel/crash_core.c | 1
mm/huge_memory.c | 58 ++++++++++---------
mm/hugetlb.c | 137 +++++++++++++++++++++++++++++++++++++---------
mm/internal.h | 51 ++++++++++++-----
mm/memory-failure.c | 36 +++++++++++-
mm/memory.c | 41 +++++++++++++
mm/migrate.c | 1
mm/page_vma_mapped.c | 27 +++++----
mm/pgtable-generic.c | 5 -
mm/rmap.c | 41 +++++++++----
mm/slab_common.c | 3 -
mm/slub.c | 37 +++++-------
mm/sparse.c | 13 +++-
mm/swapfile.c | 2
mm/truncate.c | 43 ++++++--------
22 files changed, 388 insertions(+), 154 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-06-05 3:00 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-06-05 3:00 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
13 patches, based on 16f0596fc1d78a1f3ae4628cff962bb297dc908c.
Subsystems affected by this patch series:
mips
mm/kfence
init
mm/debug
mm/pagealloc
mm/memory-hotplug
mm/hugetlb
proc
mm/kasan
mm/hugetlb
lib
ocfs2
mailmap
Subsystem: mips
Thomas Bogendoerfer <tsbogend@alpha.franken.de>:
Revert "MIPS: make userspace mapping young by default"
Subsystem: mm/kfence
Marco Elver <elver@google.com>:
kfence: use TASK_IDLE when awaiting allocation
Subsystem: init
Mark Rutland <mark.rutland@arm.com>:
pid: take a reference when initializing `cad_pid`
Subsystem: mm/debug
Gerald Schaefer <gerald.schaefer@linux.ibm.com>:
mm/debug_vm_pgtable: fix alignment for pmd/pud_advanced_tests()
Subsystem: mm/pagealloc
Ding Hui <dinghui@sangfor.com.cn>:
mm/page_alloc: fix counting of free pages after take off from buddy
Subsystem: mm/memory-hotplug
David Hildenbrand <david@redhat.com>:
drivers/base/memory: fix trying offlining memory blocks with memory holes on aarch64
Subsystem: mm/hugetlb
Naoya Horiguchi <naoya.horiguchi@nec.com>:
hugetlb: pass head page to remove_hugetlb_page()
Subsystem: proc
David Matlack <dmatlack@google.com>:
proc: add .gitignore for proc-subset-pid selftest
Subsystem: mm/kasan
Yu Kuai <yukuai3@huawei.com>:
mm/kasan/init.c: fix doc warning
Subsystem: mm/hugetlb
Mina Almasry <almasrymina@google.com>:
mm, hugetlb: fix simple resv_huge_pages underflow on UFFDIO_COPY
Subsystem: lib
YueHaibing <yuehaibing@huawei.com>:
lib: crc64: fix kernel-doc warning
Subsystem: ocfs2
Junxiao Bi <junxiao.bi@oracle.com>:
ocfs2: fix data corruption by fallocate
Subsystem: mailmap
Michel Lespinasse <michel@lespinasse.org>:
mailmap: use private address for Michel Lespinasse
.mailmap | 3 +
arch/mips/mm/cache.c | 30 ++++++++---------
drivers/base/memory.c | 6 +--
fs/ocfs2/file.c | 55 +++++++++++++++++++++++++++++---
include/linux/pgtable.h | 8 ++++
init/main.c | 2 -
lib/crc64.c | 2 -
mm/debug_vm_pgtable.c | 4 +-
mm/hugetlb.c | 16 +++++++--
mm/kasan/init.c | 4 +-
mm/kfence/core.c | 6 +--
mm/memory.c | 4 ++
mm/page_alloc.c | 2 +
tools/testing/selftests/proc/.gitignore | 1
14 files changed, 107 insertions(+), 36 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-05-23 0:41 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-05-23 0:41 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
10 patches, based on 4ff2473bdb4cf2bb7d208ccf4418d3d7e6b1652c.
Subsystems affected by this patch series:
mm/pagealloc
mm/gup
ipc
selftests
mm/kasan
kernel/watchdog
bitmap
procfs
lib
mm/userfaultfd
Subsystem: mm/pagealloc
Arnd Bergmann <arnd@arndb.de>:
mm/shuffle: fix section mismatch warning
Subsystem: mm/gup
Michal Hocko <mhocko@suse.com>:
Revert "mm/gup: check page posion status for coredump."
Subsystem: ipc
Varad Gautam <varad.gautam@suse.com>:
ipc/mqueue, msg, sem: avoid relying on a stack reference past its expiry
Subsystem: selftests
Yang Yingliang <yangyingliang@huawei.com>:
tools/testing/selftests/exec: fix link error
Subsystem: mm/kasan
Alexander Potapenko <glider@google.com>:
kasan: slab: always reset the tag in get_freepointer_safe()
Subsystem: kernel/watchdog
Petr Mladek <pmladek@suse.com>:
watchdog: reliable handling of timestamps
Subsystem: bitmap
Rikard Falkeborn <rikard.falkeborn@gmail.com>:
linux/bits.h: fix compilation error with GENMASK
Subsystem: procfs
Alexey Dobriyan <adobriyan@gmail.com>:
proc: remove Alexey from MAINTAINERS
Subsystem: lib
Zhen Lei <thunder.leizhen@huawei.com>:
lib: kunit: suppress a compilation warning of frame size
Subsystem: mm/userfaultfd
Mike Kravetz <mike.kravetz@oracle.com>:
userfaultfd: hugetlbfs: fix new flag usage in error path
MAINTAINERS | 1 -
fs/hugetlbfs/inode.c | 2 +-
include/linux/bits.h | 2 +-
include/linux/const.h | 8 ++++++++
include/linux/minmax.h | 10 ++--------
ipc/mqueue.c | 6 ++++--
ipc/msg.c | 6 ++++--
ipc/sem.c | 6 ++++--
kernel/watchdog.c | 34 ++++++++++++++++++++--------------
lib/Makefile | 1 +
mm/gup.c | 4 ----
mm/internal.h | 20 --------------------
mm/shuffle.h | 4 ++--
mm/slub.c | 1 +
mm/userfaultfd.c | 28 ++++++++++++++--------------
tools/include/linux/bits.h | 2 +-
tools/include/linux/const.h | 8 ++++++++
tools/testing/selftests/exec/Makefile | 6 +++---
18 files changed, 74 insertions(+), 75 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-05-15 0:26 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-05-15 0:26 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
13 patches, based on bd3c9cdb21a2674dd0db70199df884828e37abd4.
Subsystems affected by this patch series:
mm/hugetlb
mm/slub
resource
squashfs
mm/userfaultfd
mm/ksm
mm/pagealloc
mm/kasan
mm/pagemap
hfsplus
modprobe
mm/ioremap
Subsystem: mm/hugetlb
Peter Xu <peterx@redhat.com>:
Patch series "mm/hugetlb: Fix issues on file sealing and fork", v2:
mm/hugetlb: fix F_SEAL_FUTURE_WRITE
mm/hugetlb: fix cow where page writtable in child
Subsystem: mm/slub
Vlastimil Babka <vbabka@suse.cz>:
mm, slub: move slub_debug static key enabling outside slab_mutex
Subsystem: resource
Alistair Popple <apopple@nvidia.com>:
kernel/resource: fix return code check in __request_free_mem_region
Subsystem: squashfs
Phillip Lougher <phillip@squashfs.org.uk>:
squashfs: fix divide error in calculate_skip()
Subsystem: mm/userfaultfd
Axel Rasmussen <axelrasmussen@google.com>:
userfaultfd: release page in error path to avoid BUG_ON
Subsystem: mm/ksm
Hugh Dickins <hughd@google.com>:
ksm: revert "use GET_KSM_PAGE_NOLOCK to get ksm page in remove_rmap_item_from_tree()"
Subsystem: mm/pagealloc
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
mm: fix struct page layout on 32-bit systems
Subsystem: mm/kasan
Peter Collingbourne <pcc@google.com>:
kasan: fix unit tests with CONFIG_UBSAN_LOCAL_BOUNDS enabled
Subsystem: mm/pagemap
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
mm/filemap: fix readahead return types
Subsystem: hfsplus
Jouni Roivas <jouni.roivas@tuxera.com>:
hfsplus: prevent corruption in shrinking truncate
Subsystem: modprobe
Rasmus Villemoes <linux@rasmusvillemoes.dk>:
docs: admin-guide: update description for kernel.modprobe sysctl
Subsystem: mm/ioremap
Christophe Leroy <christophe.leroy@csgroup.eu>:
mm/ioremap: fix iomap_max_page_shift
Documentation/admin-guide/sysctl/kernel.rst | 9 ++++---
fs/hfsplus/extents.c | 7 +++--
fs/hugetlbfs/inode.c | 5 ++++
fs/iomap/buffered-io.c | 4 +--
fs/squashfs/file.c | 6 ++--
include/linux/mm.h | 32 ++++++++++++++++++++++++++
include/linux/mm_types.h | 4 +--
include/linux/pagemap.h | 6 ++--
include/net/page_pool.h | 12 +++++++++
kernel/resource.c | 2 -
lib/test_kasan.c | 29 ++++++++++++++++++-----
mm/hugetlb.c | 1
mm/ioremap.c | 6 ++--
mm/ksm.c | 3 +-
mm/shmem.c | 34 ++++++++++++----------------
mm/slab_common.c | 10 ++++++++
mm/slub.c | 9 -------
net/core/page_pool.c | 12 +++++----
18 files changed, 129 insertions(+), 62 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2021-05-07 1:01 incoming Andrew Morton
@ 2021-05-07 7:12 ` Linus Torvalds
0 siblings, 0 replies; 409+ messages in thread
From: Linus Torvalds @ 2021-05-07 7:12 UTC (permalink / raw)
To: Andrew Morton; +Cc: mm-commits, Linux-MM
On Thu, May 6, 2021 at 6:01 PM Andrew Morton <akpm@linux-foundation.org> wrote:
>
> I've been wobbly about the secretmem patches due to doubts about
> whether the feature is sufficiently useful to justify inclusion, but
> developers are now weighing in with helpful information and I've asked Mike
> for an extensively updated [0/n] changelog. This will take a few days
> to play out so it is possible that I will prevail upon you for a post-rc1
> merge.
Oh, much too late for this release by now.
> If that's a problem, there's always 5.13-rc1.
5.13-rc1 is two days from now, it would be for 5.14-rc1.. How time -
and version numbers - fly.
Linus
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-05-07 1:01 Andrew Morton
2021-05-07 7:12 ` incoming Linus Torvalds
0 siblings, 1 reply; 409+ messages in thread
From: Andrew Morton @ 2021-05-07 1:01 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
This is everything else from -mm for this merge window, with the
possible exception of Mike Rapoport's "secretmem" syscall patch series
(https://lkml.kernel.org/r/20210303162209.8609-1-rppt@kernel.org).
I've been wobbly about the secretmem patches due to doubts about
whether the feature is sufficiently useful to justify inclusion, but
developers are now weighing in with helpful information and I've asked Mike
for an extensively updated [0/n] changelog. This will take a few days
to play out so it is possible that I will prevail upon you for a post-rc1
merge. If that's a problem, there's always 5.13-rc1.
91 patches, based on 8ca5297e7e38f2dc8c753d33a5092e7be181fff0, plus
previously sent patches.
Thanks.
Subsystems affected by this patch series:
alpha
procfs
sysctl
misc
core-kernel
bitmap
lib
compat
checkpatch
epoll
isofs
nilfs2
hpfs
exit
fork
kexec
gcov
panic
delayacct
gdb
resource
selftests
async
initramfs
ipc
mm/cleanups
drivers/char
mm/slub
spelling
Subsystem: alpha
Randy Dunlap <rdunlap@infradead.org>:
alpha: eliminate old-style function definitions
alpha: csum_partial_copy.c: add function prototypes from <net/checksum.h>
Subsystem: procfs
Colin Ian King <colin.king@canonical.com>:
fs/proc/generic.c: fix incorrect pde_is_permanent check
Alexey Dobriyan <adobriyan@gmail.com>:
proc: save LOC in __xlate_proc_name()
proc: mandate ->proc_lseek in "struct proc_ops"
proc: delete redundant subset=pid check
selftests: proc: test subset=pid
Subsystem: sysctl
zhouchuangao <zhouchuangao@vivo.com>:
proc/sysctl: fix function name error in comments
Subsystem: misc
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
include: remove pagemap.h from blkdev.h
Andy Shevchenko <andriy.shevchenko@linux.intel.com>:
kernel.h: drop inclusion in bitmap.h
Wan Jiabing <wanjiabing@vivo.com>:
linux/profile.h: remove unnecessary declaration
Subsystem: core-kernel
Rasmus Villemoes <linux@rasmusvillemoes.dk>:
kernel/async.c: fix pr_debug statement
kernel/cred.c: make init_groups static
Subsystem: bitmap
Yury Norov <yury.norov@gmail.com>:
Patch series "lib/find_bit: fast path for small bitmaps", v6:
tools: disable -Wno-type-limits
tools: bitmap: sync function declarations with the kernel
tools: sync BITMAP_LAST_WORD_MASK() macro with the kernel
arch: rearrange headers inclusion order in asm/bitops for m68k, sh and h8300
lib: extend the scope of small_const_nbits() macro
tools: sync small_const_nbits() macro with the kernel
lib: inline _find_next_bit() wrappers
tools: sync find_next_bit implementation
lib: add fast path for find_next_*_bit()
lib: add fast path for find_first_*_bit() and find_last_bit()
tools: sync lib/find_bit implementation
MAINTAINERS: add entry for the bitmap API
Subsystem: lib
Bhaskar Chowdhury <unixbhaskar@gmail.com>:
lib/bch.c: fix a typo in the file bch.c
Wang Qing <wangqing@vivo.com>:
lib: fix inconsistent indenting in process_bit1()
ToastC <mrtoastcheng@gmail.com>:
lib/list_sort.c: fix typo in function description
Bhaskar Chowdhury <unixbhaskar@gmail.com>:
lib/genalloc.c: Fix a typo
Richard Fitzgerald <rf@opensource.cirrus.com>:
lib: crc8: pointer to data block should be const
Zqiang <qiang.zhang@windriver.com>:
lib: stackdepot: turn depot_lock spinlock to raw_spinlock
Alex Shi <alexs@kernel.org>:
lib/percpu_counter: tame kernel-doc compile warning
lib/genalloc: add parameter description to fix doc compile warning
Randy Dunlap <rdunlap@infradead.org>:
lib: parser: clean up kernel-doc
Subsystem: compat
Masahiro Yamada <masahiroy@kernel.org>:
include/linux/compat.h: remove unneeded declaration from COMPAT_SYSCALL_DEFINEx()
Subsystem: checkpatch
Joe Perches <joe@perches.com>:
checkpatch: warn when missing newline in return sysfs_emit() formats
Vincent Mailhol <mailhol.vincent@wanadoo.fr>:
checkpatch: exclude four preprocessor sub-expressions from MACRO_ARG_REUSE
Christophe JAILLET <christophe.jaillet@wanadoo.fr>:
checkpatch: improve ALLOC_ARRAY_ARGS test
Subsystem: epoll
Davidlohr Bueso <dave@stgolabs.net>:
Patch series "fs/epoll: restore user-visible behavior upon event ready":
kselftest: introduce new epoll test case
fs/epoll: restore waking from ep_done_scan()
Subsystem: isofs
"Gustavo A. R. Silva" <gustavoars@kernel.org>:
isofs: fix fall-through warnings for Clang
Subsystem: nilfs2
Liu xuzhi <liu.xuzhi@zte.com.cn>:
fs/nilfs2: fix misspellings using codespell tool
Lu Jialin <lujialin4@huawei.com>:
nilfs2: fix typos in comments
Subsystem: hpfs
"Gustavo A. R. Silva" <gustavoars@kernel.org>:
hpfs: replace one-element array with flexible-array member
Subsystem: exit
Jim Newsome <jnewsome@torproject.org>:
do_wait: make PIDTYPE_PID case O(1) instead of O(n)
Subsystem: fork
Rolf Eike Beer <eb@emlix.com>:
kernel/fork.c: simplify copy_mm()
Xiaofeng Cao <cxfcosmos@gmail.com>:
kernel/fork.c: fix typos
Subsystem: kexec
Saeed Mirzamohammadi <saeed.mirzamohammadi@oracle.com>:
kernel/crash_core: add crashkernel=auto for vmcore creation
Joe LeVeque <jolevequ@microsoft.com>:
kexec: Add kexec reboot string
Jia-Ju Bai <baijiaju1990@gmail.com>:
kernel: kexec_file: fix error return code of kexec_calculate_store_digests()
Pavel Tatashin <pasha.tatashin@soleen.com>:
kexec: dump kmessage before machine_kexec
Subsystem: gcov
Johannes Berg <johannes.berg@intel.com>:
gcov: combine common code
gcov: simplify buffer allocation
gcov: use kvmalloc()
Nick Desaulniers <ndesaulniers@google.com>:
gcov: clang: drop support for clang-10 and older
Subsystem: panic
He Ying <heying24@huawei.com>:
smp: kernel/panic.c - silence warnings
Subsystem: delayacct
Yafang Shao <laoar.shao@gmail.com>:
delayacct: clear right task's flag after blkio completes
Subsystem: gdb
Johannes Berg <johannes.berg@intel.com>:
gdb: lx-symbols: store the abspath()
Barry Song <song.bao.hua@hisilicon.com>:
Patch series "scripts/gdb: clarify the platforms supporting lx_current and add arm64 support", v2:
scripts/gdb: document lx_current is only supported by x86
scripts/gdb: add lx_current support for arm64
Subsystem: resource
David Hildenbrand <david@redhat.com>:
Patch series "kernel/resource: make walk_system_ram_res() and walk_mem_res() search the whole tree", v2:
kernel/resource: make walk_system_ram_res() find all busy IORESOURCE_SYSTEM_RAM resources
kernel/resource: make walk_mem_res() find all busy IORESOURCE_MEM resources
kernel/resource: remove first_lvl / siblings_only logic
Alistair Popple <apopple@nvidia.com>:
kernel/resource: allow region_intersects users to hold resource_lock
kernel/resource: refactor __request_region to allow external locking
kernel/resource: fix locking in request_free_mem_region
Subsystem: selftests
Zhang Yunkai <zhang.yunkai@zte.com.cn>:
selftests: remove duplicate include
Subsystem: async
Rasmus Villemoes <linux@rasmusvillemoes.dk>:
kernel/async.c: stop guarding pr_debug() statements
kernel/async.c: remove async_unregister_domain()
Subsystem: initramfs
Rasmus Villemoes <linux@rasmusvillemoes.dk>:
Patch series "background initramfs unpacking, and CONFIG_MODPROBE_PATH", v3:
init/initramfs.c: do unpacking asynchronously
modules: add CONFIG_MODPROBE_PATH
Subsystem: ipc
Bhaskar Chowdhury <unixbhaskar@gmail.com>:
ipc/sem.c: mundane typo fixes
Subsystem: mm/cleanups
Shijie Luo <luoshijie1@huawei.com>:
mm: fix some typos and code style problems
Subsystem: drivers/char
David Hildenbrand <david@redhat.com>:
Patch series "drivers/char: remove /dev/kmem for good":
drivers/char: remove /dev/kmem for good
mm: remove xlate_dev_kmem_ptr()
mm/vmalloc: remove vwrite()
Subsystem: mm/slub
Maninder Singh <maninder1.s@samsung.com>:
arm: print alloc free paths for address in registers
Subsystem: spelling
Drew Fustini <drew@beagleboard.org>:
scripts/spelling.txt: add "overlfow"
zuoqilin <zuoqilin@yulong.com>:
scripts/spelling.txt: Add "diabled" typo
Drew Fustini <drew@beagleboard.org>:
scripts/spelling.txt: add "overflw"
Colin Ian King <colin.king@canonical.com>:
mm/slab.c: fix spelling mistake "disired" -> "desired"
Bhaskar Chowdhury <unixbhaskar@gmail.com>:
include/linux/pgtable.h: few spelling fixes
zhouchuangao <zhouchuangao@vivo.com>:
kernel/umh.c: fix some spelling mistakes
Xiaofeng Cao <cxfcosmos@gmail.com>:
kernel/user_namespace.c: fix typos
Bhaskar Chowdhury <unixbhaskar@gmail.com>:
kernel/up.c: fix typo
Xiaofeng Cao <caoxiaofeng@yulong.com>:
kernel/sys.c: fix typo
dingsenjie <dingsenjie@yulong.com>:
fs: fat: fix spelling typo of values
Bhaskar Chowdhury <unixbhaskar@gmail.com>:
ipc/sem.c: spelling fix
Masahiro Yamada <masahiroy@kernel.org>:
treewide: remove editor modelines and cruft
Ingo Molnar <mingo@kernel.org>:
mm: fix typos in comments
Lu Jialin <lujialin4@huawei.com>:
mm: fix typos in comments
Documentation/admin-guide/devices.txt | 2
Documentation/admin-guide/kdump/kdump.rst | 3
Documentation/admin-guide/kernel-parameters.txt | 18
Documentation/dev-tools/gdb-kernel-debugging.rst | 4
MAINTAINERS | 16
arch/Kconfig | 20
arch/alpha/include/asm/io.h | 5
arch/alpha/kernel/pc873xx.c | 4
arch/alpha/lib/csum_partial_copy.c | 1
arch/arm/configs/dove_defconfig | 1
arch/arm/configs/magician_defconfig | 1
arch/arm/configs/moxart_defconfig | 1
arch/arm/configs/mps2_defconfig | 1
arch/arm/configs/mvebu_v5_defconfig | 1
arch/arm/configs/xcep_defconfig | 1
arch/arm/include/asm/bug.h | 1
arch/arm/include/asm/io.h | 5
arch/arm/kernel/process.c | 11
arch/arm/kernel/traps.c | 1
arch/h8300/include/asm/bitops.h | 8
arch/hexagon/configs/comet_defconfig | 1
arch/hexagon/include/asm/io.h | 1
arch/ia64/include/asm/io.h | 1
arch/ia64/include/asm/uaccess.h | 18
arch/m68k/atari/time.c | 7
arch/m68k/configs/amcore_defconfig | 1
arch/m68k/include/asm/bitops.h | 6
arch/m68k/include/asm/io_mm.h | 5
arch/mips/include/asm/io.h | 5
arch/openrisc/configs/or1ksim_defconfig | 1
arch/parisc/include/asm/io.h | 5
arch/parisc/include/asm/pdc_chassis.h | 1
arch/powerpc/include/asm/io.h | 5
arch/s390/include/asm/io.h | 5
arch/sh/configs/edosk7705_defconfig | 1
arch/sh/configs/se7206_defconfig | 1
arch/sh/configs/sh2007_defconfig | 1
arch/sh/configs/sh7724_generic_defconfig | 1
arch/sh/configs/sh7770_generic_defconfig | 1
arch/sh/configs/sh7785lcr_32bit_defconfig | 1
arch/sh/include/asm/bitops.h | 5
arch/sh/include/asm/io.h | 5
arch/sparc/configs/sparc64_defconfig | 1
arch/sparc/include/asm/io_64.h | 5
arch/um/drivers/cow.h | 7
arch/xtensa/configs/xip_kc705_defconfig | 1
block/blk-settings.c | 1
drivers/auxdisplay/panel.c | 7
drivers/base/firmware_loader/main.c | 2
drivers/block/brd.c | 1
drivers/block/loop.c | 1
drivers/char/Kconfig | 10
drivers/char/mem.c | 231 --------
drivers/gpu/drm/qxl/qxl_drv.c | 1
drivers/isdn/capi/kcapi_proc.c | 1
drivers/md/bcache/super.c | 1
drivers/media/usb/pwc/pwc-uncompress.c | 3
drivers/net/ethernet/adaptec/starfire.c | 8
drivers/net/ethernet/amd/atarilance.c | 8
drivers/net/ethernet/amd/pcnet32.c | 7
drivers/net/wireless/intersil/hostap/hostap_proc.c | 1
drivers/net/wireless/intersil/orinoco/orinoco_nortel.c | 8
drivers/net/wireless/intersil/orinoco/orinoco_pci.c | 8
drivers/net/wireless/intersil/orinoco/orinoco_plx.c | 8
drivers/net/wireless/intersil/orinoco/orinoco_tmd.c | 8
drivers/nvdimm/btt.c | 1
drivers/nvdimm/pmem.c | 1
drivers/parport/parport_ip32.c | 12
drivers/platform/x86/dell/dell_rbu.c | 3
drivers/scsi/53c700.c | 1
drivers/scsi/53c700.h | 1
drivers/scsi/ch.c | 6
drivers/scsi/esas2r/esas2r_main.c | 1
drivers/scsi/ips.c | 20
drivers/scsi/ips.h | 20
drivers/scsi/lasi700.c | 1
drivers/scsi/megaraid/mbox_defs.h | 2
drivers/scsi/megaraid/mega_common.h | 2
drivers/scsi/megaraid/megaraid_mbox.c | 2
drivers/scsi/megaraid/megaraid_mbox.h | 2
drivers/scsi/qla1280.c | 12
drivers/scsi/scsicam.c | 1
drivers/scsi/sni_53c710.c | 1
drivers/video/fbdev/matrox/matroxfb_base.c | 9
drivers/video/fbdev/vga16fb.c | 10
fs/configfs/configfs_internal.h | 4
fs/configfs/dir.c | 4
fs/configfs/file.c | 4
fs/configfs/inode.c | 4
fs/configfs/item.c | 4
fs/configfs/mount.c | 4
fs/configfs/symlink.c | 4
fs/eventpoll.c | 6
fs/fat/fatent.c | 2
fs/hpfs/hpfs.h | 3
fs/isofs/rock.c | 1
fs/nfs/dir.c | 7
fs/nfs/nfs4proc.c | 6
fs/nfs/nfs4renewd.c | 6
fs/nfs/nfs4state.c | 6
fs/nfs/nfs4xdr.c | 6
fs/nfsd/nfs4proc.c | 6
fs/nfsd/nfs4xdr.c | 6
fs/nfsd/xdr4.h | 6
fs/nilfs2/cpfile.c | 2
fs/nilfs2/ioctl.c | 4
fs/nilfs2/segment.c | 4
fs/nilfs2/the_nilfs.c | 2
fs/ocfs2/acl.c | 4
fs/ocfs2/acl.h | 4
fs/ocfs2/alloc.c | 4
fs/ocfs2/alloc.h | 4
fs/ocfs2/aops.c | 4
fs/ocfs2/aops.h | 4
fs/ocfs2/blockcheck.c | 4
fs/ocfs2/blockcheck.h | 4
fs/ocfs2/buffer_head_io.c | 4
fs/ocfs2/buffer_head_io.h | 4
fs/ocfs2/cluster/heartbeat.c | 4
fs/ocfs2/cluster/heartbeat.h | 4
fs/ocfs2/cluster/masklog.c | 4
fs/ocfs2/cluster/masklog.h | 4
fs/ocfs2/cluster/netdebug.c | 4
fs/ocfs2/cluster/nodemanager.c | 4
fs/ocfs2/cluster/nodemanager.h | 4
fs/ocfs2/cluster/ocfs2_heartbeat.h | 4
fs/ocfs2/cluster/ocfs2_nodemanager.h | 4
fs/ocfs2/cluster/quorum.c | 4
fs/ocfs2/cluster/quorum.h | 4
fs/ocfs2/cluster/sys.c | 4
fs/ocfs2/cluster/sys.h | 4
fs/ocfs2/cluster/tcp.c | 4
fs/ocfs2/cluster/tcp.h | 4
fs/ocfs2/cluster/tcp_internal.h | 4
fs/ocfs2/dcache.c | 4
fs/ocfs2/dcache.h | 4
fs/ocfs2/dir.c | 4
fs/ocfs2/dir.h | 4
fs/ocfs2/dlm/dlmapi.h | 4
fs/ocfs2/dlm/dlmast.c | 4
fs/ocfs2/dlm/dlmcommon.h | 4
fs/ocfs2/dlm/dlmconvert.c | 4
fs/ocfs2/dlm/dlmconvert.h | 4
fs/ocfs2/dlm/dlmdebug.c | 4
fs/ocfs2/dlm/dlmdebug.h | 4
fs/ocfs2/dlm/dlmdomain.c | 4
fs/ocfs2/dlm/dlmdomain.h | 4
fs/ocfs2/dlm/dlmlock.c | 4
fs/ocfs2/dlm/dlmmaster.c | 4
fs/ocfs2/dlm/dlmrecovery.c | 4
fs/ocfs2/dlm/dlmthread.c | 4
fs/ocfs2/dlm/dlmunlock.c | 4
fs/ocfs2/dlmfs/dlmfs.c | 4
fs/ocfs2/dlmfs/userdlm.c | 4
fs/ocfs2/dlmfs/userdlm.h | 4
fs/ocfs2/dlmglue.c | 4
fs/ocfs2/dlmglue.h | 4
fs/ocfs2/export.c | 4
fs/ocfs2/export.h | 4
fs/ocfs2/extent_map.c | 4
fs/ocfs2/extent_map.h | 4
fs/ocfs2/file.c | 4
fs/ocfs2/file.h | 4
fs/ocfs2/filecheck.c | 4
fs/ocfs2/filecheck.h | 4
fs/ocfs2/heartbeat.c | 4
fs/ocfs2/heartbeat.h | 4
fs/ocfs2/inode.c | 4
fs/ocfs2/inode.h | 4
fs/ocfs2/journal.c | 4
fs/ocfs2/journal.h | 4
fs/ocfs2/localalloc.c | 4
fs/ocfs2/localalloc.h | 4
fs/ocfs2/locks.c | 4
fs/ocfs2/locks.h | 4
fs/ocfs2/mmap.c | 4
fs/ocfs2/move_extents.c | 4
fs/ocfs2/move_extents.h | 4
fs/ocfs2/namei.c | 4
fs/ocfs2/namei.h | 4
fs/ocfs2/ocfs1_fs_compat.h | 4
fs/ocfs2/ocfs2.h | 4
fs/ocfs2/ocfs2_fs.h | 4
fs/ocfs2/ocfs2_ioctl.h | 4
fs/ocfs2/ocfs2_lockid.h | 4
fs/ocfs2/ocfs2_lockingver.h | 4
fs/ocfs2/refcounttree.c | 4
fs/ocfs2/refcounttree.h | 4
fs/ocfs2/reservations.c | 4
fs/ocfs2/reservations.h | 4
fs/ocfs2/resize.c | 4
fs/ocfs2/resize.h | 4
fs/ocfs2/slot_map.c | 4
fs/ocfs2/slot_map.h | 4
fs/ocfs2/stack_o2cb.c | 4
fs/ocfs2/stack_user.c | 4
fs/ocfs2/stackglue.c | 4
fs/ocfs2/stackglue.h | 4
fs/ocfs2/suballoc.c | 4
fs/ocfs2/suballoc.h | 4
fs/ocfs2/super.c | 4
fs/ocfs2/super.h | 4
fs/ocfs2/symlink.c | 4
fs/ocfs2/symlink.h | 4
fs/ocfs2/sysfile.c | 4
fs/ocfs2/sysfile.h | 4
fs/ocfs2/uptodate.c | 4
fs/ocfs2/uptodate.h | 4
fs/ocfs2/xattr.c | 4
fs/ocfs2/xattr.h | 4
fs/proc/generic.c | 13
fs/proc/inode.c | 18
fs/proc/proc_sysctl.c | 2
fs/reiserfs/procfs.c | 10
include/asm-generic/bitops/find.h | 108 +++
include/asm-generic/bitops/le.h | 38 +
include/asm-generic/bitsperlong.h | 12
include/asm-generic/io.h | 11
include/linux/align.h | 15
include/linux/async.h | 1
include/linux/bitmap.h | 11
include/linux/bitops.h | 12
include/linux/blkdev.h | 1
include/linux/compat.h | 1
include/linux/configfs.h | 4
include/linux/crc8.h | 2
include/linux/cred.h | 1
include/linux/delayacct.h | 20
include/linux/fs.h | 2
include/linux/genl_magic_func.h | 1
include/linux/genl_magic_struct.h | 1
include/linux/gfp.h | 2
include/linux/init_task.h | 1
include/linux/initrd.h | 2
include/linux/kernel.h | 9
include/linux/mm.h | 2
include/linux/mmzone.h | 2
include/linux/pgtable.h | 10
include/linux/proc_fs.h | 1
include/linux/profile.h | 3
include/linux/smp.h | 8
include/linux/swap.h | 1
include/linux/vmalloc.h | 7
include/uapi/linux/if_bonding.h | 11
include/uapi/linux/nfs4.h | 6
include/xen/interface/elfnote.h | 10
include/xen/interface/hvm/hvm_vcpu.h | 10
include/xen/interface/io/xenbus.h | 10
init/Kconfig | 12
init/initramfs.c | 38 +
init/main.c | 1
ipc/sem.c | 12
kernel/async.c | 68 --
kernel/configs/android-base.config | 1
kernel/crash_core.c | 7
kernel/cred.c | 2
kernel/exit.c | 67 ++
kernel/fork.c | 23
kernel/gcov/Kconfig | 1
kernel/gcov/base.c | 49 +
kernel/gcov/clang.c | 282 ----------
kernel/gcov/fs.c | 146 ++++-
kernel/gcov/gcc_4_7.c | 173 ------
kernel/gcov/gcov.h | 14
kernel/kexec_core.c | 4
kernel/kexec_file.c | 4
kernel/kmod.c | 2
kernel/resource.c | 198 ++++---
kernel/sys.c | 14
kernel/umh.c | 8
kernel/up.c | 2
kernel/user_namespace.c | 6
lib/bch.c | 2
lib/crc8.c | 2
lib/decompress_unlzma.c | 2
lib/find_bit.c | 68 --
lib/genalloc.c | 7
lib/list_sort.c | 2
lib/parser.c | 61 +-
lib/percpu_counter.c | 2
lib/stackdepot.c | 6
mm/balloon_compaction.c | 4
mm/compaction.c | 4
mm/filemap.c | 2
mm/gup.c | 2
mm/highmem.c | 2
mm/huge_memory.c | 6
mm/hugetlb.c | 6
mm/internal.h | 2
mm/kasan/kasan.h | 8
mm/kasan/quarantine.c | 4
mm/kasan/shadow.c | 4
mm/kfence/report.c | 2
mm/khugepaged.c | 2
mm/ksm.c | 6
mm/madvise.c | 4
mm/memcontrol.c | 18
mm/memory-failure.c | 2
mm/memory.c | 18
mm/mempolicy.c | 6
mm/migrate.c | 8
mm/mmap.c | 4
mm/mprotect.c | 2
mm/mremap.c | 2
mm/nommu.c | 10
mm/oom_kill.c | 2
mm/page-writeback.c | 4
mm/page_alloc.c | 16
mm/page_owner.c | 2
mm/page_vma_mapped.c | 2
mm/percpu-internal.h | 2
mm/percpu.c | 2
mm/pgalloc-track.h | 6
mm/rmap.c | 2
mm/slab.c | 8
mm/slub.c | 2
mm/swap.c | 4
mm/swap_slots.c | 2
mm/swap_state.c | 2
mm/vmalloc.c | 124 ----
mm/vmstat.c | 2
mm/z3fold.c | 2
mm/zpool.c | 2
mm/zsmalloc.c | 6
samples/configfs/configfs_sample.c | 2
scripts/checkpatch.pl | 15
scripts/gdb/linux/cpus.py | 23
scripts/gdb/linux/symbols.py | 3
scripts/spelling.txt | 3
tools/include/asm-generic/bitops/find.h | 85 ++-
tools/include/asm-generic/bitsperlong.h | 3
tools/include/linux/bitmap.h | 18
tools/lib/bitmap.c | 4
tools/lib/find_bit.c | 56 -
tools/scripts/Makefile.include | 1
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c | 44 +
tools/testing/selftests/kvm/lib/sparsebit.c | 1
tools/testing/selftests/mincore/mincore_selftest.c | 1
tools/testing/selftests/powerpc/mm/tlbie_test.c | 1
tools/testing/selftests/proc/Makefile | 1
tools/testing/selftests/proc/proc-subset-pid.c | 121 ++++
tools/testing/selftests/proc/read.c | 4
tools/usb/hcd-tests.sh | 2
343 files changed, 1383 insertions(+), 2119 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2021-05-05 17:44 ` incoming Andrew Morton
@ 2021-05-06 3:19 ` Anshuman Khandual
0 siblings, 0 replies; 409+ messages in thread
From: Anshuman Khandual @ 2021-05-06 3:19 UTC (permalink / raw)
To: Andrew Morton, Linus Torvalds; +Cc: Konstantin Ryabitsev, Linux-MM, mm-commits
On 5/5/21 11:14 PM, Andrew Morton wrote:
> On Wed, 5 May 2021 10:10:33 -0700 Linus Torvalds <torvalds@linux-foundation.org> wrote:
>
>> On Tue, May 4, 2021 at 8:16 PM Andrew Morton <akpm@linux-foundation.org> wrote:
>>> Let me resend right now with the same in-reply-to. Hopefully they will
>>> land in the correct place.
>> Well, you re-sent it twice, and I have three copies in my own mailbox,
>> bot they still don't show up on the mm-commits mailing list.
>>
>> So the list hates them for some odd reason.
>>
>> I've picked them up locally, but adding Konstantin to the participants
>> to see if he can see what's up.
>>
>> Konstantin: patches 103/106/107 are missing on lore out of Andrew's
>> series of 143. Odd.
> It's weird. They don't turn up on linux-mm either, and that's running
> at kvack.org, also majordomo. They don't get through when sent with
> either heirloom-mailx or with sylpheed.
>
> Also, it seems that when Anshuman originally sent the patch, linux-mm
> and linux-kernel didn't send it back out. So perhaps a spam filter
> triggered?
>
> I'm seeing
>
> https://lore.kernel.org/linux-arm-kernel/1615278790-18053-3-git-send-email-anshuman.khandual@arm.com/
>
> which is via linux-arm-kernel@lists.infradead.org but the linux-kernel
> server massacred that patch series. Searching
> https://lkml.org/lkml/2021/3/9 for "anshuman" only shows 3 of the 7
> email series.
Yeah these patches faced problem from the very beginning getting
into the MM/LKML list for some strange reason.
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2021-05-05 17:10 ` incoming Linus Torvalds
@ 2021-05-05 17:44 ` Andrew Morton
2021-05-06 3:19 ` incoming Anshuman Khandual
0 siblings, 1 reply; 409+ messages in thread
From: Andrew Morton @ 2021-05-05 17:44 UTC (permalink / raw)
To: Linus Torvalds; +Cc: Konstantin Ryabitsev, Linux-MM, mm-commits
[-- Attachment #1: Type: text/plain, Size: 1387 bytes --]
On Wed, 5 May 2021 10:10:33 -0700 Linus Torvalds <torvalds@linux-foundation.org> wrote:
> On Tue, May 4, 2021 at 8:16 PM Andrew Morton <akpm@linux-foundation.org> wrote:
> >
> > Let me resend right now with the same in-reply-to. Hopefully they will
> > land in the correct place.
>
> Well, you re-sent it twice, and I have three copies in my own mailbox,
> bot they still don't show up on the mm-commits mailing list.
>
> So the list hates them for some odd reason.
>
> I've picked them up locally, but adding Konstantin to the participants
> to see if he can see what's up.
>
> Konstantin: patches 103/106/107 are missing on lore out of Andrew's
> series of 143. Odd.
It's weird. They don't turn up on linux-mm either, and that's running
at kvack.org, also majordomo. They don't get through when sent with
either heirloom-mailx or with sylpheed.
Also, it seems that when Anshuman originally sent the patch, linux-mm
and linux-kernel didn't send it back out. So perhaps a spam filter
triggered?
I'm seeing
https://lore.kernel.org/linux-arm-kernel/1615278790-18053-3-git-send-email-anshuman.khandual@arm.com/
which is via linux-arm-kernel@lists.infradead.org but the linux-kernel
server massacred that patch series. Searching
https://lkml.org/lkml/2021/3/9 for "anshuman" only shows 3 of the 7
email series.
One of the emails (as sent my me) is attached, if that helps.
[-- Attachment #2: x.txt --]
[-- Type: text/plain, Size: 21048 bytes --]
Return-Path: <akpm@linux-foundation.org>
X-Spam-Checker-Version: SpamAssassin 3.4.1 (2015-04-28) on y
X-Spam-Level: (none)
X-Spam-Status: No, score=-101.5 required=2.5 tests=BAYES_00,T_DKIM_INVALID,
USER_IN_WHITELIST autolearn=ham autolearn_force=no version=3.4.1
Received: from localhost.localdomain (localhost.localdomain [127.0.0.1])
by localhost.localdomain (8.15.2/8.15.2/Debian-8ubuntu1) with ESMTP id 1453H2fk032202
for <akpm@localhost>; Tue, 4 May 2021 20:17:03 -0700
Received: from imap.fastmail.com [66.111.4.135]
by localhost.localdomain with IMAP (fetchmail-6.3.26)
for <akpm@localhost> (single-drop); Tue, 04 May 2021 20:17:03 -0700 (PDT)
Received: from compute1.internal (compute1.nyi.internal [10.202.2.41])
by sloti11d1t06 (Cyrus 3.5.0-alpha0-442-g5daca166b9-fm-20210428.001-g5daca166) with LMTPA;
Tue, 04 May 2021 23:16:31 -0400
X-Cyrus-Session-Id: sloti11d1t06-1620184591-1699471-2-6359664467419938249
X-Sieve: CMU Sieve 3.0
X-Resolved-to: akpm@mbx.kernel.org
X-Delivered-to: akpm@mbx.kernel.org
X-Mail-from: akpm@linux-foundation.org
Received: from mx6 ([10.202.2.205])
by compute1.internal (LMTPProxy); Tue, 04 May 2021 23:16:31 -0400
Received: from mx6.messagingengine.com (localhost [127.0.0.1])
by mailmx.nyi.internal (Postfix) with ESMTP id 40796C800E1
for <akpm@mbx.kernel.org>; Tue, 4 May 2021 23:16:31 -0400 (EDT)
Received: from mx6.messagingengine.com (localhost [127.0.0.1])
by mx6.messagingengine.com (Authentication Milter) with ESMTP
id 14870833D7F;
Tue, 4 May 2021 23:16:31 -0400
ARC-Seal: i=2; a=rsa-sha256; cv=pass; d=messagingengine.com; s=fm2; t=
1620184591; b=FBo7Gf3JFN+4QYg5Byan0oNm6RESv+sIf5HcaslVNsUd9SOTGS
yI0+IsXr1CUpGH783hE6fmgEq9SyfOwQVZjdikLaJS1+7u0JtfAYQFU3RORCtXlr
djJWrScfjVa8nAHX4rQCtzvtPYuzx5w7cTgGgeILgoJMxgLj7EC9xcT8BIf68+9W
Lw+ohAmcuiKhL2ez+de4SMuwdh3dh2FwAIHQOsSjEU1/NV+WGxMLwYbxWgTrqQGH
RQIzFNdq30qslW9huK47+e80uHOX2tXwxtshwbThFEn458bdV5LL6Y8Oh4ZWMbv1
tFgTt515DVedonZknxc07XsXtAjaJyB8bfHw==
ARC-Message-Signature: i=2; a=rsa-sha256; c=relaxed/relaxed; d=
messagingengine.com; h=date:from:to:subject:message-id
:in-reply-to; s=fm2; t=1620184591; bh=LuH7mbm3+zp863vKBEqKeoZtnp
uFxYpIb5oTVwf56Es=; b=m5E1fbz2b+an/X406oY3BuG0Zm4/W05vWAki8Lsnud
gPCc1LfPUFSuXaMppcEDPbLKprp4hH3T52itK4pivXMQCLEOyme7kVStaLMVTiky
Xxqh5ZdhOWvygBfda/GjfuLBSbbj2gfm8HPKpbL7CA5foelknIBhJHDzGkJyxetZ
YagZfVvtdo2OEwnC1mmjUCpKPO5+m5kaZO0ol6rPdl+TV0MKGhjLg+/i6Ia+0nFp
zDwV4VeACvVcGb2xY7KG5Z+BtqVxeVFn+w5JcqpWUtxEKoSBR4bWARzjwHg6eouh
7psOOKPTt/NzDKk+3f49lso5KlPiTF2xEU/+5SIttCkQ==
ARC-Authentication-Results: i=2; mx6.messagingengine.com;
arc=pass (as.1.google.com=pass, ams.1.google.com=pass)
smtp.remote-ip=209.85.215.198;
bimi=skipped (DMARC did not pass);
dkim=pass (1024-bit rsa key sha256) header.d=linux-foundation.org
header.i=@linux-foundation.org header.b=Gdz/3wY9 header.a=rsa-sha256
header.s=korg x-bits=1024;
dmarc=none policy.published-domain-policy=none
policy.applied-disposition=none policy.evaluated-disposition=none
(p=none,d=none,d.eval=none) policy.policy-from=p
header.from=linux-foundation.org;
iprev=pass smtp.remote-ip=209.85.215.198 (mail-pg1-f198.google.com);
spf=pass smtp.mailfrom=akpm@linux-foundation.org
smtp.helo=mail-pg1-f198.google.com;
x-aligned-from=pass (Address match);
x-arc-spf=pass
(google.com: domain of akpm@linux-foundation.org designates 198.145.29.99 as permitted sender)
smtp.mailfrom=akpm@linux-foundation.org x-arc-instance=1
x-arc-domain=google.com (Trusted from aar.1.google.com);
x-csa=none;
x-google-dkim=fail (message has been altered, 2048-bit rsa key)
header.d=1e100.net header.i=@1e100.net header.b=VZuDOxUf;
x-me-sender=none;
x-ptr=pass smtp.helo=mail-pg1-f198.google.com
policy.ptr=mail-pg1-f198.google.com;
x-return-mx=pass header.domain=linux-foundation.org policy.is_org=yes
(MX Records found: ASPMX.L.GOOGLE.COM,ALT1.ASPMX.L.GOOGLE.COM,ALT2.ASPMX.L.GOOGLE.COM,ALT3.ASPMX.L.GOOGLE.COM,ALT4.ASPMX.L.GOOGLE.COM);
x-return-mx=pass smtp.domain=linux-foundation.org policy.is_org=yes
(MX Records found: ASPMX.L.GOOGLE.COM,ALT1.ASPMX.L.GOOGLE.COM,ALT2.ASPMX.L.GOOGLE.COM,ALT3.ASPMX.L.GOOGLE.COM,ALT4.ASPMX.L.GOOGLE.COM);
x-tls=pass smtp.version=TLSv1.3 smtp.cipher=TLS_AES_256_GCM_SHA384
smtp.bits=256/256;
x-vs=clean score=40 state=0
Authentication-Results: mx6.messagingengine.com;
arc=pass (as.1.google.com=pass, ams.1.google.com=pass)
smtp.remote-ip=209.85.215.198;
bimi=skipped (DMARC did not pass);
dkim=pass (1024-bit rsa key sha256) header.d=linux-foundation.org
header.i=@linux-foundation.org header.b=Gdz/3wY9 header.a=rsa-sha256
header.s=korg x-bits=1024;
dmarc=none policy.published-domain-policy=none
policy.applied-disposition=none policy.evaluated-disposition=none
(p=none,d=none,d.eval=none) policy.policy-from=p
header.from=linux-foundation.org;
iprev=pass smtp.remote-ip=209.85.215.198 (mail-pg1-f198.google.com);
spf=pass smtp.mailfrom=akpm@linux-foundation.org
smtp.helo=mail-pg1-f198.google.com;
x-aligned-from=pass (Address match);
x-arc-spf=pass
(google.com: domain of akpm@linux-foundation.org designates 198.145.29.99 as permitted sender)
smtp.mailfrom=akpm@linux-foundation.org x-arc-instance=1
x-arc-domain=google.com (Trusted from aar.1.google.com);
x-csa=none;
x-google-dkim=fail (message has been altered, 2048-bit rsa key)
header.d=1e100.net header.i=@1e100.net header.b=VZuDOxUf;
x-me-sender=none;
x-ptr=pass smtp.helo=mail-pg1-f198.google.com
policy.ptr=mail-pg1-f198.google.com;
x-return-mx=pass header.domain=linux-foundation.org policy.is_org=yes
(MX Records found: ASPMX.L.GOOGLE.COM,ALT1.ASPMX.L.GOOGLE.COM,ALT2.ASPMX.L.GOOGLE.COM,ALT3.ASPMX.L.GOOGLE.COM,ALT4.ASPMX.L.GOOGLE.COM);
x-return-mx=pass smtp.domain=linux-foundation.org policy.is_org=yes
(MX Records found: ASPMX.L.GOOGLE.COM,ALT1.ASPMX.L.GOOGLE.COM,ALT2.ASPMX.L.GOOGLE.COM,ALT3.ASPMX.L.GOOGLE.COM,ALT4.ASPMX.L.GOOGLE.COM);
x-tls=pass smtp.version=TLSv1.3 smtp.cipher=TLS_AES_256_GCM_SHA384
smtp.bits=256/256;
x-vs=clean score=40 state=0
X-ME-VSCause: gggruggvucftvghtrhhoucdtuddrgeduledrvdefjedgieegucetufdoteggodetrfdotf
fvucfrrhhofhhilhgvmecuhfgrshhtofgrihhlpdggtfgfnhhsuhgsshgtrhhisggvpdfu
rfetoffkrfgpnffqhgenuceurghilhhouhhtmecufedttdenucgoufhorhhtvggutfgvtg
hiphdvucdlgedtmdenucfjughrpeffhffvuffkjggfsedttdertddtredtnecuhfhrohhm
peetnhgurhgvficuofhorhhtohhnuceorghkphhmsehlihhnuhigqdhfohhunhgurghtih
honhdrohhrgheqnecuggftrfgrthhtvghrnhepjeevfeduveffvddvudetkefhgeduveeu
geevvdfhhfevhfekkedtieefgfduheeinecuffhomhgrihhnpehkvghrnhgvlhdrohhrgh
enucfkphepvddtledrkeehrddvudehrdduleekpdduleekrddugeehrddvledrleelnecu
uegrugftvghpuhhtkfhppeduleekrddugeehrddvledrleelnecuvehluhhsthgvrhfuih
iivgeptdenucfrrghrrghmpehinhgvthepvddtledrkeehrddvudehrdduleekpdhhvghl
ohepmhgrihhlqdhpghduqdhfudelkedrghhoohhglhgvrdgtohhmpdhmrghilhhfrhhomh
epoegrkhhpmheslhhinhhugidqfhhouhhnuggrthhiohhnrdhorhhgqe
X-ME-VSScore: 40
X-ME-VSCategory: clean
X-ME-CSA: none
Received-SPF: pass
(linux-foundation.org: Sender is authorized to use 'akpm@linux-foundation.org' in 'mfrom' identity (mechanism 'include:_spf.google.com' matched))
receiver=mx6.messagingengine.com;
identity=mailfrom;
envelope-from="akpm@linux-foundation.org";
helo=mail-pg1-f198.google.com;
client-ip=209.85.215.198
Received: from mail-pg1-f198.google.com (mail-pg1-f198.google.com [209.85.215.198])
(using TLSv1.3 with cipher TLS_AES_256_GCM_SHA384 (256/256 bits)
key-exchange X25519 server-signature RSA-PSS (2048 bits) server-digest SHA256)
(No client certificate requested)
by mx6.messagingengine.com (Postfix) with ESMTPS
for <akpm@mbx.kernel.org>; Tue, 4 May 2021 23:16:31 -0400 (EDT)
Received: by mail-pg1-f198.google.com with SMTP id g5-20020a63f4050000b02901f6c7b9a6d0so593624pgi.5
for <akpm@mbx.kernel.org>; Tue, 04 May 2021 20:16:30 -0700 (PDT)
X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed;
d=1e100.net; s=20161025;
h=x-gm-message-state:dkim-signature:date:from:to:subject:message-id
:in-reply-to:user-agent;
bh=LuH7mbm3+zp863vKBEqKeoZtnpuFxYpIb5oTVwf56Es=;
b=VZuDOxUfeHXJz1/CiFfcxuMVHkmW5RznvqYS+Py8Ub6nHHXprQJGE9Ze3WgH+1ylSe
NJLEC7xgv15SR9A+e/MT4RTj3OVOwtd1Zi2vPav39a9K4tP+2uL2Ei+5d7FtT3LLZsjo
feek/DqCGSkJ/EC5woLyU9BBkfLUuQ9/2HiDCk10BMetEfWdor69Slb39NOXES8br02X
25Btabu9ZCWroyjQj7W5gwGr5Z6Hs2nbnnfAb+e92FalcUD/4ql77lNzRcWGi4/9TT8s
ntqI2g46Xv+k5LURaRH5CRBpxkkKgzcrioRPYFUHkEgOEWy1hPzg9QPk8ZO35Xm9R9d2
vl3Q==
X-Gm-Message-State: AOAM531IlYUTVWcMrsTunnxZWB7SKeeOmoZj5mZ1A5tl7N/JlZUueN8L
tvyRKnvxHr6a5mDaGHN9Tb1N/iCzT0U5oQgRVTxTnj1qFGibRa9+leLQNKX0aGlNg9JiaMfromb
xyOlCUpVXOlVvchuwTUSTn7rXum+Hh3PWQZm5II/EX+0AkzKqez62Z8U=
X-Received: by 2002:a17:90a:a581:: with SMTP id b1mr32203271pjq.53.1620184589161;
Tue, 04 May 2021 20:16:29 -0700 (PDT)
X-Google-Smtp-Source: ABdhPJxffoGdRqAjUagWoMVD5p/Lk1KTEDftEhkWh8ewatgDmZLlxh0lO1hxYIdYYwoO5dsJ/i0z
X-Received: by 2002:a17:90a:a581:: with SMTP id b1mr32203198pjq.53.1620184588109;
Tue, 04 May 2021 20:16:28 -0700 (PDT)
ARC-Seal: i=1; a=rsa-sha256; t=1620184588; cv=none;
d=google.com; s=arc-20160816;
b=Fr2b2AMXJr6OeNpSql45tq1korkuDOunp7t+DpARuEBnwvQnKfagyipQ93jywsRf/c
/i/mP2eTmJwOLWNORClh1MGF/0VfBx1ULoB9W4CI3LpVgGFXGGFis8LTcvUYD5yvhlsV
50rm2j34iS9lyo04FB/hbhGkwLtUhz2PGkLGuqHspTd+pUpUCf5SLxGJbZC5uCcUEsbO
8WSDBWyvaCPjFzJQZK60gK70ticKW+fCG1xHtOG4qsFCbqEpFKBy8eVK83OBazo/dQDr
DOheWNWyw2o/WMP4GpZMvZuj30dx3j8xnBahIpnMIQJaog6wLMcVX9pkQ8UJym3/PGNm
pO/g==
ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816;
h=user-agent:in-reply-to:message-id:subject:to:from:date
:dkim-signature;
bh=LuH7mbm3+zp863vKBEqKeoZtnpuFxYpIb5oTVwf56Es=;
b=vVN16NPMKjoxSJQ6b36VXFCkZqnmG7wABfilgE069txZqmHpEMyZb8lRStkHy557LM
Kn7UfJFP3xwsP8ZTCipVDZ6tpFW/hYFU9o4th9G8asWs+MOf9xpWX2LQZ1FTmaao2Fg5
uCHypz39cnAh0Z1EJfNsTcaTGIrkbBd6zje+mtBgs8hnfH8HcWBYTPCHCCx950Z928tb
XOPd/Igs7yzD1ioBiGXZj/ciwPbWVTaZXBg4JOZSApxkDMfuMyfyLLOs++EVkyxJHUme
TmgwvLkixcwEtKF7gIeqEhwvOUSVvilLuJLFVaLumwTcjJ1amVfGcJhBE7LIM9C3SMpA
rOOg==
ARC-Authentication-Results: i=1; mx.google.com;
dkim=pass header.i=@linux-foundation.org header.s=korg header.b="Gdz/3wY9";
spf=pass (google.com: domain of akpm@linux-foundation.org designates 198.145.29.99 as permitted sender) smtp.mailfrom=akpm@linux-foundation.org
Received: from mail.kernel.org (mail.kernel.org. [198.145.29.99])
by mx.google.com with ESMTPS id c85si20173199pfb.8.2021.05.04.20.16.27
(version=TLS1_2 cipher=ECDHE-ECDSA-AES128-GCM-SHA256 bits=128/128);
Tue, 04 May 2021 20:16:28 -0700 (PDT)
Received-SPF: pass (google.com: domain of akpm@linux-foundation.org designates 198.145.29.99 as permitted sender) client-ip=198.145.29.99;
Authentication-Results: mx.google.com;
dkim=pass header.i=@linux-foundation.org header.s=korg header.b="Gdz/3wY9";
spf=pass (google.com: domain of akpm@linux-foundation.org designates 198.145.29.99 as permitted sender) smtp.mailfrom=akpm@linux-foundation.org
Received: by mail.kernel.org (Postfix) with ESMTPSA id A4DB4610D2;
Wed, 5 May 2021 03:16:26 +0000 (UTC)
DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=linux-foundation.org;
s=korg; t=1620184587;
bh=TxN4wgKcKf2UUem+5pL09m9GL/7U592mEalo2U6vwAU=;
h=Date:From:To:Subject:In-Reply-To:From;
b=Gdz/3wY9ktH3hOmn2DAOkfh0JXwPdMJ8xsNQFa9eI25K39Z3iHdRGo9jX3QtMDtog
D4Zakt52CQCYsV91c9oCai8KnCTkkAjJq/Ez7p8UHpz97Go3yYYxqg6DDl6d8HCQvN
H47dTaZAgeH2sw29bjB9fRzNuTx7k4RAPlqZIpiE=
Date: Tue, 04 May 2021 20:16:26 -0700
From: Andrew Morton <akpm@linux-foundation.org>
To: akpm@linux-foundation.org, anshuman.khandual@arm.com, aou@eecs.berkeley.edu, arnd@arndb.de, benh@kernel.crashing.org, borntraeger@de.ibm.com, bp@alien8.de, catalin.marinas@arm.com, dalias@libc.org, deller@gmx.de, gor@linux.ibm.com, hca@linux.ibm.com, hpa@zytor.com, James.Bottomley@HansenPartnership.com, linux-mm@kvack.org, linux@armlinux.org.uk, mingo@redhat.com, mm-commits@vger.kernel.org, mpe@ellerman.id.au, palmerdabbelt@google.com, paul.walmsley@sifive.com, paulus@samba.org, tglx@linutronix.de, torvalds@linux-foundation.org, tsbogend@alpha.franken.de, vgupta@synopsys.com, viro@zeniv.linux.org.uk, will@kernel.org, ysato@users.osdn.me
Subject: [patch 103/143] mm: generalize SYS_SUPPORTS_HUGETLBFS (rename as ARCH_SUPPORTS_HUGETLBFS)
Message-ID: <20210505031626.c8o4WL7KE%akpm@linux-foundation.org>
In-Reply-To: <20210504183219.a3cc46aee4013d77402276c5@linux-foundation.org>
User-Agent: s-nail v14.8.16
X-Gm-Original-To: akpm@linux-foundation.org
From: Anshuman Khandual <anshuman.khandual@arm.com>
Subject: mm: generalize SYS_SUPPORTS_HUGETLBFS (rename as ARCH_SUPPORTS_HUGETLBFS)
SYS_SUPPORTS_HUGETLBFS config has duplicate definitions on platforms that
subscribe it. Instead, just make it a generic option which can be
selected on applicable platforms. Also rename it as
ARCH_SUPPORTS_HUGETLBFS instead. This reduces code duplication and makes
it cleaner.
Link: https://lkml.kernel.org/r/1617259448-22529-3-git-send-email-anshuman.khandual@arm.com
Signed-off-by: Anshuman Khandual <anshuman.khandual@arm.com>
Acked-by: Catalin Marinas <catalin.marinas@arm.com> [arm64]
Acked-by: Palmer Dabbelt <palmerdabbelt@google.com> [riscv]
Acked-by: Michael Ellerman <mpe@ellerman.id.au> [powerpc]
Cc: Russell King <linux@armlinux.org.uk>
Cc: Will Deacon <will@kernel.org>
Cc: Thomas Bogendoerfer <tsbogend@alpha.franken.de>
Cc: "James E.J. Bottomley" <James.Bottomley@HansenPartnership.com>
Cc: Helge Deller <deller@gmx.de>
Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
Cc: Paul Mackerras <paulus@samba.org>
Cc: Paul Walmsley <paul.walmsley@sifive.com>
Cc: Albert Ou <aou@eecs.berkeley.edu>
Cc: Yoshinori Sato <ysato@users.sourceforge.jp>
Cc: Rich Felker <dalias@libc.org>
Cc: Alexander Viro <viro@zeniv.linux.org.uk>
Cc: Arnd Bergmann <arnd@arndb.de>
Cc: Borislav Petkov <bp@alien8.de>
Cc: Christian Borntraeger <borntraeger@de.ibm.com>
Cc: Heiko Carstens <hca@linux.ibm.com>
Cc: "H. Peter Anvin" <hpa@zytor.com>
Cc: Ingo Molnar <mingo@redhat.com>
Cc: Thomas Gleixner <tglx@linutronix.de>
Cc: Vasily Gorbik <gor@linux.ibm.com>
Cc: Vineet Gupta <vgupta@synopsys.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
arch/arm/Kconfig | 5 +----
arch/arm64/Kconfig | 4 +---
arch/mips/Kconfig | 6 +-----
arch/parisc/Kconfig | 5 +----
arch/powerpc/Kconfig | 3 ---
arch/powerpc/platforms/Kconfig.cputype | 6 +++---
arch/riscv/Kconfig | 5 +----
arch/sh/Kconfig | 5 +----
fs/Kconfig | 5 ++++-
9 files changed, 13 insertions(+), 31 deletions(-)
--- a/arch/arm64/Kconfig~mm-generalize-sys_supports_hugetlbfs-rename-as-arch_supports_hugetlbfs
+++ a/arch/arm64/Kconfig
@@ -73,6 +73,7 @@ config ARM64
select ARCH_USE_QUEUED_SPINLOCKS
select ARCH_USE_SYM_ANNOTATIONS
select ARCH_SUPPORTS_DEBUG_PAGEALLOC
+ select ARCH_SUPPORTS_HUGETLBFS
select ARCH_SUPPORTS_MEMORY_FAILURE
select ARCH_SUPPORTS_SHADOW_CALL_STACK if CC_HAVE_SHADOW_CALL_STACK
select ARCH_SUPPORTS_LTO_CLANG if CPU_LITTLE_ENDIAN
@@ -1072,9 +1073,6 @@ config HW_PERF_EVENTS
def_bool y
depends on ARM_PMU
-config SYS_SUPPORTS_HUGETLBFS
- def_bool y
-
config ARCH_HAS_FILTER_PGPROT
def_bool y
--- a/arch/arm/Kconfig~mm-generalize-sys_supports_hugetlbfs-rename-as-arch_supports_hugetlbfs
+++ a/arch/arm/Kconfig
@@ -31,6 +31,7 @@ config ARM
select ARCH_OPTIONAL_KERNEL_RWX if ARCH_HAS_STRICT_KERNEL_RWX
select ARCH_OPTIONAL_KERNEL_RWX_DEFAULT if CPU_V7
select ARCH_SUPPORTS_ATOMIC_RMW
+ select ARCH_SUPPORTS_HUGETLBFS if ARM_LPAE
select ARCH_USE_BUILTIN_BSWAP
select ARCH_USE_CMPXCHG_LOCKREF
select ARCH_USE_MEMTEST
@@ -1511,10 +1512,6 @@ config HW_PERF_EVENTS
def_bool y
depends on ARM_PMU
-config SYS_SUPPORTS_HUGETLBFS
- def_bool y
- depends on ARM_LPAE
-
config HAVE_ARCH_TRANSPARENT_HUGEPAGE
def_bool y
depends on ARM_LPAE
--- a/arch/mips/Kconfig~mm-generalize-sys_supports_hugetlbfs-rename-as-arch_supports_hugetlbfs
+++ a/arch/mips/Kconfig
@@ -19,6 +19,7 @@ config MIPS
select ARCH_USE_MEMTEST
select ARCH_USE_QUEUED_RWLOCKS
select ARCH_USE_QUEUED_SPINLOCKS
+ select ARCH_SUPPORTS_HUGETLBFS if CPU_SUPPORTS_HUGEPAGES
select ARCH_WANT_DEFAULT_TOPDOWN_MMAP_LAYOUT if MMU
select ARCH_WANT_IPC_PARSE_VERSION
select ARCH_WANT_LD_ORPHAN_WARN
@@ -1287,11 +1288,6 @@ config SYS_SUPPORTS_BIG_ENDIAN
config SYS_SUPPORTS_LITTLE_ENDIAN
bool
-config SYS_SUPPORTS_HUGETLBFS
- bool
- depends on CPU_SUPPORTS_HUGEPAGES
- default y
-
config MIPS_HUGE_TLB_SUPPORT
def_bool HUGETLB_PAGE || TRANSPARENT_HUGEPAGE
--- a/arch/parisc/Kconfig~mm-generalize-sys_supports_hugetlbfs-rename-as-arch_supports_hugetlbfs
+++ a/arch/parisc/Kconfig
@@ -12,6 +12,7 @@ config PARISC
select ARCH_HAS_STRICT_KERNEL_RWX
select ARCH_HAS_UBSAN_SANITIZE_ALL
select ARCH_NO_SG_CHAIN
+ select ARCH_SUPPORTS_HUGETLBFS if PA20
select ARCH_SUPPORTS_MEMORY_FAILURE
select DMA_OPS
select RTC_CLASS
@@ -138,10 +139,6 @@ config PGTABLE_LEVELS
default 3 if 64BIT && PARISC_PAGE_SIZE_4KB
default 2
-config SYS_SUPPORTS_HUGETLBFS
- def_bool y if PA20
-
-
menu "Processor type and features"
choice
--- a/arch/powerpc/Kconfig~mm-generalize-sys_supports_hugetlbfs-rename-as-arch_supports_hugetlbfs
+++ a/arch/powerpc/Kconfig
@@ -697,9 +697,6 @@ config ARCH_SPARSEMEM_DEFAULT
def_bool y
depends on PPC_BOOK3S_64
-config SYS_SUPPORTS_HUGETLBFS
- bool
-
config ILLEGAL_POINTER_VALUE
hex
# This is roughly half way between the top of user space and the bottom
--- a/arch/powerpc/platforms/Kconfig.cputype~mm-generalize-sys_supports_hugetlbfs-rename-as-arch_supports_hugetlbfs
+++ a/arch/powerpc/platforms/Kconfig.cputype
@@ -40,8 +40,8 @@ config PPC_85xx
config PPC_8xx
bool "Freescale 8xx"
+ select ARCH_SUPPORTS_HUGETLBFS
select FSL_SOC
- select SYS_SUPPORTS_HUGETLBFS
select PPC_HAVE_KUEP
select PPC_HAVE_KUAP
select HAVE_ARCH_VMAP_STACK
@@ -95,9 +95,9 @@ config PPC_BOOK3S_64
bool "Server processors"
select PPC_FPU
select PPC_HAVE_PMU_SUPPORT
- select SYS_SUPPORTS_HUGETLBFS
select HAVE_ARCH_TRANSPARENT_HUGEPAGE
select ARCH_ENABLE_THP_MIGRATION if TRANSPARENT_HUGEPAGE
+ select ARCH_SUPPORTS_HUGETLBFS
select ARCH_SUPPORTS_NUMA_BALANCING
select IRQ_WORK
select PPC_MM_SLICES
@@ -278,9 +278,9 @@ config FSL_BOOKE
# this is for common code between PPC32 & PPC64 FSL BOOKE
config PPC_FSL_BOOK3E
bool
+ select ARCH_SUPPORTS_HUGETLBFS if PHYS_64BIT || PPC64
select FSL_EMB_PERFMON
select PPC_SMP_MUXED_IPI
- select SYS_SUPPORTS_HUGETLBFS if PHYS_64BIT || PPC64
select PPC_DOORBELL
default y if FSL_BOOKE
--- a/arch/riscv/Kconfig~mm-generalize-sys_supports_hugetlbfs-rename-as-arch_supports_hugetlbfs
+++ a/arch/riscv/Kconfig
@@ -30,6 +30,7 @@ config RISCV
select ARCH_HAS_STRICT_KERNEL_RWX if MMU
select ARCH_OPTIONAL_KERNEL_RWX if ARCH_HAS_STRICT_KERNEL_RWX
select ARCH_OPTIONAL_KERNEL_RWX_DEFAULT
+ select ARCH_SUPPORTS_HUGETLBFS if MMU
select ARCH_WANT_DEFAULT_TOPDOWN_MMAP_LAYOUT if MMU
select ARCH_WANT_FRAME_POINTERS
select ARCH_WANT_HUGE_PMD_SHARE if 64BIT
@@ -165,10 +166,6 @@ config ARCH_WANT_GENERAL_HUGETLB
config ARCH_SUPPORTS_UPROBES
def_bool y
-config SYS_SUPPORTS_HUGETLBFS
- depends on MMU
- def_bool y
-
config STACKTRACE_SUPPORT
def_bool y
--- a/arch/sh/Kconfig~mm-generalize-sys_supports_hugetlbfs-rename-as-arch_supports_hugetlbfs
+++ a/arch/sh/Kconfig
@@ -101,9 +101,6 @@ config SYS_SUPPORTS_APM_EMULATION
bool
select ARCH_SUSPEND_POSSIBLE
-config SYS_SUPPORTS_HUGETLBFS
- bool
-
config SYS_SUPPORTS_SMP
bool
@@ -175,12 +172,12 @@ config CPU_SH3
config CPU_SH4
bool
+ select ARCH_SUPPORTS_HUGETLBFS if MMU
select CPU_HAS_INTEVT
select CPU_HAS_SR_RB
select CPU_HAS_FPU if !CPU_SH4AL_DSP
select SH_INTC
select SYS_SUPPORTS_SH_TMU
- select SYS_SUPPORTS_HUGETLBFS if MMU
config CPU_SH4A
bool
--- a/fs/Kconfig~mm-generalize-sys_supports_hugetlbfs-rename-as-arch_supports_hugetlbfs
+++ a/fs/Kconfig
@@ -223,10 +223,13 @@ config TMPFS_INODE64
If unsure, say N.
+config ARCH_SUPPORTS_HUGETLBFS
+ def_bool n
+
config HUGETLBFS
bool "HugeTLB file system support"
depends on X86 || IA64 || SPARC64 || (S390 && 64BIT) || \
- SYS_SUPPORTS_HUGETLBFS || BROKEN
+ ARCH_SUPPORTS_HUGETLBFS || BROKEN
help
hugetlbfs is a filesystem backing for HugeTLB pages, based on
ramfs. For architectures that support it, say Y here and read
_
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2021-05-05 3:16 ` incoming Andrew Morton
@ 2021-05-05 17:10 ` Linus Torvalds
2021-05-05 17:44 ` incoming Andrew Morton
0 siblings, 1 reply; 409+ messages in thread
From: Linus Torvalds @ 2021-05-05 17:10 UTC (permalink / raw)
To: Andrew Morton, Konstantin Ryabitsev; +Cc: Linux-MM, mm-commits
On Tue, May 4, 2021 at 8:16 PM Andrew Morton <akpm@linux-foundation.org> wrote:
>
> Let me resend right now with the same in-reply-to. Hopefully they will
> land in the correct place.
Well, you re-sent it twice, and I have three copies in my own mailbox,
bot they still don't show up on the mm-commits mailing list.
So the list hates them for some odd reason.
I've picked them up locally, but adding Konstantin to the participants
to see if he can see what's up.
Konstantin: patches 103/106/107 are missing on lore out of Andrew's
series of 143. Odd.
Linus
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2021-05-05 1:47 ` incoming Linus Torvalds
@ 2021-05-05 3:16 ` Andrew Morton
2021-05-05 17:10 ` incoming Linus Torvalds
0 siblings, 1 reply; 409+ messages in thread
From: Andrew Morton @ 2021-05-05 3:16 UTC (permalink / raw)
To: Linus Torvalds; +Cc: Linux-MM, mm-commits
On Tue, 4 May 2021 18:47:19 -0700 Linus Torvalds <torvalds@linux-foundation.org> wrote:
> On Tue, May 4, 2021 at 6:32 PM Andrew Morton <akpm@linux-foundation.org> wrote:
> >
> > 143 patches
>
> Hmm. Only 140 seem to have made it to the list, with 103, 106 and 107 missing.
>
> Maybe just some mail delay? But at least right now
>
> https://lore.kernel.org/mm-commits/
>
> doesn't show them (and thus 'b4' doesn't work).
>
> I'll check again later.
>
Well that's strange. I see all three via cc:me, but not on linux-mm or
mm-commits.
Let me resend right now with the same in-reply-to. Hopefully they will
land in the correct place.
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2021-05-05 1:32 incoming Andrew Morton
@ 2021-05-05 1:47 ` Linus Torvalds
2021-05-05 3:16 ` incoming Andrew Morton
0 siblings, 1 reply; 409+ messages in thread
From: Linus Torvalds @ 2021-05-05 1:47 UTC (permalink / raw)
To: Andrew Morton; +Cc: Linux-MM, mm-commits
On Tue, May 4, 2021 at 6:32 PM Andrew Morton <akpm@linux-foundation.org> wrote:
>
> 143 patches
Hmm. Only 140 seem to have made it to the list, with 103, 106 and 107 missing.
Maybe just some mail delay? But at least right now
https://lore.kernel.org/mm-commits/
doesn't show them (and thus 'b4' doesn't work).
I'll check again later.
Linus
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-05-05 1:32 Andrew Morton
2021-05-05 1:47 ` incoming Linus Torvalds
0 siblings, 1 reply; 409+ messages in thread
From: Andrew Morton @ 2021-05-05 1:32 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
The remainder of the main mm/ queue.
143 patches, based on 8ca5297e7e38f2dc8c753d33a5092e7be181fff0, plus
previously sent patches.
Subsystems affected by this patch series:
mm/pagecache
mm/hugetlb
mm/userfaultfd
mm/vmscan
mm/compaction
mm/migration
mm/cma
mm/ksm
mm/vmstat
mm/mmap
mm/kconfig
mm/util
mm/memory-hotplug
mm/zswap
mm/zsmalloc
mm/highmem
mm/cleanups
mm/kfence
Subsystem: mm/pagecache
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
Patch series "Remove nrexceptional tracking", v2:
mm: introduce and use mapping_empty()
mm: stop accounting shadow entries
dax: account DAX entries as nrpages
mm: remove nrexceptional from inode
Hugh Dickins <hughd@google.com>:
mm: remove nrexceptional from inode: remove BUG_ON
Subsystem: mm/hugetlb
Peter Xu <peterx@redhat.com>:
Patch series "hugetlb: Disable huge pmd unshare for uffd-wp", v4:
hugetlb: pass vma into huge_pte_alloc() and huge_pmd_share()
hugetlb/userfaultfd: forbid huge pmd sharing when uffd enabled
mm/hugetlb: move flush_hugetlb_tlb_range() into hugetlb.h
hugetlb/userfaultfd: unshare all pmds for hugetlbfs when register wp
Miaohe Lin <linmiaohe@huawei.com>:
mm/hugetlb: remove redundant reservation check condition in alloc_huge_page()
Anshuman Khandual <anshuman.khandual@arm.com>:
mm: generalize HUGETLB_PAGE_SIZE_VARIABLE
Miaohe Lin <linmiaohe@huawei.com>:
Patch series "Some cleanups for hugetlb":
mm/hugetlb: use some helper functions to cleanup code
mm/hugetlb: optimize the surplus state transfer code in move_hugetlb_state()
mm/hugetlb_cgroup: remove unnecessary VM_BUG_ON_PAGE in hugetlb_cgroup_migrate()
mm/hugetlb: simplify the code when alloc_huge_page() failed in hugetlb_no_page()
mm/hugetlb: avoid calculating fault_mutex_hash in truncate_op case
Patch series "Cleanup and fixup for khugepaged", v2:
khugepaged: remove unneeded return value of khugepaged_collapse_pte_mapped_thps()
khugepaged: reuse the smp_wmb() inside __SetPageUptodate()
khugepaged: use helper khugepaged_test_exit() in __khugepaged_enter()
khugepaged: fix wrong result value for trace_mm_collapse_huge_page_isolate()
mm/huge_memory.c: remove unnecessary local variable ret2
Patch series "Some cleanups for huge_memory", v3:
mm/huge_memory.c: rework the function vma_adjust_trans_huge()
mm/huge_memory.c: make get_huge_zero_page() return bool
mm/huge_memory.c: rework the function do_huge_pmd_numa_page() slightly
mm/huge_memory.c: remove redundant PageCompound() check
mm/huge_memory.c: remove unused macro TRANSPARENT_HUGEPAGE_DEBUG_COW_FLAG
mm/huge_memory.c: use helper function migration_entry_to_page()
Yanfei Xu <yanfei.xu@windriver.com>:
mm/khugepaged.c: replace barrier() with READ_ONCE() for a selective variable
Miaohe Lin <linmiaohe@huawei.com>:
Patch series "Cleanup for khugepaged":
khugepaged: use helper function range_in_vma() in collapse_pte_mapped_thp()
khugepaged: remove unnecessary out label in collapse_huge_page()
khugepaged: remove meaningless !pte_present() check in khugepaged_scan_pmd()
Zi Yan <ziy@nvidia.com>:
mm: huge_memory: a new debugfs interface for splitting THP tests
mm: huge_memory: debugfs for file-backed THP split
Miaohe Lin <linmiaohe@huawei.com>:
Patch series "Cleanup and fixup for hugetlb", v2:
mm/hugeltb: remove redundant VM_BUG_ON() in region_add()
mm/hugeltb: simplify the return code of __vma_reservation_common()
mm/hugeltb: clarify (chg - freed) won't go negative in hugetlb_unreserve_pages()
mm/hugeltb: handle the error case in hugetlb_fix_reserve_counts()
mm/hugetlb: remove unused variable pseudo_vma in remove_inode_hugepages()
Mike Kravetz <mike.kravetz@oracle.com>:
Patch series "make hugetlb put_page safe for all calling contexts", v5:
mm/cma: change cma mutex to irq safe spinlock
hugetlb: no need to drop hugetlb_lock to call cma_release
hugetlb: add per-hstate mutex to synchronize user adjustments
hugetlb: create remove_hugetlb_page() to separate functionality
hugetlb: call update_and_free_page without hugetlb_lock
hugetlb: change free_pool_huge_page to remove_pool_huge_page
hugetlb: make free_huge_page irq safe
hugetlb: add lockdep_assert_held() calls for hugetlb_lock
Oscar Salvador <osalvador@suse.de>:
Patch series "Make alloc_contig_range handle Hugetlb pages", v10:
mm,page_alloc: bail out earlier on -ENOMEM in alloc_contig_migrate_range
mm,compaction: let isolate_migratepages_{range,block} return error codes
mm,hugetlb: drop clearing of flag from prep_new_huge_page
mm,hugetlb: split prep_new_huge_page functionality
mm: make alloc_contig_range handle free hugetlb pages
mm: make alloc_contig_range handle in-use hugetlb pages
mm,page_alloc: drop unnecessary checks from pfn_range_valid_contig
Subsystem: mm/userfaultfd
Axel Rasmussen <axelrasmussen@google.com>:
Patch series "userfaultfd: add minor fault handling", v9:
userfaultfd: add minor fault registration mode
userfaultfd: disable huge PMD sharing for MINOR registered VMAs
userfaultfd: hugetlbfs: only compile UFFD helpers if config enabled
userfaultfd: add UFFDIO_CONTINUE ioctl
userfaultfd: update documentation to describe minor fault handling
userfaultfd/selftests: add test exercising minor fault handling
Subsystem: mm/vmscan
Dave Hansen <dave.hansen@linux.intel.com>:
mm/vmscan: move RECLAIM* bits to uapi header
mm/vmscan: replace implicit RECLAIM_ZONE checks with explicit checks
Yang Shi <shy828301@gmail.com>:
Patch series "Make shrinker's nr_deferred memcg aware", v10:
mm: vmscan: use nid from shrink_control for tracepoint
mm: vmscan: consolidate shrinker_maps handling code
mm: vmscan: use shrinker_rwsem to protect shrinker_maps allocation
mm: vmscan: remove memcg_shrinker_map_size
mm: vmscan: use kvfree_rcu instead of call_rcu
mm: memcontrol: rename shrinker_map to shrinker_info
mm: vmscan: add shrinker_info_protected() helper
mm: vmscan: use a new flag to indicate shrinker is registered
mm: vmscan: add per memcg shrinker nr_deferred
mm: vmscan: use per memcg nr_deferred of shrinker
mm: vmscan: don't need allocate shrinker->nr_deferred for memcg aware shrinkers
mm: memcontrol: reparent nr_deferred when memcg offline
mm: vmscan: shrink deferred objects proportional to priority
Subsystem: mm/compaction
Pintu Kumar <pintu@codeaurora.org>:
mm/compaction: remove unused variable sysctl_compact_memory
Charan Teja Reddy <charante@codeaurora.org>:
mm: compaction: update the COMPACT[STALL|FAIL] events properly
Subsystem: mm/migration
Minchan Kim <minchan@kernel.org>:
mm: disable LRU pagevec during the migration temporarily
mm: replace migrate_[prep|finish] with lru_cache_[disable|enable]
mm: fs: invalidate BH LRU during page migration
Miaohe Lin <linmiaohe@huawei.com>:
Patch series "Cleanup and fixup for mm/migrate.c", v3:
mm/migrate.c: make putback_movable_page() static
mm/migrate.c: remove unnecessary rc != MIGRATEPAGE_SUCCESS check in 'else' case
mm/migrate.c: fix potential indeterminate pte entry in migrate_vma_insert_page()
mm/migrate.c: use helper migrate_vma_collect_skip() in migrate_vma_collect_hole()
Revert "mm: migrate: skip shared exec THP for NUMA balancing"
Subsystem: mm/cma
Minchan Kim <minchan@kernel.org>:
mm: vmstat: add cma statistics
Baolin Wang <baolin.wang@linux.alibaba.com>:
mm: cma: use pr_err_ratelimited for CMA warning
Liam Mark <lmark@codeaurora.org>:
mm: cma: add trace events for CMA alloc perf testing
Minchan Kim <minchan@kernel.org>:
mm: cma: support sysfs
mm: cma: add the CMA instance name to cma trace events
mm: use proper type for cma_[alloc|release]
Subsystem: mm/ksm
Miaohe Lin <linmiaohe@huawei.com>:
Patch series "Cleanup and fixup for ksm":
ksm: remove redundant VM_BUG_ON_PAGE() on stable_tree_search()
ksm: use GET_KSM_PAGE_NOLOCK to get ksm page in remove_rmap_item_from_tree()
ksm: remove dedicated macro KSM_FLAG_MASK
ksm: fix potential missing rmap_item for stable_node
Chengyang Fan <cy.fan@huawei.com>:
mm/ksm: remove unused parameter from remove_trailing_rmap_items()
Subsystem: mm/vmstat
Hugh Dickins <hughd@google.com>:
mm: restore node stat checking in /proc/sys/vm/stat_refresh
mm: no more EINVAL from /proc/sys/vm/stat_refresh
mm: /proc/sys/vm/stat_refresh skip checking known negative stats
mm: /proc/sys/vm/stat_refresh stop checking monotonic numa stats
Saravanan D <saravanand@fb.com>:
x86/mm: track linear mapping split events
Subsystem: mm/mmap
Liam Howlett <liam.howlett@oracle.com>:
mm/mmap.c: don't unlock VMAs in remap_file_pages()
Subsystem: mm/kconfig
Anshuman Khandual <anshuman.khandual@arm.com>:
Patch series "mm: some config cleanups", v2:
mm: generalize ARCH_HAS_CACHE_LINE_SIZE
mm: generalize SYS_SUPPORTS_HUGETLBFS (rename as ARCH_SUPPORTS_HUGETLBFS)
mm: generalize ARCH_ENABLE_MEMORY_[HOTPLUG|HOTREMOVE]
mm: drop redundant ARCH_ENABLE_[HUGEPAGE|THP]_MIGRATION
mm: drop redundant ARCH_ENABLE_SPLIT_PMD_PTLOCK
mm: drop redundant HAVE_ARCH_TRANSPARENT_HUGEPAGE
Subsystem: mm/util
Joe Perches <joe@perches.com>:
mm/util.c: reduce mem_dump_obj() object size
Bhaskar Chowdhury <unixbhaskar@gmail.com>:
mm/util.c: fix typo
Subsystem: mm/memory-hotplug
Pavel Tatashin <pasha.tatashin@soleen.com>:
Patch series "prohibit pinning pages in ZONE_MOVABLE", v11:
mm/gup: don't pin migrated cma pages in movable zone
mm/gup: check every subpage of a compound page during isolation
mm/gup: return an error on migration failure
mm/gup: check for isolation errors
mm cma: rename PF_MEMALLOC_NOCMA to PF_MEMALLOC_PIN
mm: apply per-task gfp constraints in fast path
mm: honor PF_MEMALLOC_PIN for all movable pages
mm/gup: do not migrate zero page
mm/gup: migrate pinned pages out of movable zone
memory-hotplug.rst: add a note about ZONE_MOVABLE and page pinning
mm/gup: change index type to long as it counts pages
mm/gup: longterm pin migration cleanup
selftests/vm: gup_test: fix test flag
selftests/vm: gup_test: test faulting in kernel, and verify pinnable pages
Mel Gorman <mgorman@techsingularity.net>:
mm/memory_hotplug: remove broken locking of zone PCP structures during hot remove
Oscar Salvador <osalvador@suse.de>:
Patch series "Allocate memmap from hotadded memory (per device)", v10:
drivers/base/memory: introduce memory_block_{online,offline}
mm,memory_hotplug: relax fully spanned sections check
David Hildenbrand <david@redhat.com>:
mm,memory_hotplug: factor out adjusting present pages into adjust_present_page_count()
Oscar Salvador <osalvador@suse.de>:
mm,memory_hotplug: allocate memmap from the added memory range
acpi,memhotplug: enable MHP_MEMMAP_ON_MEMORY when supported
mm,memory_hotplug: add kernel boot option to enable memmap_on_memory
x86/Kconfig: introduce ARCH_MHP_MEMMAP_ON_MEMORY_ENABLE
arm64/Kconfig: introduce ARCH_MHP_MEMMAP_ON_MEMORY_ENABLE
Subsystem: mm/zswap
Zhiyuan Dai <daizhiyuan@phytium.com.cn>:
mm/zswap.c: switch from strlcpy to strscpy
Subsystem: mm/zsmalloc
zhouchuangao <zhouchuangao@vivo.com>:
mm/zsmalloc: use BUG_ON instead of if condition followed by BUG.
Subsystem: mm/highmem
Ira Weiny <ira.weiny@intel.com>:
Patch series "btrfs: Convert kmap/memset/kunmap to memzero_user()":
iov_iter: lift memzero_page() to highmem.h
btrfs: use memzero_page() instead of open coded kmap pattern
songqiang <songqiang@uniontech.com>:
mm/highmem.c: fix coding style issue
Subsystem: mm/cleanups
Zhiyuan Dai <daizhiyuan@phytium.com.cn>:
mm/mempool: minor coding style tweaks
Zhang Yunkai <zhang.yunkai@zte.com.cn>:
mm/process_vm_access.c: remove duplicate include
Subsystem: mm/kfence
Marco Elver <elver@google.com>:
kfence: zero guard page after out-of-bounds access
Patch series "kfence: optimize timer scheduling", v2:
kfence: await for allocation using wait_event
kfence: maximize allocation wait timeout duration
kfence: use power-efficient work queue to run delayed work
Documentation/ABI/testing/sysfs-kernel-mm-cma | 25
Documentation/admin-guide/kernel-parameters.txt | 17
Documentation/admin-guide/mm/memory-hotplug.rst | 9
Documentation/admin-guide/mm/userfaultfd.rst | 105 +-
arch/arc/Kconfig | 9
arch/arm/Kconfig | 10
arch/arm64/Kconfig | 34
arch/arm64/mm/hugetlbpage.c | 7
arch/ia64/Kconfig | 14
arch/ia64/mm/hugetlbpage.c | 3
arch/mips/Kconfig | 6
arch/mips/mm/hugetlbpage.c | 4
arch/parisc/Kconfig | 5
arch/parisc/mm/hugetlbpage.c | 2
arch/powerpc/Kconfig | 17
arch/powerpc/mm/hugetlbpage.c | 3
arch/powerpc/platforms/Kconfig.cputype | 16
arch/riscv/Kconfig | 5
arch/s390/Kconfig | 12
arch/s390/mm/hugetlbpage.c | 2
arch/sh/Kconfig | 7
arch/sh/mm/Kconfig | 8
arch/sh/mm/hugetlbpage.c | 2
arch/sparc/mm/hugetlbpage.c | 2
arch/x86/Kconfig | 33
arch/x86/mm/pat/set_memory.c | 8
drivers/acpi/acpi_memhotplug.c | 5
drivers/base/memory.c | 105 ++
fs/Kconfig | 5
fs/block_dev.c | 2
fs/btrfs/compression.c | 5
fs/btrfs/extent_io.c | 22
fs/btrfs/inode.c | 33
fs/btrfs/reflink.c | 6
fs/btrfs/zlib.c | 5
fs/btrfs/zstd.c | 5
fs/buffer.c | 36
fs/dax.c | 8
fs/gfs2/glock.c | 3
fs/hugetlbfs/inode.c | 9
fs/inode.c | 11
fs/proc/task_mmu.c | 3
fs/userfaultfd.c | 149 +++
include/linux/buffer_head.h | 4
include/linux/cma.h | 4
include/linux/compaction.h | 1
include/linux/fs.h | 2
include/linux/gfp.h | 2
include/linux/highmem.h | 7
include/linux/huge_mm.h | 3
include/linux/hugetlb.h | 37
include/linux/memcontrol.h | 27
include/linux/memory.h | 8
include/linux/memory_hotplug.h | 15
include/linux/memremap.h | 2
include/linux/migrate.h | 11
include/linux/mm.h | 28
include/linux/mmzone.h | 20
include/linux/pagemap.h | 5
include/linux/pgtable.h | 12
include/linux/sched.h | 2
include/linux/sched/mm.h | 27
include/linux/shrinker.h | 7
include/linux/swap.h | 21
include/linux/userfaultfd_k.h | 55 +
include/linux/vm_event_item.h | 8
include/trace/events/cma.h | 92 +-
include/trace/events/migrate.h | 25
include/trace/events/mmflags.h | 7
include/uapi/linux/mempolicy.h | 7
include/uapi/linux/userfaultfd.h | 36
init/Kconfig | 5
kernel/sysctl.c | 2
lib/Kconfig.kfence | 1
lib/iov_iter.c | 8
mm/Kconfig | 28
mm/Makefile | 6
mm/cma.c | 70 +
mm/cma.h | 25
mm/cma_debug.c | 8
mm/cma_sysfs.c | 112 ++
mm/compaction.c | 113 ++
mm/filemap.c | 24
mm/frontswap.c | 12
mm/gup.c | 264 +++---
mm/gup_test.c | 29
mm/gup_test.h | 3
mm/highmem.c | 11
mm/huge_memory.c | 326 +++++++-
mm/hugetlb.c | 843 ++++++++++++++--------
mm/hugetlb_cgroup.c | 9
mm/internal.h | 10
mm/kfence/core.c | 61 +
mm/khugepaged.c | 63 -
mm/ksm.c | 17
mm/list_lru.c | 6
mm/memcontrol.c | 137 ---
mm/memory_hotplug.c | 220 +++++
mm/mempolicy.c | 16
mm/mempool.c | 2
mm/migrate.c | 103 --
mm/mlock.c | 4
mm/mmap.c | 18
mm/oom_kill.c | 2
mm/page_alloc.c | 83 +-
mm/process_vm_access.c | 1
mm/shmem.c | 2
mm/sparse.c | 4
mm/swap.c | 69 +
mm/swap_state.c | 4
mm/swapfile.c | 4
mm/truncate.c | 19
mm/userfaultfd.c | 39 -
mm/util.c | 26
mm/vmalloc.c | 2
mm/vmscan.c | 543 +++++++++-----
mm/vmstat.c | 45 -
mm/workingset.c | 1
mm/zsmalloc.c | 6
mm/zswap.c | 2
tools/testing/selftests/vm/.gitignore | 1
tools/testing/selftests/vm/Makefile | 1
tools/testing/selftests/vm/gup_test.c | 38
tools/testing/selftests/vm/split_huge_page_test.c | 400 ++++++++++
tools/testing/selftests/vm/userfaultfd.c | 164 ++++
125 files changed, 3596 insertions(+), 1668 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-04-30 5:52 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-04-30 5:52 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
A few misc subsystems and some of MM.
178 patches, based on 8ca5297e7e38f2dc8c753d33a5092e7be181fff0.
Subsystems affected by this patch series:
ia64
kbuild
scripts
sh
ocfs2
kfifo
vfs
kernel/watchdog
mm/slab-generic
mm/slub
mm/kmemleak
mm/debug
mm/pagecache
mm/msync
mm/gup
mm/memremap
mm/memcg
mm/pagemap
mm/mremap
mm/dma
mm/sparsemem
mm/vmalloc
mm/documentation
mm/kasan
mm/initialization
mm/pagealloc
mm/memory-failure
Subsystem: ia64
Zhang Yunkai <zhang.yunkai@zte.com.cn>:
arch/ia64/kernel/head.S: remove duplicate include
Bhaskar Chowdhury <unixbhaskar@gmail.com>:
arch/ia64/kernel/fsys.S: fix typos
arch/ia64/include/asm/pgtable.h: minor typo fixes
Valentin Schneider <valentin.schneider@arm.com>:
ia64: ensure proper NUMA distance and possible map initialization
Sergei Trofimovich <slyfox@gentoo.org>:
ia64: drop unused IA64_FW_EMU ifdef
ia64: simplify code flow around swiotlb init
Bhaskar Chowdhury <unixbhaskar@gmail.com>:
ia64: trivial spelling fixes
Sergei Trofimovich <slyfox@gentoo.org>:
ia64: fix EFI_DEBUG build
ia64: mca: always make IA64_MCA_DEBUG an expression
ia64: drop marked broken DISCONTIGMEM and VIRTUAL_MEM_MAP
ia64: module: fix symbolizer crash on fdescr
Subsystem: kbuild
Luc Van Oostenryck <luc.vanoostenryck@gmail.com>:
include/linux/compiler-gcc.h: sparse can do constant folding of __builtin_bswap*()
Subsystem: scripts
Tom Saeger <tom.saeger@oracle.com>:
scripts/spelling.txt: add entries for recent discoveries
Wan Jiabing <wanjiabing@vivo.com>:
scripts: a new script for checking duplicate struct declaration
Subsystem: sh
Zhang Yunkai <zhang.yunkai@zte.com.cn>:
arch/sh/include/asm/tlb.h: remove duplicate include
Subsystem: ocfs2
Yang Li <yang.lee@linux.alibaba.com>:
ocfs2: replace DEFINE_SIMPLE_ATTRIBUTE with DEFINE_DEBUGFS_ATTRIBUTE
Joseph Qi <joseph.qi@linux.alibaba.com>:
ocfs2: map flags directly in flags_to_o2dlm()
Bhaskar Chowdhury <unixbhaskar@gmail.com>:
ocfs2: fix a typo
Jiapeng Chong <jiapeng.chong@linux.alibaba.com>:
ocfs2/dlm: remove unused function
Subsystem: kfifo
Dan Carpenter <dan.carpenter@oracle.com>:
kfifo: fix ternary sign extension bugs
Subsystem: vfs
Randy Dunlap <rdunlap@infradead.org>:
vfs: fs_parser: clean up kernel-doc warnings
Subsystem: kernel/watchdog
Petr Mladek <pmladek@suse.com>:
Patch series "watchdog/softlockup: Report overall time and some cleanup", v2:
watchdog: rename __touch_watchdog() to a better descriptive name
watchdog: explicitly update timestamp when reporting softlockup
watchdog/softlockup: report the overall time of softlockups
watchdog/softlockup: remove logic that tried to prevent repeated reports
watchdog: fix barriers when printing backtraces from all CPUs
watchdog: cleanup handling of false positives
Subsystem: mm/slab-generic
Rafael Aquini <aquini@redhat.com>:
mm/slab_common: provide "slab_merge" option for !IS_ENABLED(CONFIG_SLAB_MERGE_DEFAULT) builds
Subsystem: mm/slub
Vlastimil Babka <vbabka@suse.cz>:
mm, slub: enable slub_debug static key when creating cache with explicit debug flags
Oliver Glitta <glittao@gmail.com>:
kunit: add a KUnit test for SLUB debugging functionality
slub: remove resiliency_test() function
Bhaskar Chowdhury <unixbhaskar@gmail.com>:
mm/slub.c: trivial typo fixes
Subsystem: mm/kmemleak
Bhaskar Chowdhury <unixbhaskar@gmail.com>:
mm/kmemleak.c: fix a typo
Subsystem: mm/debug
Georgi Djakov <georgi.djakov@linaro.org>:
mm/page_owner: record the timestamp of all pages during free
zhongjiang-ali <zhongjiang-ali@linux.alibaba.com>:
mm, page_owner: remove unused parameter in __set_page_owner_handle
Sergei Trofimovich <slyfox@gentoo.org>:
mm: page_owner: fetch backtrace only for tracked pages
mm: page_owner: use kstrtobool() to parse bool option
mm: page_owner: detect page_owner recursion via task_struct
mm: page_poison: print page info when corruption is caught
Anshuman Khandual <anshuman.khandual@arm.com>:
mm/memtest: add ARCH_USE_MEMTEST
Subsystem: mm/pagecache
Jens Axboe <axboe@kernel.dk>:
Patch series "Improve IOCB_NOWAIT O_DIRECT reads", v3:
mm: provide filemap_range_needs_writeback() helper
mm: use filemap_range_needs_writeback() for O_DIRECT reads
iomap: use filemap_range_needs_writeback() for O_DIRECT reads
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
mm/filemap: use filemap_read_page in filemap_fault
mm/filemap: drop check for truncated page after I/O
Johannes Weiner <hannes@cmpxchg.org>:
mm: page-writeback: simplify memcg handling in test_clear_page_writeback()
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
mm: move page_mapping_file to pagemap.h
Rui Sun <sunrui26@huawei.com>:
mm/filemap: update stale comment
Subsystem: mm/msync
Nikita Ermakov <sh1r4s3@mail.si-head.nl>:
mm/msync: exit early when the flags is an MS_ASYNC and start < vm_start
Subsystem: mm/gup
Joao Martins <joao.m.martins@oracle.com>:
Patch series "mm/gup: page unpining improvements", v4:
mm/gup: add compound page list iterator
mm/gup: decrement head page once for group of subpages
mm/gup: add a range variant of unpin_user_pages_dirty_lock()
RDMA/umem: batch page unpin in __ib_umem_release()
Yang Shi <shy828301@gmail.com>:
mm: gup: remove FOLL_SPLIT
Subsystem: mm/memremap
Zhiyuan Dai <daizhiyuan@phytium.com.cn>:
mm/memremap.c: fix improper SPDX comment style
Subsystem: mm/memcg
Muchun Song <songmuchun@bytedance.com>:
mm: memcontrol: fix kernel stack account
Shakeel Butt <shakeelb@google.com>:
memcg: cleanup root memcg checks
memcg: enable memcg oom-kill for __GFP_NOFAIL
Johannes Weiner <hannes@cmpxchg.org>:
Patch series "mm: memcontrol: switch to rstat", v3:
mm: memcontrol: fix cpuhotplug statistics flushing
mm: memcontrol: kill mem_cgroup_nodeinfo()
mm: memcontrol: privatize memcg_page_state query functions
cgroup: rstat: support cgroup1
cgroup: rstat: punt root-level optimization to individual controllers
mm: memcontrol: switch to rstat
mm: memcontrol: consolidate lruvec stat flushing
kselftests: cgroup: update kmem test for new vmstat implementation
Shakeel Butt <shakeelb@google.com>:
memcg: charge before adding to swapcache on swapin
Muchun Song <songmuchun@bytedance.com>:
Patch series "Use obj_cgroup APIs to charge kmem pages", v5:
mm: memcontrol: slab: fix obtain a reference to a freeing memcg
mm: memcontrol: introduce obj_cgroup_{un}charge_pages
mm: memcontrol: directly access page->memcg_data in mm/page_alloc.c
mm: memcontrol: change ug->dummy_page only if memcg changed
mm: memcontrol: use obj_cgroup APIs to charge kmem pages
mm: memcontrol: inline __memcg_kmem_{un}charge() into obj_cgroup_{un}charge_pages()
mm: memcontrol: move PageMemcgKmem to the scope of CONFIG_MEMCG_KMEM
Wan Jiabing <wanjiabing@vivo.com>:
linux/memcontrol.h: remove duplicate struct declaration
Johannes Weiner <hannes@cmpxchg.org>:
mm: page_counter: mitigate consequences of a page_counter underflow
Subsystem: mm/pagemap
Wang Qing <wangqing@vivo.com>:
mm/memory.c: do_numa_page(): delete bool "migrated"
Zhiyuan Dai <daizhiyuan@phytium.com.cn>:
mm/interval_tree: add comments to improve code readability
Oscar Salvador <osalvador@suse.de>:
Patch series "Cleanup and fixups for vmemmap handling", v6:
x86/vmemmap: drop handling of 4K unaligned vmemmap range
x86/vmemmap: drop handling of 1GB vmemmap ranges
x86/vmemmap: handle unpopulated sub-pmd ranges
x86/vmemmap: optimize for consecutive sections in partial populated PMDs
Ovidiu Panait <ovidiu.panait@windriver.com>:
mm, tracing: improve rss_stat tracepoint message
Christoph Hellwig <hch@lst.de>:
Patch series "add remap_pfn_range_notrack instead of reinventing it in i915", v2:
mm: add remap_pfn_range_notrack
mm: add a io_mapping_map_user helper
i915: use io_mapping_map_user
i915: fix remap_io_sg to verify the pgprot
Huang Ying <ying.huang@intel.com>:
NUMA balancing: reduce TLB flush via delaying mapping on hint page fault
Subsystem: mm/mremap
Brian Geffon <bgeffon@google.com>:
Patch series "mm: Extend MREMAP_DONTUNMAP to non-anonymous mappings", v5:
mm: extend MREMAP_DONTUNMAP to non-anonymous mappings
Revert "mremap: don't allow MREMAP_DONTUNMAP on special_mappings and aio"
selftests: add a MREMAP_DONTUNMAP selftest for shmem
Subsystem: mm/dma
Zhiyuan Dai <daizhiyuan@phytium.com.cn>:
mm/dmapool: switch from strlcpy to strscpy
Subsystem: mm/sparsemem
Wang Wensheng <wangwensheng4@huawei.com>:
mm/sparse: add the missing sparse_buffer_fini() in error branch
Subsystem: mm/vmalloc
Christoph Hellwig <hch@lst.de>:
Patch series "remap_vmalloc_range cleanups":
samples/vfio-mdev/mdpy: use remap_vmalloc_range
mm: unexport remap_vmalloc_range_partial
Serapheim Dimitropoulos <serapheim.dimitro@delphix.com>:
mm/vmalloc: use rb_tree instead of list for vread() lookups
Nicholas Piggin <npiggin@gmail.com>:
Patch series "huge vmalloc mappings", v13:
ARM: mm: add missing pud_page define to 2-level page tables
mm/vmalloc: fix HUGE_VMAP regression by enabling huge pages in vmalloc_to_page
mm: apply_to_pte_range warn and fail if a large pte is encountered
mm/vmalloc: rename vmap_*_range vmap_pages_*_range
mm/ioremap: rename ioremap_*_range to vmap_*_range
mm: HUGE_VMAP arch support cleanup
powerpc: inline huge vmap supported functions
arm64: inline huge vmap supported functions
x86: inline huge vmap supported functions
mm/vmalloc: provide fallback arch huge vmap support functions
mm: move vmap_range from mm/ioremap.c to mm/vmalloc.c
mm/vmalloc: add vmap_range_noflush variant
mm/vmalloc: hugepage vmalloc mappings
Patch series "mm/vmalloc: cleanup after hugepage series", v2:
mm/vmalloc: remove map_kernel_range
kernel/dma: remove unnecessary unmap_kernel_range
powerpc/xive: remove unnecessary unmap_kernel_range
mm/vmalloc: remove unmap_kernel_range
mm/vmalloc: improve allocation failure error messages
Vijayanand Jitta <vjitta@codeaurora.org>:
mm: vmalloc: prevent use after free in _vm_unmap_aliases
"Uladzislau Rezki (Sony)" <urezki@gmail.com>:
lib/test_vmalloc.c: remove two kvfree_rcu() tests
lib/test_vmalloc.c: add a new 'nr_threads' parameter
vm/test_vmalloc.sh: adapt for updated driver interface
mm/vmalloc: refactor the preloading loagic
mm/vmalloc: remove an empty line
Subsystem: mm/documentation
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
mm/doc: fix fault_flag_allow_retry_first kerneldoc
mm/doc: fix page_maybe_dma_pinned kerneldoc
mm/doc: turn fault flags into an enum
mm/doc: add mm.h and mm_types.h to the mm-api document
Lukas Bulwahn <lukas.bulwahn@gmail.com>:
Patch series "kernel-doc and MAINTAINERS clean-up":
MAINTAINERS: assign pagewalk.h to MEMORY MANAGEMENT
pagewalk: prefix struct kernel-doc descriptions
Subsystem: mm/kasan
Zhiyuan Dai <daizhiyuan@phytium.com.cn>:
mm/kasan: switch from strlcpy to strscpy
Peter Collingbourne <pcc@google.com>:
kasan: fix kasan_byte_accessible() to be consistent with actual checks
Andrey Konovalov <andreyknvl@google.com>:
kasan: initialize shadow to TAG_INVALID for SW_TAGS
mm, kasan: don't poison boot memory with tag-based modes
Patch series "kasan: integrate with init_on_alloc/free", v3:
arm64: kasan: allow to init memory when setting tags
kasan: init memory in kasan_(un)poison for HW_TAGS
kasan, mm: integrate page_alloc init with HW_TAGS
kasan, mm: integrate slab init_on_alloc with HW_TAGS
kasan, mm: integrate slab init_on_free with HW_TAGS
kasan: docs: clean up sections
kasan: docs: update overview section
kasan: docs: update usage section
kasan: docs: update error reports section
kasan: docs: update boot parameters section
kasan: docs: update GENERIC implementation details section
kasan: docs: update SW_TAGS implementation details section
kasan: docs: update HW_TAGS implementation details section
kasan: docs: update shadow memory section
kasan: docs: update ignoring accesses section
kasan: docs: update tests section
Walter Wu <walter-zh.wu@mediatek.com>:
kasan: record task_work_add() call stack
Andrey Konovalov <andreyknvl@google.com>:
kasan: detect false-positives in tests
Zqiang <qiang.zhang@windriver.com>:
irq_work: record irq_work_queue() call stack
Subsystem: mm/initialization
Kefeng Wang <wangkefeng.wang@huawei.com>:
mm: move mem_init_print_info() into mm_init()
Subsystem: mm/pagealloc
David Hildenbrand <david@redhat.com>:
mm/page_alloc: drop pr_info_ratelimited() in alloc_contig_range()
Minchan Kim <minchan@kernel.org>:
mm: remove lru_add_drain_all in alloc_contig_range
Yu Zhao <yuzhao@google.com>:
include/linux/page-flags-layout.h: correctly determine LAST_CPUPID_WIDTH
include/linux/page-flags-layout.h: cleanups
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
Patch series "Rationalise __alloc_pages wrappers", v3:
mm/page_alloc: rename alloc_mask to alloc_gfp
mm/page_alloc: rename gfp_mask to gfp
mm/page_alloc: combine __alloc_pages and __alloc_pages_nodemask
mm/mempolicy: rename alloc_pages_current to alloc_pages
mm/mempolicy: rewrite alloc_pages documentation
mm/mempolicy: rewrite alloc_pages_vma documentation
mm/mempolicy: fix mpol_misplaced kernel-doc
Minchan Kim <minchan@kernel.org>:
mm: page_alloc: dump migrate-failed pages
Geert Uytterhoeven <geert@linux-m68k.org>:
mm/Kconfig: remove default DISCONTIGMEM_MANUAL
Kefeng Wang <wangkefeng.wang@huawei.com>:
mm, page_alloc: avoid page_to_pfn() in move_freepages()
zhouchuangao <zhouchuangao@vivo.com>:
mm/page_alloc: duplicate include linux/vmalloc.h
Mel Gorman <mgorman@techsingularity.net>:
Patch series "Introduce a bulk order-0 page allocator with two in-tree users", v6:
mm/page_alloc: rename alloced to allocated
mm/page_alloc: add a bulk page allocator
mm/page_alloc: add an array-based interface to the bulk page allocator
Jesper Dangaard Brouer <brouer@redhat.com>:
mm/page_alloc: optimize code layout for __alloc_pages_bulk
mm/page_alloc: inline __rmqueue_pcplist
Chuck Lever <chuck.lever@oracle.com>:
Patch series "SUNRPC consumer for the bulk page allocator":
SUNRPC: set rq_page_end differently
SUNRPC: refresh rq_pages using a bulk page allocator
Jesper Dangaard Brouer <brouer@redhat.com>:
net: page_pool: refactor dma_map into own function page_pool_dma_map
net: page_pool: use alloc_pages_bulk in refill code path
Sergei Trofimovich <slyfox@gentoo.org>:
mm: page_alloc: ignore init_on_free=1 for debug_pagealloc=1
huxiang <huxiang@uniontech.com>:
mm/page_alloc: redundant definition variables of pfn in for loop
Mike Rapoport <rppt@linux.ibm.com>:
mm/mmzone.h: fix existing kernel-doc comments and link them to core-api
Subsystem: mm/memory-failure
Jane Chu <jane.chu@oracle.com>:
mm/memory-failure: unnecessary amount of unmapping
Documentation/admin-guide/kernel-parameters.txt | 7
Documentation/admin-guide/mm/transhuge.rst | 2
Documentation/core-api/cachetlb.rst | 4
Documentation/core-api/mm-api.rst | 6
Documentation/dev-tools/kasan.rst | 355 +++++-----
Documentation/vm/page_owner.rst | 2
Documentation/vm/transhuge.rst | 5
MAINTAINERS | 1
arch/Kconfig | 11
arch/alpha/mm/init.c | 1
arch/arc/mm/init.c | 1
arch/arm/Kconfig | 1
arch/arm/include/asm/pgtable-3level.h | 2
arch/arm/include/asm/pgtable.h | 3
arch/arm/mm/copypage-v4mc.c | 1
arch/arm/mm/copypage-v6.c | 1
arch/arm/mm/copypage-xscale.c | 1
arch/arm/mm/init.c | 2
arch/arm64/Kconfig | 1
arch/arm64/include/asm/memory.h | 4
arch/arm64/include/asm/mte-kasan.h | 39 -
arch/arm64/include/asm/vmalloc.h | 38 -
arch/arm64/mm/init.c | 4
arch/arm64/mm/mmu.c | 36 -
arch/csky/abiv1/cacheflush.c | 1
arch/csky/mm/init.c | 1
arch/h8300/mm/init.c | 2
arch/hexagon/mm/init.c | 1
arch/ia64/Kconfig | 23
arch/ia64/configs/bigsur_defconfig | 1
arch/ia64/include/asm/meminit.h | 11
arch/ia64/include/asm/module.h | 6
arch/ia64/include/asm/page.h | 25
arch/ia64/include/asm/pgtable.h | 7
arch/ia64/kernel/Makefile | 2
arch/ia64/kernel/acpi.c | 7
arch/ia64/kernel/efi.c | 11
arch/ia64/kernel/fsys.S | 4
arch/ia64/kernel/head.S | 6
arch/ia64/kernel/ia64_ksyms.c | 12
arch/ia64/kernel/machine_kexec.c | 2
arch/ia64/kernel/mca.c | 4
arch/ia64/kernel/module.c | 29
arch/ia64/kernel/pal.S | 6
arch/ia64/mm/Makefile | 1
arch/ia64/mm/contig.c | 4
arch/ia64/mm/discontig.c | 21
arch/ia64/mm/fault.c | 15
arch/ia64/mm/init.c | 221 ------
arch/m68k/mm/init.c | 1
arch/microblaze/mm/init.c | 1
arch/mips/Kconfig | 1
arch/mips/loongson64/numa.c | 1
arch/mips/mm/cache.c | 1
arch/mips/mm/init.c | 1
arch/mips/sgi-ip27/ip27-memory.c | 1
arch/nds32/mm/init.c | 1
arch/nios2/mm/cacheflush.c | 1
arch/nios2/mm/init.c | 1
arch/openrisc/mm/init.c | 2
arch/parisc/mm/init.c | 2
arch/powerpc/Kconfig | 1
arch/powerpc/include/asm/vmalloc.h | 34 -
arch/powerpc/kernel/isa-bridge.c | 4
arch/powerpc/kernel/pci_64.c | 2
arch/powerpc/mm/book3s64/radix_pgtable.c | 29
arch/powerpc/mm/ioremap.c | 2
arch/powerpc/mm/mem.c | 1
arch/powerpc/sysdev/xive/common.c | 4
arch/riscv/mm/init.c | 1
arch/s390/mm/init.c | 2
arch/sh/include/asm/tlb.h | 10
arch/sh/mm/cache-sh4.c | 1
arch/sh/mm/cache-sh7705.c | 1
arch/sh/mm/init.c | 1
arch/sparc/include/asm/pgtable_32.h | 3
arch/sparc/mm/init_32.c | 2
arch/sparc/mm/init_64.c | 1
arch/sparc/mm/tlb.c | 1
arch/um/kernel/mem.c | 1
arch/x86/Kconfig | 1
arch/x86/include/asm/vmalloc.h | 42 -
arch/x86/kernel/cpu/resctrl/pseudo_lock.c | 2
arch/x86/mm/init_32.c | 2
arch/x86/mm/init_64.c | 222 ++++--
arch/x86/mm/ioremap.c | 33
arch/x86/mm/pgtable.c | 13
arch/xtensa/Kconfig | 1
arch/xtensa/mm/init.c | 1
block/blk-cgroup.c | 17
drivers/gpu/drm/i915/Kconfig | 1
drivers/gpu/drm/i915/gem/i915_gem_mman.c | 9
drivers/gpu/drm/i915/i915_drv.h | 3
drivers/gpu/drm/i915/i915_mm.c | 117 ---
drivers/infiniband/core/umem.c | 12
drivers/pci/pci.c | 2
fs/aio.c | 5
fs/fs_parser.c | 2
fs/iomap/direct-io.c | 24
fs/ocfs2/blockcheck.c | 2
fs/ocfs2/dlm/dlmrecovery.c | 7
fs/ocfs2/stack_o2cb.c | 36 -
fs/ocfs2/stackglue.c | 2
include/linux/compiler-gcc.h | 8
include/linux/fs.h | 2
include/linux/gfp.h | 45 -
include/linux/io-mapping.h | 3
include/linux/io.h | 9
include/linux/kasan.h | 51 +
include/linux/memcontrol.h | 271 ++++----
include/linux/mm.h | 50 -
include/linux/mmzone.h | 43 -
include/linux/page-flags-layout.h | 64 -
include/linux/pagemap.h | 10
include/linux/pagewalk.h | 4
include/linux/sched.h | 4
include/linux/slab.h | 2
include/linux/slub_def.h | 2
include/linux/vmalloc.h | 73 +-
include/linux/vmstat.h | 24
include/net/page_pool.h | 2
include/trace/events/kmem.h | 24
init/main.c | 2
kernel/cgroup/cgroup.c | 34 -
kernel/cgroup/rstat.c | 61 +
kernel/dma/remap.c | 1
kernel/fork.c | 13
kernel/irq_work.c | 7
kernel/task_work.c | 3
kernel/watchdog.c | 102 +--
lib/Kconfig.debug | 14
lib/Makefile | 1
lib/test_kasan.c | 59 -
lib/test_slub.c | 124 +++
lib/test_vmalloc.c | 128 +--
mm/Kconfig | 4
mm/Makefile | 1
mm/debug_vm_pgtable.c | 4
mm/dmapool.c | 2
mm/filemap.c | 61 +
mm/gup.c | 145 +++-
mm/hugetlb.c | 2
mm/internal.h | 25
mm/interval_tree.c | 2
mm/io-mapping.c | 29
mm/ioremap.c | 361 ++--------
mm/kasan/common.c | 53 -
mm/kasan/generic.c | 12
mm/kasan/kasan.h | 28
mm/kasan/report_generic.c | 2
mm/kasan/shadow.c | 10
mm/kasan/sw_tags.c | 12
mm/kmemleak.c | 2
mm/memcontrol.c | 798 ++++++++++++------------
mm/memory-failure.c | 2
mm/memory.c | 191 +++--
mm/mempolicy.c | 78 --
mm/mempool.c | 4
mm/memremap.c | 2
mm/migrate.c | 2
mm/mm_init.c | 4
mm/mmap.c | 6
mm/mremap.c | 6
mm/msync.c | 6
mm/page-writeback.c | 9
mm/page_alloc.c | 430 +++++++++---
mm/page_counter.c | 8
mm/page_owner.c | 68 --
mm/page_poison.c | 6
mm/percpu-vm.c | 7
mm/slab.c | 43 -
mm/slab.h | 24
mm/slab_common.c | 10
mm/slub.c | 215 ++----
mm/sparse.c | 1
mm/swap_state.c | 13
mm/util.c | 10
mm/vmalloc.c | 728 ++++++++++++++++-----
net/core/page_pool.c | 127 ++-
net/sunrpc/svc_xprt.c | 38 -
samples/kfifo/bytestream-example.c | 8
samples/kfifo/inttype-example.c | 8
samples/kfifo/record-example.c | 8
samples/vfio-mdev/mdpy.c | 4
scripts/checkdeclares.pl | 53 +
scripts/spelling.txt | 26
tools/testing/selftests/cgroup/test_kmem.c | 22
tools/testing/selftests/vm/mremap_dontunmap.c | 52 +
tools/testing/selftests/vm/test_vmalloc.sh | 21
189 files changed, 3642 insertions(+), 3013 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-04-23 21:28 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-04-23 21:28 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
5 patches, based on 5bfc75d92efd494db37f5c4c173d3639d4772966.
Subsystems affected by this patch series:
coda
overlayfs
mm/pagecache
mm/memcg
Subsystem: coda
Christian König <christian.koenig@amd.com>:
coda: fix reference counting in coda_file_mmap error path
Subsystem: overlayfs
Christian König <christian.koenig@amd.com>:
ovl: fix reference counting in ovl_mmap error path
Subsystem: mm/pagecache
Hugh Dickins <hughd@google.com>:
mm/filemap: fix find_lock_entries hang on 32-bit THP
mm/filemap: fix mapping_seek_hole_data on THP & 32-bit
Subsystem: mm/memcg
Vasily Averin <vvs@virtuozzo.com>:
tools/cgroup/slabinfo.py: updated to work on current kernel
fs/coda/file.c | 6 +++---
fs/overlayfs/file.c | 11 +----------
mm/filemap.c | 31 +++++++++++++++++++------------
tools/cgroup/memcg_slabinfo.py | 8 ++++----
4 files changed, 27 insertions(+), 29 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-04-16 22:45 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-04-16 22:45 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
12 patches, based on 06c2aac4014c38247256fe49c61b7f55890271e7.
Subsystems affected by this patch series:
mm/documentation
mm/kasan
csky
ia64
mm/pagemap
gcov
lib
Subsystem: mm/documentation
Randy Dunlap <rdunlap@infradead.org>:
mm: eliminate "expecting prototype" kernel-doc warnings
Subsystem: mm/kasan
Arnd Bergmann <arnd@arndb.de>:
kasan: fix hwasan build for gcc
Walter Wu <walter-zh.wu@mediatek.com>:
kasan: remove redundant config option
Subsystem: csky
Randy Dunlap <rdunlap@infradead.org>:
csky: change a Kconfig symbol name to fix e1000 build error
Subsystem: ia64
Randy Dunlap <rdunlap@infradead.org>:
ia64: remove duplicate entries in generic_defconfig
ia64: fix discontig.c section mismatches
John Paul Adrian Glaubitz <glaubitz () physik ! fu-berlin ! de>:
ia64: tools: remove inclusion of ia64-specific version of errno.h header
John Paul Adrian Glaubitz <glaubitz@physik.fu-berlin.de>:
ia64: tools: remove duplicate definition of ia64_mf() on ia64
Subsystem: mm/pagemap
Zack Rusin <zackr@vmware.com>:
mm/mapping_dirty_helpers: guard hugepage pud's usage
Christophe Leroy <christophe.leroy@csgroup.eu>:
mm: ptdump: fix build failure
Subsystem: gcov
Johannes Berg <johannes.berg@intel.com>:
gcov: clang: fix clang-11+ build
Subsystem: lib
Randy Dunlap <rdunlap@infradead.org>:
lib: remove "expecting prototype" kernel-doc warnings
arch/arm64/kernel/sleep.S | 2 +-
arch/csky/Kconfig | 2 +-
arch/csky/include/asm/page.h | 2 +-
arch/ia64/configs/generic_defconfig | 2 --
arch/ia64/mm/discontig.c | 6 +++---
arch/x86/kernel/acpi/wakeup_64.S | 2 +-
include/linux/kasan.h | 2 +-
kernel/gcov/clang.c | 2 +-
lib/Kconfig.kasan | 9 ++-------
lib/earlycpio.c | 4 ++--
lib/lru_cache.c | 3 ++-
lib/parman.c | 4 ++--
lib/radix-tree.c | 11 ++++++-----
mm/kasan/common.c | 2 +-
mm/kasan/kasan.h | 2 +-
mm/kasan/report_generic.c | 2 +-
mm/mapping_dirty_helpers.c | 2 ++
mm/mmu_gather.c | 29 +++++++++++++++++++----------
mm/oom_kill.c | 2 +-
mm/ptdump.c | 2 +-
mm/shuffle.c | 4 ++--
scripts/Makefile.kasan | 22 ++++++++++++++--------
security/Kconfig.hardening | 4 ++--
tools/arch/ia64/include/asm/barrier.h | 3 ---
tools/include/uapi/asm/errno.h | 2 --
25 files changed, 67 insertions(+), 60 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-04-09 20:26 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-04-09 20:26 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
16 patches, based on 17e7124aad766b3f158943acb51467f86220afe9.
Subsystems affected by this patch series:
MAINTAINERS
mailmap
mm/kasan
mm/gup
nds32
gcov
ocfs2
ia64
mm/pagecache
mm/kasan
mm/kfence
lib
Subsystem: MAINTAINERS
Marek Behún <kabel@kernel.org>:
MAINTAINERS: update CZ.NIC's Turris information
treewide: change my e-mail address, fix my name
Subsystem: mailmap
Jordan Crouse <jordan@cosmicpenguin.net>:
mailmap: update email address for Jordan Crouse
Matthew Wilcox <willy@infradead.org>:
.mailmap: fix old email addresses
Subsystem: mm/kasan
Arnd Bergmann <arnd@arndb.de>:
kasan: fix hwasan build for gcc
Walter Wu <walter-zh.wu@mediatek.com>:
kasan: remove redundant config option
Subsystem: mm/gup
Aili Yao <yaoaili@kingsoft.com>:
mm/gup: check page posion status for coredump.
Subsystem: nds32
Mike Rapoport <rppt@linux.ibm.com>:
nds32: flush_dcache_page: use page_mapping_file to avoid races with swapoff
Subsystem: gcov
Nick Desaulniers <ndesaulniers@google.com>:
gcov: re-fix clang-11+ support
Subsystem: ocfs2
Wengang Wang <wen.gang.wang@oracle.com>:
ocfs2: fix deadlock between setattr and dio_end_io_write
Subsystem: ia64
Sergei Trofimovich <slyfox@gentoo.org>:
ia64: fix user_stack_pointer() for ptrace()
Subsystem: mm/pagecache
Jack Qiu <jack.qiu@huawei.com>:
fs: direct-io: fix missing sdio->boundary
Subsystem: mm/kasan
Andrey Konovalov <andreyknvl@google.com>:
kasan: fix conflict with page poisoning
Andrew Morton <akpm@linux-foundation.org>:
lib/test_kasan_module.c: suppress unused var warning
Subsystem: mm/kfence
Marco Elver <elver@google.com>:
kfence, x86: fix preemptible warning on KPTI-enabled systems
Subsystem: lib
Julian Braha <julianbraha@gmail.com>:
lib: fix kconfig dependency on ARCH_WANT_FRAME_POINTERS
.mailmap | 7 ++
Documentation/ABI/testing/debugfs-moxtet | 4 -
Documentation/ABI/testing/debugfs-turris-mox-rwtm | 2
Documentation/ABI/testing/sysfs-bus-moxtet-devices | 6 +-
Documentation/ABI/testing/sysfs-class-led-driver-turris-omnia | 2
Documentation/ABI/testing/sysfs-firmware-turris-mox-rwtm | 10 +--
Documentation/devicetree/bindings/leds/cznic,turris-omnia-leds.yaml | 2
MAINTAINERS | 13 +++-
arch/arm64/boot/dts/marvell/armada-3720-turris-mox.dts | 2
arch/arm64/kernel/sleep.S | 2
arch/ia64/include/asm/ptrace.h | 8 --
arch/nds32/mm/cacheflush.c | 2
arch/x86/include/asm/kfence.h | 7 ++
arch/x86/kernel/acpi/wakeup_64.S | 2
drivers/bus/moxtet.c | 4 -
drivers/firmware/turris-mox-rwtm.c | 4 -
drivers/gpio/gpio-moxtet.c | 4 -
drivers/leds/leds-turris-omnia.c | 4 -
drivers/mailbox/armada-37xx-rwtm-mailbox.c | 4 -
drivers/watchdog/armada_37xx_wdt.c | 4 -
fs/direct-io.c | 5 +
fs/ocfs2/aops.c | 11 ---
fs/ocfs2/file.c | 8 ++
include/dt-bindings/bus/moxtet.h | 2
include/linux/armada-37xx-rwtm-mailbox.h | 2
include/linux/kasan.h | 2
include/linux/moxtet.h | 2
kernel/gcov/clang.c | 29 ++++++----
lib/Kconfig.debug | 6 +-
lib/Kconfig.kasan | 9 ---
lib/test_kasan_module.c | 2
mm/gup.c | 4 +
mm/internal.h | 20 ++++++
mm/kasan/common.c | 2
mm/kasan/kasan.h | 2
mm/kasan/report_generic.c | 2
mm/page_poison.c | 4 +
scripts/Makefile.kasan | 18 ++++--
security/Kconfig.hardening | 4 -
39 files changed, 136 insertions(+), 91 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-03-25 4:36 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-03-25 4:36 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
14 patches, based on 7acac4b3196caee5e21fb5ea53f8bc124e6a16fc.
Subsystems affected by this patch series:
mm/hugetlb
mm/kasan
mm/gup
mm/selftests
mm/z3fold
squashfs
ia64
gcov
mm/kfence
mm/memblock
mm/highmem
mailmap
Subsystem: mm/hugetlb
Miaohe Lin <linmiaohe@huawei.com>:
hugetlb_cgroup: fix imbalanced css_get and css_put pair for shared mappings
Subsystem: mm/kasan
Andrey Konovalov <andreyknvl@google.com>:
kasan: fix per-page tags for non-page_alloc pages
Subsystem: mm/gup
Sean Christopherson <seanjc@google.com>:
mm/mmu_notifiers: ensure range_end() is paired with range_start()
Subsystem: mm/selftests
Rong Chen <rong.a.chen@intel.com>:
selftests/vm: fix out-of-tree build
Subsystem: mm/z3fold
Thomas Hebb <tommyhebb@gmail.com>:
z3fold: prevent reclaim/free race for headless pages
Subsystem: squashfs
Sean Nyekjaer <sean@geanix.com>:
squashfs: fix inode lookup sanity checks
Phillip Lougher <phillip@squashfs.org.uk>:
squashfs: fix xattr id and id lookup sanity checks
Subsystem: ia64
Sergei Trofimovich <slyfox@gentoo.org>:
ia64: mca: allocate early mca with GFP_ATOMIC
ia64: fix format strings for err_inject
Subsystem: gcov
Nick Desaulniers <ndesaulniers@google.com>:
gcov: fix clang-11+ support
Subsystem: mm/kfence
Marco Elver <elver@google.com>:
kfence: make compatible with kmemleak
Subsystem: mm/memblock
Mike Rapoport <rppt@linux.ibm.com>:
mm: memblock: fix section mismatch warning again
Subsystem: mm/highmem
Ira Weiny <ira.weiny@intel.com>:
mm/highmem: fix CONFIG_DEBUG_KMAP_LOCAL_FORCE_MAP
Subsystem: mailmap
Andrey Konovalov <andreyknvl@google.com>:
mailmap: update Andrey Konovalov's email address
.mailmap | 1
arch/ia64/kernel/err_inject.c | 22 +++++------
arch/ia64/kernel/mca.c | 2 -
fs/squashfs/export.c | 8 +++-
fs/squashfs/id.c | 6 ++-
fs/squashfs/squashfs_fs.h | 1
fs/squashfs/xattr_id.c | 6 ++-
include/linux/hugetlb_cgroup.h | 15 ++++++-
include/linux/memblock.h | 4 +-
include/linux/mm.h | 18 +++++++--
include/linux/mmu_notifier.h | 10 ++---
kernel/gcov/clang.c | 69 ++++++++++++++++++++++++++++++++++++
mm/highmem.c | 4 +-
mm/hugetlb.c | 41 +++++++++++++++++++--
mm/hugetlb_cgroup.c | 10 ++++-
mm/kfence/core.c | 9 ++++
mm/kmemleak.c | 3 +
mm/mmu_notifier.c | 23 ++++++++++++
mm/z3fold.c | 16 +++++++-
tools/testing/selftests/vm/Makefile | 4 +-
20 files changed, 230 insertions(+), 42 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-03-13 5:06 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-03-13 5:06 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
29 patches, based on f78d76e72a4671ea52d12752d92077788b4f5d50.
Subsystems affected by this patch series:
mm/memblock
core-kernel
kconfig
mm/pagealloc
fork
mm/hugetlb
mm/highmem
binfmt
MAINTAINERS
kbuild
mm/kfence
mm/oom-kill
mm/madvise
mm/kasan
mm/userfaultfd
mm/memory-failure
ia64
mm/memcg
mm/zram
Subsystem: mm/memblock
Arnd Bergmann <arnd@arndb.de>:
memblock: fix section mismatch warning
Subsystem: core-kernel
Arnd Bergmann <arnd@arndb.de>:
stop_machine: mark helpers __always_inline
Subsystem: kconfig
Masahiro Yamada <masahiroy@kernel.org>:
init/Kconfig: make COMPILE_TEST depend on HAS_IOMEM
Subsystem: mm/pagealloc
Mike Rapoport <rppt@linux.ibm.com>:
mm/page_alloc.c: refactor initialization of struct page for holes in memory layout
Subsystem: fork
Fenghua Yu <fenghua.yu@intel.com>:
mm/fork: clear PASID for new mm
Subsystem: mm/hugetlb
Peter Xu <peterx@redhat.com>:
Patch series "mm/hugetlb: Early cow on fork, and a few cleanups", v5:
hugetlb: dedup the code to add a new file_region
hugetlb: break earlier in add_reservation_in_range() when we can
mm: introduce page_needs_cow_for_dma() for deciding whether cow
mm: use is_cow_mapping() across tree where proper
hugetlb: do early cow when page pinned on src mm
Subsystem: mm/highmem
OGAWA Hirofumi <hirofumi@mail.parknet.co.jp>:
mm/highmem.c: fix zero_user_segments() with start > end
Subsystem: binfmt
Lior Ribak <liorribak@gmail.com>:
binfmt_misc: fix possible deadlock in bm_register_write
Subsystem: MAINTAINERS
Vlastimil Babka <vbabka@suse.cz>:
MAINTAINERS: exclude uapi directories in API/ABI section
Subsystem: kbuild
Arnd Bergmann <arnd@arndb.de>:
linux/compiler-clang.h: define HAVE_BUILTIN_BSWAP*
Subsystem: mm/kfence
Marco Elver <elver@google.com>:
kfence: fix printk format for ptrdiff_t
kfence, slab: fix cache_alloc_debugcheck_after() for bulk allocations
kfence: fix reports if constant function prefixes exist
Subsystem: mm/oom-kill
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
include/linux/sched/mm.h: use rcu_dereference in in_vfork()
Subsystem: mm/madvise
Suren Baghdasaryan <surenb@google.com>:
mm/madvise: replace ptrace attach requirement for process_madvise
Subsystem: mm/kasan
Andrey Konovalov <andreyknvl@google.com>:
kasan, mm: fix crash with HW_TAGS and DEBUG_PAGEALLOC
kasan: fix KASAN_STACK dependency for HW_TAGS
Subsystem: mm/userfaultfd
Nadav Amit <namit@vmware.com>:
mm/userfaultfd: fix memory corruption due to writeprotect
Subsystem: mm/memory-failure
Naoya Horiguchi <naoya.horiguchi@nec.com>:
mm, hwpoison: do not lock page again when me_huge_page() successfully recovers
Subsystem: ia64
Sergei Trofimovich <slyfox@gentoo.org>:
ia64: fix ia64_syscall_get_set_arguments() for break-based syscalls
ia64: fix ptrace(PTRACE_SYSCALL_INFO_EXIT) sign
Subsystem: mm/memcg
Zhou Guanghui <zhouguanghui1@huawei.com>:
mm/memcg: rename mem_cgroup_split_huge_fixup to split_page_memcg and add nr_pages argument
mm/memcg: set memcg when splitting page
Subsystem: mm/zram
Minchan Kim <minchan@kernel.org>:
zram: fix return value on writeback_store
zram: fix broken page writeback
MAINTAINERS | 4
arch/ia64/include/asm/syscall.h | 2
arch/ia64/kernel/ptrace.c | 24 +++-
drivers/block/zram/zram_drv.c | 17 +-
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c | 4
drivers/gpu/drm/vmwgfx/vmwgfx_ttm_glue.c | 2
fs/binfmt_misc.c | 29 ++---
fs/proc/task_mmu.c | 2
include/linux/compiler-clang.h | 6 +
include/linux/memblock.h | 4
include/linux/memcontrol.h | 6 -
include/linux/mm.h | 21 +++
include/linux/mm_types.h | 1
include/linux/sched/mm.h | 3
include/linux/stop_machine.h | 11 +
init/Kconfig | 3
kernel/fork.c | 8 +
lib/Kconfig.kasan | 1
mm/highmem.c | 17 ++
mm/huge_memory.c | 10 -
mm/hugetlb.c | 123 +++++++++++++++------
mm/internal.h | 5
mm/kfence/report.c | 30 +++--
mm/madvise.c | 13 ++
mm/memcontrol.c | 15 +-
mm/memory-failure.c | 4
mm/memory.c | 16 +-
mm/page_alloc.c | 167 ++++++++++++++---------------
mm/slab.c | 2
29 files changed, 334 insertions(+), 216 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2021-02-26 17:55 ` incoming Linus Torvalds
@ 2021-02-26 19:16 ` Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-02-26 19:16 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, Linux-MM
On Fri, 26 Feb 2021 09:55:27 -0800 Linus Torvalds <torvalds@linux-foundation.org> wrote:
> On Thu, Feb 25, 2021 at 5:14 PM Andrew Morton <akpm@linux-foundation.org> wrote:
> >
> > - The rest of MM.
> >
> > Includes kfence - another runtime memory validator. Not as
> > thorough as KASAN, but it has unmeasurable overhead and is intended
> > to be usable in production builds.
> >
> > - Everything else
>
> Just to clarify: you have nothing else really pending?
Yes, that's it from me for -rc1.
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2021-02-26 1:14 incoming Andrew Morton
@ 2021-02-26 17:55 ` Linus Torvalds
2021-02-26 19:16 ` incoming Andrew Morton
0 siblings, 1 reply; 409+ messages in thread
From: Linus Torvalds @ 2021-02-26 17:55 UTC (permalink / raw)
To: Andrew Morton; +Cc: mm-commits, Linux-MM
On Thu, Feb 25, 2021 at 5:14 PM Andrew Morton <akpm@linux-foundation.org> wrote:
>
> - The rest of MM.
>
> Includes kfence - another runtime memory validator. Not as
> thorough as KASAN, but it has unmeasurable overhead and is intended
> to be usable in production builds.
>
> - Everything else
Just to clarify: you have nothing else really pending?
I'm hoping to just do -rc1 this weekend after all - despite my late
start due to loss of power for several days.
I'll allow late stragglers with good reason through, but the fewer of
those there are, the better, of course.
Thanks,
Linus
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-02-26 1:14 Andrew Morton
2021-02-26 17:55 ` incoming Linus Torvalds
0 siblings, 1 reply; 409+ messages in thread
From: Andrew Morton @ 2021-02-26 1:14 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
- The rest of MM.
Includes kfence - another runtime memory validator. Not as
thorough as KASAN, but it has unmeasurable overhead and is intended
to be usable in production builds.
- Everything else
118 patches, based on 6fbd6cf85a3be127454a1ad58525a3adcf8612ab.
Subsystems affected by this patch series:
mm/thp
mm/cma
mm/vmstat
mm/memory-hotplug
mm/mlock
mm/rmap
mm/zswap
mm/zsmalloc
mm/cleanups
mm/kfence
mm/kasan2
alpha
procfs
sysctl
misc
core-kernel
MAINTAINERS
lib
bitops
checkpatch
init
coredump
seq_file
gdb
ubsan
initramfs
mm/pagemap2
Subsystem: mm/thp
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
Patch series "Overhaul multi-page lookups for THP", v4:
mm: make pagecache tagged lookups return only head pages
mm/shmem: use pagevec_lookup in shmem_unlock_mapping
mm/swap: optimise get_shadow_from_swap_cache
mm: add FGP_ENTRY
mm/filemap: rename find_get_entry to mapping_get_entry
mm/filemap: add helper for finding pages
mm/filemap: add mapping_seek_hole_data
iomap: use mapping_seek_hole_data
mm: add and use find_lock_entries
mm: add an 'end' parameter to find_get_entries
mm: add an 'end' parameter to pagevec_lookup_entries
mm: remove nr_entries parameter from pagevec_lookup_entries
mm: pass pvec directly to find_get_entries
mm: remove pagevec_lookup_entries
Rik van Riel <riel@surriel.com>:
Patch series "mm,thp,shm: limit shmem THP alloc gfp_mask", v6:
mm,thp,shmem: limit shmem THP alloc gfp_mask
mm,thp,shm: limit gfp mask to no more than specified
mm,thp,shmem: make khugepaged obey tmpfs mount flags
mm,shmem,thp: limit shmem THP allocations to requested zones
Subsystem: mm/cma
Roman Gushchin <guro@fb.com>:
mm: cma: allocate cma areas bottom-up
David Hildenbrand <david@redhat.com>:
mm/cma: expose all pages to the buddy if activation of an area fails
mm/page_alloc: count CMA pages per zone and print them in /proc/zoneinfo
Patrick Daly <pdaly@codeaurora.org>:
mm: cma: print region name on failure
Subsystem: mm/vmstat
Johannes Weiner <hannes@cmpxchg.org>:
mm: vmstat: fix NOHZ wakeups for node stat changes
mm: vmstat: add some comments on internal storage of byte items
Jiang Biao <benbjiang@tencent.com>:
mm/vmstat.c: erase latency in vmstat_shepherd
Subsystem: mm/memory-hotplug
Dan Williams <dan.j.williams@intel.com>:
Patch series "mm: Fix pfn_to_online_page() with respect to ZONE_DEVICE", v4:
mm: move pfn_to_online_page() out of line
mm: teach pfn_to_online_page() to consider subsection validity
mm: teach pfn_to_online_page() about ZONE_DEVICE section collisions
mm: fix memory_failure() handling of dax-namespace metadata
Anshuman Khandual <anshuman.khandual@arm.com>:
mm/memory_hotplug: rename all existing 'memhp' into 'mhp'
David Hildenbrand <david@redhat.com>:
mm/memory_hotplug: MEMHP_MERGE_RESOURCE -> MHP_MERGE_RESOURCE
Miaohe Lin <linmiaohe@huawei.com>:
mm/memory_hotplug: use helper function zone_end_pfn() to get end_pfn
David Hildenbrand <david@redhat.com>:
drivers/base/memory: don't store phys_device in memory blocks
Documentation: sysfs/memory: clarify some memory block device properties
Anshuman Khandual <anshuman.khandual@arm.com>:
Patch series "mm/memory_hotplug: Pre-validate the address range with platform", v5:
mm/memory_hotplug: prevalidate the address range being added with platform
arm64/mm: define arch_get_mappable_range()
s390/mm: define arch_get_mappable_range()
David Hildenbrand <david@redhat.com>:
virtio-mem: check against mhp_get_pluggable_range() which memory we can hotplug
Subsystem: mm/mlock
Miaohe Lin <linmiaohe@huawei.com>:
mm/mlock: stop counting mlocked pages when none vma is found
Subsystem: mm/rmap
Miaohe Lin <linmiaohe@huawei.com>:
mm/rmap: correct some obsolete comments of anon_vma
mm/rmap: remove unneeded semicolon in page_not_mapped()
mm/rmap: fix obsolete comment in __page_check_anon_rmap()
mm/rmap: use page_not_mapped in try_to_unmap()
mm/rmap: correct obsolete comment of page_get_anon_vma()
mm/rmap: fix potential pte_unmap on an not mapped pte
Subsystem: mm/zswap
Randy Dunlap <rdunlap@infradead.org>:
mm: zswap: clean up confusing comment
Tian Tao <tiantao6@hisilicon.com>:
Patch series "Fix the compatibility of zsmalloc and zswap":
mm/zswap: add the flag can_sleep_mapped
mm: set the sleep_mapped to true for zbud and z3fold
Subsystem: mm/zsmalloc
Miaohe Lin <linmiaohe@huawei.com>:
mm/zsmalloc.c: convert to use kmem_cache_zalloc in cache_alloc_zspage()
Rokudo Yan <wu-yan@tcl.com>:
zsmalloc: account the number of compacted pages correctly
Miaohe Lin <linmiaohe@huawei.com>:
mm/zsmalloc.c: use page_private() to access page->private
Subsystem: mm/cleanups
Guo Ren <guoren@linux.alibaba.com>:
mm: page-flags.h: Typo fix (It -> If)
Daniel Vetter <daniel.vetter@ffwll.ch>:
mm/dmapool: use might_alloc()
mm/backing-dev.c: use might_alloc()
Stephen Zhang <stephenzhangzsd@gmail.com>:
mm/early_ioremap.c: use __func__ instead of function name
Subsystem: mm/kfence
Alexander Potapenko <glider@google.com>:
Patch series "KFENCE: A low-overhead sampling-based memory safety error detector", v7:
mm: add Kernel Electric-Fence infrastructure
x86, kfence: enable KFENCE for x86
Marco Elver <elver@google.com>:
arm64, kfence: enable KFENCE for ARM64
kfence: use pt_regs to generate stack trace on faults
Alexander Potapenko <glider@google.com>:
mm, kfence: insert KFENCE hooks for SLAB
mm, kfence: insert KFENCE hooks for SLUB
kfence, kasan: make KFENCE compatible with KASAN
Marco Elver <elver@google.com>:
kfence, Documentation: add KFENCE documentation
kfence: add test suite
MAINTAINERS: add entry for KFENCE
kfence: report sensitive information based on no_hash_pointers
Alexander Potapenko <glider@google.com>:
Patch series "Add error_report_end tracepoint to KFENCE and KASAN", v3:
tracing: add error_report_end trace point
kfence: use error_report_end tracepoint
kasan: use error_report_end tracepoint
Subsystem: mm/kasan2
Andrey Konovalov <andreyknvl@google.com>:
Patch series "kasan: optimizations and fixes for HW_TAGS", v4:
kasan, mm: don't save alloc stacks twice
kasan, mm: optimize kmalloc poisoning
kasan: optimize large kmalloc poisoning
kasan: clean up setting free info in kasan_slab_free
kasan: unify large kfree checks
kasan: rework krealloc tests
kasan, mm: fail krealloc on freed objects
kasan, mm: optimize krealloc poisoning
kasan: ensure poisoning size alignment
arm64: kasan: simplify and inline MTE functions
kasan: inline HW_TAGS helper functions
kasan: clarify that only first bug is reported in HW_TAGS
Subsystem: alpha
Randy Dunlap <rdunlap@infradead.org>:
alpha: remove CONFIG_EXPERIMENTAL from defconfigs
Subsystem: procfs
Helge Deller <deller@gmx.de>:
proc/wchan: use printk format instead of lookup_symbol_name()
Josef Bacik <josef@toxicpanda.com>:
proc: use kvzalloc for our kernel buffer
Subsystem: sysctl
Lin Feng <linf@wangsu.com>:
sysctl.c: fix underflow value setting risk in vm_table
Subsystem: misc
Randy Dunlap <rdunlap@infradead.org>:
include/linux: remove repeated words
Miguel Ojeda <ojeda@kernel.org>:
treewide: Miguel has moved
Subsystem: core-kernel
Hubert Jasudowicz <hubert.jasudowicz@gmail.com>:
groups: use flexible-array member in struct group_info
groups: simplify struct group_info allocation
Randy Dunlap <rdunlap@infradead.org>:
kernel: delete repeated words in comments
Subsystem: MAINTAINERS
Vlastimil Babka <vbabka@suse.cz>:
MAINTAINERS: add uapi directories to API/ABI section
Subsystem: lib
Huang Shijie <sjhuang@iluvatar.ai>:
lib/genalloc.c: change return type to unsigned long for bitmap_set_ll
Francis Laniel <laniel_francis@privacyrequired.com>:
string.h: move fortified functions definitions in a dedicated header.
Yogesh Lal <ylal@codeaurora.org>:
lib: stackdepot: add support to configure STACK_HASH_SIZE
Vijayanand Jitta <vjitta@codeaurora.org>:
lib: stackdepot: add support to disable stack depot
lib: stackdepot: fix ignoring return value warning
Masahiro Yamada <masahiroy@kernel.org>:
lib/cmdline: remove an unneeded local variable in next_arg()
Subsystem: bitops
Geert Uytterhoeven <geert+renesas@glider.be>:
include/linux/bitops.h: spelling s/synomyn/synonym/
Subsystem: checkpatch
Joe Perches <joe@perches.com>:
checkpatch: improve blank line after declaration test
Peng Wang <rocking@linux.alibaba.com>:
checkpatch: ignore warning designated initializers using NR_CPUS
Dwaipayan Ray <dwaipayanray1@gmail.com>:
checkpatch: trivial style fixes
Joe Perches <joe@perches.com>:
checkpatch: prefer ftrace over function entry/exit printks
checkpatch: improve TYPECAST_INT_CONSTANT test message
Aditya Srivastava <yashsri421@gmail.com>:
checkpatch: add warning for avoiding .L prefix symbols in assembly files
Joe Perches <joe@perches.com>:
checkpatch: add kmalloc_array_node to unnecessary OOM message check
Chris Down <chris@chrisdown.name>:
checkpatch: don't warn about colon termination in linker scripts
Song Liu <songliubraving@fb.com>:
checkpatch: do not apply "initialise globals to 0" check to BPF progs
Subsystem: init
Masahiro Yamada <masahiroy@kernel.org>:
init/version.c: remove Version_<LINUX_VERSION_CODE> symbol
init: clean up early_param_on_off() macro
Bhaskar Chowdhury <unixbhaskar@gmail.com>:
init/Kconfig: fix a typo in CC_VERSION_TEXT help text
Subsystem: coredump
Ira Weiny <ira.weiny@intel.com>:
fs/coredump: use kmap_local_page()
Subsystem: seq_file
NeilBrown <neilb@suse.de>:
Patch series "Fix some seq_file users that were recently broken":
seq_file: document how per-entry resources are managed.
x86: fix seq_file iteration for pat/memtype.c
Subsystem: gdb
George Prekas <prekageo@amazon.com>:
scripts/gdb: fix list_for_each
Sumit Garg <sumit.garg@linaro.org>:
kgdb: fix to kill breakpoints on initmem after boot
Subsystem: ubsan
Andrey Ryabinin <ryabinin.a.a@gmail.com>:
ubsan: remove overflow checks
Subsystem: initramfs
Florian Fainelli <f.fainelli@gmail.com>:
initramfs: panic with memory information
Subsystem: mm/pagemap2
Huang Pei <huangpei@loongson.cn>:
MIPS: make userspace mapping young by default
.mailmap | 1
CREDITS | 9
Documentation/ABI/testing/sysfs-devices-memory | 58 -
Documentation/admin-guide/auxdisplay/cfag12864b.rst | 2
Documentation/admin-guide/auxdisplay/ks0108.rst | 2
Documentation/admin-guide/kernel-parameters.txt | 6
Documentation/admin-guide/mm/memory-hotplug.rst | 20
Documentation/dev-tools/index.rst | 1
Documentation/dev-tools/kasan.rst | 8
Documentation/dev-tools/kfence.rst | 318 +++++++
Documentation/filesystems/seq_file.rst | 6
MAINTAINERS | 26
arch/alpha/configs/defconfig | 1
arch/arm64/Kconfig | 1
arch/arm64/include/asm/cache.h | 1
arch/arm64/include/asm/kasan.h | 1
arch/arm64/include/asm/kfence.h | 26
arch/arm64/include/asm/mte-def.h | 2
arch/arm64/include/asm/mte-kasan.h | 65 +
arch/arm64/include/asm/mte.h | 2
arch/arm64/kernel/mte.c | 46 -
arch/arm64/lib/mte.S | 16
arch/arm64/mm/fault.c | 8
arch/arm64/mm/mmu.c | 23
arch/mips/mm/cache.c | 30
arch/s390/mm/init.c | 1
arch/s390/mm/vmem.c | 14
arch/x86/Kconfig | 1
arch/x86/include/asm/kfence.h | 76 +
arch/x86/mm/fault.c | 10
arch/x86/mm/pat/memtype.c | 4
drivers/auxdisplay/cfag12864b.c | 4
drivers/auxdisplay/cfag12864bfb.c | 4
drivers/auxdisplay/ks0108.c | 4
drivers/base/memory.c | 35
drivers/block/zram/zram_drv.c | 2
drivers/hv/hv_balloon.c | 2
drivers/virtio/virtio_mem.c | 43
drivers/xen/balloon.c | 2
fs/coredump.c | 4
fs/iomap/seek.c | 125 --
fs/proc/base.c | 21
fs/proc/proc_sysctl.c | 4
include/linux/bitops.h | 2
include/linux/cfag12864b.h | 2
include/linux/cred.h | 2
include/linux/fortify-string.h | 302 ++++++
include/linux/gfp.h | 2
include/linux/init.h | 4
include/linux/kasan.h | 25
include/linux/kfence.h | 230 +++++
include/linux/kgdb.h | 2
include/linux/khugepaged.h | 2
include/linux/ks0108.h | 2
include/linux/mdev.h | 2
include/linux/memory.h | 3
include/linux/memory_hotplug.h | 33
include/linux/memremap.h | 6
include/linux/mmzone.h | 49 -
include/linux/page-flags.h | 4
include/linux/pagemap.h | 10
include/linux/pagevec.h | 10
include/linux/pgtable.h | 8
include/linux/ptrace.h | 2
include/linux/rmap.h | 3
include/linux/slab_def.h | 3
include/linux/slub_def.h | 3
include/linux/stackdepot.h | 9
include/linux/string.h | 282 ------
include/linux/vmstat.h | 6
include/linux/zpool.h | 3
include/linux/zsmalloc.h | 2
include/trace/events/error_report.h | 74 +
include/uapi/linux/firewire-cdev.h | 2
include/uapi/linux/input.h | 2
init/Kconfig | 2
init/initramfs.c | 19
init/main.c | 6
init/version.c | 8
kernel/debug/debug_core.c | 11
kernel/events/core.c | 8
kernel/events/uprobes.c | 2
kernel/groups.c | 7
kernel/locking/rtmutex.c | 4
kernel/locking/rwsem.c | 2
kernel/locking/semaphore.c | 2
kernel/sched/fair.c | 2
kernel/sched/membarrier.c | 2
kernel/sysctl.c | 8
kernel/trace/Makefile | 1
kernel/trace/error_report-traces.c | 12
lib/Kconfig | 9
lib/Kconfig.debug | 1
lib/Kconfig.kfence | 84 +
lib/Kconfig.ubsan | 17
lib/cmdline.c | 7
lib/genalloc.c | 3
lib/stackdepot.c | 41
lib/test_kasan.c | 111 ++
lib/test_ubsan.c | 49 -
lib/ubsan.c | 68 -
mm/Makefile | 1
mm/backing-dev.c | 3
mm/cma.c | 64 -
mm/dmapool.c | 3
mm/early_ioremap.c | 12
mm/filemap.c | 361 +++++---
mm/huge_memory.c | 6
mm/internal.h | 6
mm/kasan/common.c | 213 +++-
mm/kasan/generic.c | 3
mm/kasan/hw_tags.c | 2
mm/kasan/kasan.h | 97 +-
mm/kasan/report.c | 8
mm/kasan/shadow.c | 78 +
mm/kfence/Makefile | 6
mm/kfence/core.c | 875 +++++++++++++++++++-
mm/kfence/kfence.h | 126 ++
mm/kfence/kfence_test.c | 860 +++++++++++++++++++
mm/kfence/report.c | 350 ++++++--
mm/khugepaged.c | 22
mm/memory-failure.c | 6
mm/memory.c | 4
mm/memory_hotplug.c | 178 +++-
mm/memremap.c | 23
mm/mlock.c | 2
mm/page_alloc.c | 1
mm/rmap.c | 24
mm/shmem.c | 160 +--
mm/slab.c | 38
mm/slab_common.c | 29
mm/slub.c | 63 +
mm/swap.c | 54 -
mm/swap_state.c | 7
mm/truncate.c | 141 ---
mm/vmstat.c | 35
mm/z3fold.c | 1
mm/zbud.c | 1
mm/zpool.c | 13
mm/zsmalloc.c | 22
mm/zswap.c | 57 +
samples/auxdisplay/cfag12864b-example.c | 2
scripts/Makefile.ubsan | 2
scripts/checkpatch.pl | 152 ++-
scripts/gdb/linux/lists.py | 5
145 files changed, 5046 insertions(+), 1682 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2021-02-25 9:12 ` incoming Andrey Ryabinin
@ 2021-02-25 11:07 ` Walter Wu
0 siblings, 0 replies; 409+ messages in thread
From: Walter Wu @ 2021-02-25 11:07 UTC (permalink / raw)
To: Andrey Ryabinin
Cc: Arnd Bergmann, Linus Torvalds, Andrew Morton, Dmitry Vyukov,
Nathan Chancellor, Arnd Bergmann, Andrey Konovalov, Linux-MM,
mm-commits, Andrey Ryabinin, Alexander Potapenko
Hi Andrey,
On Thu, 2021-02-25 at 12:12 +0300, Andrey Ryabinin wrote:
> On Thu, Feb 25, 2021 at 11:53 AM Arnd Bergmann <arnd@kernel.org> wrote:
> >
> > On Wed, Feb 24, 2021 at 10:37 PM Linus Torvalds
> > <torvalds@linux-foundation.org> wrote:
> > >
> > > On Wed, Feb 24, 2021 at 1:30 PM Linus Torvalds
> > > <torvalds@linux-foundation.org> wrote:
> > > >
> > > > Hmm. I haven't bisected things yet, but I suspect it's something with
> > > > the KASAN patches. With this all applied, I get:
> > > >
> > > > lib/crypto/curve25519-hacl64.c: In function ‘ladder_cmult.constprop’:
> > > > lib/crypto/curve25519-hacl64.c:601:1: warning: the frame size of
> > > > 2288 bytes is larger than 2048 bytes [-Wframe-larger-than=]
> > > >
> > > > and
> > > >
> > > > lib/bitfield_kunit.c: In function ‘test_bitfields_constants’:
> > > > lib/bitfield_kunit.c:93:1: warning: the frame size of 11200 bytes is
> > > > larger than 2048 bytes [-Wframe-larger-than=]
> > > >
> > > > which is obviously not really acceptable. A 11kB stack frame _will_
> > > > cause issues.
> > >
> > > A quick bisect shoes that this was introduced by "[patch 101/173]
> > > kasan: remove redundant config option".
> > >
> > > I didn't check what part of that patch screws up, but it's definitely
> > > doing something bad.
> >
> > I'm not sure why that patch surfaced the bug, but it's worth pointing
> > out that the underlying problem is asan-stack in combination
> > with the structleak plugin. This will happen for every user of kunit.
> >
>
> The patch didn't update KASAN_STACK dependency in kconfig:
> config GCC_PLUGIN_STRUCTLEAK_BYREF
> ....
> depends on !(KASAN && KASAN_STACK=1)
>
> This 'depends on' stopped working with the patch
Thanks for pointing out this problem. I will re-send that patch.
Walter
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2021-02-25 8:53 ` incoming Arnd Bergmann
@ 2021-02-25 9:12 ` Andrey Ryabinin
2021-02-25 11:07 ` incoming Walter Wu
0 siblings, 1 reply; 409+ messages in thread
From: Andrey Ryabinin @ 2021-02-25 9:12 UTC (permalink / raw)
To: Arnd Bergmann
Cc: Linus Torvalds, Andrew Morton, Walter Wu, Dmitry Vyukov,
Nathan Chancellor, Arnd Bergmann, Andrey Konovalov, Linux-MM,
mm-commits, Andrey Ryabinin, Alexander Potapenko
On Thu, Feb 25, 2021 at 11:53 AM Arnd Bergmann <arnd@kernel.org> wrote:
>
> On Wed, Feb 24, 2021 at 10:37 PM Linus Torvalds
> <torvalds@linux-foundation.org> wrote:
> >
> > On Wed, Feb 24, 2021 at 1:30 PM Linus Torvalds
> > <torvalds@linux-foundation.org> wrote:
> > >
> > > Hmm. I haven't bisected things yet, but I suspect it's something with
> > > the KASAN patches. With this all applied, I get:
> > >
> > > lib/crypto/curve25519-hacl64.c: In function ‘ladder_cmult.constprop’:
> > > lib/crypto/curve25519-hacl64.c:601:1: warning: the frame size of
> > > 2288 bytes is larger than 2048 bytes [-Wframe-larger-than=]
> > >
> > > and
> > >
> > > lib/bitfield_kunit.c: In function ‘test_bitfields_constants’:
> > > lib/bitfield_kunit.c:93:1: warning: the frame size of 11200 bytes is
> > > larger than 2048 bytes [-Wframe-larger-than=]
> > >
> > > which is obviously not really acceptable. A 11kB stack frame _will_
> > > cause issues.
> >
> > A quick bisect shoes that this was introduced by "[patch 101/173]
> > kasan: remove redundant config option".
> >
> > I didn't check what part of that patch screws up, but it's definitely
> > doing something bad.
>
> I'm not sure why that patch surfaced the bug, but it's worth pointing
> out that the underlying problem is asan-stack in combination
> with the structleak plugin. This will happen for every user of kunit.
>
The patch didn't update KASAN_STACK dependency in kconfig:
config GCC_PLUGIN_STRUCTLEAK_BYREF
....
depends on !(KASAN && KASAN_STACK=1)
This 'depends on' stopped working with the patch
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2021-02-24 21:37 ` incoming Linus Torvalds
@ 2021-02-25 8:53 ` Arnd Bergmann
2021-02-25 9:12 ` incoming Andrey Ryabinin
0 siblings, 1 reply; 409+ messages in thread
From: Arnd Bergmann @ 2021-02-25 8:53 UTC (permalink / raw)
To: Linus Torvalds
Cc: Andrew Morton, Walter Wu, Dmitry Vyukov, Nathan Chancellor,
Arnd Bergmann, Andrey Konovalov, Linux-MM, mm-commits,
Andrey Ryabinin, Alexander Potapenko
On Wed, Feb 24, 2021 at 10:37 PM Linus Torvalds
<torvalds@linux-foundation.org> wrote:
>
> On Wed, Feb 24, 2021 at 1:30 PM Linus Torvalds
> <torvalds@linux-foundation.org> wrote:
> >
> > Hmm. I haven't bisected things yet, but I suspect it's something with
> > the KASAN patches. With this all applied, I get:
> >
> > lib/crypto/curve25519-hacl64.c: In function ‘ladder_cmult.constprop’:
> > lib/crypto/curve25519-hacl64.c:601:1: warning: the frame size of
> > 2288 bytes is larger than 2048 bytes [-Wframe-larger-than=]
> >
> > and
> >
> > lib/bitfield_kunit.c: In function ‘test_bitfields_constants’:
> > lib/bitfield_kunit.c:93:1: warning: the frame size of 11200 bytes is
> > larger than 2048 bytes [-Wframe-larger-than=]
> >
> > which is obviously not really acceptable. A 11kB stack frame _will_
> > cause issues.
>
> A quick bisect shoes that this was introduced by "[patch 101/173]
> kasan: remove redundant config option".
>
> I didn't check what part of that patch screws up, but it's definitely
> doing something bad.
I'm not sure why that patch surfaced the bug, but it's worth pointing
out that the underlying problem is asan-stack in combination
with the structleak plugin. This will happen for every user of kunit.
I sent a series[1] out earlier this year to turn off the structleak
plugin as an alternative workaround, but need to follow up on
the remaining patches. Someone suggested adding a more
generic way to turn off the plugin for a file instead of open-coding
the CLFAGS_REMOVE_*.o Makefile bit, which would help.
I am also still hoping that someone can come up with a way
to make kunit work better with the structleak plugin, as there
shouldn't be a fundamental reason why it can't work, just that
it the code pattern triggers a particularly bad case in the compiler.
Arnd
[1] https://lore.kernel.org/lkml/20210125124533.101339-1-arnd@kernel.org/
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2021-02-24 21:30 ` incoming Linus Torvalds
@ 2021-02-24 21:37 ` Linus Torvalds
2021-02-25 8:53 ` incoming Arnd Bergmann
0 siblings, 1 reply; 409+ messages in thread
From: Linus Torvalds @ 2021-02-24 21:37 UTC (permalink / raw)
To: Andrew Morton, Walter Wu, Dmitry Vyukov, Nathan Chancellor,
Arnd Bergmann, Andrey Konovalov
Cc: Linux-MM, mm-commits, Andrey Ryabinin, Alexander Potapenko
On Wed, Feb 24, 2021 at 1:30 PM Linus Torvalds
<torvalds@linux-foundation.org> wrote:
>
> Hmm. I haven't bisected things yet, but I suspect it's something with
> the KASAN patches. With this all applied, I get:
>
> lib/crypto/curve25519-hacl64.c: In function ‘ladder_cmult.constprop’:
> lib/crypto/curve25519-hacl64.c:601:1: warning: the frame size of
> 2288 bytes is larger than 2048 bytes [-Wframe-larger-than=]
>
> and
>
> lib/bitfield_kunit.c: In function ‘test_bitfields_constants’:
> lib/bitfield_kunit.c:93:1: warning: the frame size of 11200 bytes is
> larger than 2048 bytes [-Wframe-larger-than=]
>
> which is obviously not really acceptable. A 11kB stack frame _will_
> cause issues.
A quick bisect shoes that this was introduced by "[patch 101/173]
kasan: remove redundant config option".
I didn't check what part of that patch screws up, but it's definitely
doing something bad.
I will drop that patch.
Linus
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2021-02-24 19:58 incoming Andrew Morton
@ 2021-02-24 21:30 ` Linus Torvalds
2021-02-24 21:37 ` incoming Linus Torvalds
0 siblings, 1 reply; 409+ messages in thread
From: Linus Torvalds @ 2021-02-24 21:30 UTC (permalink / raw)
To: Andrew Morton; +Cc: Linux-MM, mm-commits
On Wed, Feb 24, 2021 at 11:58 AM Andrew Morton
<akpm@linux-foundation.org> wrote:
>
> A few small subsystems and some of MM.
Hmm. I haven't bisected things yet, but I suspect it's something with
the KASAN patches. With this all applied, I get:
lib/crypto/curve25519-hacl64.c: In function ‘ladder_cmult.constprop’:
lib/crypto/curve25519-hacl64.c:601:1: warning: the frame size of
2288 bytes is larger than 2048 bytes [-Wframe-larger-than=]
and
lib/bitfield_kunit.c: In function ‘test_bitfields_constants’:
lib/bitfield_kunit.c:93:1: warning: the frame size of 11200 bytes is
larger than 2048 bytes [-Wframe-larger-than=]
which is obviously not really acceptable. A 11kB stack frame _will_
cause issues.
Linus
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-02-24 19:58 Andrew Morton
2021-02-24 21:30 ` incoming Linus Torvalds
0 siblings, 1 reply; 409+ messages in thread
From: Andrew Morton @ 2021-02-24 19:58 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
A few small subsystems and some of MM.
173 patches, based on c03c21ba6f4e95e406a1a7b4c34ef334b977c194.
Subsystems affected by this patch series:
hexagon
scripts
ntfs
ocfs2
vfs
mm/slab-generic
mm/slab
mm/slub
mm/debug
mm/pagecache
mm/swap
mm/memcg
mm/pagemap
mm/mprotect
mm/mremap
mm/page-reporting
mm/vmalloc
mm/kasan
mm/pagealloc
mm/memory-failure
mm/hugetlb
mm/vmscan
mm/z3fold
mm/compaction
mm/mempolicy
mm/oom-kill
mm/hugetlbfs
mm/migration
Subsystem: hexagon
Randy Dunlap <rdunlap@infradead.org>:
hexagon: remove CONFIG_EXPERIMENTAL from defconfigs
Subsystem: scripts
tangchunyou <tangchunyou@yulong.com>:
scripts/spelling.txt: increase error-prone spell checking
zuoqilin <zuoqilin@yulong.com>:
scripts/spelling.txt: check for "exeeds"
dingsenjie <dingsenjie@yulong.com>:
scripts/spelling.txt: add "allocted" and "exeeds" typo
Colin Ian King <colin.king@canonical.com>:
scripts/spelling.txt: add more spellings to spelling.txt
Subsystem: ntfs
Randy Dunlap <rdunlap@infradead.org>:
ntfs: layout.h: delete duplicated words
Rustam Kovhaev <rkovhaev@gmail.com>:
ntfs: check for valid standard information attribute
Subsystem: ocfs2
Yi Li <yili@winhong.com>:
ocfs2: remove redundant conditional before iput
guozh <guozh88@chinatelecom.cn>:
ocfs2: clean up some definitions which are not used any more
Dan Carpenter <dan.carpenter@oracle.com>:
ocfs2: fix a use after free on error
Jiapeng Chong <jiapeng.chong@linux.alibaba.com>:
ocfs2: simplify the calculation of variables
Subsystem: vfs
Randy Dunlap <rdunlap@infradead.org>:
fs: delete repeated words in comments
Alexey Dobriyan <adobriyan@gmail.com>:
ramfs: support O_TMPFILE
Subsystem: mm/slab-generic
Jacob Wen <jian.w.wen@oracle.com>:
mm, tracing: record slab name for kmem_cache_free()
Nikolay Borisov <nborisov@suse.com>:
mm/sl?b.c: remove ctor argument from kmem_cache_flags
Subsystem: mm/slab
Zhiyuan Dai <daizhiyuan@phytium.com.cn>:
mm/slab: minor coding style tweaks
Subsystem: mm/slub
Johannes Berg <johannes.berg@intel.com>:
mm/slub: disable user tracing for kmemleak caches by default
Vlastimil Babka <vbabka@suse.cz>:
Patch series "mm, slab, slub: remove cpu and memory hotplug locks":
mm, slub: stop freeing kmem_cache_node structures on node offline
mm, slab, slub: stop taking memory hotplug lock
mm, slab, slub: stop taking cpu hotplug lock
mm, slub: splice cpu and page freelists in deactivate_slab()
mm, slub: remove slub_memcg_sysfs boot param and CONFIG_SLUB_MEMCG_SYSFS_ON
Zhiyuan Dai <daizhiyuan@phytium.com.cn>:
mm/slub: minor coding style tweaks
Subsystem: mm/debug
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
mm/debug: improve memcg debugging
Anshuman Khandual <anshuman.khandual@arm.com>:
Patch series "mm/debug_vm_pgtable: Some minor updates", v3:
mm/debug_vm_pgtable/basic: add validation for dirtiness after write protect
mm/debug_vm_pgtable/basic: iterate over entire protection_map[]
Miaohe Lin <linmiaohe@huawei.com>:
mm/page_owner: use helper function zone_end_pfn() to get end_pfn
Subsystem: mm/pagecache
Baolin Wang <baolin.wang@linux.alibaba.com>:
mm/filemap: remove unused parameter and change to void type for replace_page_cache_page()
Pavel Begunkov <asml.silence@gmail.com>:
mm/filemap: don't revert iter on -EIOCBQUEUED
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
Patch series "Refactor generic_file_buffered_read", v5:
mm/filemap: rename generic_file_buffered_read subfunctions
mm/filemap: remove dynamically allocated array from filemap_read
mm/filemap: convert filemap_get_pages to take a pagevec
mm/filemap: use head pages in generic_file_buffered_read
mm/filemap: pass a sleep state to put_and_wait_on_page_locked
mm/filemap: support readpage splitting a page
mm/filemap: inline __wait_on_page_locked_async into caller
mm/filemap: don't call ->readpage if IOCB_WAITQ is set
mm/filemap: change filemap_read_page calling conventions
mm/filemap: change filemap_create_page calling conventions
mm/filemap: convert filemap_update_page to return an errno
mm/filemap: move the iocb checks into filemap_update_page
mm/filemap: add filemap_range_uptodate
mm/filemap: split filemap_readahead out of filemap_get_pages
mm/filemap: restructure filemap_get_pages
mm/filemap: don't relock the page after calling readpage
Christoph Hellwig <hch@lst.de>:
mm/filemap: rename generic_file_buffered_read to filemap_read
mm/filemap: simplify generic_file_read_iter
Yang Guo <guoyang2@huawei.com>:
fs/buffer.c: add checking buffer head stat before clear
Baolin Wang <baolin.wang@linux.alibaba.com>:
mm: backing-dev: Remove duplicated macro definition
Subsystem: mm/swap
Yang Li <abaci-bugfix@linux.alibaba.com>:
mm/swap_slots.c: remove redundant NULL check
Stephen Zhang <stephenzhangzsd@gmail.com>:
mm/swapfile.c: fix debugging information problem
Georgi Djakov <georgi.djakov@linaro.org>:
mm/page_io: use pr_alert_ratelimited for swap read/write errors
Rikard Falkeborn <rikard.falkeborn@gmail.com>:
mm/swap_state: constify static struct attribute_group
Yu Zhao <yuzhao@google.com>:
mm/swap: don't SetPageWorkingset unconditionally during swapin
Subsystem: mm/memcg
Roman Gushchin <guro@fb.com>:
mm: memcg/slab: pre-allocate obj_cgroups for slab caches with SLAB_ACCOUNT
Muchun Song <songmuchun@bytedance.com>:
mm: memcontrol: optimize per-lruvec stats counter memory usage
Patch series "Convert all THP vmstat counters to pages", v6:
mm: memcontrol: fix NR_ANON_THPS accounting in charge moving
mm: memcontrol: convert NR_ANON_THPS account to pages
mm: memcontrol: convert NR_FILE_THPS account to pages
mm: memcontrol: convert NR_SHMEM_THPS account to pages
mm: memcontrol: convert NR_SHMEM_PMDMAPPED account to pages
mm: memcontrol: convert NR_FILE_PMDMAPPED account to pages
mm: memcontrol: make the slab calculation consistent
Alex Shi <alex.shi@linux.alibaba.com>:
mm/memcg: revise the using condition of lock_page_lruvec function series
mm/memcg: remove rcu locking for lock_page_lruvec function series
Shakeel Butt <shakeelb@google.com>:
mm: memcg: add swapcache stat for memcg v2
Roman Gushchin <guro@fb.com>:
mm: kmem: make __memcg_kmem_(un)charge static
Feng Tang <feng.tang@intel.com>:
mm: page_counter: re-layout structure to reduce false sharing
Yang Li <abaci-bugfix@linux.alibaba.com>:
mm/memcontrol: remove redundant NULL check
Muchun Song <songmuchun@bytedance.com>:
mm: memcontrol: replace the loop with a list_for_each_entry()
Shakeel Butt <shakeelb@google.com>:
mm/list_lru.c: remove kvfree_rcu_local()
Johannes Weiner <hannes@cmpxchg.org>:
fs: buffer: use raw page_memcg() on locked page
Muchun Song <songmuchun@bytedance.com>:
mm: memcontrol: fix swap undercounting in cgroup2
mm: memcontrol: fix get_active_memcg return value
mm: memcontrol: fix slub memory accounting
Subsystem: mm/pagemap
Adrian Huang <ahuang12@lenovo.com>:
mm/mmap.c: remove unnecessary local variable
Miaohe Lin <linmiaohe@huawei.com>:
mm/memory.c: fix potential pte_unmap_unlock pte error
mm/pgtable-generic.c: simplify the VM_BUG_ON condition in pmdp_huge_clear_flush()
mm/pgtable-generic.c: optimize the VM_BUG_ON condition in pmdp_huge_clear_flush()
mm/memory.c: fix potential pte_unmap_unlock pte error
Subsystem: mm/mprotect
Tianjia Zhang <tianjia.zhang@linux.alibaba.com>:
mm/mprotect.c: optimize error detection in do_mprotect_pkey()
Subsystem: mm/mremap
Li Xinhai <lixinhai.lxh@gmail.com>:
mm: rmap: explicitly reset vma->anon_vma in unlink_anon_vmas()
mm: mremap: unlink anon_vmas when mremap with MREMAP_DONTUNMAP success
Subsystem: mm/page-reporting
sh <sh_def@163.com>:
mm/page_reporting: use list_entry_is_head() in page_reporting_cycle()
Subsystem: mm/vmalloc
Yang Li <abaci-bugfix@linux.alibaba.com>:
vmalloc: remove redundant NULL check
Subsystem: mm/kasan
Andrey Konovalov <andreyknvl@google.com>:
Patch series "kasan: HW_TAGS tests support and fixes", v4:
kasan: prefix global functions with kasan_
kasan: clarify HW_TAGS impact on TBI
kasan: clean up comments in tests
kasan: add macros to simplify checking test constraints
kasan: add match-all tag tests
kasan, arm64: allow using KUnit tests with HW_TAGS mode
kasan: rename CONFIG_TEST_KASAN_MODULE
kasan: add compiler barriers to KUNIT_EXPECT_KASAN_FAIL
kasan: adapt kmalloc_uaf2 test to HW_TAGS mode
kasan: fix memory corruption in kasan_bitops_tags test
kasan: move _RET_IP_ to inline wrappers
kasan: fix bug detection via ksize for HW_TAGS mode
kasan: add proper page allocator tests
kasan: add a test for kmem_cache_alloc/free_bulk
kasan: don't run tests when KASAN is not enabled
Walter Wu <walter-zh.wu@mediatek.com>:
kasan: remove redundant config option
Subsystem: mm/pagealloc
Baoquan He <bhe@redhat.com>:
Patch series "mm: clean up names and parameters of memmap_init_xxxx functions", v5:
mm: fix prototype warning from kernel test robot
mm: rename memmap_init() and memmap_init_zone()
mm: simplify parater of function memmap_init_zone()
mm: simplify parameter of setup_usemap()
mm: remove unneeded local variable in free_area_init_core
David Hildenbrand <david@redhat.com>:
Patch series "mm: simplify free_highmem_page() and free_reserved_page()":
video: fbdev: acornfb: remove free_unused_pages()
mm: simplify free_highmem_page() and free_reserved_page()
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
mm/gfp: add kernel-doc for gfp_t
Subsystem: mm/memory-failure
Aili Yao <yaoaili@kingsoft.com>:
mm,hwpoison: send SIGBUS to PF_MCE_EARLY processes on action required events
Subsystem: mm/hugetlb
Bibo Mao <maobibo@loongson.cn>:
mm/huge_memory.c: update tlb entry if pmd is changed
MIPS: do not call flush_tlb_all when setting pmd entry
Miaohe Lin <linmiaohe@huawei.com>:
mm/hugetlb: fix potential double free in hugetlb_register_node() error path
Li Xinhai <lixinhai.lxh@gmail.com>:
mm/hugetlb.c: fix unnecessary address expansion of pmd sharing
Miaohe Lin <linmiaohe@huawei.com>:
mm/hugetlb: avoid unnecessary hugetlb_acct_memory() call
mm/hugetlb: use helper huge_page_order and pages_per_huge_page
mm/hugetlb: fix use after free when subpool max_hpages accounting is not enabled
Jiapeng Zhong <abaci-bugfix@linux.alibaba.com>:
mm/hugetlb: simplify the calculation of variables
Joao Martins <joao.m.martins@oracle.com>:
Patch series "mm/hugetlb: follow_hugetlb_page() improvements", v2:
mm/hugetlb: grab head page refcount once for group of subpages
mm/hugetlb: refactor subpage recording
Miaohe Lin <linmiaohe@huawei.com>:
mm/hugetlb: fix some comment typos
Yanfei Xu <yanfei.xu@windriver.com>:
mm/hugetlb: remove redundant check in preparing and destroying gigantic page
Zhiyuan Dai <daizhiyuan@phytium.com.cn>:
mm/hugetlb.c: fix typos in comments
Miaohe Lin <linmiaohe@huawei.com>:
mm/huge_memory.c: remove unused return value of set_huge_zero_page()
"Aneesh Kumar K.V" <aneesh.kumar@linux.ibm.com>:
mm/pmem: avoid inserting hugepage PTE entry with fsdax if hugepage support is disabled
Miaohe Lin <linmiaohe@huawei.com>:
hugetlb_cgroup: use helper pages_per_huge_page() in hugetlb_cgroup
mm/hugetlb: use helper function range_in_vma() in page_table_shareable()
mm/hugetlb: remove unnecessary VM_BUG_ON_PAGE on putback_active_hugepage()
mm/hugetlb: use helper huge_page_size() to get hugepage size
Mike Kravetz <mike.kravetz@oracle.com>:
hugetlb: fix update_and_free_page contig page struct assumption
hugetlb: fix copy_huge_page_from_user contig page struct assumption
Chen Wandun <chenwandun@huawei.com>:
mm/hugetlb: suppress wrong warning info when alloc gigantic page
Subsystem: mm/vmscan
Alex Shi <alex.shi@linux.alibaba.com>:
mm/vmscan: __isolate_lru_page_prepare() cleanup
Miaohe Lin <linmiaohe@huawei.com>:
mm/workingset.c: avoid unnecessary max_nodes estimation in count_shadow_nodes()
Yu Zhao <yuzhao@google.com>:
Patch series "mm: lru related cleanups", v2:
mm/vmscan.c: use add_page_to_lru_list()
include/linux/mm_inline.h: shuffle lru list addition and deletion functions
mm: don't pass "enum lru_list" to lru list addition functions
mm/swap.c: don't pass "enum lru_list" to trace_mm_lru_insertion()
mm/swap.c: don't pass "enum lru_list" to del_page_from_lru_list()
mm: add __clear_page_lru_flags() to replace page_off_lru()
mm: VM_BUG_ON lru page flags
include/linux/mm_inline.h: fold page_lru_base_type() into its sole caller
include/linux/mm_inline.h: fold __update_lru_size() into its sole caller
mm/vmscan.c: make lruvec_lru_size() static
Oscar Salvador <osalvador@suse.de>:
mm: workingset: clarify eviction order and distance calculation
Mike Kravetz <mike.kravetz@oracle.com>:
Patch series "create hugetlb flags to consolidate state", v3:
hugetlb: use page.private for hugetlb specific page flags
hugetlb: convert page_huge_active() HPageMigratable flag
hugetlb: convert PageHugeTemporary() to HPageTemporary flag
hugetlb: convert PageHugeFreed to HPageFreed flag
include/linux/hugetlb.h: add synchronization information for new hugetlb specific flags
hugetlb: fix uninitialized subpool pointer
Dave Hansen <dave.hansen@linux.intel.com>:
mm/vmscan: restore zone_reclaim_mode ABI
Subsystem: mm/z3fold
Miaohe Lin <linmiaohe@huawei.com>:
z3fold: remove unused attribute for release_z3fold_page
z3fold: simplify the zhdr initialization code in init_z3fold_page()
Subsystem: mm/compaction
Alex Shi <alex.shi@linux.alibaba.com>:
mm/compaction: remove rcu_read_lock during page compaction
Miaohe Lin <linmiaohe@huawei.com>:
mm/compaction: remove duplicated VM_BUG_ON_PAGE !PageLocked
Charan Teja Reddy <charante@codeaurora.org>:
mm/compaction: correct deferral logic for proactive compaction
Wonhyuk Yang <vvghjk1234@gmail.com>:
mm/compaction: fix misbehaviors of fast_find_migrateblock()
Vlastimil Babka <vbabka@suse.cz>:
mm, compaction: make fast_isolate_freepages() stay within zone
Subsystem: mm/mempolicy
Huang Ying <ying.huang@intel.com>:
numa balancing: migrate on fault among multiple bound nodes
Miaohe Lin <linmiaohe@huawei.com>:
mm/mempolicy: use helper range_in_vma() in queue_pages_test_walk()
Subsystem: mm/oom-kill
Tang Yizhou <tangyizhou@huawei.com>:
mm, oom: fix a comment in dump_task()
Subsystem: mm/hugetlbfs
Mike Kravetz <mike.kravetz@oracle.com>:
mm/hugetlb: change hugetlb_reserve_pages() to type bool
hugetlbfs: remove special hugetlbfs_set_page_dirty()
Miaohe Lin <linmiaohe@huawei.com>:
hugetlbfs: remove useless BUG_ON(!inode) in hugetlbfs_setattr()
hugetlbfs: use helper macro default_hstate in init_hugetlbfs_fs
hugetlbfs: correct obsolete function name in hugetlbfs_read_iter()
hugetlbfs: remove meaningless variable avoid_reserve
hugetlbfs: make hugepage size conversion more readable
hugetlbfs: correct some obsolete comments about inode i_mutex
hugetlbfs: fix some comment typos
hugetlbfs: remove unneeded return value of hugetlb_vmtruncate()
Subsystem: mm/migration
Chengyang Fan <cy.fan@huawei.com>:
mm/migrate: remove unneeded semicolons
Documentation/admin-guide/cgroup-v2.rst | 4
Documentation/admin-guide/kernel-parameters.txt | 8
Documentation/admin-guide/sysctl/vm.rst | 10
Documentation/core-api/mm-api.rst | 7
Documentation/dev-tools/kasan.rst | 24
Documentation/vm/arch_pgtable_helpers.rst | 8
arch/arm64/include/asm/memory.h | 1
arch/arm64/include/asm/mte-kasan.h | 12
arch/arm64/kernel/mte.c | 12
arch/arm64/kernel/sleep.S | 2
arch/arm64/mm/fault.c | 20
arch/hexagon/configs/comet_defconfig | 1
arch/ia64/include/asm/pgtable.h | 6
arch/ia64/mm/init.c | 18
arch/mips/mm/pgtable-32.c | 1
arch/mips/mm/pgtable-64.c | 1
arch/x86/kernel/acpi/wakeup_64.S | 2
drivers/base/node.c | 33
drivers/video/fbdev/acornfb.c | 34
fs/block_dev.c | 2
fs/btrfs/file.c | 2
fs/buffer.c | 7
fs/dcache.c | 4
fs/direct-io.c | 4
fs/exec.c | 4
fs/fhandle.c | 2
fs/fuse/dev.c | 6
fs/hugetlbfs/inode.c | 72 --
fs/ntfs/inode.c | 6
fs/ntfs/layout.h | 4
fs/ocfs2/cluster/heartbeat.c | 8
fs/ocfs2/dlm/dlmast.c | 10
fs/ocfs2/dlm/dlmcommon.h | 4
fs/ocfs2/refcounttree.c | 2
fs/ocfs2/super.c | 2
fs/pipe.c | 2
fs/proc/meminfo.c | 10
fs/proc/vmcore.c | 7
fs/ramfs/inode.c | 13
include/linux/fs.h | 4
include/linux/gfp.h | 14
include/linux/highmem-internal.h | 5
include/linux/huge_mm.h | 15
include/linux/hugetlb.h | 98 ++
include/linux/kasan-checks.h | 6
include/linux/kasan.h | 39 -
include/linux/memcontrol.h | 43 -
include/linux/migrate.h | 2
include/linux/mm.h | 28
include/linux/mm_inline.h | 123 +--
include/linux/mmzone.h | 30
include/linux/page-flags.h | 6
include/linux/page_counter.h | 9
include/linux/pagemap.h | 5
include/linux/swap.h | 8
include/trace/events/kmem.h | 24
include/trace/events/pagemap.h | 11
include/uapi/linux/mempolicy.h | 4
init/Kconfig | 14
lib/Kconfig.kasan | 14
lib/Makefile | 2
lib/test_kasan.c | 446 ++++++++----
lib/test_kasan_module.c | 5
mm/backing-dev.c | 6
mm/compaction.c | 73 +-
mm/debug.c | 10
mm/debug_vm_pgtable.c | 86 ++
mm/filemap.c | 859 +++++++++++-------------
mm/gup.c | 5
mm/huge_memory.c | 28
mm/hugetlb.c | 376 ++++------
mm/hugetlb_cgroup.c | 6
mm/kasan/common.c | 60 -
mm/kasan/generic.c | 40 -
mm/kasan/hw_tags.c | 16
mm/kasan/kasan.h | 87 +-
mm/kasan/quarantine.c | 22
mm/kasan/report.c | 15
mm/kasan/report_generic.c | 10
mm/kasan/report_hw_tags.c | 8
mm/kasan/report_sw_tags.c | 8
mm/kasan/shadow.c | 27
mm/kasan/sw_tags.c | 22
mm/khugepaged.c | 6
mm/list_lru.c | 12
mm/memcontrol.c | 309 ++++----
mm/memory-failure.c | 34
mm/memory.c | 24
mm/memory_hotplug.c | 11
mm/mempolicy.c | 18
mm/mempool.c | 2
mm/migrate.c | 10
mm/mlock.c | 3
mm/mmap.c | 4
mm/mprotect.c | 7
mm/mremap.c | 8
mm/oom_kill.c | 5
mm/page_alloc.c | 70 -
mm/page_io.c | 12
mm/page_owner.c | 4
mm/page_reporting.c | 2
mm/pgtable-generic.c | 9
mm/rmap.c | 35
mm/shmem.c | 2
mm/slab.c | 21
mm/slab.h | 20
mm/slab_common.c | 40 -
mm/slob.c | 2
mm/slub.c | 169 ++--
mm/swap.c | 54 -
mm/swap_slots.c | 3
mm/swap_state.c | 31
mm/swapfile.c | 8
mm/vmscan.c | 100 +-
mm/vmstat.c | 14
mm/workingset.c | 7
mm/z3fold.c | 11
scripts/Makefile.kasan | 10
scripts/spelling.txt | 30
tools/objtool/check.c | 2
120 files changed, 2249 insertions(+), 1954 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-02-13 4:52 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-02-13 4:52 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
6 patches, based on dcc0b49040c70ad827a7f3d58a21b01fdb14e749.
Subsystems affected by this patch series:
mm/pagemap
scripts
MAINTAINERS
h8300
Subsystem: mm/pagemap
Mike Rapoport <rppt@linux.ibm.com>:
m68k: make __pfn_to_phys() and __phys_to_pfn() available for !MMU
Subsystem: scripts
Rong Chen <rong.a.chen@intel.com>:
scripts/recordmcount.pl: support big endian for ARCH sh
Subsystem: MAINTAINERS
Andrey Konovalov <andreyknvl@google.com>:
MAINTAINERS: update KASAN file list
MAINTAINERS: update Andrey Konovalov's email address
MAINTAINERS: add Andrey Konovalov to KASAN reviewers
Subsystem: h8300
Randy Dunlap <rdunlap@infradead.org>:
h8300: fix PREEMPTION build, TI_PRE_COUNT undefined
MAINTAINERS | 8 +++++---
arch/h8300/kernel/asm-offsets.c | 3 +++
arch/m68k/include/asm/page.h | 2 +-
scripts/recordmcount.pl | 6 +++++-
4 files changed, 14 insertions(+), 5 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2021-02-09 21:41 incoming Andrew Morton
@ 2021-02-10 19:30 ` Linus Torvalds
0 siblings, 0 replies; 409+ messages in thread
From: Linus Torvalds @ 2021-02-10 19:30 UTC (permalink / raw)
To: Andrew Morton; +Cc: Linux-MM, mm-commits
Hah. This series shows a small deficiency in your scripting wrt the diffstat:
On Tue, Feb 9, 2021 at 1:41 PM Andrew Morton <akpm@linux-foundation.org> wrote:
>
> .mailmap | 1
...
> mm/slub.c | 18 +++++++++-
> 17 files changed, 172 insertions(+), 49 deletions(-)
It actually has 18 files changed, but one of them is a pure rename (no
change to the content), and apparently your diffstat tool can't handle
that case.
It *should* have ended with
...
mm/slub.c | 18 +++++-
.../selftests/vm/{run_vmtests => run_vmtests.sh} | 0
18 files changed, 172 insertions(+), 49 deletions(-)
rename tools/testing/selftests/vm/{run_vmtests => run_vmtests.sh} (100%)
if you'd done a proper "git diff -M --stat --summary" of the series.
[ Ok, by default git would actually have said
18 files changed, 171 insertions(+), 48 deletions(-)
but it looks like you use the patience diff option, which gives that
extra insertion/deletion line because it generates the diff a bit
differently ]
Not a big deal,, but it made me briefly wonder "why doesn't my
diffstat match yours".
Linus
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-02-09 21:41 Andrew Morton
2021-02-10 19:30 ` incoming Linus Torvalds
0 siblings, 1 reply; 409+ messages in thread
From: Andrew Morton @ 2021-02-09 21:41 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
14 patches, based on e0756cfc7d7cd08c98a53b6009c091a3f6a50be6.
Subsystems affected by this patch series:
squashfs
mm/kasan
firmware
mm/mremap
mm/tmpfs
mm/selftests
MAINTAINERS
mm/memcg
mm/slub
nilfs2
Subsystem: squashfs
Phillip Lougher <phillip@squashfs.org.uk>:
Patch series "Squashfs: fix BIO migration regression and add sanity checks":
squashfs: avoid out of bounds writes in decompressors
squashfs: add more sanity checks in id lookup
squashfs: add more sanity checks in inode lookup
squashfs: add more sanity checks in xattr id lookup
Subsystem: mm/kasan
Andrey Konovalov <andreyknvl@google.com>:
kasan: fix stack traces dependency for HW_TAGS
Subsystem: firmware
Fangrui Song <maskray@google.com>:
firmware_loader: align .builtin_fw to 8
Subsystem: mm/mremap
Arnd Bergmann <arnd@arndb.de>:
mm/mremap: fix BUILD_BUG_ON() error in get_extent
Subsystem: mm/tmpfs
Seth Forshee <seth.forshee@canonical.com>:
tmpfs: disallow CONFIG_TMPFS_INODE64 on s390
tmpfs: disallow CONFIG_TMPFS_INODE64 on alpha
Subsystem: mm/selftests
Rong Chen <rong.a.chen@intel.com>:
selftests/vm: rename file run_vmtests to run_vmtests.sh
Subsystem: MAINTAINERS
Andrey Ryabinin <ryabinin.a.a@gmail.com>:
MAINTAINERS: update Andrey Ryabinin's email address
Subsystem: mm/memcg
Johannes Weiner <hannes@cmpxchg.org>:
Revert "mm: memcontrol: avoid workload stalls when lowering memory.high"
Subsystem: mm/slub
Vlastimil Babka <vbabka@suse.cz>:
mm, slub: better heuristic for number of cpus when calculating slab order
Subsystem: nilfs2
Joachim Henke <joachim.henke@t-systems.com>:
nilfs2: make splice write available again
.mailmap | 1
Documentation/dev-tools/kasan.rst | 3 -
MAINTAINERS | 2 -
fs/Kconfig | 4 +-
fs/nilfs2/file.c | 1
fs/squashfs/block.c | 8 ++++
fs/squashfs/export.c | 41 +++++++++++++++++++----
fs/squashfs/id.c | 40 ++++++++++++++++++-----
fs/squashfs/squashfs_fs_sb.h | 1
fs/squashfs/super.c | 6 +--
fs/squashfs/xattr.h | 10 +++++
fs/squashfs/xattr_id.c | 66 ++++++++++++++++++++++++++++++++------
include/asm-generic/vmlinux.lds.h | 2 -
mm/kasan/hw_tags.c | 8 +---
mm/memcontrol.c | 5 +-
mm/mremap.c | 5 +-
mm/slub.c | 18 +++++++++-
17 files changed, 172 insertions(+), 49 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-02-05 2:31 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-02-05 2:31 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
18 patches, based on 5c279c4cf206e03995e04fd3404fa95ffd243a97.
Subsystems affected by this patch series:
mm/hugetlb
mm/compaction
mm/vmalloc
gcov
mm/shmem
mm/memblock
mailmap
mm/pagecache
mm/kasan
ubsan
mm/hugetlb
MAINTAINERS
Subsystem: mm/hugetlb
Muchun Song <songmuchun@bytedance.com>:
mm: hugetlbfs: fix cannot migrate the fallocated HugeTLB page
mm: hugetlb: fix a race between freeing and dissolving the page
mm: hugetlb: fix a race between isolating and freeing page
mm: hugetlb: remove VM_BUG_ON_PAGE from page_huge_active
mm: migrate: do not migrate HugeTLB page whose refcount is one
Subsystem: mm/compaction
Rokudo Yan <wu-yan@tcl.com>:
mm, compaction: move high_pfn to the for loop scope
Subsystem: mm/vmalloc
Rick Edgecombe <rick.p.edgecombe@intel.com>:
mm/vmalloc: separate put pages and flush VM flags
Subsystem: gcov
Johannes Berg <johannes.berg@intel.com>:
init/gcov: allow CONFIG_CONSTRUCTORS on UML to fix module gcov
Subsystem: mm/shmem
Hugh Dickins <hughd@google.com>:
mm: thp: fix MADV_REMOVE deadlock on shmem THP
Subsystem: mm/memblock
Roman Gushchin <guro@fb.com>:
memblock: do not start bottom-up allocations with kernel_end
Subsystem: mailmap
Viresh Kumar <viresh.kumar@linaro.org>:
mailmap: fix name/email for Viresh Kumar
Manivannan Sadhasivam <manivannan.sadhasivam@linaro.org>:
mailmap: add entries for Manivannan Sadhasivam
Subsystem: mm/pagecache
Waiman Long <longman@redhat.com>:
mm/filemap: add missing mem_cgroup_uncharge() to __add_to_page_cache_locked()
Subsystem: mm/kasan
Vincenzo Frascino <vincenzo.frascino@arm.com>:
Patch series "kasan: Fix metadata detection for KASAN_HW_TAGS", v5:
kasan: add explicit preconditions to kasan_report()
kasan: make addr_has_metadata() return true for valid addresses
Subsystem: ubsan
Nathan Chancellor <nathan@kernel.org>:
ubsan: implement __ubsan_handle_alignment_assumption
Subsystem: mm/hugetlb
Muchun Song <songmuchun@bytedance.com>:
mm: hugetlb: fix missing put_page in gather_surplus_pages()
Subsystem: MAINTAINERS
Nathan Chancellor <nathan@kernel.org>:
MAINTAINERS/.mailmap: use my @kernel.org address
.mailmap | 5 ++++
MAINTAINERS | 2 -
fs/hugetlbfs/inode.c | 3 +-
include/linux/hugetlb.h | 2 +
include/linux/kasan.h | 7 ++++++
include/linux/vmalloc.h | 9 +-------
init/Kconfig | 1
init/main.c | 8 ++++++-
kernel/gcov/Kconfig | 2 -
lib/ubsan.c | 31 ++++++++++++++++++++++++++++
lib/ubsan.h | 6 +++++
mm/compaction.c | 3 +-
mm/filemap.c | 4 +++
mm/huge_memory.c | 37 ++++++++++++++++++++-------------
mm/hugetlb.c | 53 ++++++++++++++++++++++++++++++++++++++++++------
mm/kasan/kasan.h | 2 -
mm/memblock.c | 49 +++++---------------------------------------
mm/migrate.c | 6 +++++
18 files changed, 153 insertions(+), 77 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-01-24 5:00 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2021-01-24 5:00 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
19 patches, based on e1ae4b0be15891faf46d390e9f3dc9bd71a8cae1.
Subsystems affected by this patch series:
mm/pagealloc
mm/memcg
mm/kasan
ubsan
mm/memory-failure
mm/highmem
proc
MAINTAINERS
Subsystem: mm/pagealloc
Mike Rapoport <rppt@linux.ibm.com>:
Patch series "mm: fix initialization of struct page for holes in memory layout", v3:
x86/setup: don't remove E820_TYPE_RAM for pfn 0
mm: fix initialization of struct page for holes in memory layout
Subsystem: mm/memcg
Roman Gushchin <guro@fb.com>:
mm: memcg/slab: optimize objcg stock draining
Shakeel Butt <shakeelb@google.com>:
mm: memcg: fix memcg file_dirty numa stat
mm: fix numa stats for thp migration
Johannes Weiner <hannes@cmpxchg.org>:
mm: memcontrol: prevent starvation when writing memory.high
Subsystem: mm/kasan
Lecopzer Chen <lecopzer@gmail.com>:
kasan: fix unaligned address is unhandled in kasan_remove_zero_shadow
kasan: fix incorrect arguments passing in kasan_add_zero_shadow
Andrey Konovalov <andreyknvl@google.com>:
kasan: fix HW_TAGS boot parameters
kasan, mm: fix conflicts with init_on_alloc/free
kasan, mm: fix resetting page_alloc tags for HW_TAGS
Subsystem: ubsan
Arnd Bergmann <arnd@arndb.de>:
ubsan: disable unsigned-overflow check for i386
Subsystem: mm/memory-failure
Dan Williams <dan.j.williams@intel.com>:
mm: fix page reference leak in soft_offline_page()
Subsystem: mm/highmem
Thomas Gleixner <tglx@linutronix.de>:
Patch series "mm/highmem: Fix fallout from generic kmap_local conversions":
sparc/mm/highmem: flush cache and TLB
mm/highmem: prepare for overriding set_pte_at()
mips/mm/highmem: use set_pte() for kmap_local()
powerpc/mm/highmem: use __set_pte_at() for kmap_local()
Subsystem: proc
Xiaoming Ni <nixiaoming@huawei.com>:
proc_sysctl: fix oops caused by incorrect command parameters
Subsystem: MAINTAINERS
Nathan Chancellor <natechancellor@gmail.com>:
MAINTAINERS: add a couple more files to the Clang/LLVM section
Documentation/dev-tools/kasan.rst | 27 ++---------
MAINTAINERS | 2
arch/mips/include/asm/highmem.h | 1
arch/powerpc/include/asm/highmem.h | 2
arch/sparc/include/asm/highmem.h | 9 ++-
arch/x86/kernel/setup.c | 20 +++-----
fs/proc/proc_sysctl.c | 7 ++-
lib/Kconfig.ubsan | 1
mm/highmem.c | 7 ++-
mm/kasan/hw_tags.c | 77 +++++++++++++--------------------
mm/kasan/init.c | 23 +++++----
mm/memcontrol.c | 11 +---
mm/memory-failure.c | 20 ++++++--
mm/migrate.c | 27 ++++++-----
mm/page_alloc.c | 86 ++++++++++++++++++++++---------------
mm/slub.c | 7 +--
16 files changed, 173 insertions(+), 154 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2021-01-12 23:48 incoming Andrew Morton
@ 2021-01-15 23:32 ` Linus Torvalds
0 siblings, 0 replies; 409+ messages in thread
From: Linus Torvalds @ 2021-01-15 23:32 UTC (permalink / raw)
To: Andrew Morton; +Cc: Linux-MM, mm-commits
On Tue, Jan 12, 2021 at 3:48 PM Andrew Morton <akpm@linux-foundation.org> wrote:
>
> 10 patches, based on e609571b5ffa3528bf85292de1ceaddac342bc1c.
Whee. I had completely dropped the ball on this - I had built my usual
"akpm" branch with the patches, but then had completely forgotten
about it after doing my basic build tests.
I tend to leave it for a while to see if people send belated ACK/NAK's
for the patches, but that "for a while" is typically "overnight", not
several days.
So if you ever notice that I haven't merged your patch submission, and
you haven't seen me comment on them, feel free to ping me to remind
me.
Because it might just have gotten lost in the shuffle for some random
reason. Admittedly it's rare - I think this is the first time I just
randomly noticed three days later that I'd never done the actual merge
of the patch-series).
Linus
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2021-01-12 23:48 Andrew Morton
2021-01-15 23:32 ` incoming Linus Torvalds
0 siblings, 1 reply; 409+ messages in thread
From: Andrew Morton @ 2021-01-12 23:48 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
10 patches, based on e609571b5ffa3528bf85292de1ceaddac342bc1c.
Subsystems affected by this patch series:
mm/slub
mm/pagealloc
mm/memcg
mm/kasan
mm/vmalloc
mm/migration
mm/hugetlb
MAINTAINERS
mm/memory-failure
mm/process_vm_access
Subsystem: mm/slub
Jann Horn <jannh@google.com>:
mm, slub: consider rest of partial list if acquire_slab() fails
Subsystem: mm/pagealloc
Hailong liu <liu.hailong6@zte.com.cn>:
mm/page_alloc: add a missing mm_page_alloc_zone_locked() tracepoint
Subsystem: mm/memcg
Hugh Dickins <hughd@google.com>:
mm/memcontrol: fix warning in mem_cgroup_page_lruvec()
Subsystem: mm/kasan
Hailong Liu <liu.hailong6@zte.com.cn>:
arm/kasan: fix the array size of kasan_early_shadow_pte[]
Subsystem: mm/vmalloc
Miaohe Lin <linmiaohe@huawei.com>:
mm/vmalloc.c: fix potential memory leak
Subsystem: mm/migration
Jan Stancek <jstancek@redhat.com>:
mm: migrate: initialize err in do_migrate_pages
Subsystem: mm/hugetlb
Miaohe Lin <linmiaohe@huawei.com>:
mm/hugetlb: fix potential missing huge page size info
Subsystem: MAINTAINERS
Vlastimil Babka <vbabka@suse.cz>:
MAINTAINERS: add Vlastimil as slab allocators maintainer
Subsystem: mm/memory-failure
Oscar Salvador <osalvador@suse.de>:
mm,hwpoison: fix printing of page flags
Subsystem: mm/process_vm_access
Andrew Morton <akpm@linux-foundation.org>:
mm/process_vm_access.c: include compat.h
MAINTAINERS | 1 +
include/linux/kasan.h | 6 +++++-
include/linux/memcontrol.h | 2 +-
mm/hugetlb.c | 2 +-
mm/kasan/init.c | 3 ++-
mm/memory-failure.c | 2 +-
mm/mempolicy.c | 2 +-
mm/page_alloc.c | 31 ++++++++++++++++---------------
mm/process_vm_access.c | 1 +
mm/slub.c | 2 +-
mm/vmalloc.c | 4 +++-
11 files changed, 33 insertions(+), 23 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-12-29 23:13 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-12-29 23:13 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
16 patches, based on dea8dcf2a9fa8cc540136a6cd885c3beece16ec3.
Subsystems affected by this patch series:
mm/selftests
mm/hugetlb
kbuild
checkpatch
mm/pagecache
mm/mremap
mm/kasan
misc
lib
mm/slub
Subsystem: mm/selftests
Harish <harish@linux.ibm.com>:
selftests/vm: fix building protection keys test
Subsystem: mm/hugetlb
Mike Kravetz <mike.kravetz@oracle.com>:
mm/hugetlb: fix deadlock in hugetlb_cow error path
Subsystem: kbuild
Masahiro Yamada <masahiroy@kernel.org>:
Revert "kbuild: avoid static_assert for genksyms"
Subsystem: checkpatch
Joe Perches <joe@perches.com>:
checkpatch: prefer strscpy to strlcpy
Subsystem: mm/pagecache
Souptick Joarder <jrdr.linux@gmail.com>:
mm: add prototype for __add_to_page_cache_locked()
Baoquan He <bhe@redhat.com>:
mm: memmap defer init doesn't work as expected
Subsystem: mm/mremap
Kalesh Singh <kaleshsingh@google.com>:
mm/mremap.c: fix extent calculation
Nicholas Piggin <npiggin@gmail.com>:
mm: generalise COW SMC TLB flushing race comment
Subsystem: mm/kasan
Walter Wu <walter-zh.wu@mediatek.com>:
kasan: fix null pointer dereference in kasan_record_aux_stack
Subsystem: misc
Randy Dunlap <rdunlap@infradead.org>:
local64.h: make <asm/local64.h> mandatory
Huang Shijie <sjhuang@iluvatar.ai>:
sizes.h: add SZ_8G/SZ_16G/SZ_32G macros
Josh Poimboeuf <jpoimboe@redhat.com>:
kdev_t: always inline major/minor helper functions
Subsystem: lib
Huang Shijie <sjhuang@iluvatar.ai>:
lib/genalloc: fix the overflow when size is too big
Ilya Leoshkevich <iii@linux.ibm.com>:
lib/zlib: fix inflating zlib streams on s390
Randy Dunlap <rdunlap@infradead.org>:
zlib: move EXPORT_SYMBOL() and MODULE_LICENSE() out of dfltcc_syms.c
Subsystem: mm/slub
Roman Gushchin <guro@fb.com>:
mm: slub: call account_slab_page() after slab page initialization
arch/alpha/include/asm/local64.h | 1 -
arch/arc/include/asm/Kbuild | 1 -
arch/arm/include/asm/Kbuild | 1 -
arch/arm64/include/asm/Kbuild | 1 -
arch/csky/include/asm/Kbuild | 1 -
arch/h8300/include/asm/Kbuild | 1 -
arch/hexagon/include/asm/Kbuild | 1 -
arch/ia64/include/asm/local64.h | 1 -
arch/ia64/mm/init.c | 4 ++--
arch/m68k/include/asm/Kbuild | 1 -
arch/microblaze/include/asm/Kbuild | 1 -
arch/mips/include/asm/Kbuild | 1 -
arch/nds32/include/asm/Kbuild | 1 -
arch/openrisc/include/asm/Kbuild | 1 -
arch/parisc/include/asm/Kbuild | 1 -
arch/powerpc/include/asm/Kbuild | 1 -
arch/riscv/include/asm/Kbuild | 1 -
arch/s390/include/asm/Kbuild | 1 -
arch/sh/include/asm/Kbuild | 1 -
arch/sparc/include/asm/Kbuild | 1 -
arch/x86/include/asm/local64.h | 1 -
arch/xtensa/include/asm/Kbuild | 1 -
include/asm-generic/Kbuild | 1 +
include/linux/build_bug.h | 5 -----
include/linux/kdev_t.h | 22 +++++++++++-----------
include/linux/mm.h | 12 ++++++++++--
include/linux/sizes.h | 3 +++
lib/genalloc.c | 25 +++++++++++++------------
lib/zlib_dfltcc/Makefile | 2 +-
lib/zlib_dfltcc/dfltcc.c | 6 +++++-
lib/zlib_dfltcc/dfltcc_deflate.c | 3 +++
lib/zlib_dfltcc/dfltcc_inflate.c | 4 ++--
lib/zlib_dfltcc/dfltcc_syms.c | 17 -----------------
mm/hugetlb.c | 22 +++++++++++++++++++++-
mm/kasan/generic.c | 2 ++
mm/memory.c | 8 +++++---
mm/memory_hotplug.c | 2 +-
mm/mremap.c | 4 +++-
mm/page_alloc.c | 8 +++++---
mm/slub.c | 5 ++---
scripts/checkpatch.pl | 6 ++++++
tools/testing/selftests/vm/Makefile | 10 +++++-----
42 files changed, 101 insertions(+), 91 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2020-12-22 19:58 incoming Andrew Morton
@ 2020-12-22 21:43 ` Linus Torvalds
0 siblings, 0 replies; 409+ messages in thread
From: Linus Torvalds @ 2020-12-22 21:43 UTC (permalink / raw)
To: Andrew Morton; +Cc: Linux-MM, mm-commits
On Tue, Dec 22, 2020 at 11:58 AM Andrew Morton
<akpm@linux-foundation.org> wrote:
>
> 60 patches, based on 8653b778e454a7708847aeafe689bce07aeeb94e.
I see that you enabled renaming in the patches. Lovely.
Can you also enable it in the diffstat?
> 74 files changed, 2869 insertions(+), 1553 deletions(-)
With -M in the diffstat, you should have seen
72 files changed, 2775 insertions(+), 1460 deletions(-)
and if you add "--summary", you'll also see the rename part ofthe file
create/delete summary:
rename mm/kasan/{tags_report.c => report_sw_tags.c} (78%)
which is often nice to see in addition to the line stats..
Linus
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-12-22 19:58 Andrew Morton
2020-12-22 21:43 ` incoming Linus Torvalds
0 siblings, 1 reply; 409+ messages in thread
From: Andrew Morton @ 2020-12-22 19:58 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
60 patches, based on 8653b778e454a7708847aeafe689bce07aeeb94e.
Subsystems affected by this patch series:
mm/kasan
Subsystem: mm/kasan
Andrey Konovalov <andreyknvl@google.com>:
Patch series "kasan: add hardware tag-based mode for arm64", v11:
kasan: drop unnecessary GPL text from comment headers
kasan: KASAN_VMALLOC depends on KASAN_GENERIC
kasan: group vmalloc code
kasan: shadow declarations only for software modes
kasan: rename (un)poison_shadow to (un)poison_range
kasan: rename KASAN_SHADOW_* to KASAN_GRANULE_*
kasan: only build init.c for software modes
kasan: split out shadow.c from common.c
kasan: define KASAN_MEMORY_PER_SHADOW_PAGE
kasan: rename report and tags files
kasan: don't duplicate config dependencies
kasan: hide invalid free check implementation
kasan: decode stack frame only with KASAN_STACK_ENABLE
kasan, arm64: only init shadow for software modes
kasan, arm64: only use kasan_depth for software modes
kasan, arm64: move initialization message
kasan, arm64: rename kasan_init_tags and mark as __init
kasan: rename addr_has_shadow to addr_has_metadata
kasan: rename print_shadow_for_address to print_memory_metadata
kasan: rename SHADOW layout macros to META
kasan: separate metadata_fetch_row for each mode
kasan: introduce CONFIG_KASAN_HW_TAGS
Vincenzo Frascino <vincenzo.frascino@arm.com>:
arm64: enable armv8.5-a asm-arch option
arm64: mte: add in-kernel MTE helpers
arm64: mte: reset the page tag in page->flags
arm64: mte: add in-kernel tag fault handler
arm64: kasan: allow enabling in-kernel MTE
arm64: mte: convert gcr_user into an exclude mask
arm64: mte: switch GCR_EL1 in kernel entry and exit
kasan, mm: untag page address in free_reserved_area
Andrey Konovalov <andreyknvl@google.com>:
arm64: kasan: align allocations for HW_TAGS
arm64: kasan: add arch layer for memory tagging helpers
kasan: define KASAN_GRANULE_SIZE for HW_TAGS
kasan, x86, s390: update undef CONFIG_KASAN
kasan, arm64: expand CONFIG_KASAN checks
kasan, arm64: implement HW_TAGS runtime
kasan, arm64: print report from tag fault handler
kasan, mm: reset tags when accessing metadata
kasan, arm64: enable CONFIG_KASAN_HW_TAGS
kasan: add documentation for hardware tag-based mode
Vincenzo Frascino <vincenzo.frascino@arm.com>:
kselftest/arm64: check GCR_EL1 after context switch
Andrey Konovalov <andreyknvl@google.com>:
Patch series "kasan: boot parameters for hardware tag-based mode", v4:
kasan: simplify quarantine_put call site
kasan: rename get_alloc/free_info
kasan: introduce set_alloc_info
kasan, arm64: unpoison stack only with CONFIG_KASAN_STACK
kasan: allow VMAP_STACK for HW_TAGS mode
kasan: remove __kasan_unpoison_stack
kasan: inline kasan_reset_tag for tag-based modes
kasan: inline random_tag for HW_TAGS
kasan: open-code kasan_unpoison_slab
kasan: inline (un)poison_range and check_invalid_free
kasan: add and integrate kasan boot parameters
kasan, mm: check kasan_enabled in annotations
kasan, mm: rename kasan_poison_kfree
kasan: don't round_up too much
kasan: simplify assign_tag and set_tag calls
kasan: clarify comment in __kasan_kfree_large
kasan: sanitize objects when metadata doesn't fit
kasan, mm: allow cache merging with no metadata
kasan: update documentation
Documentation/dev-tools/kasan.rst | 274 ++-
arch/Kconfig | 8
arch/arm64/Kconfig | 9
arch/arm64/Makefile | 7
arch/arm64/include/asm/assembler.h | 2
arch/arm64/include/asm/cache.h | 3
arch/arm64/include/asm/esr.h | 1
arch/arm64/include/asm/kasan.h | 17
arch/arm64/include/asm/memory.h | 15
arch/arm64/include/asm/mte-def.h | 16
arch/arm64/include/asm/mte-kasan.h | 67
arch/arm64/include/asm/mte.h | 22
arch/arm64/include/asm/processor.h | 2
arch/arm64/include/asm/string.h | 5
arch/arm64/include/asm/uaccess.h | 23
arch/arm64/kernel/asm-offsets.c | 3
arch/arm64/kernel/cpufeature.c | 3
arch/arm64/kernel/entry.S | 41
arch/arm64/kernel/head.S | 2
arch/arm64/kernel/hibernate.c | 5
arch/arm64/kernel/image-vars.h | 2
arch/arm64/kernel/kaslr.c | 3
arch/arm64/kernel/module.c | 6
arch/arm64/kernel/mte.c | 124 +
arch/arm64/kernel/setup.c | 2
arch/arm64/kernel/sleep.S | 2
arch/arm64/kernel/smp.c | 2
arch/arm64/lib/mte.S | 16
arch/arm64/mm/copypage.c | 9
arch/arm64/mm/fault.c | 59
arch/arm64/mm/kasan_init.c | 41
arch/arm64/mm/mteswap.c | 9
arch/arm64/mm/proc.S | 23
arch/arm64/mm/ptdump.c | 6
arch/s390/boot/string.c | 1
arch/x86/boot/compressed/misc.h | 1
arch/x86/kernel/acpi/wakeup_64.S | 2
include/linux/kasan-checks.h | 2
include/linux/kasan.h | 423 ++++-
include/linux/mm.h | 24
include/linux/moduleloader.h | 3
include/linux/page-flags-layout.h | 2
include/linux/sched.h | 2
include/linux/string.h | 2
init/init_task.c | 2
kernel/fork.c | 4
lib/Kconfig.kasan | 71
lib/test_kasan.c | 2
lib/test_kasan_module.c | 2
mm/kasan/Makefile | 33
mm/kasan/common.c | 1006 +++-----------
mm/kasan/generic.c | 72 -
mm/kasan/generic_report.c | 13
mm/kasan/hw_tags.c | 276 +++
mm/kasan/init.c | 25
mm/kasan/kasan.h | 195 ++
mm/kasan/quarantine.c | 35
mm/kasan/report.c | 363 +----
mm/kasan/report_generic.c | 169 ++
mm/kasan/report_hw_tags.c | 44
mm/kasan/report_sw_tags.c | 22
mm/kasan/shadow.c | 528 +++++++
mm/kasan/sw_tags.c | 34
mm/kasan/tags.c | 7
mm/kasan/tags_report.c | 7
mm/mempool.c | 4
mm/page_alloc.c | 9
mm/page_poison.c | 2
mm/ptdump.c | 13
mm/slab_common.c | 5
mm/slub.c | 29
scripts/Makefile.lib | 2
tools/testing/selftests/arm64/mte/Makefile | 2
tools/testing/selftests/arm64/mte/check_gcr_el1_cswitch.c | 155 ++
74 files changed, 2869 insertions(+), 1553 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-12-18 22:00 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-12-18 22:00 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
78 patches, based on a409ed156a90093a03fe6a93721ddf4c591eac87.
Subsystems affected by this patch series:
mm/memcg
epoll
mm/kasan
mm/cleanups
epoll
Subsystem: mm/memcg
Alex Shi <alex.shi@linux.alibaba.com>:
Patch series "bail out early for memcg disable":
mm/memcg: bail early from swap accounting if memcg disabled
mm/memcg: warning on !memcg after readahead page charged
Wei Yang <richard.weiyang@gmail.com>:
mm/memcg: remove unused definitions
Shakeel Butt <shakeelb@google.com>:
mm, kvm: account kvm_vcpu_mmap to kmemcg
Hui Su <sh_def@163.com>:
mm/memcontrol:rewrite mem_cgroup_page_lruvec()
Subsystem: epoll
Soheil Hassas Yeganeh <soheil@google.com>:
Patch series "simplify ep_poll":
epoll: check for events when removing a timed out thread from the wait queue
epoll: simplify signal handling
epoll: pull fatal signal checks into ep_send_events()
epoll: move eavail next to the list_empty_careful check
epoll: simplify and optimize busy loop logic
epoll: pull all code between fetch_events and send_event into the loop
epoll: replace gotos with a proper loop
epoll: eliminate unnecessary lock for zero timeout
Subsystem: mm/kasan
Andrey Konovalov <andreyknvl@google.com>:
Patch series "kasan: add hardware tag-based mode for arm64", v11:
kasan: drop unnecessary GPL text from comment headers
kasan: KASAN_VMALLOC depends on KASAN_GENERIC
kasan: group vmalloc code
kasan: shadow declarations only for software modes
kasan: rename (un)poison_shadow to (un)poison_range
kasan: rename KASAN_SHADOW_* to KASAN_GRANULE_*
kasan: only build init.c for software modes
kasan: split out shadow.c from common.c
kasan: define KASAN_MEMORY_PER_SHADOW_PAGE
kasan: rename report and tags files
kasan: don't duplicate config dependencies
kasan: hide invalid free check implementation
kasan: decode stack frame only with KASAN_STACK_ENABLE
kasan, arm64: only init shadow for software modes
kasan, arm64: only use kasan_depth for software modes
kasan, arm64: move initialization message
kasan, arm64: rename kasan_init_tags and mark as __init
kasan: rename addr_has_shadow to addr_has_metadata
kasan: rename print_shadow_for_address to print_memory_metadata
kasan: rename SHADOW layout macros to META
kasan: separate metadata_fetch_row for each mode
kasan: introduce CONFIG_KASAN_HW_TAGS
Vincenzo Frascino <vincenzo.frascino@arm.com>:
arm64: enable armv8.5-a asm-arch option
arm64: mte: add in-kernel MTE helpers
arm64: mte: reset the page tag in page->flags
arm64: mte: add in-kernel tag fault handler
arm64: kasan: allow enabling in-kernel MTE
arm64: mte: convert gcr_user into an exclude mask
arm64: mte: switch GCR_EL1 in kernel entry and exit
kasan, mm: untag page address in free_reserved_area
Andrey Konovalov <andreyknvl@google.com>:
arm64: kasan: align allocations for HW_TAGS
arm64: kasan: add arch layer for memory tagging helpers
kasan: define KASAN_GRANULE_SIZE for HW_TAGS
kasan, x86, s390: update undef CONFIG_KASAN
kasan, arm64: expand CONFIG_KASAN checks
kasan, arm64: implement HW_TAGS runtime
kasan, arm64: print report from tag fault handler
kasan, mm: reset tags when accessing metadata
kasan, arm64: enable CONFIG_KASAN_HW_TAGS
kasan: add documentation for hardware tag-based mode
Vincenzo Frascino <vincenzo.frascino@arm.com>:
kselftest/arm64: check GCR_EL1 after context switch
Andrey Konovalov <andreyknvl@google.com>:
Patch series "kasan: boot parameters for hardware tag-based mode", v4:
kasan: simplify quarantine_put call site
kasan: rename get_alloc/free_info
kasan: introduce set_alloc_info
kasan, arm64: unpoison stack only with CONFIG_KASAN_STACK
kasan: allow VMAP_STACK for HW_TAGS mode
kasan: remove __kasan_unpoison_stack
kasan: inline kasan_reset_tag for tag-based modes
kasan: inline random_tag for HW_TAGS
kasan: open-code kasan_unpoison_slab
kasan: inline (un)poison_range and check_invalid_free
kasan: add and integrate kasan boot parameters
kasan, mm: check kasan_enabled in annotations
kasan, mm: rename kasan_poison_kfree
kasan: don't round_up too much
kasan: simplify assign_tag and set_tag calls
kasan: clarify comment in __kasan_kfree_large
kasan: sanitize objects when metadata doesn't fit
kasan, mm: allow cache merging with no metadata
kasan: update documentation
Subsystem: mm/cleanups
Colin Ian King <colin.king@canonical.com>:
mm/Kconfig: fix spelling mistake "whats" -> "what's"
Subsystem: epoll
Willem de Bruijn <willemb@google.com>:
Patch series "add epoll_pwait2 syscall", v4:
epoll: convert internal api to timespec64
epoll: add syscall epoll_pwait2
epoll: wire up syscall epoll_pwait2
selftests/filesystems: expand epoll with epoll_pwait2
Documentation/dev-tools/kasan.rst | 274 +-
arch/Kconfig | 8
arch/alpha/kernel/syscalls/syscall.tbl | 1
arch/arm/tools/syscall.tbl | 1
arch/arm64/Kconfig | 9
arch/arm64/Makefile | 7
arch/arm64/include/asm/assembler.h | 2
arch/arm64/include/asm/cache.h | 3
arch/arm64/include/asm/esr.h | 1
arch/arm64/include/asm/kasan.h | 17
arch/arm64/include/asm/memory.h | 15
arch/arm64/include/asm/mte-def.h | 16
arch/arm64/include/asm/mte-kasan.h | 67
arch/arm64/include/asm/mte.h | 22
arch/arm64/include/asm/processor.h | 2
arch/arm64/include/asm/string.h | 5
arch/arm64/include/asm/uaccess.h | 23
arch/arm64/include/asm/unistd.h | 2
arch/arm64/include/asm/unistd32.h | 2
arch/arm64/kernel/asm-offsets.c | 3
arch/arm64/kernel/cpufeature.c | 3
arch/arm64/kernel/entry.S | 41
arch/arm64/kernel/head.S | 2
arch/arm64/kernel/hibernate.c | 5
arch/arm64/kernel/image-vars.h | 2
arch/arm64/kernel/kaslr.c | 3
arch/arm64/kernel/module.c | 6
arch/arm64/kernel/mte.c | 124 +
arch/arm64/kernel/setup.c | 2
arch/arm64/kernel/sleep.S | 2
arch/arm64/kernel/smp.c | 2
arch/arm64/lib/mte.S | 16
arch/arm64/mm/copypage.c | 9
arch/arm64/mm/fault.c | 59
arch/arm64/mm/kasan_init.c | 41
arch/arm64/mm/mteswap.c | 9
arch/arm64/mm/proc.S | 23
arch/arm64/mm/ptdump.c | 6
arch/ia64/kernel/syscalls/syscall.tbl | 1
arch/m68k/kernel/syscalls/syscall.tbl | 1
arch/microblaze/kernel/syscalls/syscall.tbl | 1
arch/mips/kernel/syscalls/syscall_n32.tbl | 1
arch/mips/kernel/syscalls/syscall_n64.tbl | 1
arch/mips/kernel/syscalls/syscall_o32.tbl | 1
arch/parisc/kernel/syscalls/syscall.tbl | 1
arch/powerpc/kernel/syscalls/syscall.tbl | 1
arch/s390/boot/string.c | 1
arch/s390/kernel/syscalls/syscall.tbl | 1
arch/sh/kernel/syscalls/syscall.tbl | 1
arch/sparc/kernel/syscalls/syscall.tbl | 1
arch/x86/boot/compressed/misc.h | 1
arch/x86/entry/syscalls/syscall_32.tbl | 1
arch/x86/entry/syscalls/syscall_64.tbl | 1
arch/x86/kernel/acpi/wakeup_64.S | 2
arch/x86/kvm/x86.c | 2
arch/xtensa/kernel/syscalls/syscall.tbl | 1
fs/eventpoll.c | 359 ++-
include/linux/compat.h | 6
include/linux/kasan-checks.h | 2
include/linux/kasan.h | 423 ++--
include/linux/memcontrol.h | 137 -
include/linux/mm.h | 24
include/linux/mmdebug.h | 13
include/linux/moduleloader.h | 3
include/linux/page-flags-layout.h | 2
include/linux/sched.h | 2
include/linux/string.h | 2
include/linux/syscalls.h | 5
include/uapi/asm-generic/unistd.h | 4
init/init_task.c | 2
kernel/fork.c | 4
kernel/sys_ni.c | 2
lib/Kconfig.kasan | 71
lib/test_kasan.c | 2
lib/test_kasan_module.c | 2
mm/Kconfig | 2
mm/kasan/Makefile | 33
mm/kasan/common.c | 1006 ++--------
mm/kasan/generic.c | 72
mm/kasan/generic_report.c | 13
mm/kasan/hw_tags.c | 294 ++
mm/kasan/init.c | 25
mm/kasan/kasan.h | 204 +-
mm/kasan/quarantine.c | 35
mm/kasan/report.c | 363 +--
mm/kasan/report_generic.c | 169 +
mm/kasan/report_hw_tags.c | 44
mm/kasan/report_sw_tags.c | 22
mm/kasan/shadow.c | 541 +++++
mm/kasan/sw_tags.c | 34
mm/kasan/tags.c | 7
mm/kasan/tags_report.c | 7
mm/memcontrol.c | 53
mm/mempool.c | 4
mm/page_alloc.c | 9
mm/page_poison.c | 2
mm/ptdump.c | 13
mm/slab_common.c | 5
mm/slub.c | 29
scripts/Makefile.lib | 2
tools/testing/selftests/arm64/mte/Makefile | 2
tools/testing/selftests/arm64/mte/check_gcr_el1_cswitch.c | 155 +
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c | 72
virt/kvm/coalesced_mmio.c | 2
virt/kvm/kvm_main.c | 2
105 files changed, 3268 insertions(+), 1873 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-12-16 4:41 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-12-16 4:41 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
- lots of little subsystems
- a few post-linux-next MM material. Most of this awaits more merging
of other trees.
95 patches, based on 489e9fea66f31086f85d9a18e61e4791d94a56a4.
Subsystems affected by this patch series:
mm/swap
mm/memory-hotplug
alpha
procfs
misc
core-kernel
bitmap
lib
lz4
bitops
checkpatch
nilfs
kdump
rapidio
gcov
bfs
relay
resource
ubsan
reboot
fault-injection
lzo
apparmor
mm/pagemap
mm/cleanups
mm/gup
Subsystem: mm/swap
Zhaoyang Huang <huangzhaoyang@gmail.com>:
mm: fix a race on nr_swap_pages
Subsystem: mm/memory-hotplug
Laurent Dufour <ldufour@linux.ibm.com>:
mm/memory_hotplug: quieting offline operation
Subsystem: alpha
Thomas Gleixner <tglx@linutronix.de>:
alpha: replace bogus in_interrupt()
Subsystem: procfs
Randy Dunlap <rdunlap@infradead.org>:
procfs: delete duplicated words + other fixes
Anand K Mistry <amistry@google.com>:
proc: provide details on indirect branch speculation
Alexey Dobriyan <adobriyan@gmail.com>:
proc: fix lookup in /proc/net subdirectories after setns(2)
Hui Su <sh_def@163.com>:
fs/proc: make pde_get() return nothing
Subsystem: misc
Christophe Leroy <christophe.leroy@csgroup.eu>:
asm-generic: force inlining of get_order() to work around gcc10 poor decision
Andy Shevchenko <andriy.shevchenko@linux.intel.com>:
kernel.h: split out mathematical helpers
Subsystem: core-kernel
Hui Su <sh_def@163.com>:
kernel/acct.c: use #elif instead of #end and #elif
Subsystem: bitmap
Andy Shevchenko <andriy.shevchenko@linux.intel.com>:
include/linux/bitmap.h: convert bitmap_empty() / bitmap_full() to return boolean
"Ma, Jianpeng" <jianpeng.ma@intel.com>:
bitmap: remove unused function declaration
Subsystem: lib
Geert Uytterhoeven <geert@linux-m68k.org>:
lib/test_free_pages.c: add basic progress indicators
"Gustavo A. R. Silva" <gustavoars@kernel.org>:
Patch series "] lib/stackdepot.c: Replace one-element array with flexible-array member":
lib/stackdepot.c: replace one-element array with flexible-array member
lib/stackdepot.c: use flex_array_size() helper in memcpy()
lib/stackdepot.c: use array_size() helper in jhash2()
Sebastian Andrzej Siewior <bigeasy@linutronix.de>:
lib/test_lockup.c: minimum fix to get it compiled on PREEMPT_RT
Andy Shevchenko <andriy.shevchenko@linux.intel.com>:
lib/list_kunit: follow new file name convention for KUnit tests
lib/linear_ranges_kunit: follow new file name convention for KUnit tests
lib/bits_kunit: follow new file name convention for KUnit tests
lib/cmdline: fix get_option() for strings starting with hyphen
lib/cmdline: allow NULL to be an output for get_option()
lib/cmdline_kunit: add a new test suite for cmdline API
Jakub Jelinek <jakub@redhat.com>:
ilog2: improve ilog2 for constant arguments
Nick Desaulniers <ndesaulniers@google.com>:
lib/string: remove unnecessary #undefs
Daniel Axtens <dja@axtens.net>:
Patch series "Fortify strscpy()", v7:
lib: string.h: detect intra-object overflow in fortified string functions
lkdtm: tests for FORTIFY_SOURCE
Francis Laniel <laniel_francis@privacyrequired.com>:
string.h: add FORTIFY coverage for strscpy()
drivers/misc/lkdtm: add new file in LKDTM to test fortified strscpy
drivers/misc/lkdtm/lkdtm.h: correct wrong filenames in comment
Alexey Dobriyan <adobriyan@gmail.com>:
lib: cleanup kstrto*() usage
Subsystem: lz4
Gao Xiang <hsiangkao@redhat.com>:
lib/lz4: explicitly support in-place decompression
Subsystem: bitops
Syed Nayyar Waris <syednwaris@gmail.com>:
Patch series "Introduce the for_each_set_clump macro", v12:
bitops: introduce the for_each_set_clump macro
lib/test_bitmap.c: add for_each_set_clump test cases
gpio: thunderx: utilize for_each_set_clump macro
gpio: xilinx: utilize generic bitmap_get_value and _set_value
Subsystem: checkpatch
Dwaipayan Ray <dwaipayanray1@gmail.com>:
checkpatch: add new exception to repeated word check
Aditya Srivastava <yashsri421@gmail.com>:
checkpatch: fix false positives in REPEATED_WORD warning
Łukasz Stelmach <l.stelmach@samsung.com>:
checkpatch: ignore generated CamelCase defines and enum values
Joe Perches <joe@perches.com>:
checkpatch: prefer static const declarations
checkpatch: allow --fix removal of unnecessary break statements
Dwaipayan Ray <dwaipayanray1@gmail.com>:
checkpatch: extend attributes check to handle more patterns
Tom Rix <trix@redhat.com>:
checkpatch: add a fixer for missing newline at eof
Joe Perches <joe@perches.com>:
checkpatch: update __attribute__((section("name"))) quote removal
Aditya Srivastava <yashsri421@gmail.com>:
checkpatch: add fix option for GERRIT_CHANGE_ID
Joe Perches <joe@perches.com>:
checkpatch: add __alias and __weak to suggested __attribute__ conversions
Dwaipayan Ray <dwaipayanray1@gmail.com>:
checkpatch: improve email parsing
checkpatch: fix spelling errors and remove repeated word
Aditya Srivastava <yashsri421@gmail.com>:
checkpatch: avoid COMMIT_LOG_LONG_LINE warning for signature tags
Dwaipayan Ray <dwaipayanray1@gmail.com>:
checkpatch: fix unescaped left brace
Aditya Srivastava <yashsri421@gmail.com>:
checkpatch: add fix option for ASSIGNMENT_CONTINUATIONS
checkpatch: add fix option for LOGICAL_CONTINUATIONS
checkpatch: add fix and improve warning msg for non-standard signature
Dwaipayan Ray <dwaipayanray1@gmail.com>:
checkpatch: add warning for unnecessary use of %h[xudi] and %hh[xudi]
checkpatch: add warning for lines starting with a '#' in commit log
checkpatch: fix TYPO_SPELLING check for words with apostrophe
Joe Perches <joe@perches.com>:
checkpatch: add printk_once and printk_ratelimit to prefer pr_<level> warning
Subsystem: nilfs
Alex Shi <alex.shi@linux.alibaba.com>:
fs/nilfs2: remove some unused macros to tame gcc
Subsystem: kdump
Alexander Egorenkov <egorenar@linux.ibm.com>:
kdump: append uts_namespace.name offset to VMCOREINFO
Subsystem: rapidio
Sebastian Andrzej Siewior <bigeasy@linutronix.de>:
rapidio: remove unused rio_get_asm() and rio_get_device()
Subsystem: gcov
Nick Desaulniers <ndesaulniers@google.com>:
gcov: remove support for GCC < 4.9
Alex Shi <alex.shi@linux.alibaba.com>:
gcov: fix kernel-doc markup issue
Subsystem: bfs
Randy Dunlap <rdunlap@infradead.org>:
bfs: don't use WARNING: string when it's just info.
Subsystem: relay
Jani Nikula <jani.nikula@intel.com>:
Patch series "relay: cleanup and const callbacks", v2:
relay: remove unused buf_mapped and buf_unmapped callbacks
relay: require non-NULL callbacks in relay_open()
relay: make create_buf_file and remove_buf_file callbacks mandatory
relay: allow the use of const callback structs
drm/i915: make relay callbacks const
ath10k: make relay callbacks const
ath11k: make relay callbacks const
ath9k: make relay callbacks const
blktrace: make relay callbacks const
Subsystem: resource
Mauro Carvalho Chehab <mchehab+huawei@kernel.org>:
kernel/resource.c: fix kernel-doc markups
Subsystem: ubsan
Kees Cook <keescook@chromium.org>:
Patch series "Clean up UBSAN Makefile", v2:
ubsan: remove redundant -Wno-maybe-uninitialized
ubsan: move cc-option tests into Kconfig
ubsan: disable object-size sanitizer under GCC
ubsan: disable UBSAN_TRAP for all*config
ubsan: enable for all*config builds
ubsan: remove UBSAN_MISC in favor of individual options
ubsan: expand tests and reporting
Dmitry Vyukov <dvyukov@google.com>:
kcov: don't instrument with UBSAN
Zou Wei <zou_wei@huawei.com>:
lib/ubsan.c: mark type_check_kinds with static keyword
Subsystem: reboot
Matteo Croce <mcroce@microsoft.com>:
reboot: refactor and comment the cpu selection code
reboot: allow to specify reboot mode via sysfs
reboot: remove cf9_safe from allowed types and rename cf9_force
Patch series "reboot: sysfs improvements":
reboot: allow to override reboot type if quirks are found
reboot: hide from sysfs not applicable settings
Subsystem: fault-injection
Barnabás Pőcze <pobrn@protonmail.com>:
fault-injection: handle EI_ETYPE_TRUE
Subsystem: lzo
Jason Yan <yanaijie@huawei.com>:
lib/lzo/lzo1x_compress.c: make lzogeneric1x_1_compress() static
Subsystem: apparmor
Andy Shevchenko <andriy.shevchenko@linux.intel.com>:
apparmor: remove duplicate macro list_entry_is_head()
Subsystem: mm/pagemap
Christoph Hellwig <hch@lst.de>:
Patch series "simplify follow_pte a bit":
mm: unexport follow_pte_pmd
mm: simplify follow_pte{,pmd}
Subsystem: mm/cleanups
Haitao Shi <shihaitao1@huawei.com>:
mm: fix some spelling mistakes in comments
Subsystem: mm/gup
Jann Horn <jannh@google.com>:
mmap locking API: don't check locking if the mm isn't live yet
mm/gup: assert that the mmap lock is held in __get_user_pages()
Documentation/ABI/testing/sysfs-kernel-reboot | 32
Documentation/admin-guide/kdump/vmcoreinfo.rst | 6
Documentation/dev-tools/ubsan.rst | 1
Documentation/filesystems/proc.rst | 2
MAINTAINERS | 5
arch/alpha/kernel/process.c | 2
arch/powerpc/kernel/vmlinux.lds.S | 4
arch/s390/pci/pci_mmio.c | 4
drivers/gpio/gpio-thunderx.c | 11
drivers/gpio/gpio-xilinx.c | 61 -
drivers/gpu/drm/i915/gt/uc/intel_guc_log.c | 2
drivers/misc/lkdtm/Makefile | 1
drivers/misc/lkdtm/bugs.c | 50 +
drivers/misc/lkdtm/core.c | 3
drivers/misc/lkdtm/fortify.c | 82 ++
drivers/misc/lkdtm/lkdtm.h | 19
drivers/net/wireless/ath/ath10k/spectral.c | 2
drivers/net/wireless/ath/ath11k/spectral.c | 2
drivers/net/wireless/ath/ath9k/common-spectral.c | 2
drivers/rapidio/rio.c | 81 --
fs/bfs/inode.c | 2
fs/dax.c | 9
fs/exec.c | 8
fs/nfs/callback_proc.c | 5
fs/nilfs2/segment.c | 5
fs/proc/array.c | 28
fs/proc/base.c | 2
fs/proc/generic.c | 24
fs/proc/internal.h | 10
fs/proc/proc_net.c | 20
include/asm-generic/bitops/find.h | 19
include/asm-generic/getorder.h | 2
include/linux/bitmap.h | 67 +-
include/linux/bitops.h | 24
include/linux/dcache.h | 1
include/linux/iommu-helper.h | 4
include/linux/kernel.h | 173 -----
include/linux/log2.h | 3
include/linux/math.h | 177 +++++
include/linux/mm.h | 6
include/linux/mm_types.h | 10
include/linux/mmap_lock.h | 16
include/linux/proc_fs.h | 8
include/linux/rcu_node_tree.h | 2
include/linux/relay.h | 29
include/linux/rio_drv.h | 3
include/linux/string.h | 75 +-
include/linux/units.h | 2
kernel/Makefile | 3
kernel/acct.c | 7
kernel/crash_core.c | 1
kernel/fail_function.c | 6
kernel/gcov/gcc_4_7.c | 10
kernel/reboot.c | 308 ++++++++-
kernel/relay.c | 111 ---
kernel/resource.c | 24
kernel/trace/blktrace.c | 2
lib/Kconfig.debug | 11
lib/Kconfig.ubsan | 154 +++-
lib/Makefile | 7
lib/bits_kunit.c | 75 ++
lib/cmdline.c | 20
lib/cmdline_kunit.c | 100 +++
lib/errname.c | 1
lib/error-inject.c | 2
lib/errseq.c | 1
lib/find_bit.c | 17
lib/linear_ranges_kunit.c | 228 +++++++
lib/list-test.c | 748 -----------------------
lib/list_kunit.c | 748 +++++++++++++++++++++++
lib/lz4/lz4_decompress.c | 6
lib/lz4/lz4defs.h | 1
lib/lzo/lzo1x_compress.c | 2
lib/math/div64.c | 4
lib/math/int_pow.c | 2
lib/math/int_sqrt.c | 3
lib/math/reciprocal_div.c | 9
lib/stackdepot.c | 11
lib/string.c | 4
lib/test_bitmap.c | 143 ++++
lib/test_bits.c | 75 --
lib/test_firmware.c | 9
lib/test_free_pages.c | 5
lib/test_kmod.c | 26
lib/test_linear_ranges.c | 228 -------
lib/test_lockup.c | 16
lib/test_ubsan.c | 74 ++
lib/ubsan.c | 2
mm/filemap.c | 2
mm/gup.c | 2
mm/huge_memory.c | 2
mm/khugepaged.c | 2
mm/memblock.c | 2
mm/memory.c | 36 -
mm/memory_hotplug.c | 2
mm/migrate.c | 2
mm/page_ext.c | 2
mm/swapfile.c | 11
scripts/Makefile.ubsan | 49 -
scripts/checkpatch.pl | 495 +++++++++++----
security/apparmor/apparmorfs.c | 3
tools/testing/selftests/lkdtm/tests.txt | 1
102 files changed, 3022 insertions(+), 1899 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2020-12-15 22:49 ` incoming Linus Torvalds
@ 2020-12-15 22:55 ` Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-12-15 22:55 UTC (permalink / raw)
To: Linus Torvalds; +Cc: Linux-MM, mm-commits
On Tue, 15 Dec 2020 14:49:24 -0800 Linus Torvalds <torvalds@linux-foundation.org> wrote:
> On Tue, Dec 15, 2020 at 2:48 PM Linus Torvalds
> <torvalds@linux-foundation.org> wrote:
> >
> > I will try to apply it on top of my merge of your previous series instead.
>
> Yes, then it applies cleanly. So apparently we just have different
> concepts of what really constitutes a "base" for applying your series.
>
oop, sorry, yes, the "based on" thing was wrong because I had two
series in flight simultaneously. I've never tried that before..
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2020-12-15 22:48 ` incoming Linus Torvalds
@ 2020-12-15 22:49 ` Linus Torvalds
2020-12-15 22:55 ` incoming Andrew Morton
0 siblings, 1 reply; 409+ messages in thread
From: Linus Torvalds @ 2020-12-15 22:49 UTC (permalink / raw)
To: Andrew Morton; +Cc: Linux-MM, mm-commits
On Tue, Dec 15, 2020 at 2:48 PM Linus Torvalds
<torvalds@linux-foundation.org> wrote:
>
> I will try to apply it on top of my merge of your previous series instead.
Yes, then it applies cleanly. So apparently we just have different
concepts of what really constitutes a "base" for applying your series.
Linus
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2020-12-15 20:32 incoming Andrew Morton
2020-12-15 21:00 ` incoming Linus Torvalds
@ 2020-12-15 22:48 ` Linus Torvalds
2020-12-15 22:49 ` incoming Linus Torvalds
1 sibling, 1 reply; 409+ messages in thread
From: Linus Torvalds @ 2020-12-15 22:48 UTC (permalink / raw)
To: Andrew Morton; +Cc: Linux-MM, mm-commits
On Tue, Dec 15, 2020 at 12:32 PM Andrew Morton
<akpm@linux-foundation.org> wrote:
>
> - more MM work: a memcg scalability improvememt
>
> 19 patches, based on 148842c98a24e508aecb929718818fbf4c2a6ff3.
With your re-send, I get all patches, but they don't actually apply cleanly.
Is that base correct?
I get
error: patch failed: mm/huge_memory.c:2750
error: mm/huge_memory.c: patch does not apply
Patch failed at 0004 mm/thp: narrow lru locking
for that patch "[patch 04/19] mm/thp: narrow lru locking", and that's
definitely true: the patch fragment has
@@ -2750,7 +2751,7 @@ int split_huge_page_to_list(struct page
__dec_lruvec_page_state(head, NR_FILE_THPS);
}
- __split_huge_page(page, list, end, flags);
+ __split_huge_page(page, list, end);
ret = 0;
} else {
if (IS_ENABLED(CONFIG_DEBUG_VM) && mapcount) {
but that __dec_lruvec_page_state() conversion was done by your
previous commit series.
So I have the feeling that what you actually mean by "base" isn't
actually really the base for that series at all..
I will try to apply it on top of my merge of your previous series instead.
Linus
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2020-12-15 20:32 incoming Andrew Morton
@ 2020-12-15 21:00 ` Linus Torvalds
2020-12-15 22:48 ` incoming Linus Torvalds
1 sibling, 0 replies; 409+ messages in thread
From: Linus Torvalds @ 2020-12-15 21:00 UTC (permalink / raw)
To: Andrew Morton; +Cc: Linux-MM, mm-commits
On Tue, Dec 15, 2020 at 12:32 PM Andrew Morton
<akpm@linux-foundation.org> wrote:
>
> - more MM work: a memcg scalability improvememt
>
> 19 patches, based on 148842c98a24e508aecb929718818fbf4c2a6ff3.
I'm not seeing patch 10/19 at all.
And patch 19/19 is corrupted and has an attachment with a '^P'
character in it. I could fix it up, but with the missing patch in the
middle I'm not going to even try. 'b4' is also very unhappy about that
patch 19/19.
I don't know what went wrong, but I'll ignore this send - please
re-send the series at your leisure, ok?
Linus
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-12-15 20:32 Andrew Morton
2020-12-15 21:00 ` incoming Linus Torvalds
2020-12-15 22:48 ` incoming Linus Torvalds
0 siblings, 2 replies; 409+ messages in thread
From: Andrew Morton @ 2020-12-15 20:32 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
- more MM work: a memcg scalability improvememt
19 patches, based on 148842c98a24e508aecb929718818fbf4c2a6ff3.
Subsystems affected by this patch series:
Alex Shi <alex.shi@linux.alibaba.com>:
Patch series "per memcg lru lock", v21:
mm/thp: move lru_add_page_tail() to huge_memory.c
mm/thp: use head for head page in lru_add_page_tail()
mm/thp: simplify lru_add_page_tail()
mm/thp: narrow lru locking
mm/vmscan: remove unnecessary lruvec adding
mm/rmap: stop store reordering issue on page->mapping
Hugh Dickins <hughd@google.com>:
mm: page_idle_get_page() does not need lru_lock
Alex Shi <alex.shi@linux.alibaba.com>:
mm/memcg: add debug checking in lock_page_memcg
mm/swap.c: fold vm event PGROTATED into pagevec_move_tail_fn
mm/lru: move lock into lru_note_cost
mm/vmscan: remove lruvec reget in move_pages_to_lru
mm/mlock: remove lru_lock on TestClearPageMlocked
mm/mlock: remove __munlock_isolate_lru_page()
mm/lru: introduce TestClearPageLRU()
mm/compaction: do page isolation first in compaction
mm/swap.c: serialize memcg changes in pagevec_lru_move_fn
mm/lru: replace pgdat lru_lock with lruvec lock
Alexander Duyck <alexander.h.duyck@linux.intel.com>:
mm/lru: introduce relock_page_lruvec()
Hugh Dickins <hughd@google.com>:
mm/lru: revise the comments of lru_lock
Documentation/admin-guide/cgroup-v1/memcg_test.rst | 15 -
Documentation/admin-guide/cgroup-v1/memory.rst | 23 -
Documentation/trace/events-kmem.rst | 2
Documentation/vm/unevictable-lru.rst | 22 -
include/linux/memcontrol.h | 110 +++++++
include/linux/mm_types.h | 2
include/linux/mmzone.h | 6
include/linux/page-flags.h | 1
include/linux/swap.h | 4
mm/compaction.c | 98 ++++---
mm/filemap.c | 4
mm/huge_memory.c | 109 ++++---
mm/memcontrol.c | 84 +++++-
mm/mlock.c | 93 ++----
mm/mmzone.c | 1
mm/page_alloc.c | 1
mm/page_idle.c | 4
mm/rmap.c | 12
mm/swap.c | 292 ++++++++-------------
mm/vmscan.c | 239 ++++++++---------
mm/workingset.c | 2
21 files changed, 644 insertions(+), 480 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2020-12-15 3:30 ` incoming Linus Torvalds
@ 2020-12-15 14:04 ` Konstantin Ryabitsev
0 siblings, 0 replies; 409+ messages in thread
From: Konstantin Ryabitsev @ 2020-12-15 14:04 UTC (permalink / raw)
To: Linus Torvalds; +Cc: Andrew Morton, mm-commits, Linux-MM
On Mon, Dec 14, 2020 at 07:30:54PM -0800, Linus Torvalds wrote:
> > All the patches except for _one_ get a nice little green check-mark
> > next to them when I use 'git am' on this series.
> >
> > The one that did not was [patch 192/200].
> >
> > I have no idea why
>
> Hmm. It looks like that patch is the only one in the series with the
> ">From" marker in the commit message, from the silly "clarify that
> this isn't the first line in a new message in mbox format".
>
> And "b4 am" has turned the single ">" into two, making the stupid
> marker worse, and actually corrupting the end result.
It's a bug in b4 that I overlooked. Public-inbox emits mboxrd-formatted
.mbox files, while Python's mailbox.mbox consumes mboxo only. The main
distinction between the two is precisely that mboxrd will convert
">From " into ">>From " in an attempt to avoid corruption during
escape/unescape (it didn't end up fixing the problem 100% and mostly
introduced incompatibilities like this one).
I have a fix in master/stable-0.6.y and I'll release a 0.6.2 before the
end of the week.
Thanks for the report.
-K
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2020-12-15 3:25 ` incoming Linus Torvalds
@ 2020-12-15 3:30 ` Linus Torvalds
2020-12-15 14:04 ` incoming Konstantin Ryabitsev
0 siblings, 1 reply; 409+ messages in thread
From: Linus Torvalds @ 2020-12-15 3:30 UTC (permalink / raw)
To: Andrew Morton, Konstantin Ryabitsev; +Cc: mm-commits, Linux-MM
On Mon, Dec 14, 2020 at 7:25 PM Linus Torvalds
<torvalds@linux-foundation.org> wrote:
>
> All the patches except for _one_ get a nice little green check-mark
> next to them when I use 'git am' on this series.
>
> The one that did not was [patch 192/200].
>
> I have no idea why
Hmm. It looks like that patch is the only one in the series with the
">From" marker in the commit message, from the silly "clarify that
this isn't the first line in a new message in mbox format".
And "b4 am" has turned the single ">" into two, making the stupid
marker worse, and actually corrupting the end result.
Coincidence? Or cause?
Linus
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2020-12-15 3:02 incoming Andrew Morton
@ 2020-12-15 3:25 ` Linus Torvalds
2020-12-15 3:30 ` incoming Linus Torvalds
0 siblings, 1 reply; 409+ messages in thread
From: Linus Torvalds @ 2020-12-15 3:25 UTC (permalink / raw)
To: Andrew Morton, Konstantin Ryabitsev; +Cc: mm-commits, Linux-MM
On Mon, Dec 14, 2020 at 7:02 PM Andrew Morton <akpm@linux-foundation.org> wrote:
>
> 200 patches, based on 2c85ebc57b3e1817b6ce1a6b703928e113a90442.
I haven't actually processed the patches yet, but I have a question
for Konstantin wrt b4.
All the patches except for _one_ get a nice little green check-mark
next to them when I use 'git am' on this series.
The one that did not was [patch 192/200].
I have no idea why - and it doesn't matter a lot to me, it just stood
out as being different. I'm assuming Andrew has started doing patch
attestation, and that patch failed. But if so, maybe Konstantin wants
to know what went wrong.
Konstantin?
Linus
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-12-15 3:02 Andrew Morton
2020-12-15 3:25 ` incoming Linus Torvalds
0 siblings, 1 reply; 409+ messages in thread
From: Andrew Morton @ 2020-12-15 3:02 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
- a few random little subsystems
- almost all of the MM patches which are staged ahead of linux-next
material. I'll trickle to post-linux-next work in as the dependents
get merged up.
200 patches, based on 2c85ebc57b3e1817b6ce1a6b703928e113a90442.
Subsystems affected by this patch series:
kthread
kbuild
ide
ntfs
ocfs2
arch
mm/slab-generic
mm/slab
mm/slub
mm/dax
mm/debug
mm/pagecache
mm/gup
mm/swap
mm/shmem
mm/memcg
mm/pagemap
mm/mremap
mm/hmm
mm/vmalloc
mm/documentation
mm/kasan
mm/pagealloc
mm/memory-failure
mm/hugetlb
mm/vmscan
mm/z3fold
mm/compaction
mm/oom-kill
mm/migration
mm/cma
mm/page-poison
mm/userfaultfd
mm/zswap
mm/zsmalloc
mm/uaccess
mm/zram
mm/cleanups
Subsystem: kthread
Rob Clark <robdclark@chromium.org>:
kthread: add kthread_work tracepoints
Petr Mladek <pmladek@suse.com>:
kthread_worker: document CPU hotplug handling
Subsystem: kbuild
Petr Vorel <petr.vorel@gmail.com>:
uapi: move constants from <linux/kernel.h> to <linux/const.h>
Subsystem: ide
Sebastian Andrzej Siewior <bigeasy@linutronix.de>:
ide/falcon: remove in_interrupt() usage
ide: remove BUG_ON(in_interrupt() || irqs_disabled()) from ide_unregister()
Subsystem: ntfs
Alex Shi <alex.shi@linux.alibaba.com>:
fs/ntfs: remove unused varibles
fs/ntfs: remove unused variable attr_len
Subsystem: ocfs2
Tom Rix <trix@redhat.com>:
fs/ocfs2/cluster/tcp.c: remove unneeded break
Mauricio Faria de Oliveira <mfo@canonical.com>:
ocfs2: ratelimit the 'max lookup times reached' notice
Subsystem: arch
Colin Ian King <colin.king@canonical.com>:
arch/Kconfig: fix spelling mistakes
Subsystem: mm/slab-generic
Hui Su <sh_def@163.com>:
mm/slab_common.c: use list_for_each_entry in dump_unreclaimable_slab()
Bartosz Golaszewski <bgolaszewski@baylibre.com>:
Patch series "slab: provide and use krealloc_array()", v3:
mm: slab: clarify krealloc()'s behavior with __GFP_ZERO
mm: slab: provide krealloc_array()
ALSA: pcm: use krealloc_array()
vhost: vringh: use krealloc_array()
pinctrl: use krealloc_array()
edac: ghes: use krealloc_array()
drm: atomic: use krealloc_array()
hwtracing: intel: use krealloc_array()
dma-buf: use krealloc_array()
Vlastimil Babka <vbabka@suse.cz>:
mm, slab, slub: clear the slab_cache field when freeing page
Subsystem: mm/slab
Alexander Popov <alex.popov@linux.com>:
mm/slab: rerform init_on_free earlier
Subsystem: mm/slub
Vlastimil Babka <vbabka@suse.cz>:
mm, slub: use kmem_cache_debug_flags() in deactivate_slab()
Bharata B Rao <bharata@linux.ibm.com>:
mm/slub: let number of online CPUs determine the slub page order
Subsystem: mm/dax
Dan Williams <dan.j.williams@intel.com>:
device-dax/kmem: use struct_size()
Subsystem: mm/debug
Zhenhua Huang <zhenhuah@codeaurora.org>:
mm: fix page_owner initializing issue for arm32
Liam Mark <lmark@codeaurora.org>:
mm/page_owner: record timestamp and pid
Subsystem: mm/pagecache
Kent Overstreet <kent.overstreet@gmail.com>:
Patch series "generic_file_buffered_read() improvements", v2:
mm/filemap/c: break generic_file_buffered_read up into multiple functions
mm/filemap.c: generic_file_buffered_read() now uses find_get_pages_contig
Alex Shi <alex.shi@linux.alibaba.com>:
mm/truncate: add parameter explanation for invalidate_mapping_pagevec
Hailong Liu <carver4lio@163.com>:
mm/filemap.c: remove else after a return
Subsystem: mm/gup
John Hubbard <jhubbard@nvidia.com>:
Patch series "selftests/vm: gup_test, hmm-tests, assorted improvements", v3:
mm/gup_benchmark: rename to mm/gup_test
selftests/vm: use a common gup_test.h
selftests/vm: rename run_vmtests --> run_vmtests.sh
selftests/vm: minor cleanup: Makefile and gup_test.c
selftests/vm: only some gup_test items are really benchmarks
selftests/vm: gup_test: introduce the dump_pages() sub-test
selftests/vm: run_vmtests.sh: update and clean up gup_test invocation
selftests/vm: hmm-tests: remove the libhugetlbfs dependency
selftests/vm: 2x speedup for run_vmtests.sh
Barry Song <song.bao.hua@hisilicon.com>:
mm/gup_test.c: mark gup_test_init as __init function
mm/gup_test: GUP_TEST depends on DEBUG_FS
Jason Gunthorpe <jgg@nvidia.com>:
Patch series "Add a seqcount between gup_fast and copy_page_range()", v4:
mm/gup: reorganize internal_get_user_pages_fast()
mm/gup: prevent gup_fast from racing with COW during fork
mm/gup: remove the vma allocation from gup_longterm_locked()
mm/gup: combine put_compound_head() and unpin_user_page()
Subsystem: mm/swap
Ralph Campbell <rcampbell@nvidia.com>:
mm: handle zone device pages in release_pages()
Miaohe Lin <linmiaohe@huawei.com>:
mm/swapfile.c: use helper function swap_count() in add_swap_count_continuation()
mm/swap_state: skip meaningless swap cache readahead when ra_info.win == 0
mm/swapfile.c: remove unnecessary out label in __swap_duplicate()
mm/swapfile.c: use memset to fill the swap_map with SWAP_HAS_CACHE
Jeff Layton <jlayton@kernel.org>:
mm: remove pagevec_lookup_range_nr_tag()
Subsystem: mm/shmem
Hui Su <sh_def@163.com>:
mm/shmem.c: make shmem_mapping() inline
Randy Dunlap <rdunlap@infradead.org>:
tmpfs: fix Documentation nits
Subsystem: mm/memcg
Johannes Weiner <hannes@cmpxchg.org>:
mm: memcontrol: add file_thp, shmem_thp to memory.stat
Muchun Song <songmuchun@bytedance.com>:
mm: memcontrol: remove unused mod_memcg_obj_state()
Miaohe Lin <linmiaohe@huawei.com>:
mm: memcontrol: eliminate redundant check in __mem_cgroup_insert_exceeded()
Muchun Song <songmuchun@bytedance.com>:
mm: memcg/slab: fix return of child memcg objcg for root memcg
mm: memcg/slab: fix use after free in obj_cgroup_charge
Shakeel Butt <shakeelb@google.com>:
mm/rmap: always do TTU_IGNORE_ACCESS
Alex Shi <alex.shi@linux.alibaba.com>:
mm/memcg: update page struct member in comments
Roman Gushchin <guro@fb.com>:
mm: memcg: fix obsolete code comments
Patch series "mm: memcg: deprecate cgroup v1 non-hierarchical mode", v1:
mm: memcg: deprecate the non-hierarchical mode
docs: cgroup-v1: reflect the deprecation of the non-hierarchical mode
cgroup: remove obsoleted broken_hierarchy and warned_broken_hierarchy
Hui Su <sh_def@163.com>:
mm/page_counter: use page_counter_read in page_counter_set_max
Lukas Bulwahn <lukas.bulwahn@gmail.com>:
mm: memcg: remove obsolete memcg_has_children()
Muchun Song <songmuchun@bytedance.com>:
mm: memcg/slab: rename *_lruvec_slab_state to *_lruvec_kmem_state
Kaixu Xia <kaixuxia@tencent.com>:
mm: memcontrol: sssign boolean values to a bool variable
Alex Shi <alex.shi@linux.alibaba.com>:
mm/memcg: remove incorrect comment
Shakeel Butt <shakeelb@google.com>:
Patch series "memcg: add pagetable comsumption to memory.stat", v2:
mm: move lruvec stats update functions to vmstat.h
mm: memcontrol: account pagetables per node
Subsystem: mm/pagemap
Dan Williams <dan.j.williams@intel.com>:
xen/unpopulated-alloc: consolidate pgmap manipulation
Kalesh Singh <kaleshsingh@google.com>:
Patch series "Speed up mremap on large regions", v4:
kselftests: vm: add mremap tests
mm: speedup mremap on 1GB or larger regions
arm64: mremap speedup - enable HAVE_MOVE_PUD
x86: mremap speedup - Enable HAVE_MOVE_PUD
John Hubbard <jhubbard@nvidia.com>:
mm: cleanup: remove unused tsk arg from __access_remote_vm
Alex Shi <alex.shi@linux.alibaba.com>:
mm/mapping_dirty_helpers: enhance the kernel-doc markups
mm/page_vma_mapped.c: add colon to fix kernel-doc markups error for check_pte
Axel Rasmussen <axelrasmussen@google.com>:
mm: mmap_lock: add tracepoints around lock acquisition
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
sparc: fix handling of page table constructor failure
mm: move free_unref_page to mm/internal.h
Subsystem: mm/mremap
Dmitry Safonov <dima@arista.com>:
Patch series "mremap: move_vma() fixes":
mm/mremap: account memory on do_munmap() failure
mm/mremap: for MREMAP_DONTUNMAP check security_vm_enough_memory_mm()
mremap: don't allow MREMAP_DONTUNMAP on special_mappings and aio
vm_ops: rename .split() callback to .may_split()
mremap: check if it's possible to split original vma
mm: forbid splitting special mappings
Subsystem: mm/hmm
Daniel Vetter <daniel.vetter@ffwll.ch>:
mm: track mmu notifiers in fs_reclaim_acquire/release
mm: extract might_alloc() debug check
locking/selftests: add testcases for fs_reclaim
Subsystem: mm/vmalloc
Andrew Morton <akpm@linux-foundation.org>:
mm/vmalloc.c:__vmalloc_area_node(): avoid 32-bit overflow
"Uladzislau Rezki (Sony)" <urezki@gmail.com>:
mm/vmalloc: use free_vm_area() if an allocation fails
mm/vmalloc: rework the drain logic
Alex Shi <alex.shi@linux.alibaba.com>:
mm/vmalloc: add 'align' parameter explanation for pvm_determine_end_from_reverse
Baolin Wang <baolin.wang@linux.alibaba.com>:
mm/vmalloc.c: remove unnecessary return statement
Waiman Long <longman@redhat.com>:
mm/vmalloc: Fix unlock order in s_stop()
Subsystem: mm/documentation
Alex Shi <alex.shi@linux.alibaba.com>:
docs/vm: remove unused 3 items explanation for /proc/vmstat
Subsystem: mm/kasan
Vincenzo Frascino <vincenzo.frascino@arm.com>:
mm/vmalloc.c: fix kasan shadow poisoning size
Walter Wu <walter-zh.wu@mediatek.com>:
Patch series "kasan: add workqueue stack for generic KASAN", v5:
workqueue: kasan: record workqueue stack
kasan: print workqueue stack
lib/test_kasan.c: add workqueue test case
kasan: update documentation for generic kasan
Marco Elver <elver@google.com>:
lkdtm: disable KASAN for rodata.o
Subsystem: mm/pagealloc
Mike Rapoport <rppt@linux.ibm.com>:
Patch series "arch, mm: deprecate DISCONTIGMEM", v2:
alpha: switch from DISCONTIGMEM to SPARSEMEM
ia64: remove custom __early_pfn_to_nid()
ia64: remove 'ifdef CONFIG_ZONE_DMA32' statements
ia64: discontig: paging_init(): remove local max_pfn calculation
ia64: split virtual map initialization out of paging_init()
ia64: forbid using VIRTUAL_MEM_MAP with FLATMEM
ia64: make SPARSEMEM default and disable DISCONTIGMEM
arm: remove CONFIG_ARCH_HAS_HOLES_MEMORYMODEL
arm, arm64: move free_unused_memmap() to generic mm
arc: use FLATMEM with freeing of unused memory map instead of DISCONTIGMEM
m68k/mm: make node data and node setup depend on CONFIG_DISCONTIGMEM
m68k/mm: enable use of generic memory_model.h for !DISCONTIGMEM
m68k: deprecate DISCONTIGMEM
Patch series "arch, mm: improve robustness of direct map manipulation", v7:
mm: introduce debug_pagealloc_{map,unmap}_pages() helpers
PM: hibernate: make direct map manipulations more explicit
arch, mm: restore dependency of __kernel_map_pages() on DEBUG_PAGEALLOC
arch, mm: make kernel_page_present() always available
Vlastimil Babka <vbabka@suse.cz>:
Patch series "disable pcplists during memory offline", v3:
mm, page_alloc: clean up pageset high and batch update
mm, page_alloc: calculate pageset high and batch once per zone
mm, page_alloc: remove setup_pageset()
mm, page_alloc: simplify pageset_update()
mm, page_alloc: cache pageset high and batch in struct zone
mm, page_alloc: move draining pcplists to page isolation users
mm, page_alloc: disable pcplists during memory offline
Miaohe Lin <linmiaohe@huawei.com>:
include/linux/page-flags.h: remove unused __[Set|Clear]PagePrivate
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
mm/page-flags: fix comment
mm/page_alloc: add __free_pages() documentation
Zou Wei <zou_wei@huawei.com>:
mm/page_alloc: mark some symbols with static keyword
David Hildenbrand <david@redhat.com>:
mm/page_alloc: clear all pages in post_alloc_hook() with init_on_alloc=1
Lin Feng <linf@wangsu.com>:
init/main: fix broken buffer_init when DEFERRED_STRUCT_PAGE_INIT set
Lorenzo Stoakes <lstoakes@gmail.com>:
mm: page_alloc: refactor setup_per_zone_lowmem_reserve()
Muchun Song <songmuchun@bytedance.com>:
mm/page_alloc: speed up the iteration of max_order
Subsystem: mm/memory-failure
Oscar Salvador <osalvador@suse.de>:
Patch series "HWpoison: further fixes and cleanups", v5:
mm,hwpoison: drain pcplists before bailing out for non-buddy zero-refcount page
mm,hwpoison: take free pages off the buddy freelists
mm,hwpoison: drop unneeded pcplist draining
Patch series "HWPoison: Refactor get page interface", v2:
mm,hwpoison: refactor get_any_page
mm,hwpoison: disable pcplists before grabbing a refcount
mm,hwpoison: remove drain_all_pages from shake_page
mm,memory_failure: always pin the page in madvise_inject_error
mm,hwpoison: return -EBUSY when migration fails
Subsystem: mm/hugetlb
Hui Su <sh_def@163.com>:
mm/hugetlb.c: just use put_page_testzero() instead of page_count()
Ralph Campbell <rcampbell@nvidia.com>:
include/linux/huge_mm.h: remove extern keyword
Alex Shi <alex.shi@linux.alibaba.com>:
khugepaged: add parameter explanations for kernel-doc markup
Liu Xiang <liu.xiang@zlingsmart.com>:
mm: hugetlb: fix type of delta parameter and related local variables in gather_surplus_pages()
Oscar Salvador <osalvador@suse.de>:
mm,hugetlb: remove unneeded initialization
Dan Carpenter <dan.carpenter@oracle.com>:
hugetlb: fix an error code in hugetlb_reserve_pages()
Subsystem: mm/vmscan
Johannes Weiner <hannes@cmpxchg.org>:
mm: don't wake kswapd prematurely when watermark boosting is disabled
Lukas Bulwahn <lukas.bulwahn@gmail.com>:
mm/vmscan: drop unneeded assignment in kswapd()
"logic.yu" <hymmsx.yu@gmail.com>:
mm/vmscan.c: remove the filename in the top of file comment
Muchun Song <songmuchun@bytedance.com>:
mm/page_isolation: do not isolate the max order page
Subsystem: mm/z3fold
Vitaly Wool <vitaly.wool@konsulko.com>:
Patch series "z3fold: stability / rt fixes":
z3fold: simplify freeing slots
z3fold: stricter locking and more careful reclaim
z3fold: remove preempt disabled sections for RT
Subsystem: mm/compaction
Yanfei Xu <yanfei.xu@windriver.com>:
mm/compaction: rename 'start_pfn' to 'iteration_start_pfn' in compact_zone()
Hui Su <sh_def@163.com>:
mm/compaction: move compaction_suitable's comment to right place
mm/compaction: make defer_compaction and compaction_deferred static
Subsystem: mm/oom-kill
Hui Su <sh_def@163.com>:
mm/oom_kill: change comment and rename is_dump_unreclaim_slabs()
Subsystem: mm/migration
Long Li <lonuxli.64@gmail.com>:
mm/migrate.c: fix comment spelling
Ralph Campbell <rcampbell@nvidia.com>:
mm/migrate.c: optimize migrate_vma_pages() mmu notifier
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
mm: support THPs in zero_user_segments
Yang Shi <shy828301@gmail.com>:
Patch series "mm: misc migrate cleanup and improvement", v3:
mm: truncate_complete_page() does not exist any more
mm: migrate: simplify the logic for handling permanent failure
mm: migrate: skip shared exec THP for NUMA balancing
mm: migrate: clean up migrate_prep{_local}
mm: migrate: return -ENOSYS if THP migration is unsupported
Stephen Zhang <starzhangzsd@gmail.com>:
mm: migrate: remove unused parameter in migrate_vma_insert_page()
Subsystem: mm/cma
Lecopzer Chen <lecopzer.chen@mediatek.com>:
mm/cma.c: remove redundant cma_mutex lock
Charan Teja Reddy <charante@codeaurora.org>:
mm: cma: improve pr_debug log in cma_release()
Subsystem: mm/page-poison
Vlastimil Babka <vbabka@suse.cz>:
Patch series "cleanup page poisoning", v3:
mm, page_alloc: do not rely on the order of page_poison and init_on_alloc/free parameters
mm, page_poison: use static key more efficiently
kernel/power: allow hibernation with page_poison sanity checking
mm, page_poison: remove CONFIG_PAGE_POISONING_NO_SANITY
mm, page_poison: remove CONFIG_PAGE_POISONING_ZERO
Subsystem: mm/userfaultfd
Lokesh Gidra <lokeshgidra@google.com>:
Patch series "Control over userfaultfd kernel-fault handling", v6:
userfaultfd: add UFFD_USER_MODE_ONLY
userfaultfd: add user-mode only option to unprivileged_userfaultfd sysctl knob
Axel Rasmussen <axelrasmussen@google.com>:
userfaultfd: selftests: make __{s,u}64 format specifiers portable
Peter Xu <peterx@redhat.com>:
Patch series "userfaultfd: selftests: Small fixes":
userfaultfd/selftests: always dump something in modes
userfaultfd/selftests: fix retval check for userfaultfd_open()
userfaultfd/selftests: hint the test runner on required privilege
Subsystem: mm/zswap
Joe Perches <joe@perches.com>:
mm/zswap: make struct kernel_param_ops definitions const
YueHaibing <yuehaibing@huawei.com>:
mm/zswap: fix passing zero to 'PTR_ERR' warning
Barry Song <song.bao.hua@hisilicon.com>:
mm/zswap: move to use crypto_acomp API for hardware acceleration
Subsystem: mm/zsmalloc
Miaohe Lin <linmiaohe@huawei.com>:
mm/zsmalloc.c: rework the list_add code in insert_zspage()
Subsystem: mm/uaccess
Colin Ian King <colin.king@canonical.com>:
mm/process_vm_access: remove redundant initialization of iov_r
Subsystem: mm/zram
Minchan Kim <minchan@kernel.org>:
zram: support page writeback
zram: add stat to gather incompressible pages since zram set up
Rui Salvaterra <rsalvaterra@gmail.com>:
zram: break the strict dependency from lzo
Subsystem: mm/cleanups
Mauro Carvalho Chehab <mchehab+huawei@kernel.org>:
mm: fix kernel-doc markups
Joe Perches <joe@perches.com>:
Patch series "mm: Convert sysfs sprintf family to sysfs_emit", v2:
mm: use sysfs_emit for struct kobject * uses
mm: huge_memory: convert remaining use of sprintf to sysfs_emit and neatening
mm:backing-dev: use sysfs_emit in macro defining functions
mm: shmem: convert shmem_enabled_show to use sysfs_emit_at
mm: slub: convert sysfs sprintf family to sysfs_emit/sysfs_emit_at
"Gustavo A. R. Silva" <gustavoars@kernel.org>:
mm: fix fall-through warnings for Clang
Alexey Dobriyan <adobriyan@gmail.com>:
mm: cleanup kstrto*() usage
/mmap_lock.h | 107 ++
a/Documentation/admin-guide/blockdev/zram.rst | 6
a/Documentation/admin-guide/cgroup-v1/memcg_test.rst | 8
a/Documentation/admin-guide/cgroup-v1/memory.rst | 42
a/Documentation/admin-guide/cgroup-v2.rst | 11
a/Documentation/admin-guide/mm/transhuge.rst | 15
a/Documentation/admin-guide/sysctl/vm.rst | 15
a/Documentation/core-api/memory-allocation.rst | 4
a/Documentation/core-api/pin_user_pages.rst | 8
a/Documentation/dev-tools/kasan.rst | 5
a/Documentation/filesystems/tmpfs.rst | 8
a/Documentation/vm/memory-model.rst | 3
a/Documentation/vm/page_owner.rst | 12
a/arch/Kconfig | 21
a/arch/alpha/Kconfig | 8
a/arch/alpha/include/asm/mmzone.h | 14
a/arch/alpha/include/asm/page.h | 7
a/arch/alpha/include/asm/pgtable.h | 12
a/arch/alpha/include/asm/sparsemem.h | 18
a/arch/alpha/kernel/setup.c | 1
a/arch/arc/Kconfig | 3
a/arch/arc/include/asm/page.h | 20
a/arch/arc/mm/init.c | 29
a/arch/arm/Kconfig | 12
a/arch/arm/kernel/vdso.c | 9
a/arch/arm/mach-bcm/Kconfig | 1
a/arch/arm/mach-davinci/Kconfig | 1
a/arch/arm/mach-exynos/Kconfig | 1
a/arch/arm/mach-highbank/Kconfig | 1
a/arch/arm/mach-omap2/Kconfig | 1
a/arch/arm/mach-s5pv210/Kconfig | 1
a/arch/arm/mach-tango/Kconfig | 1
a/arch/arm/mm/init.c | 78 -
a/arch/arm64/Kconfig | 9
a/arch/arm64/include/asm/cacheflush.h | 1
a/arch/arm64/include/asm/pgtable.h | 1
a/arch/arm64/kernel/vdso.c | 41
a/arch/arm64/mm/init.c | 68 -
a/arch/arm64/mm/pageattr.c | 12
a/arch/ia64/Kconfig | 11
a/arch/ia64/include/asm/meminit.h | 2
a/arch/ia64/mm/contig.c | 88 --
a/arch/ia64/mm/discontig.c | 44 -
a/arch/ia64/mm/init.c | 14
a/arch/ia64/mm/numa.c | 30
a/arch/m68k/Kconfig.cpu | 31
a/arch/m68k/include/asm/page.h | 2
a/arch/m68k/include/asm/page_mm.h | 7
a/arch/m68k/include/asm/virtconvert.h | 7
a/arch/m68k/mm/init.c | 10
a/arch/mips/vdso/genvdso.c | 4
a/arch/nds32/mm/mm-nds32.c | 6
a/arch/powerpc/Kconfig | 5
a/arch/riscv/Kconfig | 4
a/arch/riscv/include/asm/pgtable.h | 2
a/arch/riscv/include/asm/set_memory.h | 1
a/arch/riscv/mm/pageattr.c | 31
a/arch/s390/Kconfig | 4
a/arch/s390/configs/debug_defconfig | 2
a/arch/s390/configs/defconfig | 2
a/arch/s390/kernel/vdso.c | 11
a/arch/sparc/Kconfig | 4
a/arch/sparc/mm/init_64.c | 2
a/arch/x86/Kconfig | 5
a/arch/x86/entry/vdso/vma.c | 17
a/arch/x86/include/asm/set_memory.h | 1
a/arch/x86/kernel/cpu/resctrl/pseudo_lock.c | 2
a/arch/x86/kernel/tboot.c | 1
a/arch/x86/mm/pat/set_memory.c | 6
a/drivers/base/node.c | 2
a/drivers/block/zram/Kconfig | 42
a/drivers/block/zram/zcomp.c | 2
a/drivers/block/zram/zram_drv.c | 29
a/drivers/block/zram/zram_drv.h | 1
a/drivers/dax/device.c | 4
a/drivers/dax/kmem.c | 2
a/drivers/dma-buf/sync_file.c | 3
a/drivers/edac/ghes_edac.c | 4
a/drivers/firmware/efi/efi.c | 1
a/drivers/gpu/drm/drm_atomic.c | 3
a/drivers/hwtracing/intel_th/msu.c | 2
a/drivers/ide/falconide.c | 2
a/drivers/ide/ide-probe.c | 3
a/drivers/misc/lkdtm/Makefile | 1
a/drivers/pinctrl/pinctrl-utils.c | 2
a/drivers/vhost/vringh.c | 3
a/drivers/virtio/virtio_balloon.c | 6
a/drivers/xen/unpopulated-alloc.c | 14
a/fs/aio.c | 5
a/fs/ntfs/file.c | 5
a/fs/ntfs/inode.c | 2
a/fs/ntfs/logfile.c | 3
a/fs/ocfs2/cluster/tcp.c | 1
a/fs/ocfs2/namei.c | 4
a/fs/proc/kcore.c | 2
a/fs/proc/meminfo.c | 2
a/fs/userfaultfd.c | 20
a/include/linux/cgroup-defs.h | 15
a/include/linux/compaction.h | 12
a/include/linux/fs.h | 2
a/include/linux/gfp.h | 2
a/include/linux/highmem.h | 19
a/include/linux/huge_mm.h | 93 --
a/include/linux/memcontrol.h | 148 ---
a/include/linux/migrate.h | 4
a/include/linux/mm.h | 118 +-
a/include/linux/mm_types.h | 8
a/include/linux/mmap_lock.h | 94 ++
a/include/linux/mmzone.h | 50 -
a/include/linux/page-flags.h | 6
a/include/linux/page_ext.h | 8
a/include/linux/pagevec.h | 3
a/include/linux/poison.h | 4
a/include/linux/rmap.h | 1
a/include/linux/sched/mm.h | 16
a/include/linux/set_memory.h | 5
a/include/linux/shmem_fs.h | 6
a/include/linux/slab.h | 18
a/include/linux/vmalloc.h | 8
a/include/linux/vmstat.h | 104 ++
a/include/trace/events/sched.h | 84 +
a/include/uapi/linux/const.h | 5
a/include/uapi/linux/ethtool.h | 2
a/include/uapi/linux/kernel.h | 9
a/include/uapi/linux/lightnvm.h | 2
a/include/uapi/linux/mroute6.h | 2
a/include/uapi/linux/netfilter/x_tables.h | 2
a/include/uapi/linux/netlink.h | 2
a/include/uapi/linux/sysctl.h | 2
a/include/uapi/linux/userfaultfd.h | 9
a/init/main.c | 6
a/ipc/shm.c | 8
a/kernel/cgroup/cgroup.c | 12
a/kernel/fork.c | 3
a/kernel/kthread.c | 29
a/kernel/power/hibernate.c | 2
a/kernel/power/power.h | 2
a/kernel/power/snapshot.c | 52 +
a/kernel/ptrace.c | 2
a/kernel/workqueue.c | 3
a/lib/locking-selftest.c | 47 +
a/lib/test_kasan_module.c | 29
a/mm/Kconfig | 25
a/mm/Kconfig.debug | 28
a/mm/Makefile | 4
a/mm/backing-dev.c | 8
a/mm/cma.c | 6
a/mm/compaction.c | 29
a/mm/filemap.c | 823 ++++++++++---------
a/mm/gup.c | 329 ++-----
a/mm/gup_benchmark.c | 210 ----
a/mm/gup_test.c | 299 ++++++
a/mm/gup_test.h | 40
a/mm/highmem.c | 52 +
a/mm/huge_memory.c | 86 +
a/mm/hugetlb.c | 28
a/mm/init-mm.c | 1
a/mm/internal.h | 5
a/mm/kasan/generic.c | 3
a/mm/kasan/report.c | 4
a/mm/khugepaged.c | 58 -
a/mm/ksm.c | 50 -
a/mm/madvise.c | 14
a/mm/mapping_dirty_helpers.c | 6
a/mm/memblock.c | 80 +
a/mm/memcontrol.c | 170 +--
a/mm/memory-failure.c | 322 +++----
a/mm/memory.c | 24
a/mm/memory_hotplug.c | 44 -
a/mm/mempolicy.c | 8
a/mm/migrate.c | 183 ++--
a/mm/mm_init.c | 1
a/mm/mmap.c | 22
a/mm/mmap_lock.c | 230 +++++
a/mm/mmu_notifier.c | 7
a/mm/mmzone.c | 14
a/mm/mremap.c | 282 ++++--
a/mm/nommu.c | 8
a/mm/oom_kill.c | 14
a/mm/page_alloc.c | 517 ++++++-----
a/mm/page_counter.c | 4
a/mm/page_ext.c | 10
a/mm/page_isolation.c | 18
a/mm/page_owner.c | 17
a/mm/page_poison.c | 56 -
a/mm/page_vma_mapped.c | 9
a/mm/process_vm_access.c | 2
a/mm/rmap.c | 9
a/mm/shmem.c | 39
a/mm/slab.c | 10
a/mm/slab.h | 9
a/mm/slab_common.c | 10
a/mm/slob.c | 6
a/mm/slub.c | 156 +--
a/mm/swap.c | 12
a/mm/swap_state.c | 7
a/mm/swapfile.c | 14
a/mm/truncate.c | 18
a/mm/vmalloc.c | 105 +-
a/mm/vmscan.c | 21
a/mm/vmstat.c | 6
a/mm/workingset.c | 8
a/mm/z3fold.c | 215 ++--
a/mm/zsmalloc.c | 11
a/mm/zswap.c | 193 +++-
a/sound/core/pcm_lib.c | 4
a/tools/include/linux/poison.h | 6
a/tools/testing/selftests/vm/.gitignore | 4
a/tools/testing/selftests/vm/Makefile | 41
a/tools/testing/selftests/vm/check_config.sh | 31
a/tools/testing/selftests/vm/config | 2
a/tools/testing/selftests/vm/gup_benchmark.c | 143 ---
a/tools/testing/selftests/vm/gup_test.c | 258 +++++
a/tools/testing/selftests/vm/hmm-tests.c | 10
a/tools/testing/selftests/vm/mremap_test.c | 344 +++++++
a/tools/testing/selftests/vm/run_vmtests | 51 -
a/tools/testing/selftests/vm/userfaultfd.c | 94 --
217 files changed, 4817 insertions(+), 3369 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-12-11 21:35 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-12-11 21:35 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
8 patches, based on 33dc9614dc208291d0c4bcdeb5d30d481dcd2c4c.
Subsystems affected by this patch series:
mm/pagecache
proc
selftests
kbuild
mm/kasan
mm/hugetlb
Subsystem: mm/pagecache
Andrew Morton <akpm@linux-foundation.org>:
revert "mm/filemap: add static for function __add_to_page_cache_locked"
Subsystem: proc
Miles Chen <miles.chen@mediatek.com>:
proc: use untagged_addr() for pagemap_read addresses
Subsystem: selftests
Arnd Bergmann <arnd@arndb.de>:
selftest/fpu: avoid clang warning
Subsystem: kbuild
Arnd Bergmann <arnd@arndb.de>:
kbuild: avoid static_assert for genksyms
initramfs: fix clang build failure
elfcore: fix building with clang
Subsystem: mm/kasan
Kuan-Ying Lee <Kuan-Ying.Lee@mediatek.com>:
kasan: fix object remaining in offline per-cpu quarantine
Subsystem: mm/hugetlb
Gerald Schaefer <gerald.schaefer@linux.ibm.com>:
mm/hugetlb: clear compound_nr before freeing gigantic pages
fs/proc/task_mmu.c | 8 ++++++--
include/linux/build_bug.h | 5 +++++
include/linux/elfcore.h | 22 ++++++++++++++++++++++
init/initramfs.c | 2 +-
kernel/Makefile | 1 -
kernel/elfcore.c | 26 --------------------------
lib/Makefile | 3 ++-
mm/filemap.c | 2 +-
mm/hugetlb.c | 1 +
mm/kasan/quarantine.c | 39 +++++++++++++++++++++++++++++++++++++++
10 files changed, 77 insertions(+), 32 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-12-06 6:14 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-12-06 6:14 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
12 patches, based on 33256ce194110874d4bc90078b577c59f9076c59.
Subsystems affected by this patch series:
lib
coredump
mm/memcg
mm/zsmalloc
mm/swap
mailmap
mm/selftests
mm/pagecache
mm/hugetlb
mm/pagemap
Subsystem: lib
Randy Dunlap <rdunlap@infradead.org>:
zlib: export S390 symbols for zlib modules
Subsystem: coredump
Menglong Dong <dong.menglong@zte.com.cn>:
coredump: fix core_pattern parse error
Subsystem: mm/memcg
Roman Gushchin <guro@fb.com>:
mm: memcg/slab: fix obj_cgroup_charge() return value handling
Yang Shi <shy828301@gmail.com>:
mm: list_lru: set shrinker map bit when child nr_items is not zero
Subsystem: mm/zsmalloc
Minchan Kim <minchan@kernel.org>:
mm/zsmalloc.c: drop ZSMALLOC_PGTABLE_MAPPING
Subsystem: mm/swap
Qian Cai <qcai@redhat.com>:
mm/swapfile: do not sleep with a spin lock held
Subsystem: mailmap
Uwe Kleine-König <u.kleine-koenig@pengutronix.de>:
mailmap: add two more addresses of Uwe Kleine-König
Subsystem: mm/selftests
Xingxing Su <suxingxing@loongson.cn>:
tools/testing/selftests/vm: fix build error
Axel Rasmussen <axelrasmussen@google.com>:
userfaultfd: selftests: fix SIGSEGV if huge mmap fails
Subsystem: mm/pagecache
Alex Shi <alex.shi@linux.alibaba.com>:
mm/filemap: add static for function __add_to_page_cache_locked
Subsystem: mm/hugetlb
Mike Kravetz <mike.kravetz@oracle.com>:
hugetlb_cgroup: fix offline of hugetlb cgroup with reservations
Subsystem: mm/pagemap
Liu Zixian <liuzixian4@huawei.com>:
mm/mmap.c: fix mmap return value when vma is merged after call_mmap()
.mailmap | 2 +
arch/arm/configs/omap2plus_defconfig | 1
fs/coredump.c | 3 +
include/linux/zsmalloc.h | 1
lib/zlib_dfltcc/dfltcc_inflate.c | 3 +
mm/Kconfig | 13 -------
mm/filemap.c | 2 -
mm/hugetlb_cgroup.c | 8 +---
mm/list_lru.c | 10 ++---
mm/mmap.c | 26 ++++++--------
mm/slab.h | 40 +++++++++++++---------
mm/swapfile.c | 4 +-
mm/zsmalloc.c | 54 -------------------------------
tools/testing/selftests/vm/Makefile | 4 ++
tools/testing/selftests/vm/userfaultfd.c | 25 +++++++++-----
15 files changed, 75 insertions(+), 121 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-11-22 6:16 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-11-22 6:16 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
8 patches, based on a349e4c659609fd20e4beea89e5c4a4038e33a95.
Subsystems affected by this patch series:
mm/madvise
kbuild
mm/pagemap
mm/readahead
mm/memcg
mm/userfaultfd
vfs-akpm
mm/madvise
Subsystem: mm/madvise
Eric Dumazet <edumazet@google.com>:
mm/madvise: fix memory leak from process_madvise
Subsystem: kbuild
Nick Desaulniers <ndesaulniers@google.com>:
compiler-clang: remove version check for BPF Tracing
Subsystem: mm/pagemap
Dan Williams <dan.j.williams@intel.com>:
mm: fix phys_to_target_node() and memory_add_physaddr_to_nid() exports
Subsystem: mm/readahead
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
mm: fix readahead_page_batch for retry entries
Subsystem: mm/memcg
Muchun Song <songmuchun@bytedance.com>:
mm: memcg/slab: fix root memcg vmstats
Subsystem: mm/userfaultfd
Gerald Schaefer <gerald.schaefer@linux.ibm.com>:
mm/userfaultfd: do not access vma->vm_mm after calling handle_userfault()
Subsystem: vfs-akpm
Yicong Yang <yangyicong@hisilicon.com>:
libfs: fix error cast of negative value in simple_attr_write()
Subsystem: mm/madvise
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
mm: fix madvise WILLNEED performance problem
arch/ia64/include/asm/sparsemem.h | 6 ++++++
arch/powerpc/include/asm/mmzone.h | 5 +++++
arch/powerpc/include/asm/sparsemem.h | 5 ++---
arch/powerpc/mm/mem.c | 1 +
arch/x86/include/asm/sparsemem.h | 10 ++++++++++
arch/x86/mm/numa.c | 2 ++
drivers/dax/Kconfig | 1 -
fs/libfs.c | 6 ++++--
include/linux/compiler-clang.h | 2 ++
include/linux/memory_hotplug.h | 14 --------------
include/linux/numa.h | 30 +++++++++++++++++++++++++++++-
include/linux/pagemap.h | 2 ++
mm/huge_memory.c | 9 ++++-----
mm/madvise.c | 4 +---
mm/memcontrol.c | 9 +++++++--
mm/memory_hotplug.c | 18 ------------------
16 files changed, 75 insertions(+), 49 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-11-14 6:51 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-11-14 6:51 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
14 patches, based on 9e6a39eae450b81c8b2c8cbbfbdf8218e9b40c81.
Subsystems affected by this patch series:
mm/migration
mm/vmscan
mailmap
mm/slub
mm/gup
kbuild
reboot
kernel/watchdog
mm/memcg
mm/hugetlbfs
panic
ocfs2
Subsystem: mm/migration
Zi Yan <ziy@nvidia.com>:
mm/compaction: count pages and stop correctly during page isolation
mm/compaction: stop isolation if too many pages are isolated and we have pages to migrate
Subsystem: mm/vmscan
Nicholas Piggin <npiggin@gmail.com>:
mm/vmscan: fix NR_ISOLATED_FILE corruption on 64-bit
Subsystem: mailmap
Dmitry Baryshkov <dbaryshkov@gmail.com>:
mailmap: fix entry for Dmitry Baryshkov/Eremin-Solenikov
Subsystem: mm/slub
Laurent Dufour <ldufour@linux.ibm.com>:
mm/slub: fix panic in slab_alloc_node()
Subsystem: mm/gup
Jason Gunthorpe <jgg@nvidia.com>:
mm/gup: use unpin_user_pages() in __gup_longterm_locked()
Subsystem: kbuild
Arvind Sankar <nivedita@alum.mit.edu>:
compiler.h: fix barrier_data() on clang
Subsystem: reboot
Matteo Croce <mcroce@microsoft.com>:
Patch series "fix parsing of reboot= cmdline", v3:
Revert "kernel/reboot.c: convert simple_strtoul to kstrtoint"
reboot: fix overflow parsing reboot cpu number
Subsystem: kernel/watchdog
Santosh Sivaraj <santosh@fossix.org>:
kernel/watchdog: fix watchdog_allowed_mask not used warning
Subsystem: mm/memcg
Muchun Song <songmuchun@bytedance.com>:
mm: memcontrol: fix missing wakeup polling thread
Subsystem: mm/hugetlbfs
Mike Kravetz <mike.kravetz@oracle.com>:
hugetlbfs: fix anon huge page migration race
Subsystem: panic
Christophe Leroy <christophe.leroy@csgroup.eu>:
panic: don't dump stack twice on warn
Subsystem: ocfs2
Wengang Wang <wen.gang.wang@oracle.com>:
ocfs2: initialize ip_next_orphan
.mailmap | 5 +-
fs/ocfs2/super.c | 1
include/asm-generic/barrier.h | 1
include/linux/compiler-clang.h | 6 --
include/linux/compiler-gcc.h | 19 --------
include/linux/compiler.h | 18 +++++++-
include/linux/memcontrol.h | 11 ++++-
kernel/panic.c | 3 -
kernel/reboot.c | 28 ++++++------
kernel/watchdog.c | 4 -
mm/compaction.c | 12 +++--
mm/gup.c | 14 ++++--
mm/hugetlb.c | 90 ++---------------------------------------
mm/memory-failure.c | 36 +++++++---------
mm/migrate.c | 46 +++++++++++---------
mm/rmap.c | 5 --
mm/slub.c | 2
mm/vmscan.c | 5 +-
18 files changed, 119 insertions(+), 187 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-11-02 1:06 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-11-02 1:06 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
15 patches, based on 3cea11cd5e3b00d91caf0b4730194039b45c5891.
Subsystems affected by this patch series:
mm/memremap
mm/memcg
mm/slab-generic
mm/kasan
mm/mempolicy
signals
lib
mm/pagecache
kthread
mm/oom-kill
mm/pagemap
epoll
core-kernel
Subsystem: mm/memremap
Ralph Campbell <rcampbell@nvidia.com>:
mm/mremap_pages: fix static key devmap_managed_key updates
Subsystem: mm/memcg
Mike Kravetz <mike.kravetz@oracle.com>:
hugetlb_cgroup: fix reservation accounting
zhongjiang-ali <zhongjiang-ali@linux.alibaba.com>:
mm: memcontrol: correct the NR_ANON_THPS counter of hierarchical memcg
Roman Gushchin <guro@fb.com>:
mm: memcg: link page counters to root if use_hierarchy is false
Subsystem: mm/slab-generic
Subsystem: mm/kasan
Andrey Konovalov <andreyknvl@google.com>:
kasan: adopt KUNIT tests to SW_TAGS mode
Subsystem: mm/mempolicy
Shijie Luo <luoshijie1@huawei.com>:
mm: mempolicy: fix potential pte_unmap_unlock pte error
Subsystem: signals
Oleg Nesterov <oleg@redhat.com>:
ptrace: fix task_join_group_stop() for the case when current is traced
Subsystem: lib
Vasily Gorbik <gor@linux.ibm.com>:
lib/crc32test: remove extra local_irq_disable/enable
Subsystem: mm/pagecache
Jason Yan <yanaijie@huawei.com>:
mm/truncate.c: make __invalidate_mapping_pages() static
Subsystem: kthread
Zqiang <qiang.zhang@windriver.com>:
kthread_worker: prevent queuing delayed work from timer_fn when it is being canceled
Subsystem: mm/oom-kill
Charles Haithcock <chaithco@redhat.com>:
mm, oom: keep oom_adj under or at upper limit when printing
Subsystem: mm/pagemap
Jason Gunthorpe <jgg@nvidia.com>:
mm: always have io_remap_pfn_range() set pgprot_decrypted()
Subsystem: epoll
Soheil Hassas Yeganeh <soheil@google.com>:
epoll: check ep_events_available() upon timeout
epoll: add a selftest for epoll timeout race
Subsystem: core-kernel
Lukas Bulwahn <lukas.bulwahn@gmail.com>:
kernel/hung_task.c: make type annotations consistent
fs/eventpoll.c | 16 +
fs/proc/base.c | 2
include/linux/mm.h | 9
include/linux/pgtable.h | 4
kernel/hung_task.c | 3
kernel/kthread.c | 3
kernel/signal.c | 19 -
lib/crc32test.c | 4
lib/test_kasan.c | 149 +++++++---
mm/hugetlb.c | 20 -
mm/memcontrol.c | 25 +
mm/mempolicy.c | 6
mm/memremap.c | 39 +-
mm/truncate.c | 2
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c | 95 ++++++
15 files changed, 290 insertions(+), 106 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-10-17 23:13 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-10-17 23:13 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
40 patches, based on 9d9af1007bc08971953ae915d88dc9bb21344b53.
Subsystems affected by this patch series:
ia64
mm/memcg
mm/migration
mm/pagemap
mm/gup
mm/madvise
mm/vmalloc
misc
Subsystem: ia64
Krzysztof Kozlowski <krzk@kernel.org>:
ia64: fix build error with !COREDUMP
Subsystem: mm/memcg
Roman Gushchin <guro@fb.com>:
mm, memcg: rework remote charging API to support nesting
Patch series "mm: kmem: kernel memory accounting in an interrupt context":
mm: kmem: move memcg_kmem_bypass() calls to get_mem/obj_cgroup_from_current()
mm: kmem: remove redundant checks from get_obj_cgroup_from_current()
mm: kmem: prepare remote memcg charging infra for interrupt contexts
mm: kmem: enable kernel memcg accounting from interrupt contexts
Subsystem: mm/migration
Joonsoo Kim <iamjoonsoo.kim@lge.com>:
mm/memory-failure: remove a wrapper for alloc_migration_target()
mm/memory_hotplug: remove a wrapper for alloc_migration_target()
Miaohe Lin <linmiaohe@huawei.com>:
mm/migrate: avoid possible unnecessary process right check in kernel_move_pages()
Subsystem: mm/pagemap
"Liam R. Howlett" <Liam.Howlett@Oracle.com>:
mm/mmap: add inline vma_next() for readability of mmap code
mm/mmap: add inline munmap_vma_range() for code readability
Subsystem: mm/gup
Jann Horn <jannh@google.com>:
mm/gup_benchmark: take the mmap lock around GUP
binfmt_elf: take the mmap lock around find_extend_vma()
mm/gup: assert that the mmap lock is held in __get_user_pages()
John Hubbard <jhubbard@nvidia.com>:
Patch series "selftests/vm: gup_test, hmm-tests, assorted improvements", v2:
mm/gup_benchmark: rename to mm/gup_test
selftests/vm: use a common gup_test.h
selftests/vm: rename run_vmtests --> run_vmtests.sh
selftests/vm: minor cleanup: Makefile and gup_test.c
selftests/vm: only some gup_test items are really benchmarks
selftests/vm: gup_test: introduce the dump_pages() sub-test
selftests/vm: run_vmtests.sh: update and clean up gup_test invocation
selftests/vm: hmm-tests: remove the libhugetlbfs dependency
selftests/vm: 10x speedup for hmm-tests
Subsystem: mm/madvise
Minchan Kim <minchan@kernel.org>:
Patch series "introduce memory hinting API for external process", v9:
mm/madvise: pass mm to do_madvise
pid: move pidfd_get_pid() to pid.c
mm/madvise: introduce process_madvise() syscall: an external memory hinting API
Subsystem: mm/vmalloc
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
Patch series "remove alloc_vm_area", v4:
mm: update the documentation for vfree
Christoph Hellwig <hch@lst.de>:
mm: add a VM_MAP_PUT_PAGES flag for vmap
mm: add a vmap_pfn function
mm: allow a NULL fn callback in apply_to_page_range
zsmalloc: switch from alloc_vm_area to get_vm_area
drm/i915: use vmap in shmem_pin_map
drm/i915: stop using kmap in i915_gem_object_map
drm/i915: use vmap in i915_gem_object_map
xen/xenbus: use apply_to_page_range directly in xenbus_map_ring_pv
x86/xen: open code alloc_vm_area in arch_gnttab_valloc
mm: remove alloc_vm_area
Patch series "two small vmalloc cleanups":
mm: cleanup the gfp_mask handling in __vmalloc_area_node
mm: remove the filename in the top of file comment in vmalloc.c
Subsystem: misc
Tian Tao <tiantao6@hisilicon.com>:
mm: remove duplicate include statement in mmu.c
Documentation/core-api/pin_user_pages.rst | 8
arch/alpha/kernel/syscalls/syscall.tbl | 1
arch/arm/mm/mmu.c | 1
arch/arm/tools/syscall.tbl | 1
arch/arm64/include/asm/unistd.h | 2
arch/arm64/include/asm/unistd32.h | 2
arch/ia64/kernel/Makefile | 2
arch/ia64/kernel/syscalls/syscall.tbl | 1
arch/m68k/kernel/syscalls/syscall.tbl | 1
arch/microblaze/kernel/syscalls/syscall.tbl | 1
arch/mips/kernel/syscalls/syscall_n32.tbl | 1
arch/mips/kernel/syscalls/syscall_n64.tbl | 1
arch/mips/kernel/syscalls/syscall_o32.tbl | 1
arch/parisc/kernel/syscalls/syscall.tbl | 1
arch/powerpc/kernel/syscalls/syscall.tbl | 1
arch/s390/configs/debug_defconfig | 2
arch/s390/configs/defconfig | 2
arch/s390/kernel/syscalls/syscall.tbl | 1
arch/sh/kernel/syscalls/syscall.tbl | 1
arch/sparc/kernel/syscalls/syscall.tbl | 1
arch/x86/entry/syscalls/syscall_32.tbl | 1
arch/x86/entry/syscalls/syscall_64.tbl | 1
arch/x86/xen/grant-table.c | 27 +-
arch/xtensa/kernel/syscalls/syscall.tbl | 1
drivers/gpu/drm/i915/Kconfig | 1
drivers/gpu/drm/i915/gem/i915_gem_pages.c | 136 ++++------
drivers/gpu/drm/i915/gt/shmem_utils.c | 78 +-----
drivers/xen/xenbus/xenbus_client.c | 30 +-
fs/binfmt_elf.c | 3
fs/buffer.c | 6
fs/io_uring.c | 2
fs/notify/fanotify/fanotify.c | 5
fs/notify/inotify/inotify_fsnotify.c | 5
include/linux/memcontrol.h | 12
include/linux/mm.h | 2
include/linux/pid.h | 1
include/linux/sched/mm.h | 43 +--
include/linux/syscalls.h | 2
include/linux/vmalloc.h | 7
include/uapi/asm-generic/unistd.h | 4
kernel/exit.c | 19 -
kernel/pid.c | 19 +
kernel/sys_ni.c | 1
mm/Kconfig | 24 +
mm/Makefile | 2
mm/gup.c | 2
mm/gup_benchmark.c | 225 ------------------
mm/gup_test.c | 295 +++++++++++++++++++++--
mm/gup_test.h | 40 ++-
mm/madvise.c | 125 ++++++++--
mm/memcontrol.c | 83 ++++--
mm/memory-failure.c | 18 -
mm/memory.c | 16 -
mm/memory_hotplug.c | 46 +--
mm/migrate.c | 71 +++--
mm/mmap.c | 74 ++++-
mm/nommu.c | 7
mm/percpu.c | 3
mm/slab.h | 3
mm/vmalloc.c | 147 +++++------
mm/zsmalloc.c | 10
tools/testing/selftests/vm/.gitignore | 3
tools/testing/selftests/vm/Makefile | 40 ++-
tools/testing/selftests/vm/check_config.sh | 31 ++
tools/testing/selftests/vm/config | 2
tools/testing/selftests/vm/gup_benchmark.c | 143 -----------
tools/testing/selftests/vm/gup_test.c | 260 ++++++++++++++++++--
tools/testing/selftests/vm/hmm-tests.c | 12
tools/testing/selftests/vm/run_vmtests | 334 --------------------------
tools/testing/selftests/vm/run_vmtests.sh | 350 +++++++++++++++++++++++++++-
70 files changed, 1580 insertions(+), 1224 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2020-10-16 2:40 incoming Andrew Morton
@ 2020-10-16 3:03 ` Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-10-16 3:03 UTC (permalink / raw)
To: Linus Torvalds, mm-commits, linux-mm
And... I forgot to set in-reply-to :(
Shall resend, omitting linux-mm.
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-10-16 2:40 Andrew Morton
2020-10-16 3:03 ` incoming Andrew Morton
0 siblings, 1 reply; 409+ messages in thread
From: Andrew Morton @ 2020-10-16 2:40 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
- most of the rest of mm/
- various other subsystems
156 patches, based on 578a7155c5a1894a789d4ece181abf9d25dc6b0d.
Subsystems affected by this patch series:
mm/dax
mm/debug
mm/thp
mm/readahead
mm/page-poison
mm/util
mm/memory-hotplug
mm/zram
mm/cleanups
misc
core-kernel
get_maintainer
MAINTAINERS
lib
bitops
checkpatch
binfmt
ramfs
autofs
nilfs
rapidio
panic
relay
kgdb
ubsan
romfs
fault-injection
Subsystem: mm/dax
Dan Williams <dan.j.williams@intel.com>:
device-dax/kmem: fix resource release
Subsystem: mm/debug
"Aneesh Kumar K.V" <aneesh.kumar@linux.ibm.com>:
Patch series "mm/debug_vm_pgtable fixes", v4:
powerpc/mm: add DEBUG_VM WARN for pmd_clear
powerpc/mm: move setting pte specific flags to pfn_pte
mm/debug_vm_pgtable/ppc64: avoid setting top bits in radom value
mm/debug_vm_pgtables/hugevmap: use the arch helper to identify huge vmap support.
mm/debug_vm_pgtable/savedwrite: enable savedwrite test with CONFIG_NUMA_BALANCING
mm/debug_vm_pgtable/THP: mark the pte entry huge before using set_pmd/pud_at
mm/debug_vm_pgtable/set_pte/pmd/pud: don't use set_*_at to update an existing pte entry
mm/debug_vm_pgtable/locks: move non page table modifying test together
mm/debug_vm_pgtable/locks: take correct page table lock
mm/debug_vm_pgtable/thp: use page table depost/withdraw with THP
mm/debug_vm_pgtable/pmd_clear: don't use pmd/pud_clear on pte entries
mm/debug_vm_pgtable/hugetlb: disable hugetlb test on ppc64
mm/debug_vm_pgtable: avoid none pte in pte_clear_test
mm/debug_vm_pgtable: avoid doing memory allocation with pgtable_t mapped.
Subsystem: mm/thp
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
Patch series "Fix read-only THP for non-tmpfs filesystems":
XArray: add xa_get_order
XArray: add xas_split
mm/filemap: fix storing to a THP shadow entry
Patch series "Remove assumptions of THP size":
mm/filemap: fix page cache removal for arbitrary sized THPs
mm/memory: remove page fault assumption of compound page size
mm/page_owner: change split_page_owner to take a count
"Kirill A. Shutemov" <kirill@shutemov.name>:
mm/huge_memory: fix total_mapcount assumption of page size
mm/huge_memory: fix split assumption of page size
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
mm/huge_memory: fix page_trans_huge_mapcount assumption of THP size
mm/huge_memory: fix can_split_huge_page assumption of THP size
mm/rmap: fix assumptions of THP size
mm/truncate: fix truncation for pages of arbitrary size
mm/page-writeback: support tail pages in wait_for_stable_page
mm/vmscan: allow arbitrary sized pages to be paged out
fs: add a filesystem flag for THPs
fs: do not update nr_thps for mappings which support THPs
Huang Ying <ying.huang@intel.com>:
mm: fix a race during THP splitting
Subsystem: mm/readahead
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
Patch series "Readahead patches for 5.9/5.10":
mm/readahead: add DEFINE_READAHEAD
mm/readahead: make page_cache_ra_unbounded take a readahead_control
mm/readahead: make do_page_cache_ra take a readahead_control
David Howells <dhowells@redhat.com>:
mm/readahead: make ondemand_readahead take a readahead_control
mm/readahead: pass readahead_control to force_page_cache_ra
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
mm/readahead: add page_cache_sync_ra and page_cache_async_ra
David Howells <dhowells@redhat.com>:
mm/filemap: fold ra_submit into do_sync_mmap_readahead
mm/readahead: pass a file_ra_state into force_page_cache_ra
Subsystem: mm/page-poison
Naoya Horiguchi <naoya.horiguchi@nec.com>:
Patch series "HWPOISON: soft offline rework", v7:
mm,hwpoison: cleanup unused PageHuge() check
mm, hwpoison: remove recalculating hpage
mm,hwpoison-inject: don't pin for hwpoison_filter
Oscar Salvador <osalvador@suse.de>:
mm,hwpoison: unexport get_hwpoison_page and make it static
mm,hwpoison: refactor madvise_inject_error
mm,hwpoison: kill put_hwpoison_page
mm,hwpoison: unify THP handling for hard and soft offline
mm,hwpoison: rework soft offline for free pages
mm,hwpoison: rework soft offline for in-use pages
mm,hwpoison: refactor soft_offline_huge_page and __soft_offline_page
mm,hwpoison: return 0 if the page is already poisoned in soft-offline
Naoya Horiguchi <naoya.horiguchi@nec.com>:
mm,hwpoison: introduce MF_MSG_UNSPLIT_THP
mm,hwpoison: double-check page count in __get_any_page()
Oscar Salvador <osalvador@suse.de>:
mm,hwpoison: try to narrow window race for free pages
Mateusz Nosek <mateusznosek0@gmail.com>:
mm/page_poison.c: replace bool variable with static key
Miaohe Lin <linmiaohe@huawei.com>:
mm/vmstat.c: use helper macro abs()
Subsystem: mm/util
Bartosz Golaszewski <bgolaszewski@baylibre.com>:
mm/util.c: update the kerneldoc for kstrdup_const()
Jann Horn <jannh@google.com>:
mm/mmu_notifier: fix mmget() assert in __mmu_interval_notifier_insert
Subsystem: mm/memory-hotplug
David Hildenbrand <david@redhat.com>:
Patch series "mm/memory_hotplug: online_pages()/offline_pages() cleanups", v2:
mm/memory_hotplug: inline __offline_pages() into offline_pages()
mm/memory_hotplug: enforce section granularity when onlining/offlining
mm/memory_hotplug: simplify page offlining
mm/page_alloc: simplify __offline_isolated_pages()
mm/memory_hotplug: drop nr_isolate_pageblock in offline_pages()
mm/page_isolation: simplify return value of start_isolate_page_range()
mm/memory_hotplug: simplify page onlining
mm/page_alloc: drop stale pageblock comment in memmap_init_zone*()
mm: pass migratetype into memmap_init_zone() and move_pfn_range_to_zone()
mm/memory_hotplug: mark pageblocks MIGRATE_ISOLATE while onlining memory
Patch series "selective merging of system ram resources", v4:
kernel/resource: make release_mem_region_adjustable() never fail
kernel/resource: move and rename IORESOURCE_MEM_DRIVER_MANAGED
mm/memory_hotplug: guard more declarations by CONFIG_MEMORY_HOTPLUG
mm/memory_hotplug: prepare passing flags to add_memory() and friends
mm/memory_hotplug: MEMHP_MERGE_RESOURCE to specify merging of System RAM resources
virtio-mem: try to merge system ram resources
xen/balloon: try to merge system ram resources
hv_balloon: try to merge system ram resources
kernel/resource: make iomem_resource implicit in release_mem_region_adjustable()
Laurent Dufour <ldufour@linux.ibm.com>:
mm: don't panic when links can't be created in sysfs
David Hildenbrand <david@redhat.com>:
Patch series "mm: place pages to the freelist tail when onlining and undoing isolation", v2:
mm/page_alloc: convert "report" flag of __free_one_page() to a proper flag
mm/page_alloc: place pages to tail in __putback_isolated_page()
mm/page_alloc: move pages to tail in move_to_free_list()
mm/page_alloc: place pages to tail in __free_pages_core()
mm/memory_hotplug: update comment regarding zone shuffling
Subsystem: mm/zram
Douglas Anderson <dianders@chromium.org>:
zram: failing to decompress is WARN_ON worthy
Subsystem: mm/cleanups
YueHaibing <yuehaibing@huawei.com>:
mm/slab.h: remove duplicate include
Wei Yang <richard.weiyang@linux.alibaba.com>:
mm/page_reporting.c: drop stale list head check in page_reporting_cycle
Ira Weiny <ira.weiny@intel.com>:
mm/highmem.c: clean up endif comments
Yu Zhao <yuzhao@google.com>:
mm: use self-explanatory macros rather than "2"
Miaohe Lin <linmiaohe@huawei.com>:
mm: fix some broken comments
Chen Tao <chentao3@hotmail.com>:
mm: fix some comments formatting
Xiaofei Tan <tanxiaofei@huawei.com>:
mm/workingset.c: fix some doc warnings
Miaohe Lin <linmiaohe@huawei.com>:
mm: use helper function put_write_access()
Mike Rapoport <rppt@linux.ibm.com>:
include/linux/mmzone.h: remove unused early_pfn_valid()
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
mm: rename page_order() to buddy_order()
Subsystem: misc
Randy Dunlap <rdunlap@infradead.org>:
fs: configfs: delete repeated words in comments
Andy Shevchenko <andriy.shevchenko@linux.intel.com>:
kernel.h: split out min()/max() et al. helpers
Subsystem: core-kernel
Liao Pingfang <liao.pingfang@zte.com.cn>:
kernel/sys.c: replace do_brk with do_brk_flags in comment of prctl_set_mm_map()
Randy Dunlap <rdunlap@infradead.org>:
kernel/: fix repeated words in comments
kernel: acct.c: fix some kernel-doc nits
Subsystem: get_maintainer
Joe Perches <joe@perches.com>:
get_maintainer: add test for file in VCS
Subsystem: MAINTAINERS
Joe Perches <joe@perches.com>:
get_maintainer: exclude MAINTAINERS file(s) from --git-fallback
Jarkko Sakkinen <jarkko.sakkinen@linux.intel.com>:
MAINTAINERS: jarkko.sakkinen@linux.intel.com -> jarkko@kernel.org
Subsystem: lib
Randy Dunlap <rdunlap@infradead.org>:
lib: bitmap: delete duplicated words
lib: libcrc32c: delete duplicated words
lib: decompress_bunzip2: delete duplicated words
lib: dynamic_queue_limits: delete duplicated words + fix typo
lib: earlycpio: delete duplicated words
lib: radix-tree: delete duplicated words
lib: syscall: delete duplicated words
lib: test_sysctl: delete duplicated words
lib/mpi/mpi-bit.c: fix spello of "functions"
Stephen Boyd <swboyd@chromium.org>:
lib/idr.c: document calling context for IDA APIs mustn't use locks
lib/idr.c: document that ida_simple_{get,remove}() are deprecated
Christophe JAILLET <christophe.jaillet@wanadoo.fr>:
lib/scatterlist.c: avoid a double memset
Miaohe Lin <linmiaohe@huawei.com>:
lib/percpu_counter.c: use helper macro abs()
Andy Shevchenko <andriy.shevchenko@linux.intel.com>:
include/linux/list.h: add a macro to test if entry is pointing to the head
Dan Carpenter <dan.carpenter@oracle.com>:
lib/test_hmm.c: fix an error code in dmirror_allocate_chunk()
Tobias Jordan <kernel@cdqe.de>:
lib/crc32.c: fix trivial typo in preprocessor condition
Subsystem: bitops
Wei Yang <richard.weiyang@linux.alibaba.com>:
bitops: simplify get_count_order_long()
bitops: use the same mechanism for get_count_order[_long]
Subsystem: checkpatch
Jerome Forissier <jerome@forissier.org>:
checkpatch: add --kconfig-prefix
Joe Perches <joe@perches.com>:
checkpatch: move repeated word test
checkpatch: add test for comma use that should be semicolon
Rikard Falkeborn <rikard.falkeborn@gmail.com>:
const_structs.checkpatch: add phy_ops
Nicolas Boichat <drinkcat@chromium.org>:
checkpatch: warn if trace_printk and friends are called
Rikard Falkeborn <rikard.falkeborn@gmail.com>:
const_structs.checkpatch: add pinctrl_ops and pinmux_ops
Joe Perches <joe@perches.com>:
checkpatch: warn on self-assignments
checkpatch: allow not using -f with files that are in git
Dwaipayan Ray <dwaipayanray1@gmail.com>:
checkpatch: extend author Signed-off-by check for split From: header
Joe Perches <joe@perches.com>:
checkpatch: emit a warning on embedded filenames
Dwaipayan Ray <dwaipayanray1@gmail.com>:
checkpatch: fix multi-statement macro checks for while blocks.
Łukasz Stelmach <l.stelmach@samsung.com>:
checkpatch: fix false positive on empty block comment lines
Dwaipayan Ray <dwaipayanray1@gmail.com>:
checkpatch: add new warnings to author signoff checks.
Subsystem: binfmt
Chris Kennelly <ckennelly@google.com>:
Patch series "Selecting Load Addresses According to p_align", v3:
fs/binfmt_elf: use PT_LOAD p_align values for suitable start address
tools/testing/selftests: add self-test for verifying load alignment
Jann Horn <jannh@google.com>:
Patch series "Fix ELF / FDPIC ELF core dumping, and use mmap_lock properly in there", v5:
binfmt_elf_fdpic: stop using dump_emit() on user pointers on !MMU
coredump: let dump_emit() bail out on short writes
coredump: refactor page range dumping into common helper
coredump: rework elf/elf_fdpic vma_dump_size() into common helper
binfmt_elf, binfmt_elf_fdpic: use a VMA list snapshot
mm/gup: take mmap_lock in get_dump_page()
mm: remove the now-unnecessary mmget_still_valid() hack
Subsystem: ramfs
Matthew Wilcox (Oracle) <willy@infradead.org>:
ramfs: fix nommu mmap with gaps in the page cache
Subsystem: autofs
Matthew Wilcox <willy@infradead.org>:
autofs: harden ioctl table
Subsystem: nilfs
Wang Hai <wanghai38@huawei.com>:
nilfs2: fix some kernel-doc warnings for nilfs2
Subsystem: rapidio
Souptick Joarder <jrdr.linux@gmail.com>:
rapidio: fix error handling path
Jing Xiangfeng <jingxiangfeng@huawei.com>:
rapidio: fix the missed put_device() for rio_mport_add_riodev
Subsystem: panic
Alexey Kardashevskiy <aik@ozlabs.ru>:
panic: dump registers on panic_on_warn
Subsystem: relay
Sudip Mukherjee <sudipm.mukherjee@gmail.com>:
kernel/relay.c: drop unneeded initialization
Subsystem: kgdb
Ritesh Harjani <riteshh@linux.ibm.com>:
scripts/gdb/proc: add struct mount & struct super_block addr in lx-mounts command
scripts/gdb/tasks: add headers and improve spacing format
Subsystem: ubsan
Elena Petrova <lenaptr@google.com>:
sched.h: drop in_ubsan field when UBSAN is in trap mode
George Popescu <georgepope@android.com>:
ubsan: introduce CONFIG_UBSAN_LOCAL_BOUNDS for Clang
Subsystem: romfs
Libing Zhou <libing.zhou@nokia-sbell.com>:
ROMFS: support inode blocks calculation
Subsystem: fault-injection
Albert van der Linde <alinde@google.com>:
Patch series "add fault injection to user memory access", v3:
lib, include/linux: add usercopy failure capability
lib, uaccess: add failure injection to usercopy functions
.mailmap | 1
Documentation/admin-guide/kernel-parameters.txt | 1
Documentation/core-api/xarray.rst | 14
Documentation/fault-injection/fault-injection.rst | 7
MAINTAINERS | 6
arch/ia64/mm/init.c | 4
arch/powerpc/include/asm/book3s/64/pgtable.h | 29 +
arch/powerpc/include/asm/nohash/pgtable.h | 5
arch/powerpc/mm/pgtable.c | 5
arch/powerpc/platforms/powernv/memtrace.c | 2
arch/powerpc/platforms/pseries/hotplug-memory.c | 2
drivers/acpi/acpi_memhotplug.c | 3
drivers/base/memory.c | 3
drivers/base/node.c | 33 +-
drivers/block/zram/zram_drv.c | 2
drivers/dax/kmem.c | 50 ++-
drivers/hv/hv_balloon.c | 4
drivers/infiniband/core/uverbs_main.c | 3
drivers/rapidio/devices/rio_mport_cdev.c | 18 -
drivers/s390/char/sclp_cmd.c | 2
drivers/vfio/pci/vfio_pci.c | 38 +-
drivers/virtio/virtio_mem.c | 5
drivers/xen/balloon.c | 4
fs/autofs/dev-ioctl.c | 8
fs/binfmt_elf.c | 267 +++-------------
fs/binfmt_elf_fdpic.c | 176 ++--------
fs/configfs/dir.c | 2
fs/configfs/file.c | 2
fs/coredump.c | 238 +++++++++++++-
fs/ext4/verity.c | 4
fs/f2fs/verity.c | 4
fs/inode.c | 2
fs/nilfs2/bmap.c | 2
fs/nilfs2/cpfile.c | 6
fs/nilfs2/page.c | 1
fs/nilfs2/sufile.c | 4
fs/proc/task_mmu.c | 18 -
fs/ramfs/file-nommu.c | 2
fs/romfs/super.c | 1
fs/userfaultfd.c | 28 -
include/linux/bitops.h | 13
include/linux/blkdev.h | 1
include/linux/bvec.h | 6
include/linux/coredump.h | 13
include/linux/fault-inject-usercopy.h | 22 +
include/linux/fs.h | 28 -
include/linux/idr.h | 13
include/linux/ioport.h | 15
include/linux/jiffies.h | 3
include/linux/kernel.h | 150 ---------
include/linux/list.h | 29 +
include/linux/memory_hotplug.h | 42 +-
include/linux/minmax.h | 153 +++++++++
include/linux/mm.h | 5
include/linux/mmzone.h | 17 -
include/linux/node.h | 16
include/linux/nodemask.h | 2
include/linux/page-flags.h | 6
include/linux/page_owner.h | 6
include/linux/pagemap.h | 111 ++++++
include/linux/sched.h | 2
include/linux/sched/mm.h | 25 -
include/linux/uaccess.h | 12
include/linux/vmstat.h | 2
include/linux/xarray.h | 22 +
include/ras/ras_event.h | 3
kernel/acct.c | 10
kernel/cgroup/cpuset.c | 2
kernel/dma/direct.c | 2
kernel/fork.c | 4
kernel/futex.c | 2
kernel/irq/timings.c | 2
kernel/jump_label.c | 2
kernel/kcsan/encoding.h | 2
kernel/kexec_core.c | 2
kernel/kexec_file.c | 2
kernel/kthread.c | 2
kernel/livepatch/state.c | 2
kernel/panic.c | 12
kernel/pid_namespace.c | 2
kernel/power/snapshot.c | 2
kernel/range.c | 3
kernel/relay.c | 2
kernel/resource.c | 114 +++++--
kernel/smp.c | 2
kernel/sys.c | 2
kernel/user_namespace.c | 2
lib/Kconfig.debug | 7
lib/Kconfig.ubsan | 14
lib/Makefile | 1
lib/bitmap.c | 2
lib/crc32.c | 2
lib/decompress_bunzip2.c | 2
lib/dynamic_queue_limits.c | 4
lib/earlycpio.c | 2
lib/fault-inject-usercopy.c | 39 ++
lib/find_bit.c | 1
lib/hexdump.c | 1
lib/idr.c | 9
lib/iov_iter.c | 5
lib/libcrc32c.c | 2
lib/math/rational.c | 2
lib/math/reciprocal_div.c | 1
lib/mpi/mpi-bit.c | 2
lib/percpu_counter.c | 2
lib/radix-tree.c | 2
lib/scatterlist.c | 2
lib/strncpy_from_user.c | 3
lib/syscall.c | 2
lib/test_hmm.c | 2
lib/test_sysctl.c | 2
lib/test_xarray.c | 65 ++++
lib/usercopy.c | 5
lib/xarray.c | 208 ++++++++++++
mm/Kconfig | 2
mm/compaction.c | 6
mm/debug_vm_pgtable.c | 267 ++++++++--------
mm/filemap.c | 58 ++-
mm/gup.c | 73 ++--
mm/highmem.c | 4
mm/huge_memory.c | 47 +-
mm/hwpoison-inject.c | 18 -
mm/internal.h | 47 +-
mm/khugepaged.c | 2
mm/madvise.c | 52 ---
mm/memory-failure.c | 357 ++++++++++------------
mm/memory.c | 7
mm/memory_hotplug.c | 223 +++++--------
mm/memremap.c | 3
mm/migrate.c | 11
mm/mmap.c | 7
mm/mmu_notifier.c | 2
mm/page-writeback.c | 1
mm/page_alloc.c | 289 +++++++++++------
mm/page_isolation.c | 16
mm/page_owner.c | 10
mm/page_poison.c | 20 -
mm/page_reporting.c | 4
mm/readahead.c | 174 ++++------
mm/rmap.c | 10
mm/shmem.c | 2
mm/shuffle.c | 2
mm/slab.c | 2
mm/slab.h | 1
mm/slub.c | 2
mm/sparse.c | 2
mm/swap_state.c | 2
mm/truncate.c | 6
mm/util.c | 3
mm/vmscan.c | 5
mm/vmstat.c | 8
mm/workingset.c | 2
scripts/Makefile.ubsan | 10
scripts/checkpatch.pl | 238 ++++++++++----
scripts/const_structs.checkpatch | 3
scripts/gdb/linux/proc.py | 15
scripts/gdb/linux/tasks.py | 9
scripts/get_maintainer.pl | 9
tools/testing/selftests/exec/.gitignore | 1
tools/testing/selftests/exec/Makefile | 9
tools/testing/selftests/exec/load_address.c | 68 ++++
161 files changed, 2532 insertions(+), 1864 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-10-13 23:46 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-10-13 23:46 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
181 patches, based on 029f56db6ac248769f2c260bfaf3c3c0e23e904c.
Subsystems affected by this patch series:
kbuild
scripts
ntfs
ocfs2
vfs
mm/slab
mm/slub
mm/kmemleak
mm/dax
mm/debug
mm/pagecache
mm/fadvise
mm/gup
mm/swap
mm/memremap
mm/memcg
mm/selftests
mm/pagemap
mm/mincore
mm/hmm
mm/dma
mm/memory-failure
mm/vmalloc
mm/documentation
mm/kasan
mm/pagealloc
mm/hugetlb
mm/vmscan
mm/z3fold
mm/zbud
mm/compaction
mm/mempolicy
mm/mempool
mm/memblock
mm/oom-kill
mm/migration
Subsystem: kbuild
Nick Desaulniers <ndesaulniers@google.com>:
Patch series "set clang minimum version to 10.0.1", v3:
compiler-clang: add build check for clang 10.0.1
Revert "kbuild: disable clang's default use of -fmerge-all-constants"
Revert "arm64: bti: Require clang >= 10.0.1 for in-kernel BTI support"
Revert "arm64: vdso: Fix compilation with clang older than 8"
Partially revert "ARM: 8905/1: Emit __gnu_mcount_nc when using Clang 10.0.0 or newer"
Marco Elver <elver@google.com>:
kasan: remove mentions of unsupported Clang versions
Nick Desaulniers <ndesaulniers@google.com>:
compiler-gcc: improve version error
compiler.h: avoid escaped section names
export.h: fix section name for CONFIG_TRIM_UNUSED_KSYMS for Clang
Lukas Bulwahn <lukas.bulwahn@gmail.com>:
kbuild: doc: describe proper script invocation
Subsystem: scripts
Wang Qing <wangqing@vivo.com>:
scripts/spelling.txt: increase error-prone spell checking
Naoki Hayama <naoki.hayama@lineo.co.jp>:
scripts/spelling.txt: add "arbitrary" typo
Borislav Petkov <bp@suse.de>:
scripts/decodecode: add the capability to supply the program counter
Subsystem: ntfs
Rustam Kovhaev <rkovhaev@gmail.com>:
ntfs: add check for mft record size in superblock
Subsystem: ocfs2
Randy Dunlap <rdunlap@infradead.org>:
ocfs2: delete repeated words in comments
Gang He <ghe@suse.com>:
ocfs2: fix potential soft lockup during fstrim
Subsystem: vfs
Randy Dunlap <rdunlap@infradead.org>:
fs/xattr.c: fix kernel-doc warnings for setxattr & removexattr
Luo Jiaxing <luojiaxing@huawei.com>:
fs_parse: mark fs_param_bad_value() as static
Subsystem: mm/slab
Mateusz Nosek <mateusznosek0@gmail.com>:
mm/slab.c: clean code by removing redundant if condition
tangjianqiang <wyqt1985@gmail.com>:
include/linux/slab.h: fix a typo error in comment
Subsystem: mm/slub
Abel Wu <wuyun.wu@huawei.com>:
mm/slub.c: branch optimization in free slowpath
mm/slub: fix missing ALLOC_SLOWPATH stat when bulk alloc
mm/slub: make add_full() condition more explicit
Subsystem: mm/kmemleak
Davidlohr Bueso <dave@stgolabs.net>:
mm/kmemleak: rely on rcu for task stack scanning
Hui Su <sh_def@163.com>:
mm,kmemleak-test.c: move kmemleak-test.c to samples dir
Subsystem: mm/dax
Dan Williams <dan.j.williams@intel.com>:
Patch series "device-dax: Support sub-dividing soft-reserved ranges", v5:
x86/numa: cleanup configuration dependent command-line options
x86/numa: add 'nohmat' option
efi/fake_mem: arrange for a resource entry per efi_fake_mem instance
ACPI: HMAT: refactor hmat_register_target_device to hmem_register_device
resource: report parent to walk_iomem_res_desc() callback
mm/memory_hotplug: introduce default phys_to_target_node() implementation
ACPI: HMAT: attach a device for each soft-reserved range
device-dax: drop the dax_region.pfn_flags attribute
device-dax: move instance creation parameters to 'struct dev_dax_data'
device-dax: make pgmap optional for instance creation
device-dax/kmem: introduce dax_kmem_range()
device-dax/kmem: move resource name tracking to drvdata
device-dax/kmem: replace release_resource() with release_mem_region()
device-dax: add an allocation interface for device-dax instances
device-dax: introduce 'struct dev_dax' typed-driver operations
device-dax: introduce 'seed' devices
drivers/base: make device_find_child_by_name() compatible with sysfs inputs
device-dax: add resize support
mm/memremap_pages: convert to 'struct range'
mm/memremap_pages: support multiple ranges per invocation
device-dax: add dis-contiguous resource support
device-dax: introduce 'mapping' devices
Joao Martins <joao.m.martins@oracle.com>:
device-dax: make align a per-device property
Dan Williams <dan.j.williams@intel.com>:
device-dax: add an 'align' attribute
Joao Martins <joao.m.martins@oracle.com>:
dax/hmem: introduce dax_hmem.region_idle parameter
device-dax: add a range mapping allocation attribute
Subsystem: mm/debug
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
mm/debug.c: do not dereference i_ino blindly
John Hubbard <jhubbard@nvidia.com>:
mm, dump_page: rename head_mapcount() --> head_compound_mapcount()
Subsystem: mm/pagecache
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
Patch series "Return head pages from find_*_entry", v2:
mm: factor find_get_incore_page out of mincore_page
mm: use find_get_incore_page in memcontrol
mm: optimise madvise WILLNEED
proc: optimise smaps for shmem entries
i915: use find_lock_page instead of find_lock_entry
mm: convert find_get_entry to return the head page
mm/shmem: return head page from find_lock_entry
mm: add find_lock_head
mm/filemap: fix filemap_map_pages for THP
Subsystem: mm/fadvise
Yafang Shao <laoar.shao@gmail.com>:
mm, fadvise: improve the expensive remote LRU cache draining after FADV_DONTNEED
Subsystem: mm/gup
Barry Song <song.bao.hua@hisilicon.com>:
mm/gup_benchmark: update the documentation in Kconfig
mm/gup_benchmark: use pin_user_pages for FOLL_LONGTERM flag
mm/gup: don't permit users to call get_user_pages with FOLL_LONGTERM
John Hubbard <jhubbard@nvidia.com>:
mm/gup: protect unpin_user_pages() against npages==-ERRNO
Subsystem: mm/swap
Gao Xiang <hsiangkao@redhat.com>:
swap: rename SWP_FS to SWAP_FS_OPS to avoid ambiguity
Yu Zhao <yuzhao@google.com>:
mm: remove activate_page() from unuse_pte()
mm: remove superfluous __ClearPageActive()
Miaohe Lin <linmiaohe@huawei.com>:
mm/swap.c: fix confusing comment in release_pages()
mm/swap_slots.c: remove always zero and unused return value of enable_swap_slots_cache()
mm/page_io.c: remove useless out label in __swap_writepage()
mm/swap.c: fix incomplete comment in lru_cache_add_inactive_or_unevictable()
mm/swapfile.c: remove unnecessary goto out in _swap_info_get()
mm/swapfile.c: fix potential memory leak in sys_swapon
Subsystem: mm/memremap
Ira Weiny <ira.weiny@intel.com>:
mm/memremap.c: convert devmap static branch to {inc,dec}
Subsystem: mm/memcg
"Gustavo A. R. Silva" <gustavoars@kernel.org>:
mm: memcontrol: use flex_array_size() helper in memcpy()
mm: memcontrol: use the preferred form for passing the size of a structure type
Roman Gushchin <guro@fb.com>:
mm: memcg/slab: fix racy access to page->mem_cgroup in mem_cgroup_from_obj()
Miaohe Lin <linmiaohe@huawei.com>:
mm: memcontrol: correct the comment of mem_cgroup_iter()
Waiman Long <longman@redhat.com>:
Patch series "mm/memcg: Miscellaneous cleanups and streamlining", v2:
mm/memcg: clean up obsolete enum charge_type
mm/memcg: simplify mem_cgroup_get_max()
mm/memcg: unify swap and memsw page counters
Muchun Song <songmuchun@bytedance.com>:
mm: memcontrol: add the missing numa_stat interface for cgroup v2
Miaohe Lin <linmiaohe@huawei.com>:
mm/page_counter: correct the obsolete func name in the comment of page_counter_try_charge()
mm: memcontrol: reword obsolete comment of mem_cgroup_unmark_under_oom()
Bharata B Rao <bharata@linux.ibm.com>:
mm: memcg/slab: uncharge during kmem_cache_free_bulk()
Ralph Campbell <rcampbell@nvidia.com>:
mm/memcg: fix device private memcg accounting
Subsystem: mm/selftests
John Hubbard <jhubbard@nvidia.com>:
Patch series "selftests/vm: fix some minor aggravating factors in the Makefile":
selftests/vm: fix false build success on the second and later attempts
selftests/vm: fix incorrect gcc invocation in some cases
Subsystem: mm/pagemap
Matthew Wilcox <willy@infradead.org>:
mm: account PMD tables like PTE tables
Yanfei Xu <yanfei.xu@windriver.com>:
mm/memory.c: fix typo in __do_fault() comment
mm/memory.c: replace vmf->vma with variable vma
Wei Yang <richard.weiyang@linux.alibaba.com>:
mm/mmap: rename __vma_unlink_common() to __vma_unlink()
mm/mmap: leverage vma_rb_erase_ignore() to implement vma_rb_erase()
Chinwen Chang <chinwen.chang@mediatek.com>:
Patch series "Try to release mmap_lock temporarily in smaps_rollup", v4:
mmap locking API: add mmap_lock_is_contended()
mm: smaps*: extend smap_gather_stats to support specified beginning
mm: proc: smaps_rollup: do not stall write attempts on mmap_lock
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
Patch series "Fix PageDoubleMap":
mm: move PageDoubleMap bit
mm: simplify PageDoubleMap with PF_SECOND policy
Wei Yang <richard.weiyang@linux.alibaba.com>:
mm/mmap: leave adjust_next as virtual address instead of page frame number
Randy Dunlap <rdunlap@infradead.org>:
mm/memory.c: fix spello of "function"
Wei Yang <richard.weiyang@linux.alibaba.com>:
mm/mmap: not necessary to check mapping separately
mm/mmap: check on file instead of the rb_root_cached of its address_space
Miaohe Lin <linmiaohe@huawei.com>:
mm: use helper function mapping_allow_writable()
mm/mmap.c: use helper function allow_write_access() in __remove_shared_vm_struct()
Liao Pingfang <liao.pingfang@zte.com.cn>:
mm/mmap.c: replace do_brk with do_brk_flags in comment of insert_vm_struct()
Peter Xu <peterx@redhat.com>:
mm: remove src/dst mm parameter in copy_page_range()
Subsystem: mm/mincore
yuleixzhang <yulei.kernel@gmail.com>:
include/linux/huge_mm.h: remove mincore_huge_pmd declaration
Subsystem: mm/hmm
Ralph Campbell <rcampbell@nvidia.com>:
tools/testing/selftests/vm/hmm-tests.c: use the new SKIP() macro
lib/test_hmm.c: remove unused dmirror_zero_page
Subsystem: mm/dma
Andy Shevchenko <andriy.shevchenko@linux.intel.com>:
mm/dmapool.c: replace open-coded list_for_each_entry_safe()
mm/dmapool.c: replace hard coded function name with __func__
Subsystem: mm/memory-failure
Xianting Tian <tian.xianting@h3c.com>:
mm/memory-failure: do pgoff calculation before for_each_process()
Alex Shi <alex.shi@linux.alibaba.com>:
mm/memory-failure.c: remove unused macro `writeback'
Subsystem: mm/vmalloc
Hui Su <sh_def@163.com>:
mm/vmalloc.c: update the comment in __vmalloc_area_node()
mm/vmalloc.c: fix the comment of find_vm_area
Subsystem: mm/documentation
Alexander Gordeev <agordeev@linux.ibm.com>:
docs/vm: fix 'mm_count' vs 'mm_users' counter confusion
Subsystem: mm/kasan
Patricia Alfonso <trishalfonso@google.com>:
Patch series "KASAN-KUnit Integration", v14:
kasan/kunit: add KUnit Struct to Current Task
KUnit: KASAN Integration
KASAN: port KASAN Tests to KUnit
KASAN: Testing Documentation
David Gow <davidgow@google.com>:
mm: kasan: do not panic if both panic_on_warn and kasan_multishot set
Subsystem: mm/pagealloc
David Hildenbrand <david@redhat.com>:
Patch series "mm / virtio-mem: support ZONE_MOVABLE", v5:
mm/page_alloc: tweak comments in has_unmovable_pages()
mm/page_isolation: exit early when pageblock is isolated in set_migratetype_isolate()
mm/page_isolation: drop WARN_ON_ONCE() in set_migratetype_isolate()
mm/page_isolation: cleanup set_migratetype_isolate()
virtio-mem: don't special-case ZONE_MOVABLE
mm: document semantics of ZONE_MOVABLE
Li Xinhai <lixinhai.lxh@gmail.com>:
mm, isolation: avoid checking unmovable pages across pageblock boundary
Mateusz Nosek <mateusznosek0@gmail.com>:
mm/page_alloc.c: clean code by removing unnecessary initialization
mm/page_alloc.c: micro-optimization remove unnecessary branch
mm/page_alloc.c: fix early params garbage value accesses
mm/page_alloc.c: clean code by merging two functions
Yanfei Xu <yanfei.xu@windriver.com>:
mm/page_alloc.c: __perform_reclaim should return 'unsigned long'
Mateusz Nosek <mateusznosek0@gmail.com>:
mmzone: clean code by removing unused macro parameter
Ralph Campbell <rcampbell@nvidia.com>:
mm: move call to compound_head() in release_pages()
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
mm/page_alloc.c: fix freeing non-compound pages
Michal Hocko <mhocko@suse.com>:
include/linux/gfp.h: clarify usage of GFP_ATOMIC in !preemptible contexts
Subsystem: mm/hugetlb
Baoquan He <bhe@redhat.com>:
Patch series "mm/hugetlb: Small cleanup and improvement", v2:
mm/hugetlb.c: make is_hugetlb_entry_hwpoisoned return bool
mm/hugetlb.c: remove the unnecessary non_swap_entry()
doc/vm: fix typo in the hugetlb admin documentation
Wei Yang <richard.weiyang@linux.alibaba.com>:
Patch series "mm/hugetlb: code refine and simplification", v4:
mm/hugetlb: not necessary to coalesce regions recursively
mm/hugetlb: remove VM_BUG_ON(!nrg) in get_file_region_entry_from_cache()
mm/hugetlb: use list_splice to merge two list at once
mm/hugetlb: count file_region to be added when regions_needed != NULL
mm/hugetlb: a page from buddy is not on any list
mm/hugetlb: narrow the hugetlb_lock protection area during preparing huge page
mm/hugetlb: take the free hpage during the iteration directly
Mike Kravetz <mike.kravetz@oracle.com>:
hugetlb: add lockdep check for i_mmap_rwsem held in huge_pmd_share
Subsystem: mm/vmscan
Chunxin Zang <zangchunxin@bytedance.com>:
mm/vmscan: fix infinite loop in drop_slab_node
Hui Su <sh_def@163.com>:
mm/vmscan: fix comments for isolate_lru_page()
Subsystem: mm/z3fold
Hui Su <sh_def@163.com>:
mm/z3fold.c: use xx_zalloc instead xx_alloc and memset
Subsystem: mm/zbud
Xiang Chen <chenxiang66@hisilicon.com>:
mm/zbud: remove redundant initialization
Subsystem: mm/compaction
Mateusz Nosek <mateusznosek0@gmail.com>:
mm/compaction.c: micro-optimization remove unnecessary branch
include/linux/compaction.h: clean code by removing unused enum value
John Hubbard <jhubbard@nvidia.com>:
selftests/vm: 8x compaction_test speedup
Subsystem: mm/mempolicy
Wei Yang <richard.weiyang@linux.alibaba.com>:
mm/mempolicy: remove or narrow the lock on current
mm: remove unused alloc_page_vma_node()
Subsystem: mm/mempool
Miaohe Lin <linmiaohe@huawei.com>:
mm/mempool: add 'else' to split mutually exclusive case
Subsystem: mm/memblock
Mike Rapoport <rppt@linux.ibm.com>:
Patch series "memblock: seasonal cleaning^w cleanup", v3:
KVM: PPC: Book3S HV: simplify kvm_cma_reserve()
dma-contiguous: simplify cma_early_percent_memory()
arm, xtensa: simplify initialization of high memory pages
arm64: numa: simplify dummy_numa_init()
h8300, nds32, openrisc: simplify detection of memory extents
riscv: drop unneeded node initialization
mircoblaze: drop unneeded NUMA and sparsemem initializations
memblock: make for_each_memblock_type() iterator private
memblock: make memblock_debug and related functionality private
memblock: reduce number of parameters in for_each_mem_range()
arch, mm: replace for_each_memblock() with for_each_mem_pfn_range()
arch, drivers: replace for_each_membock() with for_each_mem_range()
x86/setup: simplify initrd relocation and reservation
x86/setup: simplify reserve_crashkernel()
memblock: remove unused memblock_mem_size()
memblock: implement for_each_reserved_mem_region() using __next_mem_region()
memblock: use separate iterators for memory and reserved regions
Subsystem: mm/oom-kill
Suren Baghdasaryan <surenb@google.com>:
mm, oom_adj: don't loop through tasks in __set_oom_adj when not necessary
Subsystem: mm/migration
Ralph Campbell <rcampbell@nvidia.com>:
mm/migrate: remove cpages-- in migrate_vma_finalize()
mm/migrate: remove obsolete comment about device public
.clang-format | 7
Documentation/admin-guide/cgroup-v2.rst | 69 +
Documentation/admin-guide/mm/hugetlbpage.rst | 2
Documentation/dev-tools/kasan.rst | 74 +
Documentation/dev-tools/kmemleak.rst | 2
Documentation/kbuild/makefiles.rst | 20
Documentation/vm/active_mm.rst | 2
Documentation/x86/x86_64/boot-options.rst | 4
MAINTAINERS | 2
Makefile | 9
arch/arm/Kconfig | 2
arch/arm/include/asm/tlb.h | 1
arch/arm/kernel/setup.c | 18
arch/arm/mm/init.c | 59 -
arch/arm/mm/mmu.c | 39
arch/arm/mm/pmsa-v7.c | 23
arch/arm/mm/pmsa-v8.c | 17
arch/arm/xen/mm.c | 7
arch/arm64/Kconfig | 2
arch/arm64/kernel/machine_kexec_file.c | 6
arch/arm64/kernel/setup.c | 4
arch/arm64/kernel/vdso/Makefile | 7
arch/arm64/mm/init.c | 11
arch/arm64/mm/kasan_init.c | 10
arch/arm64/mm/mmu.c | 11
arch/arm64/mm/numa.c | 15
arch/c6x/kernel/setup.c | 9
arch/h8300/kernel/setup.c | 8
arch/microblaze/mm/init.c | 23
arch/mips/cavium-octeon/dma-octeon.c | 14
arch/mips/kernel/setup.c | 31
arch/mips/netlogic/xlp/setup.c | 2
arch/nds32/kernel/setup.c | 8
arch/openrisc/kernel/setup.c | 9
arch/openrisc/mm/init.c | 8
arch/powerpc/kernel/fadump.c | 61 -
arch/powerpc/kexec/file_load_64.c | 16
arch/powerpc/kvm/book3s_hv_builtin.c | 12
arch/powerpc/kvm/book3s_hv_uvmem.c | 14
arch/powerpc/mm/book3s64/hash_utils.c | 16
arch/powerpc/mm/book3s64/radix_pgtable.c | 10
arch/powerpc/mm/kasan/kasan_init_32.c | 8
arch/powerpc/mm/mem.c | 31
arch/powerpc/mm/numa.c | 7
arch/powerpc/mm/pgtable_32.c | 8
arch/riscv/mm/init.c | 36
arch/riscv/mm/kasan_init.c | 10
arch/s390/kernel/setup.c | 27
arch/s390/mm/page-states.c | 6
arch/s390/mm/vmem.c | 7
arch/sh/mm/init.c | 9
arch/sparc/mm/init_64.c | 12
arch/x86/include/asm/numa.h | 8
arch/x86/kernel/e820.c | 16
arch/x86/kernel/setup.c | 56 -
arch/x86/mm/numa.c | 13
arch/x86/mm/numa_emulation.c | 3
arch/x86/xen/enlighten_pv.c | 2
arch/xtensa/mm/init.c | 55 -
drivers/acpi/numa/hmat.c | 76 -
drivers/acpi/numa/srat.c | 9
drivers/base/core.c | 2
drivers/bus/mvebu-mbus.c | 12
drivers/dax/Kconfig | 6
drivers/dax/Makefile | 3
drivers/dax/bus.c | 1237 +++++++++++++++++++++++----
drivers/dax/bus.h | 34
drivers/dax/dax-private.h | 74 +
drivers/dax/device.c | 164 +--
drivers/dax/hmem.c | 56 -
drivers/dax/hmem/Makefile | 8
drivers/dax/hmem/device.c | 100 ++
drivers/dax/hmem/hmem.c | 93 +-
drivers/dax/kmem.c | 236 ++---
drivers/dax/pmem/compat.c | 2
drivers/dax/pmem/core.c | 36
drivers/firmware/efi/x86_fake_mem.c | 12
drivers/gpu/drm/i915/gem/i915_gem_shmem.c | 4
drivers/gpu/drm/nouveau/nouveau_dmem.c | 15
drivers/irqchip/irq-gic-v3-its.c | 2
drivers/nvdimm/badrange.c | 26
drivers/nvdimm/claim.c | 13
drivers/nvdimm/nd.h | 3
drivers/nvdimm/pfn_devs.c | 13
drivers/nvdimm/pmem.c | 27
drivers/nvdimm/region.c | 21
drivers/pci/p2pdma.c | 12
drivers/virtio/virtio_mem.c | 47 -
drivers/xen/unpopulated-alloc.c | 45
fs/fs_parser.c | 2
fs/ntfs/inode.c | 6
fs/ocfs2/alloc.c | 6
fs/ocfs2/localalloc.c | 2
fs/proc/base.c | 3
fs/proc/task_mmu.c | 104 +-
fs/xattr.c | 22
include/acpi/acpi_numa.h | 14
include/kunit/test.h | 5
include/linux/acpi.h | 2
include/linux/compaction.h | 3
include/linux/compiler-clang.h | 8
include/linux/compiler-gcc.h | 2
include/linux/compiler.h | 2
include/linux/dax.h | 8
include/linux/export.h | 2
include/linux/fs.h | 4
include/linux/gfp.h | 6
include/linux/huge_mm.h | 3
include/linux/kasan.h | 6
include/linux/memblock.h | 90 +
include/linux/memcontrol.h | 13
include/linux/memory_hotplug.h | 23
include/linux/memremap.h | 15
include/linux/mm.h | 36
include/linux/mmap_lock.h | 5
include/linux/mmzone.h | 37
include/linux/numa.h | 11
include/linux/oom.h | 1
include/linux/page-flags.h | 42
include/linux/pagemap.h | 43
include/linux/range.h | 6
include/linux/sched.h | 4
include/linux/sched/coredump.h | 1
include/linux/slab.h | 2
include/linux/swap.h | 10
include/linux/swap_slots.h | 2
kernel/dma/contiguous.c | 11
kernel/fork.c | 25
kernel/resource.c | 11
lib/Kconfig.debug | 9
lib/Kconfig.kasan | 31
lib/Makefile | 5
lib/kunit/test.c | 13
lib/test_free_pages.c | 42
lib/test_hmm.c | 65 -
lib/test_kasan.c | 732 ++++++---------
lib/test_kasan_module.c | 111 ++
mm/Kconfig | 4
mm/Makefile | 1
mm/compaction.c | 5
mm/debug.c | 18
mm/dmapool.c | 46 -
mm/fadvise.c | 9
mm/filemap.c | 78 -
mm/gup.c | 44
mm/gup_benchmark.c | 23
mm/huge_memory.c | 4
mm/hugetlb.c | 100 +-
mm/internal.h | 3
mm/kasan/report.c | 34
mm/kmemleak-test.c | 99 --
mm/kmemleak.c | 8
mm/madvise.c | 21
mm/memblock.c | 102 --
mm/memcontrol.c | 262 +++--
mm/memory-failure.c | 5
mm/memory.c | 147 +--
mm/memory_hotplug.c | 10
mm/mempolicy.c | 8
mm/mempool.c | 18
mm/memremap.c | 344 ++++---
mm/migrate.c | 3
mm/mincore.c | 28
mm/mmap.c | 45
mm/oom_kill.c | 2
mm/page_alloc.c | 82 -
mm/page_counter.c | 2
mm/page_io.c | 14
mm/page_isolation.c | 41
mm/shmem.c | 19
mm/slab.c | 4
mm/slab.h | 50 -
mm/slub.c | 33
mm/sparse.c | 10
mm/swap.c | 14
mm/swap_slots.c | 3
mm/swap_state.c | 38
mm/swapfile.c | 12
mm/truncate.c | 58 -
mm/vmalloc.c | 6
mm/vmscan.c | 5
mm/z3fold.c | 3
mm/zbud.c | 1
samples/Makefile | 1
samples/kmemleak/Makefile | 3
samples/kmemleak/kmemleak-test.c | 99 ++
scripts/decodecode | 29
scripts/spelling.txt | 4
tools/testing/nvdimm/dax-dev.c | 28
tools/testing/nvdimm/test/iomap.c | 2
tools/testing/selftests/vm/Makefile | 17
tools/testing/selftests/vm/compaction_test.c | 11
tools/testing/selftests/vm/gup_benchmark.c | 14
tools/testing/selftests/vm/hmm-tests.c | 4
194 files changed, 4273 insertions(+), 2777 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-10-11 6:15 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-10-11 6:15 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
5 patches, based on da690031a5d6d50a361e3f19f3eeabd086a6f20d.
Subsystems affected by this patch series:
MAINTAINERS
mm/pagemap
mm/swap
mm/hugetlb
Subsystem: MAINTAINERS
Kees Cook <keescook@chromium.org>:
MAINTAINERS: change hardening mailing list
Antoine Tenart <atenart@kernel.org>:
MAINTAINERS: Antoine Tenart's email address
Subsystem: mm/pagemap
Miaohe Lin <linmiaohe@huawei.com>:
mm: mmap: Fix general protection fault in unlink_file_vma()
Subsystem: mm/swap
Minchan Kim <minchan@kernel.org>:
mm: validate inode in mapping_set_error()
Subsystem: mm/hugetlb
Vijay Balakrishna <vijayb@linux.microsoft.com>:
mm: khugepaged: recalculate min_free_kbytes after memory hotplug as expected by khugepaged
.mailmap | 4 +++-
MAINTAINERS | 8 ++++----
include/linux/khugepaged.h | 5 +++++
include/linux/pagemap.h | 3 ++-
mm/khugepaged.c | 13 +++++++++++--
mm/mmap.c | 6 +++++-
mm/page_alloc.c | 3 +++
7 files changed, 33 insertions(+), 9 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-10-03 5:20 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-10-03 5:20 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
3 patches, based on d3d45f8220d60a0b2aaaacf8fb2be4e6ffd9008e.
Subsystems affected by this patch series:
mm/slub
mm/cma
scripts
Subsystem: mm/slub
Eric Farman <farman@linux.ibm.com>:
mm, slub: restore initial kmem_cache flags
Subsystem: mm/cma
Joonsoo Kim <iamjoonsoo.kim@lge.com>:
mm/page_alloc: handle a missing case for memalloc_nocma_{save/restore} APIs
Subsystem: scripts
Eric Biggers <ebiggers@google.com>:
scripts/spelling.txt: fix malformed entry
mm/page_alloc.c | 19 ++++++++++++++++---
mm/slub.c | 6 +-----
scripts/spelling.txt | 2 +-
3 files changed, 18 insertions(+), 9 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-09-26 4:17 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-09-26 4:17 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
9 patches, based on 7c7ec3226f5f33f9c050d85ec20f18419c622ad6.
Subsystems affected by this patch series:
mm/thp
mm/memcg
mm/gup
mm/migration
lib
x86
mm/memory-hotplug
Subsystem: mm/thp
Gao Xiang <hsiangkao@redhat.com>:
mm, THP, swap: fix allocating cluster for swapfile by mistake
Subsystem: mm/memcg
Muchun Song <songmuchun@bytedance.com>:
mm: memcontrol: fix missing suffix of workingset_restore
Subsystem: mm/gup
Vasily Gorbik <gor@linux.ibm.com>:
mm/gup: fix gup_fast with dynamic page table folding
Subsystem: mm/migration
Zi Yan <ziy@nvidia.com>:
mm/migrate: correct thp migration stats
Subsystem: lib
Nick Desaulniers <ndesaulniers@google.com>:
lib/string.c: implement stpcpy
Jason Yan <yanaijie@huawei.com>:
lib/memregion.c: include memregion.h
Subsystem: x86
Mikulas Patocka <mpatocka@redhat.com>:
arch/x86/lib/usercopy_64.c: fix __copy_user_flushcache() cache writeback
Subsystem: mm/memory-hotplug
Laurent Dufour <ldufour@linux.ibm.com>:
Patch series "mm: fix memory to node bad links in sysfs", v3:
mm: replace memmap_context by meminit_context
mm: don't rely on system state to detect hot-plug operations
Documentation/admin-guide/cgroup-v2.rst | 25 ++++++---
arch/ia64/mm/init.c | 6 +-
arch/s390/include/asm/pgtable.h | 42 +++++++++++----
arch/x86/lib/usercopy_64.c | 2
drivers/base/node.c | 85 ++++++++++++++++++++------------
include/linux/mm.h | 2
include/linux/mmzone.h | 11 +++-
include/linux/node.h | 11 ++--
include/linux/pgtable.h | 10 +++
lib/memregion.c | 1
lib/string.c | 24 +++++++++
mm/gup.c | 18 +++---
mm/memcontrol.c | 4 -
mm/memory_hotplug.c | 5 +
mm/migrate.c | 7 +-
mm/page_alloc.c | 10 +--
mm/swapfile.c | 2
17 files changed, 181 insertions(+), 84 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-09-19 4:19 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-09-19 4:19 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
15 patches, based on 92ab97adeefccf375de7ebaad9d5b75d4125fe8b.
Subsystems affected by this patch series:
mailmap
mm/hotfixes
mm/thp
mm/memory-hotplug
misc
kcsan
Subsystem: mailmap
Kees Cook <keescook@chromium.org>:
mailmap: add older email addresses for Kees Cook
Subsystem: mm/hotfixes
Hugh Dickins <hughd@google.com>:
Patch series "mm: fixes to past from future testing":
ksm: reinstate memcg charge on copied pages
mm: migration of hugetlbfs page skip memcg
shmem: shmem_writepage() split unlikely i915 THP
mm: fix check_move_unevictable_pages() on THP
mlock: fix unevictable_pgs event counts on THP
Byron Stanoszek <gandalf@winds.org>:
tmpfs: restore functionality of nr_inodes=0
Muchun Song <songmuchun@bytedance.com>:
kprobes: fix kill kprobe which has been marked as gone
Subsystem: mm/thp
Ralph Campbell <rcampbell@nvidia.com>:
mm/thp: fix __split_huge_pmd_locked() for migration PMD
Christophe Leroy <christophe.leroy@csgroup.eu>:
selftests/vm: fix display of page size in map_hugetlb
Subsystem: mm/memory-hotplug
Pavel Tatashin <pasha.tatashin@soleen.com>:
mm/memory_hotplug: drain per-cpu pages again during memory offline
Subsystem: misc
Tobias Klauser <tklauser@distanz.ch>:
ftrace: let ftrace_enable_sysctl take a kernel pointer buffer
stackleak: let stack_erasing_sysctl take a kernel pointer buffer
fs/fs-writeback.c: adjust dirtytime_interval_handler definition to match prototype
Subsystem: kcsan
Changbin Du <changbin.du@gmail.com>:
kcsan: kconfig: move to menu 'Generic Kernel Debugging Instruments'
.mailmap | 4 ++
fs/fs-writeback.c | 2 -
include/linux/ftrace.h | 3 --
include/linux/stackleak.h | 2 -
kernel/kprobes.c | 9 +++++-
kernel/stackleak.c | 2 -
kernel/trace/ftrace.c | 3 --
lib/Kconfig.debug | 4 --
mm/huge_memory.c | 42 ++++++++++++++++---------------
mm/ksm.c | 4 ++
mm/memory_hotplug.c | 14 ++++++++++
mm/migrate.c | 3 +-
mm/mlock.c | 24 +++++++++++------
mm/page_isolation.c | 8 +++++
mm/shmem.c | 20 +++++++++++---
mm/swap.c | 6 ++--
mm/vmscan.c | 10 +++++--
tools/testing/selftests/vm/map_hugetlb.c | 2 -
18 files changed, 111 insertions(+), 51 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-09-04 23:34 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-09-04 23:34 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
19 patches, based on 59126901f200f5fc907153468b03c64e0081b6e6.
Subsystems affected by this patch series:
mm/memcg
mm/slub
MAINTAINERS
mm/pagemap
ipc
fork
checkpatch
mm/madvise
mm/migration
mm/hugetlb
lib
Subsystem: mm/memcg
Michal Hocko <mhocko@suse.com>:
memcg: fix use-after-free in uncharge_batch
Xunlei Pang <xlpang@linux.alibaba.com>:
mm: memcg: fix memcg reclaim soft lockup
Subsystem: mm/slub
Eugeniu Rosca <erosca@de.adit-jv.com>:
mm: slub: fix conversion of freelist_corrupted()
Subsystem: MAINTAINERS
Robert Richter <rric@kernel.org>:
MAINTAINERS: update Cavium/Marvell entries
Nick Desaulniers <ndesaulniers@google.com>:
MAINTAINERS: add LLVM maintainers
Randy Dunlap <rdunlap@infradead.org>:
MAINTAINERS: IA64: mark Status as Odd Fixes only
Subsystem: mm/pagemap
Joerg Roedel <jroedel@suse.de>:
mm: track page table modifications in __apply_to_page_range()
Subsystem: ipc
Tobias Klauser <tklauser@distanz.ch>:
ipc: adjust proc_ipc_sem_dointvec definition to match prototype
Subsystem: fork
Tobias Klauser <tklauser@distanz.ch>:
fork: adjust sysctl_max_threads definition to match prototype
Subsystem: checkpatch
Mrinal Pandey <mrinalmni@gmail.com>:
checkpatch: fix the usage of capture group ( ... )
Subsystem: mm/madvise
Yang Shi <shy828301@gmail.com>:
mm: madvise: fix vma user-after-free
Subsystem: mm/migration
Alistair Popple <alistair@popple.id.au>:
mm/migrate: fixup setting UFFD_WP flag
mm/rmap: fixup copying of soft dirty and uffd ptes
Ralph Campbell <rcampbell@nvidia.com>:
Patch series "mm/migrate: preserve soft dirty in remove_migration_pte()":
mm/migrate: remove unnecessary is_zone_device_page() check
mm/migrate: preserve soft dirty in remove_migration_pte()
Subsystem: mm/hugetlb
Li Xinhai <lixinhai.lxh@gmail.com>:
mm/hugetlb: try preferred node first when alloc gigantic page from cma
Muchun Song <songmuchun@bytedance.com>:
mm/hugetlb: fix a race between hugetlb sysctl handlers
David Howells <dhowells@redhat.com>:
mm/khugepaged.c: fix khugepaged's request size in collapse_file
Subsystem: lib
Jason Gunthorpe <jgg@nvidia.com>:
include/linux/log2.h: add missing () around n in roundup_pow_of_two()
MAINTAINERS | 32 ++++++++++++++++----------------
include/linux/log2.h | 2 +-
ipc/ipc_sysctl.c | 2 +-
kernel/fork.c | 2 +-
mm/hugetlb.c | 49 +++++++++++++++++++++++++++++++++++++------------
mm/khugepaged.c | 2 +-
mm/madvise.c | 2 +-
mm/memcontrol.c | 6 ++++++
mm/memory.c | 37 ++++++++++++++++++++++++-------------
mm/migrate.c | 31 +++++++++++++++++++------------
mm/rmap.c | 9 +++++++--
mm/slub.c | 12 ++++++------
mm/vmscan.c | 8 ++++++++
scripts/checkpatch.pl | 4 ++--
14 files changed, 130 insertions(+), 68 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-08-21 0:41 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-08-21 0:41 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
11 patches, based on 7eac66d0456fe12a462e5c14c68e97c7460989da.
Subsystems affected by this patch series:
misc
mm/hugetlb
mm/vmalloc
mm/misc
romfs
relay
uprobes
squashfs
mm/cma
mm/pagealloc
Subsystem: misc
Nick Desaulniers <ndesaulniers@google.com>:
mailmap: add Andi Kleen
Subsystem: mm/hugetlb
Xu Wang <vulab@iscas.ac.cn>:
hugetlb_cgroup: convert comma to semicolon
Hugh Dickins <hughd@google.com>:
khugepaged: adjust VM_BUG_ON_MM() in __khugepaged_enter()
Subsystem: mm/vmalloc
"Aneesh Kumar K.V" <aneesh.kumar@linux.ibm.com>:
mm/vunmap: add cond_resched() in vunmap_pmd_range
Subsystem: mm/misc
Leon Romanovsky <leonro@nvidia.com>:
mm/rodata_test.c: fix missing function declaration
Subsystem: romfs
Jann Horn <jannh@google.com>:
romfs: fix uninitialized memory leak in romfs_dev_read()
Subsystem: relay
Wei Yongjun <weiyongjun1@huawei.com>:
kernel/relay.c: fix memleak on destroy relay channel
Subsystem: uprobes
Hugh Dickins <hughd@google.com>:
uprobes: __replace_page() avoid BUG in munlock_vma_page()
Subsystem: squashfs
Phillip Lougher <phillip@squashfs.org.uk>:
squashfs: avoid bio_alloc() failure with 1Mbyte blocks
Subsystem: mm/cma
Doug Berger <opendmb@gmail.com>:
mm: include CMA pages in lowmem_reserve at boot
Subsystem: mm/pagealloc
Charan Teja Reddy <charante@codeaurora.org>:
mm, page_alloc: fix core hung in free_pcppages_bulk()
.mailmap | 1 +
fs/romfs/storage.c | 4 +---
fs/squashfs/block.c | 6 +++++-
kernel/events/uprobes.c | 2 +-
kernel/relay.c | 1 +
mm/hugetlb_cgroup.c | 4 ++--
mm/khugepaged.c | 2 +-
mm/page_alloc.c | 7 ++++++-
mm/rodata_test.c | 1 +
mm/vmalloc.c | 2 ++
10 files changed, 21 insertions(+), 9 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-08-15 0:29 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-08-15 0:29 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
39 patches, based on b923f1247b72fc100b87792fd2129d026bb10e66.
Subsystems affected by this patch series:
mm/hotfixes
lz4
exec
mailmap
mm/thp
autofs
mm/madvise
sysctl
mm/kmemleak
mm/misc
lib
Subsystem: mm/hotfixes
Mike Rapoport <rppt@linux.ibm.com>:
asm-generic: pgalloc.h: use correct #ifdef to enable pud_alloc_one()
Baoquan He <bhe@redhat.com>:
Revert "mm/vmstat.c: do not show lowmem reserve protection information of empty zone"
Subsystem: lz4
Nick Terrell <terrelln@fb.com>:
lz4: fix kernel decompression speed
Subsystem: exec
Kees Cook <keescook@chromium.org>:
Patch series "Fix S_ISDIR execve() errno":
exec: restore EACCES of S_ISDIR execve()
selftests/exec: add file type errno tests
Subsystem: mailmap
Greg Kurz <groug@kaod.org>:
mailmap: add entry for Greg Kurz
Subsystem: mm/thp
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
Patch series "THP prep patches":
mm: store compound_nr as well as compound_order
mm: move page-flags include to top of file
mm: add thp_order
mm: add thp_size
mm: replace hpage_nr_pages with thp_nr_pages
mm: add thp_head
mm: introduce offset_in_thp
Subsystem: autofs
Randy Dunlap <rdunlap@infradead.org>:
fs: autofs: delete repeated words in comments
Subsystem: mm/madvise
Minchan Kim <minchan@kernel.org>:
Patch series "introduce memory hinting API for external process", v8:
mm/madvise: pass task and mm to do_madvise
pid: move pidfd_get_pid() to pid.c
mm/madvise: introduce process_madvise() syscall: an external memory hinting API
mm/madvise: check fatal signal pending of target process
Subsystem: sysctl
Xiaoming Ni <nixiaoming@huawei.com>:
all arch: remove system call sys_sysctl
Subsystem: mm/kmemleak
Qian Cai <cai@lca.pw>:
mm/kmemleak: silence KCSAN splats in checksum
Subsystem: mm/misc
Qian Cai <cai@lca.pw>:
mm/frontswap: mark various intentional data races
mm/page_io: mark various intentional data races
mm/swap_state: mark various intentional data races
Kirill A. Shutemov <kirill@shutemov.name>:
mm/filemap.c: fix a data race in filemap_fault()
Qian Cai <cai@lca.pw>:
mm/swapfile: fix and annotate various data races
mm/page_counter: fix various data races at memsw
mm/memcontrol: fix a data race in scan count
mm/list_lru: fix a data race in list_lru_count_one
mm/mempool: fix a data race in mempool_free()
mm/rmap: annotate a data race at tlb_flush_batched
mm/swap.c: annotate data races for lru_rotate_pvecs
mm: annotate a data race in page_zonenum()
Romain Naour <romain.naour@gmail.com>:
include/asm-generic/vmlinux.lds.h: align ro_after_init
Kuninori Morimoto <kuninori.morimoto.gx@renesas.com>:
sh: clkfwk: remove r8/r16/r32
sh: use generic strncpy()
Subsystem: lib
Krzysztof Kozlowski <krzk@kernel.org>:
Patch series "iomap: Constify ioreadX() iomem argument", v3:
iomap: constify ioreadX() iomem argument (as in generic implementation)
rtl818x: constify ioreadX() iomem argument (as in generic implementation)
ntb: intel: constify ioreadX() iomem argument (as in generic implementation)
virtio: pci: constify ioreadX() iomem argument (as in generic implementation)
.mailmap | 1
arch/alpha/include/asm/core_apecs.h | 6
arch/alpha/include/asm/core_cia.h | 6
arch/alpha/include/asm/core_lca.h | 6
arch/alpha/include/asm/core_marvel.h | 4
arch/alpha/include/asm/core_mcpcia.h | 6
arch/alpha/include/asm/core_t2.h | 2
arch/alpha/include/asm/io.h | 12 -
arch/alpha/include/asm/io_trivial.h | 16 -
arch/alpha/include/asm/jensen.h | 2
arch/alpha/include/asm/machvec.h | 6
arch/alpha/kernel/core_marvel.c | 2
arch/alpha/kernel/io.c | 12 -
arch/alpha/kernel/syscalls/syscall.tbl | 3
arch/arm/configs/am200epdkit_defconfig | 1
arch/arm/tools/syscall.tbl | 3
arch/arm64/include/asm/unistd.h | 2
arch/arm64/include/asm/unistd32.h | 6
arch/ia64/kernel/syscalls/syscall.tbl | 3
arch/m68k/kernel/syscalls/syscall.tbl | 3
arch/microblaze/kernel/syscalls/syscall.tbl | 3
arch/mips/configs/cu1000-neo_defconfig | 1
arch/mips/kernel/syscalls/syscall_n32.tbl | 3
arch/mips/kernel/syscalls/syscall_n64.tbl | 3
arch/mips/kernel/syscalls/syscall_o32.tbl | 3
arch/parisc/include/asm/io.h | 4
arch/parisc/kernel/syscalls/syscall.tbl | 3
arch/parisc/lib/iomap.c | 72 +++---
arch/powerpc/kernel/iomap.c | 28 +-
arch/powerpc/kernel/syscalls/syscall.tbl | 3
arch/s390/kernel/syscalls/syscall.tbl | 3
arch/sh/configs/dreamcast_defconfig | 1
arch/sh/configs/espt_defconfig | 1
arch/sh/configs/hp6xx_defconfig | 1
arch/sh/configs/landisk_defconfig | 1
arch/sh/configs/lboxre2_defconfig | 1
arch/sh/configs/microdev_defconfig | 1
arch/sh/configs/migor_defconfig | 1
arch/sh/configs/r7780mp_defconfig | 1
arch/sh/configs/r7785rp_defconfig | 1
arch/sh/configs/rts7751r2d1_defconfig | 1
arch/sh/configs/rts7751r2dplus_defconfig | 1
arch/sh/configs/se7206_defconfig | 1
arch/sh/configs/se7343_defconfig | 1
arch/sh/configs/se7619_defconfig | 1
arch/sh/configs/se7705_defconfig | 1
arch/sh/configs/se7750_defconfig | 1
arch/sh/configs/se7751_defconfig | 1
arch/sh/configs/secureedge5410_defconfig | 1
arch/sh/configs/sh03_defconfig | 1
arch/sh/configs/sh7710voipgw_defconfig | 1
arch/sh/configs/sh7757lcr_defconfig | 1
arch/sh/configs/sh7763rdp_defconfig | 1
arch/sh/configs/shmin_defconfig | 1
arch/sh/configs/titan_defconfig | 1
arch/sh/include/asm/string_32.h | 26 --
arch/sh/kernel/iomap.c | 22 -
arch/sh/kernel/syscalls/syscall.tbl | 3
arch/sparc/kernel/syscalls/syscall.tbl | 3
arch/x86/entry/syscalls/syscall_32.tbl | 3
arch/x86/entry/syscalls/syscall_64.tbl | 4
arch/xtensa/kernel/syscalls/syscall.tbl | 3
drivers/mailbox/bcm-pdc-mailbox.c | 2
drivers/net/wireless/realtek/rtl818x/rtl8180/rtl8180.h | 6
drivers/ntb/hw/intel/ntb_hw_gen1.c | 2
drivers/ntb/hw/intel/ntb_hw_gen3.h | 2
drivers/ntb/hw/intel/ntb_hw_intel.h | 2
drivers/nvdimm/btt.c | 4
drivers/nvdimm/pmem.c | 6
drivers/sh/clk/cpg.c | 25 --
drivers/virtio/virtio_pci_modern.c | 6
fs/autofs/dev-ioctl.c | 4
fs/io_uring.c | 2
fs/namei.c | 4
include/asm-generic/iomap.h | 28 +-
include/asm-generic/pgalloc.h | 2
include/asm-generic/vmlinux.lds.h | 1
include/linux/compat.h | 5
include/linux/huge_mm.h | 58 ++++-
include/linux/io-64-nonatomic-hi-lo.h | 4
include/linux/io-64-nonatomic-lo-hi.h | 4
include/linux/memcontrol.h | 2
include/linux/mm.h | 16 -
include/linux/mm_inline.h | 6
include/linux/mm_types.h | 1
include/linux/pagemap.h | 6
include/linux/pid.h | 1
include/linux/syscalls.h | 4
include/linux/sysctl.h | 6
include/uapi/asm-generic/unistd.h | 4
kernel/Makefile | 2
kernel/exit.c | 17 -
kernel/pid.c | 17 +
kernel/sys_ni.c | 3
kernel/sysctl_binary.c | 171 --------------
lib/iomap.c | 30 +-
lib/lz4/lz4_compress.c | 4
lib/lz4/lz4_decompress.c | 18 -
lib/lz4/lz4defs.h | 10
lib/lz4/lz4hc_compress.c | 2
mm/compaction.c | 2
mm/filemap.c | 22 +
mm/frontswap.c | 8
mm/gup.c | 2
mm/internal.h | 4
mm/kmemleak.c | 2
mm/list_lru.c | 2
mm/madvise.c | 190 ++++++++++++++--
mm/memcontrol.c | 10
mm/memory.c | 4
mm/memory_hotplug.c | 7
mm/mempolicy.c | 2
mm/mempool.c | 2
mm/migrate.c | 18 -
mm/mlock.c | 9
mm/page_alloc.c | 5
mm/page_counter.c | 13 -
mm/page_io.c | 12 -
mm/page_vma_mapped.c | 6
mm/rmap.c | 10
mm/swap.c | 21 -
mm/swap_state.c | 10
mm/swapfile.c | 33 +-
mm/vmscan.c | 6
mm/vmstat.c | 12 -
mm/workingset.c | 6
tools/perf/arch/powerpc/entry/syscalls/syscall.tbl | 2
tools/perf/arch/s390/entry/syscalls/syscall.tbl | 2
tools/perf/arch/x86/entry/syscalls/syscall_64.tbl | 2
tools/testing/selftests/exec/.gitignore | 1
tools/testing/selftests/exec/Makefile | 5
tools/testing/selftests/exec/non-regular.c | 196 +++++++++++++++++
132 files changed, 815 insertions(+), 614 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-08-12 1:29 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-08-12 1:29 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
- Most of the rest of MM
- various other subsystems
165 patches, based on 00e4db51259a5f936fec1424b884f029479d3981.
Subsystems affected by this patch series:
mm/memcg
mm/hugetlb
mm/vmscan
mm/proc
mm/compaction
mm/mempolicy
mm/oom-kill
mm/hugetlbfs
mm/migration
mm/thp
mm/cma
mm/util
mm/memory-hotplug
mm/cleanups
mm/uaccess
alpha
misc
sparse
bitmap
lib
lz4
bitops
checkpatch
autofs
minix
nilfs
ufs
fat
signals
kmod
coredump
exec
kdump
rapidio
panic
kcov
kgdb
ipc
mm/migration
mm/gup
mm/pagemap
Subsystem: mm/memcg
Roman Gushchin <guro@fb.com>:
Patch series "mm: memcg accounting of percpu memory", v3:
percpu: return number of released bytes from pcpu_free_area()
mm: memcg/percpu: account percpu memory to memory cgroups
mm: memcg/percpu: per-memcg percpu memory statistics
mm: memcg: charge memcg percpu memory to the parent cgroup
kselftests: cgroup: add perpcu memory accounting test
Subsystem: mm/hugetlb
Muchun Song <songmuchun@bytedance.com>:
mm/hugetlb: add mempolicy check in the reservation routine
Subsystem: mm/vmscan
Joonsoo Kim <iamjoonsoo.kim@lge.com>:
Patch series "workingset protection/detection on the anonymous LRU list", v7:
mm/vmscan: make active/inactive ratio as 1:1 for anon lru
mm/vmscan: protect the workingset on anonymous LRU
mm/workingset: prepare the workingset detection infrastructure for anon LRU
mm/swapcache: support to handle the shadow entries
mm/swap: implement workingset detection for anonymous LRU
mm/vmscan: restore active/inactive ratio for anonymous LRU
Subsystem: mm/proc
Michal Koutný <mkoutny@suse.com>:
/proc/PID/smaps: consistent whitespace output format
Subsystem: mm/compaction
Nitin Gupta <nigupta@nvidia.com>:
mm: proactive compaction
mm: fix compile error due to COMPACTION_HPAGE_ORDER
mm: use unsigned types for fragmentation score
Alex Shi <alex.shi@linux.alibaba.com>:
mm/compaction: correct the comments of compact_defer_shift
Subsystem: mm/mempolicy
Krzysztof Kozlowski <krzk@kernel.org>:
mm: mempolicy: fix kerneldoc of numa_map_to_online_node()
Wenchao Hao <haowenchao22@gmail.com>:
mm/mempolicy.c: check parameters first in kernel_get_mempolicy
Yanfei Xu <yanfei.xu@windriver.com>:
include/linux/mempolicy.h: fix typo
Subsystem: mm/oom-kill
Yafang Shao <laoar.shao@gmail.com>:
mm, oom: make the calculation of oom badness more accurate
Michal Hocko <mhocko@suse.com>:
doc, mm: sync up oom_score_adj documentation
doc, mm: clarify /proc/<pid>/oom_score value range
Yafang Shao <laoar.shao@gmail.com>:
mm, oom: show process exiting information in __oom_kill_process()
Subsystem: mm/hugetlbfs
Mike Kravetz <mike.kravetz@oracle.com>:
hugetlbfs: prevent filesystem stacking of hugetlbfs
hugetlbfs: remove call to huge_pte_alloc without i_mmap_rwsem
Subsystem: mm/migration
Ralph Campbell <rcampbell@nvidia.com>:
Patch series "mm/migrate: optimize migrate_vma_setup() for holes":
mm/migrate: optimize migrate_vma_setup() for holes
mm/migrate: add migrate-shared test for migrate_vma_*()
Subsystem: mm/thp
Yang Shi <yang.shi@linux.alibaba.com>:
mm: thp: remove debug_cow switch
Anshuman Khandual <anshuman.khandual@arm.com>:
mm/vmstat: add events for THP migration without split
Subsystem: mm/cma
Jianqun Xu <jay.xu@rock-chips.com>:
mm/cma.c: fix NULL pointer dereference when cma could not be activated
Barry Song <song.bao.hua@hisilicon.com>:
Patch series "mm: fix the names of general cma and hugetlb cma", v2:
mm: cma: fix the name of CMA areas
mm: hugetlb: fix the name of hugetlb CMA
Mike Kravetz <mike.kravetz@oracle.com>:
cma: don't quit at first error when activating reserved areas
Subsystem: mm/util
Waiman Long <longman@redhat.com>:
include/linux/sched/mm.h: optimize current_gfp_context()
Krzysztof Kozlowski <krzk@kernel.org>:
mm: mmu_notifier: fix and extend kerneldoc
Subsystem: mm/memory-hotplug
Daniel Jordan <daniel.m.jordan@oracle.com>:
x86/mm: use max memory block size on bare metal
Jia He <justin.he@arm.com>:
mm/memory_hotplug: introduce default dummy memory_add_physaddr_to_nid()
mm/memory_hotplug: fix unpaired mem_hotplug_begin/done
Charan Teja Reddy <charante@codeaurora.org>:
mm, memory_hotplug: update pcp lists everytime onlining a memory block
Subsystem: mm/cleanups
Randy Dunlap <rdunlap@infradead.org>:
mm: drop duplicated words in <linux/pgtable.h>
mm: drop duplicated words in <linux/mm.h>
include/linux/highmem.h: fix duplicated words in a comment
include/linux/frontswap.h: drop duplicated word in a comment
include/linux/memcontrol.h: drop duplicate word and fix spello
Arvind Sankar <nivedita@alum.mit.edu>:
sh/mm: drop unused MAX_PHYSADDR_BITS
sparc: drop unused MAX_PHYSADDR_BITS
Randy Dunlap <rdunlap@infradead.org>:
mm/compaction.c: delete duplicated word
mm/filemap.c: delete duplicated word
mm/hmm.c: delete duplicated word
mm/hugetlb.c: delete duplicated words
mm/memcontrol.c: delete duplicated words
mm/memory.c: delete duplicated words
mm/migrate.c: delete duplicated word
mm/nommu.c: delete duplicated words
mm/page_alloc.c: delete or fix duplicated words
mm/shmem.c: delete duplicated word
mm/slab_common.c: delete duplicated word
mm/usercopy.c: delete duplicated word
mm/vmscan.c: delete or fix duplicated words
mm/zpool.c: delete duplicated word and fix grammar
mm/zsmalloc.c: fix duplicated words
Subsystem: mm/uaccess
Christoph Hellwig <hch@lst.de>:
Patch series "clean up address limit helpers", v2:
syscalls: use uaccess_kernel in addr_limit_user_check
nds32: use uaccess_kernel in show_regs
riscv: include <asm/pgtable.h> in <asm/uaccess.h>
uaccess: remove segment_eq
uaccess: add force_uaccess_{begin,end} helpers
exec: use force_uaccess_begin during exec and exit
Subsystem: alpha
Luc Van Oostenryck <luc.vanoostenryck@gmail.com>:
alpha: fix annotation of io{read,write}{16,32}be()
Subsystem: misc
Randy Dunlap <rdunlap@infradead.org>:
include/linux/compiler-clang.h: drop duplicated word in a comment
include/linux/exportfs.h: drop duplicated word in a comment
include/linux/async_tx.h: drop duplicated word in a comment
include/linux/xz.h: drop duplicated word
Christoph Hellwig <hch@lst.de>:
kernel: add a kernel_wait helper
Feng Tang <feng.tang@intel.com>:
./Makefile: add debug option to enable function aligned on 32 bytes
Arvind Sankar <nivedita@alum.mit.edu>:
kernel.h: remove duplicate include of asm/div64.h
"Alexander A. Klimov" <grandmaster@al2klimov.de>:
include/: replace HTTP links with HTTPS ones
Matthew Wilcox <willy@infradead.org>:
include/linux/poison.h: remove obsolete comment
Subsystem: sparse
Luc Van Oostenryck <luc.vanoostenryck@gmail.com>:
sparse: group the defines by functionality
Subsystem: bitmap
Stefano Brivio <sbrivio@redhat.com>:
Patch series "lib: Fix bitmap_cut() for overlaps, add test":
lib/bitmap.c: fix bitmap_cut() for partial overlapping case
lib/test_bitmap.c: add test for bitmap_cut()
Subsystem: lib
Luc Van Oostenryck <luc.vanoostenryck@gmail.com>:
lib/generic-radix-tree.c: remove unneeded __rcu
Geert Uytterhoeven <geert@linux-m68k.org>:
lib/test_bitops: do the full test during module init
Wei Yongjun <weiyongjun1@huawei.com>:
lib/test_lockup.c: make symbol 'test_works' static
Tiezhu Yang <yangtiezhu@loongson.cn>:
lib/Kconfig.debug: make TEST_LOCKUP depend on module
lib/test_lockup.c: fix return value of test_lockup_init()
"Alexander A. Klimov" <grandmaster@al2klimov.de>:
lib/: replace HTTP links with HTTPS ones
"Kars Mulder" <kerneldev@karsmulder.nl>:
kstrto*: correct documentation references to simple_strto*()
kstrto*: do not describe simple_strto*() as obsolete/replaced
Subsystem: lz4
Nick Terrell <terrelln@fb.com>:
lz4: fix kernel decompression speed
Subsystem: bitops
Rikard Falkeborn <rikard.falkeborn@gmail.com>:
lib/test_bits.c: add tests of GENMASK
Subsystem: checkpatch
Joe Perches <joe@perches.com>:
checkpatch: add test for possible misuse of IS_ENABLED() without CONFIG_
checkpatch: add --fix option for ASSIGN_IN_IF
Quentin Monnet <quentin@isovalent.com>:
checkpatch: fix CONST_STRUCT when const_structs.checkpatch is missing
Joe Perches <joe@perches.com>:
checkpatch: add test for repeated words
checkpatch: remove missing switch/case break test
Subsystem: autofs
Randy Dunlap <rdunlap@infradead.org>:
autofs: fix doubled word
Subsystem: minix
Eric Biggers <ebiggers@google.com>:
Patch series "fs/minix: fix syzbot bugs and set s_maxbytes":
fs/minix: check return value of sb_getblk()
fs/minix: don't allow getting deleted inodes
fs/minix: reject too-large maximum file size
fs/minix: set s_maxbytes correctly
fs/minix: fix block limit check for V1 filesystems
fs/minix: remove expected error message in block_to_path()
Subsystem: nilfs
Eric Biggers <ebiggers@google.com>:
Patch series "nilfs2 updates":
nilfs2: only call unlock_new_inode() if I_NEW
Joe Perches <joe@perches.com>:
nilfs2: convert __nilfs_msg to integrate the level and format
nilfs2: use a more common logging style
Subsystem: ufs
Colin Ian King <colin.king@canonical.com>:
fs/ufs: avoid potential u32 multiplication overflow
Subsystem: fat
Yubo Feng <fengyubo3@huawei.com>:
fatfs: switch write_lock to read_lock in fat_ioctl_get_attributes
"Alexander A. Klimov" <grandmaster@al2klimov.de>:
VFAT/FAT/MSDOS FILESYSTEM: replace HTTP links with HTTPS ones
OGAWA Hirofumi <hirofumi@mail.parknet.co.jp>:
fat: fix fat_ra_init() for data clusters == 0
Subsystem: signals
Helge Deller <deller@gmx.de>:
fs/signalfd.c: fix inconsistent return codes for signalfd4
Subsystem: kmod
Tiezhu Yang <yangtiezhu@loongson.cn>:
Patch series "kmod/umh: a few fixes":
selftests: kmod: use variable NAME in kmod_test_0001()
kmod: remove redundant "be an" in the comment
test_kmod: avoid potential double free in trigger_config_run_type()
Subsystem: coredump
Lepton Wu <ytht.net@gmail.com>:
coredump: add %f for executable filename
Subsystem: exec
Kees Cook <keescook@chromium.org>:
Patch series "Relocate execve() sanity checks", v2:
exec: change uselib(2) IS_SREG() failure to EACCES
exec: move S_ISREG() check earlier
exec: move path_noexec() check earlier
Subsystem: kdump
Vijay Balakrishna <vijayb@linux.microsoft.com>:
kdump: append kernel build-id string to VMCOREINFO
Subsystem: rapidio
"Gustavo A. R. Silva" <gustavoars@kernel.org>:
drivers/rapidio/devices/rio_mport_cdev.c: use struct_size() helper
drivers/rapidio/rio-scan.c: use struct_size() helper
rapidio/rio_mport_cdev: use array_size() helper in copy_{from,to}_user()
Subsystem: panic
Tiezhu Yang <yangtiezhu@loongson.cn>:
kernel/panic.c: make oops_may_print() return bool
lib/Kconfig.debug: fix typo in the help text of CONFIG_PANIC_TIMEOUT
Yue Hu <huyue2@yulong.com>:
panic: make print_oops_end_marker() static
Subsystem: kcov
Marco Elver <elver@google.com>:
kcov: unconditionally add -fno-stack-protector to compiler options
Wei Yongjun <weiyongjun1@huawei.com>:
kcov: make some symbols static
Subsystem: kgdb
Nick Desaulniers <ndesaulniers@google.com>:
scripts/gdb: fix python 3.8 SyntaxWarning
Subsystem: ipc
Alexey Dobriyan <adobriyan@gmail.com>:
ipc: uninline functions
Liao Pingfang <liao.pingfang@zte.com.cn>:
ipc/shm.c: remove the superfluous break
Subsystem: mm/migration
Joonsoo Kim <iamjoonsoo.kim@lge.com>:
Patch series "clean-up the migration target allocation functions", v5:
mm/page_isolation: prefer the node of the source page
mm/migrate: move migration helper from .h to .c
mm/hugetlb: unify migration callbacks
mm/migrate: clear __GFP_RECLAIM to make the migration callback consistent with regular THP allocations
mm/migrate: introduce a standard migration target allocation function
mm/mempolicy: use a standard migration target allocation callback
mm/page_alloc: remove a wrapper for alloc_migration_target()
Subsystem: mm/gup
Joonsoo Kim <iamjoonsoo.kim@lge.com>:
mm/gup: restrict CMA region by using allocation scope API
mm/hugetlb: make hugetlb migration callback CMA aware
mm/gup: use a standard migration target allocation callback
Subsystem: mm/pagemap
Peter Xu <peterx@redhat.com>:
Patch series "mm: Page fault accounting cleanups", v5:
mm: do page fault accounting in handle_mm_fault
mm/alpha: use general page fault accounting
mm/arc: use general page fault accounting
mm/arm: use general page fault accounting
mm/arm64: use general page fault accounting
mm/csky: use general page fault accounting
mm/hexagon: use general page fault accounting
mm/ia64: use general page fault accounting
mm/m68k: use general page fault accounting
mm/microblaze: use general page fault accounting
mm/mips: use general page fault accounting
mm/nds32: use general page fault accounting
mm/nios2: use general page fault accounting
mm/openrisc: use general page fault accounting
mm/parisc: use general page fault accounting
mm/powerpc: use general page fault accounting
mm/riscv: use general page fault accounting
mm/s390: use general page fault accounting
mm/sh: use general page fault accounting
mm/sparc32: use general page fault accounting
mm/sparc64: use general page fault accounting
mm/x86: use general page fault accounting
mm/xtensa: use general page fault accounting
mm: clean up the last pieces of page fault accountings
mm/gup: remove task_struct pointer for all gup code
Documentation/admin-guide/cgroup-v2.rst | 4
Documentation/admin-guide/sysctl/kernel.rst | 3
Documentation/admin-guide/sysctl/vm.rst | 15 +
Documentation/filesystems/proc.rst | 11 -
Documentation/vm/page_migration.rst | 27 +++
Makefile | 4
arch/alpha/include/asm/io.h | 8
arch/alpha/include/asm/uaccess.h | 2
arch/alpha/mm/fault.c | 10 -
arch/arc/include/asm/segment.h | 3
arch/arc/kernel/process.c | 2
arch/arc/mm/fault.c | 20 --
arch/arm/include/asm/uaccess.h | 4
arch/arm/kernel/signal.c | 2
arch/arm/mm/fault.c | 27 ---
arch/arm64/include/asm/uaccess.h | 2
arch/arm64/kernel/sdei.c | 2
arch/arm64/mm/fault.c | 31 ---
arch/arm64/mm/numa.c | 10 -
arch/csky/include/asm/segment.h | 2
arch/csky/mm/fault.c | 15 -
arch/h8300/include/asm/segment.h | 2
arch/hexagon/mm/vm_fault.c | 11 -
arch/ia64/include/asm/uaccess.h | 2
arch/ia64/mm/fault.c | 11 -
arch/ia64/mm/numa.c | 2
arch/m68k/include/asm/segment.h | 2
arch/m68k/include/asm/tlbflush.h | 6
arch/m68k/mm/fault.c | 16 -
arch/microblaze/include/asm/uaccess.h | 2
arch/microblaze/mm/fault.c | 11 -
arch/mips/include/asm/uaccess.h | 2
arch/mips/kernel/unaligned.c | 27 +--
arch/mips/mm/fault.c | 16 -
arch/nds32/include/asm/uaccess.h | 2
arch/nds32/kernel/process.c | 2
arch/nds32/mm/alignment.c | 7
arch/nds32/mm/fault.c | 21 --
arch/nios2/include/asm/uaccess.h | 2
arch/nios2/mm/fault.c | 16 -
arch/openrisc/include/asm/uaccess.h | 2
arch/openrisc/mm/fault.c | 11 -
arch/parisc/include/asm/uaccess.h | 2
arch/parisc/mm/fault.c | 10 -
arch/powerpc/include/asm/uaccess.h | 3
arch/powerpc/mm/copro_fault.c | 7
arch/powerpc/mm/fault.c | 13 -
arch/riscv/include/asm/uaccess.h | 6
arch/riscv/mm/fault.c | 18 --
arch/s390/include/asm/uaccess.h | 2
arch/s390/kvm/interrupt.c | 2
arch/s390/kvm/kvm-s390.c | 2
arch/s390/kvm/priv.c | 8
arch/s390/mm/fault.c | 18 --
arch/s390/mm/gmap.c | 4
arch/sh/include/asm/segment.h | 3
arch/sh/include/asm/sparsemem.h | 4
arch/sh/kernel/traps_32.c | 12 -
arch/sh/mm/fault.c | 13 -
arch/sh/mm/init.c | 9 -
arch/sparc/include/asm/sparsemem.h | 1
arch/sparc/include/asm/uaccess_32.h | 2
arch/sparc/include/asm/uaccess_64.h | 2
arch/sparc/mm/fault_32.c | 15 -
arch/sparc/mm/fault_64.c | 13 -
arch/um/kernel/trap.c | 6
arch/x86/include/asm/uaccess.h | 2
arch/x86/mm/fault.c | 19 --
arch/x86/mm/init_64.c | 9 +
arch/x86/mm/numa.c | 1
arch/xtensa/include/asm/uaccess.h | 2
arch/xtensa/mm/fault.c | 17 -
drivers/firmware/arm_sdei.c | 5
drivers/gpu/drm/i915/gem/i915_gem_userptr.c | 2
drivers/infiniband/core/umem_odp.c | 2
drivers/iommu/amd/iommu_v2.c | 2
drivers/iommu/intel/svm.c | 3
drivers/rapidio/devices/rio_mport_cdev.c | 7
drivers/rapidio/rio-scan.c | 8
drivers/vfio/vfio_iommu_type1.c | 4
fs/coredump.c | 17 +
fs/exec.c | 38 ++--
fs/fat/Kconfig | 2
fs/fat/fatent.c | 3
fs/fat/file.c | 4
fs/hugetlbfs/inode.c | 6
fs/minix/inode.c | 48 ++++-
fs/minix/itree_common.c | 8
fs/minix/itree_v1.c | 16 -
fs/minix/itree_v2.c | 15 -
fs/minix/minix.h | 1
fs/namei.c | 10 -
fs/nilfs2/alloc.c | 38 ++--
fs/nilfs2/btree.c | 42 ++--
fs/nilfs2/cpfile.c | 10 -
fs/nilfs2/dat.c | 14 -
fs/nilfs2/direct.c | 14 -
fs/nilfs2/gcinode.c | 2
fs/nilfs2/ifile.c | 4
fs/nilfs2/inode.c | 32 +--
fs/nilfs2/ioctl.c | 37 ++--
fs/nilfs2/mdt.c | 2
fs/nilfs2/namei.c | 6
fs/nilfs2/nilfs.h | 18 +-
fs/nilfs2/page.c | 11 -
fs/nilfs2/recovery.c | 32 +--
fs/nilfs2/segbuf.c | 2
fs/nilfs2/segment.c | 38 ++--
fs/nilfs2/sufile.c | 29 +--
fs/nilfs2/super.c | 73 ++++----
fs/nilfs2/sysfs.c | 29 +--
fs/nilfs2/the_nilfs.c | 85 ++++-----
fs/open.c | 6
fs/proc/base.c | 11 +
fs/proc/task_mmu.c | 4
fs/signalfd.c | 10 -
fs/ufs/super.c | 2
include/asm-generic/uaccess.h | 4
include/clocksource/timer-ti-dm.h | 2
include/linux/async_tx.h | 2
include/linux/btree.h | 2
include/linux/compaction.h | 6
include/linux/compiler-clang.h | 2
include/linux/compiler_types.h | 44 ++---
include/linux/crash_core.h | 6
include/linux/delay.h | 2
include/linux/dma/k3-psil.h | 2
include/linux/dma/k3-udma-glue.h | 2
include/linux/dma/ti-cppi5.h | 2
include/linux/exportfs.h | 2
include/linux/frontswap.h | 2
include/linux/fs.h | 10 +
include/linux/generic-radix-tree.h | 2
include/linux/highmem.h | 2
include/linux/huge_mm.h | 7
include/linux/hugetlb.h | 53 ++++--
include/linux/irqchip/irq-omap-intc.h | 2
include/linux/jhash.h | 2
include/linux/kernel.h | 12 -
include/linux/leds-ti-lmu-common.h | 2
include/linux/memcontrol.h | 12 +
include/linux/mempolicy.h | 18 +-
include/linux/migrate.h | 42 +---
include/linux/mm.h | 20 +-
include/linux/mmzone.h | 17 +
include/linux/oom.h | 4
include/linux/pgtable.h | 12 -
include/linux/platform_data/davinci-cpufreq.h | 2
include/linux/platform_data/davinci_asp.h | 2
include/linux/platform_data/elm.h | 2
include/linux/platform_data/gpio-davinci.h | 2
include/linux/platform_data/gpmc-omap.h | 2
include/linux/platform_data/mtd-davinci-aemif.h | 2
include/linux/platform_data/omap-twl4030.h | 2
include/linux/platform_data/uio_pruss.h | 2
include/linux/platform_data/usb-omap.h | 2
include/linux/poison.h | 4
include/linux/sched/mm.h | 8
include/linux/sched/task.h | 1
include/linux/soc/ti/k3-ringacc.h | 2
include/linux/soc/ti/knav_qmss.h | 2
include/linux/soc/ti/ti-msgmgr.h | 2
include/linux/swap.h | 25 ++
include/linux/syscalls.h | 2
include/linux/uaccess.h | 20 ++
include/linux/vm_event_item.h | 3
include/linux/wkup_m3_ipc.h | 2
include/linux/xxhash.h | 2
include/linux/xz.h | 4
include/linux/zlib.h | 2
include/soc/arc/aux.h | 2
include/trace/events/migrate.h | 17 +
include/uapi/linux/auto_dev-ioctl.h | 2
include/uapi/linux/elf.h | 2
include/uapi/linux/map_to_7segment.h | 2
include/uapi/linux/types.h | 2
include/uapi/linux/usb/ch9.h | 2
ipc/sem.c | 3
ipc/shm.c | 4
kernel/Makefile | 2
kernel/crash_core.c | 50 +++++
kernel/events/callchain.c | 5
kernel/events/core.c | 5
kernel/events/uprobes.c | 8
kernel/exit.c | 18 +-
kernel/futex.c | 2
kernel/kcov.c | 6
kernel/kmod.c | 5
kernel/kthread.c | 5
kernel/panic.c | 4
kernel/stacktrace.c | 5
kernel/sysctl.c | 11 +
kernel/umh.c | 29 ---
lib/Kconfig.debug | 27 ++-
lib/Makefile | 1
lib/bitmap.c | 4
lib/crc64.c | 2
lib/decompress_bunzip2.c | 2
lib/decompress_unlzma.c | 6
lib/kstrtox.c | 20 --
lib/lz4/lz4_compress.c | 4
lib/lz4/lz4_decompress.c | 18 +-
lib/lz4/lz4defs.h | 10 +
lib/lz4/lz4hc_compress.c | 2
lib/math/rational.c | 2
lib/rbtree.c | 2
lib/test_bitmap.c | 58 ++++++
lib/test_bitops.c | 18 +-
lib/test_bits.c | 75 ++++++++
lib/test_kmod.c | 2
lib/test_lockup.c | 6
lib/ts_bm.c | 2
lib/xxhash.c | 2
lib/xz/xz_crc32.c | 2
lib/xz/xz_dec_bcj.c | 2
lib/xz/xz_dec_lzma2.c | 2
lib/xz/xz_lzma2.h | 2
lib/xz/xz_stream.h | 2
mm/cma.c | 40 +---
mm/cma.h | 4
mm/compaction.c | 207 +++++++++++++++++++++--
mm/filemap.c | 2
mm/gup.c | 195 ++++++----------------
mm/hmm.c | 5
mm/huge_memory.c | 23 --
mm/hugetlb.c | 93 ++++------
mm/internal.h | 9 -
mm/khugepaged.c | 2
mm/ksm.c | 3
mm/maccess.c | 22 +-
mm/memcontrol.c | 42 +++-
mm/memory-failure.c | 7
mm/memory.c | 107 +++++++++---
mm/memory_hotplug.c | 30 ++-
mm/mempolicy.c | 49 +----
mm/migrate.c | 151 ++++++++++++++---
mm/mmu_notifier.c | 9 -
mm/nommu.c | 4
mm/oom_kill.c | 24 +-
mm/page_alloc.c | 14 +
mm/page_isolation.c | 21 --
mm/percpu-internal.h | 55 ++++++
mm/percpu-km.c | 5
mm/percpu-stats.c | 36 ++--
mm/percpu-vm.c | 5
mm/percpu.c | 208 +++++++++++++++++++++---
mm/process_vm_access.c | 2
mm/rmap.c | 2
mm/shmem.c | 5
mm/slab_common.c | 2
mm/swap.c | 13 -
mm/swap_state.c | 80 +++++++--
mm/swapfile.c | 4
mm/usercopy.c | 2
mm/userfaultfd.c | 2
mm/vmscan.c | 36 ++--
mm/vmstat.c | 32 +++
mm/workingset.c | 23 +-
mm/zpool.c | 8
mm/zsmalloc.c | 2
scripts/checkpatch.pl | 116 +++++++++----
scripts/gdb/linux/rbtree.py | 4
security/tomoyo/domain.c | 2
tools/testing/selftests/cgroup/test_kmem.c | 70 +++++++-
tools/testing/selftests/kmod/kmod.sh | 4
tools/testing/selftests/vm/hmm-tests.c | 35 ++++
virt/kvm/async_pf.c | 2
virt/kvm/kvm_main.c | 2
268 files changed, 2481 insertions(+), 1551 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-08-07 6:16 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-08-07 6:16 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
- A few MM hotfixes
- kthread, tools, scripts, ntfs and ocfs2
- Some of MM
163 patches, based on d6efb3ac3e6c19ab722b28bdb9252bae0b9676b6.
Subsystems affected by this patch series:
mm/pagemap
mm/hofixes
mm/pagealloc
kthread
tools
scripts
ntfs
ocfs2
mm/slab-generic
mm/slab
mm/slub
mm/kcsan
mm/debug
mm/pagecache
mm/gup
mm/swap
mm/shmem
mm/memcg
mm/pagemap
mm/mremap
mm/mincore
mm/sparsemem
mm/vmalloc
mm/kasan
mm/pagealloc
mm/hugetlb
mm/vmscan
Subsystem: mm/pagemap
Yang Shi <yang.shi@linux.alibaba.com>:
mm/memory.c: avoid access flag update TLB flush for retried page fault
Subsystem: mm/hofixes
Ralph Campbell <rcampbell@nvidia.com>:
mm/migrate: fix migrate_pgmap_owner w/o CONFIG_MMU_NOTIFIER
Subsystem: mm/pagealloc
David Hildenbrand <david@redhat.com>:
mm/shuffle: don't move pages between zones and don't read garbage memmaps
Subsystem: kthread
Peter Zijlstra <peterz@infradead.org>:
mm: fix kthread_use_mm() vs TLB invalidate
Ilias Stamatis <stamatis.iliass@gmail.com>:
kthread: remove incorrect comment in kthread_create_on_cpu()
Subsystem: tools
"Alexander A. Klimov" <grandmaster@al2klimov.de>:
tools/: replace HTTP links with HTTPS ones
Gaurav Singh <gaurav1086@gmail.com>:
tools/testing/selftests/cgroup/cgroup_util.c: cg_read_strcmp: fix null pointer dereference
Subsystem: scripts
Jialu Xu <xujialu@vimux.org>:
scripts/tags.sh: collect compiled source precisely
Nikolay Borisov <nborisov@suse.com>:
scripts/bloat-o-meter: Support comparing library archives
Konstantin Khlebnikov <khlebnikov@yandex-team.ru>:
scripts/decode_stacktrace.sh: skip missing symbols
scripts/decode_stacktrace.sh: guess basepath if not specified
scripts/decode_stacktrace.sh: guess path to modules
scripts/decode_stacktrace.sh: guess path to vmlinux by release name
Joe Perches <joe@perches.com>:
const_structs.checkpatch: add regulator_ops
Colin Ian King <colin.king@canonical.com>:
scripts/spelling.txt: add more spellings to spelling.txt
Subsystem: ntfs
Luca Stefani <luca.stefani.ge1@gmail.com>:
ntfs: fix ntfs_test_inode and ntfs_init_locked_inode function type
Subsystem: ocfs2
Gang He <ghe@suse.com>:
ocfs2: fix remounting needed after setfacl command
Randy Dunlap <rdunlap@infradead.org>:
ocfs2: suballoc.h: delete a duplicated word
Junxiao Bi <junxiao.bi@oracle.com>:
ocfs2: change slot number type s16 to u16
"Alexander A. Klimov" <grandmaster@al2klimov.de>:
ocfs2: replace HTTP links with HTTPS ones
Pavel Machek <pavel@ucw.cz>:
ocfs2: fix unbalanced locking
Subsystem: mm/slab-generic
Waiman Long <longman@redhat.com>:
mm, treewide: rename kzfree() to kfree_sensitive()
William Kucharski <william.kucharski@oracle.com>:
mm: ksize() should silently accept a NULL pointer
Subsystem: mm/slab
Kees Cook <keescook@chromium.org>:
Patch series "mm: Expand CONFIG_SLAB_FREELIST_HARDENED to include SLAB":
mm/slab: expand CONFIG_SLAB_FREELIST_HARDENED to include SLAB
mm/slab: add naive detection of double free
Long Li <lonuxli.64@gmail.com>:
mm, slab: check GFP_SLAB_BUG_MASK before alloc_pages in kmalloc_order
Xiao Yang <yangx.jy@cn.fujitsu.com>:
mm/slab.c: update outdated kmem_list3 in a comment
Subsystem: mm/slub
Vlastimil Babka <vbabka@suse.cz>:
Patch series "slub_debug fixes and improvements":
mm, slub: extend slub_debug syntax for multiple blocks
mm, slub: make some slub_debug related attributes read-only
mm, slub: remove runtime allocation order changes
mm, slub: make remaining slub_debug related attributes read-only
mm, slub: make reclaim_account attribute read-only
mm, slub: introduce static key for slub_debug()
mm, slub: introduce kmem_cache_debug_flags()
mm, slub: extend checks guarded by slub_debug static key
mm, slab/slub: move and improve cache_from_obj()
mm, slab/slub: improve error reporting and overhead of cache_from_obj()
Sebastian Andrzej Siewior <bigeasy@linutronix.de>:
mm/slub.c: drop lockdep_assert_held() from put_map()
Subsystem: mm/kcsan
Marco Elver <elver@google.com>:
mm, kcsan: instrument SLAB/SLUB free with "ASSERT_EXCLUSIVE_ACCESS"
Subsystem: mm/debug
Anshuman Khandual <anshuman.khandual@arm.com>:
Patch series "mm/debug_vm_pgtable: Add some more tests", v5:
mm/debug_vm_pgtable: add tests validating arch helpers for core MM features
mm/debug_vm_pgtable: add tests validating advanced arch page table helpers
mm/debug_vm_pgtable: add debug prints for individual tests
Documentation/mm: add descriptions for arch page table helpers
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
Patch series "Improvements for dump_page()", v2:
mm/debug: handle page->mapping better in dump_page
mm/debug: dump compound page information on a second line
mm/debug: print head flags in dump_page
mm/debug: switch dump_page to get_kernel_nofault
mm/debug: print the inode number in dump_page
mm/debug: print hashed address of struct page
John Hubbard <jhubbard@nvidia.com>:
mm, dump_page: do not crash with bad compound_mapcount()
Subsystem: mm/pagecache
Yang Shi <yang.shi@linux.alibaba.com>:
mm: filemap: clear idle flag for writes
mm: filemap: add missing FGP_ flags in kerneldoc comment for pagecache_get_page
Subsystem: mm/gup
Tang Yizhou <tangyizhou@huawei.com>:
mm/gup.c: fix the comment of return value for populate_vma_page_range()
Subsystem: mm/swap
Zhen Lei <thunder.leizhen@huawei.com>:
Patch series "clean up some functions in mm/swap_slots.c":
mm/swap_slots.c: simplify alloc_swap_slot_cache()
mm/swap_slots.c: simplify enable_swap_slots_cache()
mm/swap_slots.c: remove redundant check for swap_slot_cache_initialized
Krzysztof Kozlowski <krzk@kernel.org>:
mm: swap: fix kerneldoc of swap_vma_readahead()
Xianting Tian <xianting_tian@126.com>:
mm/page_io.c: use blk_io_schedule() for avoiding task hung in sync io
Subsystem: mm/shmem
Chris Down <chris@chrisdown.name>:
Patch series "tmpfs: inode: Reduce risk of inum overflow", v7:
tmpfs: per-superblock i_ino support
tmpfs: support 64-bit inums per-sb
Subsystem: mm/memcg
Roman Gushchin <guro@fb.com>:
mm: kmem: make memcg_kmem_enabled() irreversible
Patch series "The new cgroup slab memory controller", v7:
mm: memcg: factor out memcg- and lruvec-level changes out of __mod_lruvec_state()
mm: memcg: prepare for byte-sized vmstat items
mm: memcg: convert vmstat slab counters to bytes
mm: slub: implement SLUB version of obj_to_index()
Johannes Weiner <hannes@cmpxchg.org>:
mm: memcontrol: decouple reference counting from page accounting
Roman Gushchin <guro@fb.com>:
mm: memcg/slab: obj_cgroup API
mm: memcg/slab: allocate obj_cgroups for non-root slab pages
mm: memcg/slab: save obj_cgroup for non-root slab objects
mm: memcg/slab: charge individual slab objects instead of pages
mm: memcg/slab: deprecate memory.kmem.slabinfo
mm: memcg/slab: move memcg_kmem_bypass() to memcontrol.h
mm: memcg/slab: use a single set of kmem_caches for all accounted allocations
mm: memcg/slab: simplify memcg cache creation
mm: memcg/slab: remove memcg_kmem_get_cache()
mm: memcg/slab: deprecate slab_root_caches
mm: memcg/slab: remove redundant check in memcg_accumulate_slabinfo()
mm: memcg/slab: use a single set of kmem_caches for all allocations
kselftests: cgroup: add kernel memory accounting tests
tools/cgroup: add memcg_slabinfo.py tool
Shakeel Butt <shakeelb@google.com>:
mm: memcontrol: account kernel stack per node
Roman Gushchin <guro@fb.com>:
mm: memcg/slab: remove unused argument by charge_slab_page()
mm: slab: rename (un)charge_slab_page() to (un)account_slab_page()
mm: kmem: switch to static_branch_likely() in memcg_kmem_enabled()
mm: memcontrol: avoid workload stalls when lowering memory.high
Chris Down <chris@chrisdown.name>:
Patch series "mm, memcg: reclaim harder before high throttling", v2:
mm, memcg: reclaim more aggressively before high allocator throttling
mm, memcg: unify reclaim retry limits with page allocator
Yafang Shao <laoar.shao@gmail.com>:
Patch series "mm, memcg: memory.{low,min} reclaim fix & cleanup", v4:
mm, memcg: avoid stale protection values when cgroup is above protection
Chris Down <chris@chrisdown.name>:
mm, memcg: decouple e{low,min} state mutations from protection checks
Yafang Shao <laoar.shao@gmail.com>:
memcg, oom: check memcg margin for parallel oom
Johannes Weiner <hannes@cmpxchg.org>:
mm: memcontrol: restore proper dirty throttling when memory.high changes
mm: memcontrol: don't count limit-setting reclaim as memory pressure
Michal Koutný <mkoutny@suse.com>:
mm/page_counter.c: fix protection usage propagation
Subsystem: mm/pagemap
Ralph Campbell <rcampbell@nvidia.com>:
mm: remove redundant check non_swap_entry()
Alex Zhang <zhangalex@google.com>:
mm/memory.c: make remap_pfn_range() reject unaligned addr
Mike Rapoport <rppt@linux.ibm.com>:
Patch series "mm: cleanup usage of <asm/pgalloc.h>":
mm: remove unneeded includes of <asm/pgalloc.h>
opeinrisc: switch to generic version of pte allocation
xtensa: switch to generic version of pte allocation
asm-generic: pgalloc: provide generic pmd_alloc_one() and pmd_free_one()
asm-generic: pgalloc: provide generic pud_alloc_one() and pud_free_one()
asm-generic: pgalloc: provide generic pgd_free()
mm: move lib/ioremap.c to mm/
Joerg Roedel <jroedel@suse.de>:
mm: move p?d_alloc_track to separate header file
Zhen Lei <thunder.leizhen@huawei.com>:
mm/mmap: optimize a branch judgment in ksys_mmap_pgoff()
Feng Tang <feng.tang@intel.com>:
Patch series "make vm_committed_as_batch aware of vm overcommit policy", v6:
proc/meminfo: avoid open coded reading of vm_committed_as
mm/util.c: make vm_memory_committed() more accurate
percpu_counter: add percpu_counter_sync()
mm: adjust vm_committed_as_batch according to vm overcommit policy
Anshuman Khandual <anshuman.khandual@arm.com>:
Patch series "arm64: Enable vmemmap mapping from device memory", v4:
mm/sparsemem: enable vmem_altmap support in vmemmap_populate_basepages()
mm/sparsemem: enable vmem_altmap support in vmemmap_alloc_block_buf()
arm64/mm: enable vmem_altmap support for vmemmap mappings
Miaohe Lin <linmiaohe@huawei.com>:
mm: mmap: merge vma after call_mmap() if possible
Peter Collingbourne <pcc@google.com>:
mm: remove unnecessary wrapper function do_mmap_pgoff()
Subsystem: mm/mremap
Wei Yang <richard.weiyang@linux.alibaba.com>:
Patch series "mm/mremap: cleanup move_page_tables() a little", v5:
mm/mremap: it is sure to have enough space when extent meets requirement
mm/mremap: calculate extent in one place
mm/mremap: start addresses are properly aligned
Subsystem: mm/mincore
Ricardo Cañuelo <ricardo.canuelo@collabora.com>:
selftests: add mincore() tests
Subsystem: mm/sparsemem
Wei Yang <richard.weiyang@linux.alibaba.com>:
mm/sparse: never partially remove memmap for early section
mm/sparse: only sub-section aligned range would be populated
Mike Rapoport <rppt@linux.ibm.com>:
mm/sparse: cleanup the code surrounding memory_present()
Subsystem: mm/vmalloc
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
vmalloc: convert to XArray
"Uladzislau Rezki (Sony)" <urezki@gmail.com>:
mm/vmalloc: simplify merge_or_add_vmap_area()
mm/vmalloc: simplify augment_tree_propagate_check()
mm/vmalloc: switch to "propagate()" callback
mm/vmalloc: update the header about KVA rework
Mike Rapoport <rppt@linux.ibm.com>:
mm: vmalloc: remove redundant assignment in unmap_kernel_range_noflush()
"Uladzislau Rezki (Sony)" <urezki@gmail.com>:
mm/vmalloc.c: remove BUG() from the find_va_links()
Subsystem: mm/kasan
Marco Elver <elver@google.com>:
kasan: improve and simplify Kconfig.kasan
kasan: update required compiler versions in documentation
Walter Wu <walter-zh.wu@mediatek.com>:
Patch series "kasan: memorize and print call_rcu stack", v8:
rcu: kasan: record and print call_rcu() call stack
kasan: record and print the free track
kasan: add tests for call_rcu stack recording
kasan: update documentation for generic kasan
Vincenzo Frascino <vincenzo.frascino@arm.com>:
kasan: remove kasan_unpoison_stack_above_sp_to()
Walter Wu <walter-zh.wu@mediatek.com>:
lib/test_kasan.c: fix KASAN unit tests for tag-based KASAN
Andrey Konovalov <andreyknvl@google.com>:
Patch series "kasan: support stack instrumentation for tag-based mode", v2:
kasan: don't tag stacks allocated with pagealloc
efi: provide empty efi_enter_virtual_mode implementation
kasan, arm64: don't instrument functions that enable kasan
kasan: allow enabling stack tagging for tag-based mode
kasan: adjust kasan_stack_oob for tag-based mode
Subsystem: mm/pagealloc
Vlastimil Babka <vbabka@suse.cz>:
mm, page_alloc: use unlikely() in task_capc()
Jaewon Kim <jaewon31.kim@samsung.com>:
page_alloc: consider highatomic reserve in watermark fast
Charan Teja Reddy <charante@codeaurora.org>:
mm, page_alloc: skip ->waternark_boost for atomic order-0 allocations
David Hildenbrand <david@redhat.com>:
mm: remove vm_total_pages
mm/page_alloc: remove nr_free_pagecache_pages()
mm/memory_hotplug: document why shuffle_zone() is relevant
mm/shuffle: remove dynamic reconfiguration
Wei Yang <richard.weiyang@linux.alibaba.com>:
mm/page_alloc.c: replace the definition of NR_MIGRATETYPE_BITS with PB_migratetype_bits
mm/page_alloc.c: extract the common part in pfn_to_bitidx()
mm/page_alloc.c: simplify pageblock bitmap access
mm/page_alloc.c: remove unnecessary end_bitidx for [set|get]_pfnblock_flags_mask()
Qian Cai <cai@lca.pw>:
mm/page_alloc: silence a KASAN false positive
Wei Yang <richard.weiyang@linux.alibaba.com>:
mm/page_alloc: fallbacks at most has 3 elements
Muchun Song <songmuchun@bytedance.com>:
mm/page_alloc.c: skip setting nodemask when we are in interrupt
Joonsoo Kim <iamjoonsoo.kim@lge.com>:
mm/page_alloc: fix memalloc_nocma_{save/restore} APIs
Subsystem: mm/hugetlb
"Alexander A. Klimov" <grandmaster@al2klimov.de>:
mm: thp: replace HTTP links with HTTPS ones
Peter Xu <peterx@redhat.com>:
mm/hugetlb: fix calculation of adjust_range_if_pmd_sharing_possible
Hugh Dickins <hughd@google.com>:
khugepaged: collapse_pte_mapped_thp() flush the right range
khugepaged: collapse_pte_mapped_thp() protect the pmd lock
khugepaged: retract_page_tables() remember to test exit
khugepaged: khugepaged_test_exit() check mmget_still_valid()
Subsystem: mm/vmscan
dylan-meiners <spacct.spacct@gmail.com>:
mm/vmscan.c: fix typo
Shakeel Butt <shakeelb@google.com>:
mm: vmscan: consistent update to pgrefill
Documentation/admin-guide/kernel-parameters.txt | 2
Documentation/dev-tools/kasan.rst | 10
Documentation/filesystems/dlmfs.rst | 2
Documentation/filesystems/ocfs2.rst | 2
Documentation/filesystems/tmpfs.rst | 18
Documentation/vm/arch_pgtable_helpers.rst | 258 +++++
Documentation/vm/memory-model.rst | 9
Documentation/vm/slub.rst | 51 -
arch/alpha/include/asm/pgalloc.h | 21
arch/alpha/include/asm/tlbflush.h | 1
arch/alpha/kernel/core_irongate.c | 1
arch/alpha/kernel/core_marvel.c | 1
arch/alpha/kernel/core_titan.c | 1
arch/alpha/kernel/machvec_impl.h | 2
arch/alpha/kernel/smp.c | 1
arch/alpha/mm/numa.c | 1
arch/arc/mm/fault.c | 1
arch/arc/mm/init.c | 1
arch/arm/include/asm/pgalloc.h | 12
arch/arm/include/asm/tlb.h | 1
arch/arm/kernel/machine_kexec.c | 1
arch/arm/kernel/smp.c | 1
arch/arm/kernel/suspend.c | 1
arch/arm/mach-omap2/omap-mpuss-lowpower.c | 1
arch/arm/mm/hugetlbpage.c | 1
arch/arm/mm/init.c | 9
arch/arm/mm/mmu.c | 1
arch/arm64/include/asm/pgalloc.h | 39
arch/arm64/kernel/setup.c | 2
arch/arm64/kernel/smp.c | 1
arch/arm64/mm/hugetlbpage.c | 1
arch/arm64/mm/init.c | 6
arch/arm64/mm/ioremap.c | 1
arch/arm64/mm/mmu.c | 63 -
arch/csky/include/asm/pgalloc.h | 7
arch/csky/kernel/smp.c | 1
arch/hexagon/include/asm/pgalloc.h | 7
arch/ia64/include/asm/pgalloc.h | 24
arch/ia64/include/asm/tlb.h | 1
arch/ia64/kernel/process.c | 1
arch/ia64/kernel/smp.c | 1
arch/ia64/kernel/smpboot.c | 1
arch/ia64/mm/contig.c | 1
arch/ia64/mm/discontig.c | 4
arch/ia64/mm/hugetlbpage.c | 1
arch/ia64/mm/tlb.c | 1
arch/m68k/include/asm/mmu_context.h | 2
arch/m68k/include/asm/sun3_pgalloc.h | 7
arch/m68k/kernel/dma.c | 2
arch/m68k/kernel/traps.c | 3
arch/m68k/mm/cache.c | 2
arch/m68k/mm/fault.c | 1
arch/m68k/mm/kmap.c | 2
arch/m68k/mm/mcfmmu.c | 1
arch/m68k/mm/memory.c | 1
arch/m68k/sun3x/dvma.c | 2
arch/microblaze/include/asm/pgalloc.h | 6
arch/microblaze/include/asm/tlbflush.h | 1
arch/microblaze/kernel/process.c | 1
arch/microblaze/kernel/signal.c | 1
arch/microblaze/mm/init.c | 3
arch/mips/include/asm/pgalloc.h | 19
arch/mips/kernel/setup.c | 8
arch/mips/loongson64/numa.c | 1
arch/mips/sgi-ip27/ip27-memory.c | 2
arch/mips/sgi-ip32/ip32-memory.c | 1
arch/nds32/mm/mm-nds32.c | 2
arch/nios2/include/asm/pgalloc.h | 7
arch/openrisc/include/asm/pgalloc.h | 33
arch/openrisc/include/asm/tlbflush.h | 1
arch/openrisc/kernel/or32_ksyms.c | 1
arch/parisc/include/asm/mmu_context.h | 1
arch/parisc/include/asm/pgalloc.h | 12
arch/parisc/kernel/cache.c | 1
arch/parisc/kernel/pci-dma.c | 1
arch/parisc/kernel/process.c | 1
arch/parisc/kernel/signal.c | 1
arch/parisc/kernel/smp.c | 1
arch/parisc/mm/hugetlbpage.c | 1
arch/parisc/mm/init.c | 5
arch/parisc/mm/ioremap.c | 2
arch/powerpc/include/asm/tlb.h | 1
arch/powerpc/mm/book3s64/hash_hugetlbpage.c | 1
arch/powerpc/mm/book3s64/hash_pgtable.c | 1
arch/powerpc/mm/book3s64/hash_tlb.c | 1
arch/powerpc/mm/book3s64/radix_hugetlbpage.c | 1
arch/powerpc/mm/init_32.c | 1
arch/powerpc/mm/init_64.c | 4
arch/powerpc/mm/kasan/8xx.c | 1
arch/powerpc/mm/kasan/book3s_32.c | 1
arch/powerpc/mm/mem.c | 3
arch/powerpc/mm/nohash/40x.c | 1
arch/powerpc/mm/nohash/8xx.c | 1
arch/powerpc/mm/nohash/fsl_booke.c | 1
arch/powerpc/mm/nohash/kaslr_booke.c | 1
arch/powerpc/mm/nohash/tlb.c | 1
arch/powerpc/mm/numa.c | 1
arch/powerpc/mm/pgtable.c | 1
arch/powerpc/mm/pgtable_64.c | 1
arch/powerpc/mm/ptdump/hashpagetable.c | 2
arch/powerpc/mm/ptdump/ptdump.c | 1
arch/powerpc/platforms/pseries/cmm.c | 1
arch/riscv/include/asm/pgalloc.h | 18
arch/riscv/mm/fault.c | 1
arch/riscv/mm/init.c | 3
arch/s390/crypto/prng.c | 4
arch/s390/include/asm/tlb.h | 1
arch/s390/include/asm/tlbflush.h | 1
arch/s390/kernel/machine_kexec.c | 1
arch/s390/kernel/ptrace.c | 1
arch/s390/kvm/diag.c | 1
arch/s390/kvm/priv.c | 1
arch/s390/kvm/pv.c | 1
arch/s390/mm/cmm.c | 1
arch/s390/mm/init.c | 1
arch/s390/mm/mmap.c | 1
arch/s390/mm/pgtable.c | 1
arch/sh/include/asm/pgalloc.h | 4
arch/sh/kernel/idle.c | 1
arch/sh/kernel/machine_kexec.c | 1
arch/sh/mm/cache-sh3.c | 1
arch/sh/mm/cache-sh7705.c | 1
arch/sh/mm/hugetlbpage.c | 1
arch/sh/mm/init.c | 7
arch/sh/mm/ioremap_fixed.c | 1
arch/sh/mm/numa.c | 3
arch/sh/mm/tlb-sh3.c | 1
arch/sparc/include/asm/ide.h | 1
arch/sparc/include/asm/tlb_64.h | 1
arch/sparc/kernel/leon_smp.c | 1
arch/sparc/kernel/process_32.c | 1
arch/sparc/kernel/signal_32.c | 1
arch/sparc/kernel/smp_32.c | 1
arch/sparc/kernel/smp_64.c | 1
arch/sparc/kernel/sun4m_irq.c | 1
arch/sparc/mm/highmem.c | 1
arch/sparc/mm/init_64.c | 1
arch/sparc/mm/io-unit.c | 1
arch/sparc/mm/iommu.c | 1
arch/sparc/mm/tlb.c | 1
arch/um/include/asm/pgalloc.h | 9
arch/um/include/asm/pgtable-3level.h | 3
arch/um/kernel/mem.c | 17
arch/x86/ia32/ia32_aout.c | 1
arch/x86/include/asm/mmu_context.h | 1
arch/x86/include/asm/pgalloc.h | 42
arch/x86/kernel/alternative.c | 1
arch/x86/kernel/apic/apic.c | 1
arch/x86/kernel/mpparse.c | 1
arch/x86/kernel/traps.c | 1
arch/x86/mm/fault.c | 1
arch/x86/mm/hugetlbpage.c | 1
arch/x86/mm/init_32.c | 2
arch/x86/mm/init_64.c | 12
arch/x86/mm/kaslr.c | 1
arch/x86/mm/pgtable_32.c | 1
arch/x86/mm/pti.c | 1
arch/x86/platform/uv/bios_uv.c | 1
arch/x86/power/hibernate.c | 2
arch/xtensa/include/asm/pgalloc.h | 46
arch/xtensa/kernel/xtensa_ksyms.c | 1
arch/xtensa/mm/cache.c | 1
arch/xtensa/mm/fault.c | 1
crypto/adiantum.c | 2
crypto/ahash.c | 4
crypto/api.c | 2
crypto/asymmetric_keys/verify_pefile.c | 4
crypto/deflate.c | 2
crypto/drbg.c | 10
crypto/ecc.c | 8
crypto/ecdh.c | 2
crypto/gcm.c | 2
crypto/gf128mul.c | 4
crypto/jitterentropy-kcapi.c | 2
crypto/rng.c | 2
crypto/rsa-pkcs1pad.c | 6
crypto/seqiv.c | 2
crypto/shash.c | 2
crypto/skcipher.c | 2
crypto/testmgr.c | 6
crypto/zstd.c | 2
drivers/base/node.c | 10
drivers/block/xen-blkback/common.h | 1
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c | 2
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c | 2
drivers/crypto/amlogic/amlogic-gxl-cipher.c | 4
drivers/crypto/atmel-ecc.c | 2
drivers/crypto/caam/caampkc.c | 28
drivers/crypto/cavium/cpt/cptvf_main.c | 6
drivers/crypto/cavium/cpt/cptvf_reqmanager.c | 12
drivers/crypto/cavium/nitrox/nitrox_lib.c | 4
drivers/crypto/cavium/zip/zip_crypto.c | 6
drivers/crypto/ccp/ccp-crypto-rsa.c | 6
drivers/crypto/ccree/cc_aead.c | 4
drivers/crypto/ccree/cc_buffer_mgr.c | 4
drivers/crypto/ccree/cc_cipher.c | 6
drivers/crypto/ccree/cc_hash.c | 8
drivers/crypto/ccree/cc_request_mgr.c | 2
drivers/crypto/marvell/cesa/hash.c | 2
drivers/crypto/marvell/octeontx/otx_cptvf_main.c | 6
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.h | 2
drivers/crypto/nx/nx.c | 4
drivers/crypto/virtio/virtio_crypto_algs.c | 12
drivers/crypto/virtio/virtio_crypto_core.c | 2
drivers/iommu/ipmmu-vmsa.c | 1
drivers/md/dm-crypt.c | 32
drivers/md/dm-integrity.c | 6
drivers/misc/ibmvmc.c | 6
drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_mbx.c | 2
drivers/net/ethernet/intel/ixgbe/ixgbe_ipsec.c | 6
drivers/net/ppp/ppp_mppe.c | 6
drivers/net/wireguard/noise.c | 4
drivers/net/wireguard/peer.c | 2
drivers/net/wireless/intel/iwlwifi/pcie/rx.c | 2
drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c | 6
drivers/net/wireless/intel/iwlwifi/pcie/tx.c | 6
drivers/net/wireless/intersil/orinoco/wext.c | 4
drivers/s390/crypto/ap_bus.h | 4
drivers/staging/ks7010/ks_hostif.c | 2
drivers/staging/rtl8723bs/core/rtw_security.c | 2
drivers/staging/wlan-ng/p80211netdev.c | 2
drivers/target/iscsi/iscsi_target_auth.c | 2
drivers/xen/balloon.c | 1
drivers/xen/privcmd.c | 1
fs/Kconfig | 21
fs/aio.c | 6
fs/binfmt_elf_fdpic.c | 1
fs/cifs/cifsencrypt.c | 2
fs/cifs/connect.c | 10
fs/cifs/dfs_cache.c | 2
fs/cifs/misc.c | 8
fs/crypto/inline_crypt.c | 5
fs/crypto/keyring.c | 6
fs/crypto/keysetup_v1.c | 4
fs/ecryptfs/keystore.c | 4
fs/ecryptfs/messaging.c | 2
fs/hugetlbfs/inode.c | 2
fs/ntfs/dir.c | 2
fs/ntfs/inode.c | 27
fs/ntfs/inode.h | 4
fs/ntfs/mft.c | 4
fs/ocfs2/Kconfig | 6
fs/ocfs2/acl.c | 2
fs/ocfs2/blockcheck.c | 2
fs/ocfs2/dlmglue.c | 8
fs/ocfs2/ocfs2.h | 4
fs/ocfs2/suballoc.c | 4
fs/ocfs2/suballoc.h | 2
fs/ocfs2/super.c | 4
fs/proc/meminfo.c | 10
include/asm-generic/pgalloc.h | 80 +
include/asm-generic/tlb.h | 1
include/crypto/aead.h | 2
include/crypto/akcipher.h | 2
include/crypto/gf128mul.h | 2
include/crypto/hash.h | 2
include/crypto/internal/acompress.h | 2
include/crypto/kpp.h | 2
include/crypto/skcipher.h | 2
include/linux/efi.h | 4
include/linux/fs.h | 17
include/linux/huge_mm.h | 2
include/linux/kasan.h | 4
include/linux/memcontrol.h | 209 +++-
include/linux/mm.h | 86 -
include/linux/mm_types.h | 5
include/linux/mman.h | 4
include/linux/mmu_notifier.h | 13
include/linux/mmzone.h | 54 -
include/linux/pageblock-flags.h | 30
include/linux/percpu_counter.h | 4
include/linux/sched/mm.h | 8
include/linux/shmem_fs.h | 3
include/linux/slab.h | 11
include/linux/slab_def.h | 9
include/linux/slub_def.h | 31
include/linux/swap.h | 2
include/linux/vmstat.h | 14
init/Kconfig | 9
init/main.c | 2
ipc/shm.c | 2
kernel/fork.c | 54 -
kernel/kthread.c | 8
kernel/power/snapshot.c | 2
kernel/rcu/tree.c | 2
kernel/scs.c | 2
kernel/sysctl.c | 2
lib/Kconfig.kasan | 39
lib/Makefile | 1
lib/ioremap.c | 287 -----
lib/mpi/mpiutil.c | 6
lib/percpu_counter.c | 19
lib/test_kasan.c | 87 +
mm/Kconfig | 6
mm/Makefile | 2
mm/debug.c | 103 +-
mm/debug_vm_pgtable.c | 666 +++++++++++++
mm/filemap.c | 9
mm/gup.c | 3
mm/huge_memory.c | 14
mm/hugetlb.c | 25
mm/ioremap.c | 289 +++++
mm/kasan/common.c | 41
mm/kasan/generic.c | 43
mm/kasan/generic_report.c | 1
mm/kasan/kasan.h | 25
mm/kasan/quarantine.c | 1
mm/kasan/report.c | 54 -
mm/kasan/tags.c | 37
mm/khugepaged.c | 75 -
mm/memcontrol.c | 832 ++++++++++-------
mm/memory.c | 15
mm/memory_hotplug.c | 11
mm/migrate.c | 6
mm/mm_init.c | 20
mm/mmap.c | 45
mm/mremap.c | 19
mm/nommu.c | 6
mm/oom_kill.c | 2
mm/page-writeback.c | 6
mm/page_alloc.c | 226 ++--
mm/page_counter.c | 6
mm/page_io.c | 2
mm/pgalloc-track.h | 51 +
mm/shmem.c | 133 ++
mm/shuffle.c | 46
mm/shuffle.h | 17
mm/slab.c | 129 +-
mm/slab.h | 755 ++++++---------
mm/slab_common.c | 829 ++--------------
mm/slob.c | 12
mm/slub.c | 680 ++++---------
mm/sparse-vmemmap.c | 62 -
mm/sparse.c | 31
mm/swap_slots.c | 45
mm/swap_state.c | 2
mm/util.c | 52 +
mm/vmalloc.c | 176 +--
mm/vmscan.c | 39
mm/vmstat.c | 38
mm/workingset.c | 6
net/atm/mpoa_caches.c | 4
net/bluetooth/ecdh_helper.c | 6
net/bluetooth/smp.c | 24
net/core/sock.c | 2
net/ipv4/tcp_fastopen.c | 2
net/mac80211/aead_api.c | 4
net/mac80211/aes_gmac.c | 2
net/mac80211/key.c | 2
net/mac802154/llsec.c | 20
net/sctp/auth.c | 2
net/sunrpc/auth_gss/gss_krb5_crypto.c | 4
net/sunrpc/auth_gss/gss_krb5_keys.c | 6
net/sunrpc/auth_gss/gss_krb5_mech.c | 2
net/tipc/crypto.c | 10
net/wireless/core.c | 2
net/wireless/ibss.c | 4
net/wireless/lib80211_crypt_tkip.c | 2
net/wireless/lib80211_crypt_wep.c | 2
net/wireless/nl80211.c | 24
net/wireless/sme.c | 6
net/wireless/util.c | 2
net/wireless/wext-sme.c | 2
scripts/Makefile.kasan | 3
scripts/bloat-o-meter | 2
scripts/coccinelle/free/devm_free.cocci | 4
scripts/coccinelle/free/ifnullfree.cocci | 4
scripts/coccinelle/free/kfree.cocci | 6
scripts/coccinelle/free/kfreeaddr.cocci | 2
scripts/const_structs.checkpatch | 1
scripts/decode_stacktrace.sh | 85 +
scripts/spelling.txt | 19
scripts/tags.sh | 18
security/apparmor/domain.c | 4
security/apparmor/include/file.h | 2
security/apparmor/policy.c | 24
security/apparmor/policy_ns.c | 6
security/apparmor/policy_unpack.c | 14
security/keys/big_key.c | 6
security/keys/dh.c | 14
security/keys/encrypted-keys/encrypted.c | 14
security/keys/trusted-keys/trusted_tpm1.c | 34
security/keys/user_defined.c | 6
tools/cgroup/memcg_slabinfo.py | 226 ++++
tools/include/linux/jhash.h | 2
tools/lib/rbtree.c | 2
tools/lib/traceevent/event-parse.h | 2
tools/testing/ktest/examples/README | 2
tools/testing/ktest/examples/crosstests.conf | 2
tools/testing/selftests/Makefile | 1
tools/testing/selftests/cgroup/.gitignore | 1
tools/testing/selftests/cgroup/Makefile | 2
tools/testing/selftests/cgroup/cgroup_util.c | 2
tools/testing/selftests/cgroup/test_kmem.c | 382 +++++++
tools/testing/selftests/mincore/.gitignore | 2
tools/testing/selftests/mincore/Makefile | 6
tools/testing/selftests/mincore/mincore_selftest.c | 361 +++++++
397 files changed, 5547 insertions(+), 4072 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-07-24 4:14 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-07-24 4:14 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
15 patches, based on f37e99aca03f63aa3f2bd13ceaf769455d12c4b0.
Subsystems affected by this patch series:
mm/pagemap
mm/shmem
mm/hotfixes
mm/memcg
mm/hugetlb
mailmap
squashfs
scripts
io-mapping
MAINTAINERS
gdb
Subsystem: mm/pagemap
Yang Shi <yang.shi@linux.alibaba.com>:
mm/memory.c: avoid access flag update TLB flush for retried page fault
"Kirill A. Shutemov" <kirill.shutemov@linux.intel.com>:
mm/mmap.c: close race between munmap() and expand_upwards()/downwards()
Subsystem: mm/shmem
Chengguang Xu <cgxu519@mykernel.net>:
vfs/xattr: mm/shmem: kernfs: release simple xattr entry in a right way
Subsystem: mm/hotfixes
Tom Rix <trix@redhat.com>:
mm: initialize return of vm_insert_pages
Bhupesh Sharma <bhsharma@redhat.com>:
mm/memcontrol: fix OOPS inside mem_cgroup_get_nr_swap_pages()
Subsystem: mm/memcg
Hugh Dickins <hughd@google.com>:
mm/memcg: fix refcount error while moving and swapping
Muchun Song <songmuchun@bytedance.com>:
mm: memcg/slab: fix memory leak at non-root kmem_cache destroy
Subsystem: mm/hugetlb
Barry Song <song.bao.hua@hisilicon.com>:
mm/hugetlb: avoid hardcoding while checking if cma is enabled
"Kirill A. Shutemov" <kirill.shutemov@linux.intel.com>:
khugepaged: fix null-pointer dereference due to race
Subsystem: mailmap
Mike Rapoport <rppt@linux.ibm.com>:
mailmap: add entry for Mike Rapoport
Subsystem: squashfs
Phillip Lougher <phillip@squashfs.org.uk>:
squashfs: fix length field overlap check in metadata reading
Subsystem: scripts
Pi-Hsun Shih <pihsun@chromium.org>:
scripts/decode_stacktrace: strip basepath from all paths
Subsystem: io-mapping
"Michael J. Ruhl" <michael.j.ruhl@intel.com>:
io-mapping: indicate mapping failure
Subsystem: MAINTAINERS
Andrey Konovalov <andreyknvl@google.com>:
MAINTAINERS: add KCOV section
Subsystem: gdb
Stefano Garzarella <sgarzare@redhat.com>:
scripts/gdb: fix lx-symbols 'gdb.error' while loading modules
.mailmap | 3 +++
MAINTAINERS | 11 +++++++++++
fs/squashfs/block.c | 2 +-
include/linux/io-mapping.h | 5 ++++-
include/linux/xattr.h | 3 ++-
mm/hugetlb.c | 15 ++++++++++-----
mm/khugepaged.c | 3 +++
mm/memcontrol.c | 13 ++++++++++---
mm/memory.c | 9 +++++++--
mm/mmap.c | 16 ++++++++++++++--
mm/shmem.c | 2 +-
mm/slab_common.c | 35 ++++++++++++++++++++++++++++-------
scripts/decode_stacktrace.sh | 4 ++--
scripts/gdb/linux/symbols.py | 2 +-
14 files changed, 97 insertions(+), 26 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-07-03 22:14 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-07-03 22:14 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
5 patches, based on cdd3bb54332f82295ed90cd0c09c78cd0c0ee822.
Subsystems affected by this patch series:
mm/hugetlb
samples
mm/cma
mm/vmalloc
mm/pagealloc
Subsystem: mm/hugetlb
Mike Kravetz <mike.kravetz@oracle.com>:
mm/hugetlb.c: fix pages per hugetlb calculation
Subsystem: samples
Kees Cook <keescook@chromium.org>:
samples/vfs: avoid warning in statx override
Subsystem: mm/cma
Barry Song <song.bao.hua@hisilicon.com>:
mm/cma.c: use exact_nid true to fix possible per-numa cma leak
Subsystem: mm/vmalloc
Christoph Hellwig <hch@lst.de>:
vmalloc: fix the owner argument for the new __vmalloc_node_range callers
Subsystem: mm/pagealloc
Joel Savitz <jsavitz@redhat.com>:
mm/page_alloc: fix documentation error
arch/arm64/kernel/probes/kprobes.c | 2 +-
arch/x86/hyperv/hv_init.c | 3 ++-
kernel/module.c | 2 +-
mm/cma.c | 4 ++--
mm/hugetlb.c | 2 +-
mm/page_alloc.c | 2 +-
samples/vfs/test-statx.c | 2 ++
7 files changed, 10 insertions(+), 7 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2020-06-26 17:39 ` incoming Konstantin Ryabitsev
@ 2020-06-26 17:40 ` Konstantin Ryabitsev
0 siblings, 0 replies; 409+ messages in thread
From: Konstantin Ryabitsev @ 2020-06-26 17:40 UTC (permalink / raw)
To: Linus Torvalds; +Cc: Andrew Morton, Linux-MM, mm-commits
On Fri, 26 Jun 2020 at 13:39, Konstantin Ryabitsev
<konstantin@linuxfoundation.org> wrote:
> > Konstantin, maybe mm-commits could be on lore too and then they'd have
> > been caught that way?
>
> Yes, I already have a request from Kees for linux-mm addition, so that
> should show up in archives before long.
correction: mm-commits, that is
-K
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2020-06-26 6:51 ` incoming Linus Torvalds
2020-06-26 7:31 ` incoming Linus Torvalds
@ 2020-06-26 17:39 ` Konstantin Ryabitsev
2020-06-26 17:40 ` incoming Konstantin Ryabitsev
1 sibling, 1 reply; 409+ messages in thread
From: Konstantin Ryabitsev @ 2020-06-26 17:39 UTC (permalink / raw)
To: Linus Torvalds; +Cc: Andrew Morton, Linux-MM, mm-commits
On Thu, Jun 25, 2020 at 11:51:06PM -0700, Linus Torvalds wrote:
> On Thu, Jun 25, 2020 at 8:28 PM Andrew Morton <akpm@linux-foundation.org> wrote:
> >
> > 32 patches, based on 908f7d12d3ba51dfe0449b9723199b423f97ca9a.
>
> You didn't cc lkml, so now none of the nice 'b4' automation seems to
> work for this series..
>
> Yes, this cover-letter went to linux-mm (which is on lore), but the
> individual patches didn't.
>
> Konstantin, maybe mm-commits could be on lore too and then they'd have
> been caught that way?
Yes, I already have a request from Kees for linux-mm addition, so that
should show up in archives before long.
-K
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2020-06-26 6:51 ` incoming Linus Torvalds
@ 2020-06-26 7:31 ` Linus Torvalds
2020-06-26 17:39 ` incoming Konstantin Ryabitsev
1 sibling, 0 replies; 409+ messages in thread
From: Linus Torvalds @ 2020-06-26 7:31 UTC (permalink / raw)
To: Andrew Morton, Konstantin Ryabitsev; +Cc: Linux-MM, mm-commits
On Thu, Jun 25, 2020 at 11:51 PM Linus Torvalds
<torvalds@linux-foundation.org> wrote:
>
> You didn't cc lkml, so now none of the nice 'b4' automation seems to
> work for this series..
Note that I've picked them up the old-fashioned way, so don't re-send them.
So more of a note for "please, next time..."
Linus
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2020-06-26 3:28 incoming Andrew Morton
@ 2020-06-26 6:51 ` Linus Torvalds
2020-06-26 7:31 ` incoming Linus Torvalds
2020-06-26 17:39 ` incoming Konstantin Ryabitsev
0 siblings, 2 replies; 409+ messages in thread
From: Linus Torvalds @ 2020-06-26 6:51 UTC (permalink / raw)
To: Andrew Morton, Konstantin Ryabitsev; +Cc: Linux-MM, mm-commits
On Thu, Jun 25, 2020 at 8:28 PM Andrew Morton <akpm@linux-foundation.org> wrote:
>
> 32 patches, based on 908f7d12d3ba51dfe0449b9723199b423f97ca9a.
You didn't cc lkml, so now none of the nice 'b4' automation seems to
work for this series..
Yes, this cover-letter went to linux-mm (which is on lore), but the
individual patches didn't.
Konstantin, maybe mm-commits could be on lore too and then they'd have
been caught that way?
Linus
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-06-26 3:28 Andrew Morton
2020-06-26 6:51 ` incoming Linus Torvalds
0 siblings, 1 reply; 409+ messages in thread
From: Andrew Morton @ 2020-06-26 3:28 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
32 patches, based on 908f7d12d3ba51dfe0449b9723199b423f97ca9a.
Subsystems affected by this patch series:
hotfixes
mm/pagealloc
kexec
ocfs2
lib
misc
mm/slab
mm/slab
mm/slub
mm/swap
mm/pagemap
mm/vmalloc
mm/memcg
mm/gup
mm/thp
mm/vmscan
x86
mm/memory-hotplug
MAINTAINERS
Subsystem: hotfixes
Stafford Horne <shorne@gmail.com>:
openrisc: fix boot oops when DEBUG_VM is enabled
Michal Hocko <mhocko@suse.com>:
mm: do_swap_page(): fix up the error code
Subsystem: mm/pagealloc
Vlastimil Babka <vbabka@suse.cz>:
mm, compaction: make capture control handling safe wrt interrupts
Subsystem: kexec
Lianbo Jiang <lijiang@redhat.com>:
kexec: do not verify the signature without the lockdown or mandatory signature
Subsystem: ocfs2
Junxiao Bi <junxiao.bi@oracle.com>:
Patch series "ocfs2: fix nfsd over ocfs2 issues", v2:
ocfs2: avoid inode removal while nfsd is accessing it
ocfs2: load global_inode_alloc
ocfs2: fix panic on nfs server over ocfs2
ocfs2: fix value of OCFS2_INVALID_SLOT
Subsystem: lib
Randy Dunlap <rdunlap@infradead.org>:
lib: fix test_hmm.c reference after free
Subsystem: misc
Rikard Falkeborn <rikard.falkeborn@gmail.com>:
linux/bits.h: fix unsigned less than zero warnings
Subsystem: mm/slab
Waiman Long <longman@redhat.com>:
mm, slab: fix sign conversion problem in memcg_uncharge_slab()
Subsystem: mm/slab
Waiman Long <longman@redhat.com>:
mm/slab: use memzero_explicit() in kzfree()
Subsystem: mm/slub
Sebastian Andrzej Siewior <bigeasy@linutronix.de>:
slub: cure list_slab_objects() from double fix
Subsystem: mm/swap
Hugh Dickins <hughd@google.com>:
mm: fix swap cache node allocation mask
Subsystem: mm/pagemap
Arjun Roy <arjunroy@google.com>:
mm/memory.c: properly pte_offset_map_lock/unlock in vm_insert_pages()
Christophe Leroy <christophe.leroy@csgroup.eu>:
mm/debug_vm_pgtable: fix build failure with powerpc 8xx
Stephen Rothwell <sfr@canb.auug.org.au>:
make asm-generic/cacheflush.h more standalone
Nathan Chancellor <natechancellor@gmail.com>:
media: omap3isp: remove cacheflush.h
Subsystem: mm/vmalloc
Masanari Iida <standby24x7@gmail.com>:
mm/vmalloc.c: fix a warning while make xmldocs
Subsystem: mm/memcg
Johannes Weiner <hannes@cmpxchg.org>:
mm: memcontrol: handle div0 crash race condition in memory.low
Muchun Song <songmuchun@bytedance.com>:
mm/memcontrol.c: add missed css_put()
Chris Down <chris@chrisdown.name>:
mm/memcontrol.c: prevent missed memory.low load tears
Subsystem: mm/gup
Souptick Joarder <jrdr.linux@gmail.com>:
docs: mm/gup: minor documentation update
Subsystem: mm/thp
Yang Shi <yang.shi@linux.alibaba.com>:
doc: THP CoW fault no longer allocate THP
Subsystem: mm/vmscan
Johannes Weiner <hannes@cmpxchg.org>:
Patch series "fix for "mm: balance LRU lists based on relative thrashing" patchset":
mm: workingset: age nonresident information alongside anonymous pages
Joonsoo Kim <iamjoonsoo.kim@lge.com>:
mm/swap: fix for "mm: workingset: age nonresident information alongside anonymous pages"
mm/memory: fix IO cost for anonymous page
Subsystem: x86
Christoph Hellwig <hch@lst.de>:
Patch series "fix a hyperv W^X violation and remove vmalloc_exec":
x86/hyperv: allocate the hypercall page with only read and execute bits
arm64: use PAGE_KERNEL_ROX directly in alloc_insn_page
mm: remove vmalloc_exec
Subsystem: mm/memory-hotplug
Ben Widawsky <ben.widawsky@intel.com>:
mm/memory_hotplug.c: fix false softlockup during pfn range removal
Subsystem: MAINTAINERS
Luc Van Oostenryck <luc.vanoostenryck@gmail.com>:
MAINTAINERS: update info for sparse
Documentation/admin-guide/cgroup-v2.rst | 4 +-
Documentation/admin-guide/mm/transhuge.rst | 3 -
Documentation/core-api/pin_user_pages.rst | 2 -
MAINTAINERS | 4 +-
arch/arm64/kernel/probes/kprobes.c | 12 +------
arch/openrisc/kernel/dma.c | 5 +++
arch/x86/hyperv/hv_init.c | 4 +-
arch/x86/include/asm/pgtable_types.h | 2 +
drivers/media/platform/omap3isp/isp.c | 2 -
drivers/media/platform/omap3isp/ispvideo.c | 1
fs/ocfs2/dlmglue.c | 17 ++++++++++
fs/ocfs2/ocfs2.h | 1
fs/ocfs2/ocfs2_fs.h | 4 +-
fs/ocfs2/suballoc.c | 9 +++--
include/asm-generic/cacheflush.h | 5 +++
include/linux/bits.h | 3 +
include/linux/mmzone.h | 4 +-
include/linux/swap.h | 1
include/linux/vmalloc.h | 1
kernel/kexec_file.c | 36 ++++------------------
kernel/module.c | 4 +-
lib/test_hmm.c | 3 -
mm/compaction.c | 17 ++++++++--
mm/debug_vm_pgtable.c | 4 +-
mm/memcontrol.c | 18 ++++++++---
mm/memory.c | 33 +++++++++++++-------
mm/memory_hotplug.c | 13 ++++++--
mm/nommu.c | 17 ----------
mm/slab.h | 4 +-
mm/slab_common.c | 2 -
mm/slub.c | 19 ++---------
mm/swap.c | 3 -
mm/swap_state.c | 4 +-
mm/vmalloc.c | 21 -------------
mm/vmscan.c | 3 +
mm/workingset.c | 46 +++++++++++++++++------------
36 files changed, 168 insertions(+), 163 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-06-12 0:30 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-06-12 0:30 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
A few fixes and stragglers.
5 patches, based on 623f6dc593eaf98b91916836785278eddddaacf8.
Subsystems affected by this patch series:
mm/memory-failure
ocfs2
lib/lzo
misc
Subsystem: mm/memory-failure
Naoya Horiguchi <nao.horiguchi@gmail.com>:
Patch series "hwpoison: fixes signaling on memory error":
mm/memory-failure: prioritize prctl(PR_MCE_KILL) over vm.memory_failure_early_kill
mm/memory-failure: send SIGBUS(BUS_MCEERR_AR) only to current thread
Subsystem: ocfs2
Tom Seewald <tseewald@gmail.com>:
ocfs2: fix build failure when TCP/IP is disabled
Subsystem: lib/lzo
Dave Rodgman <dave.rodgman@arm.com>:
lib/lzo: fix ambiguous encoding bug in lzo-rle
Subsystem: misc
Christoph Hellwig <hch@lst.de>:
amdgpu: a NULL ->mm does not mean a thread is a kthread
Documentation/lzo.txt | 8 ++++-
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h | 2 -
fs/ocfs2/Kconfig | 2 -
lib/lzo/lzo1x_compress.c | 13 ++++++++
mm/memory-failure.c | 43 +++++++++++++++++------------
5 files changed, 47 insertions(+), 21 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-06-11 1:40 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-06-11 1:40 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
- various hotfixes and minor things
- hch's use_mm/unuse_mm clearnups
- new syscall process_madvise(): perform madvise() on a process other
than self
25 patches, based on 6f630784cc0d92fb58ea326e2bc01aa056279ecb.
Subsystems affected by this patch series:
mm/hugetlb
scripts
kcov
lib
nilfs
checkpatch
lib
mm/debug
ocfs2
lib
misc
mm/madvise
Subsystem: mm/hugetlb
Dan Carpenter <dan.carpenter@oracle.com>:
khugepaged: selftests: fix timeout condition in wait_for_scan()
Subsystem: scripts
SeongJae Park <sjpark@amazon.de>:
scripts/spelling: add a few more typos
Subsystem: kcov
Andrey Konovalov <andreyknvl@google.com>:
kcov: check kcov_softirq in kcov_remote_stop()
Subsystem: lib
Joe Perches <joe@perches.com>:
lib/lz4/lz4_decompress.c: document deliberate use of `&'
Subsystem: nilfs
Ryusuke Konishi <konishi.ryusuke@gmail.com>:
nilfs2: fix null pointer dereference at nilfs_segctor_do_construct()
Subsystem: checkpatch
Tim Froidcoeur <tim.froidcoeur@tessares.net>:
checkpatch: correct check for kernel parameters doc
Subsystem: lib
Alexander Gordeev <agordeev@linux.ibm.com>:
lib: fix bitmap_parse() on 64-bit big endian archs
Subsystem: mm/debug
"Aneesh Kumar K.V" <aneesh.kumar@linux.ibm.com>:
mm/debug_vm_pgtable: fix kernel crash by checking for THP support
Subsystem: ocfs2
Keyur Patel <iamkeyur96@gmail.com>:
ocfs2: fix spelling mistake and grammar
Ben Widawsky <ben.widawsky@intel.com>:
mm: add comments on pglist_data zones
Subsystem: lib
Wei Yang <richard.weiyang@gmail.com>:
lib: test get_count_order/long in test_bitops.c
Subsystem: misc
Walter Wu <walter-zh.wu@mediatek.com>:
stacktrace: cleanup inconsistent variable type
Christoph Hellwig <hch@lst.de>:
Patch series "improve use_mm / unuse_mm", v2:
kernel: move use_mm/unuse_mm to kthread.c
kernel: move use_mm/unuse_mm to kthread.c
kernel: better document the use_mm/unuse_mm API contract
kernel: set USER_DS in kthread_use_mm
Subsystem: mm/madvise
Minchan Kim <minchan@kernel.org>:
Patch series "introduce memory hinting API for external process", v7:
mm/madvise: pass task and mm to do_madvise
mm/madvise: introduce process_madvise() syscall: an external memory hinting API
mm/madvise: check fatal signal pending of target process
pid: move pidfd_get_pid() to pid.c
mm/madvise: support both pid and pidfd for process_madvise
Oleksandr Natalenko <oleksandr@redhat.com>:
mm/madvise: allow KSM hints for remote API
Minchan Kim <minchan@kernel.org>:
mm: support vector address ranges for process_madvise
mm: use only pidfd for process_madvise syscall
YueHaibing <yuehaibing@huawei.com>:
mm/madvise.c: remove duplicated include
arch/alpha/kernel/syscalls/syscall.tbl | 1
arch/arm/tools/syscall.tbl | 1
arch/arm64/include/asm/unistd.h | 2
arch/arm64/include/asm/unistd32.h | 4
arch/ia64/kernel/syscalls/syscall.tbl | 1
arch/m68k/kernel/syscalls/syscall.tbl | 1
arch/microblaze/kernel/syscalls/syscall.tbl | 1
arch/mips/kernel/syscalls/syscall_n32.tbl | 3
arch/mips/kernel/syscalls/syscall_n64.tbl | 1
arch/mips/kernel/syscalls/syscall_o32.tbl | 3
arch/parisc/kernel/syscalls/syscall.tbl | 3
arch/powerpc/kernel/syscalls/syscall.tbl | 3
arch/powerpc/platforms/powernv/vas-fault.c | 4
arch/s390/kernel/syscalls/syscall.tbl | 3
arch/sh/kernel/syscalls/syscall.tbl | 1
arch/sparc/kernel/syscalls/syscall.tbl | 3
arch/x86/entry/syscalls/syscall_32.tbl | 3
arch/x86/entry/syscalls/syscall_64.tbl | 5
arch/xtensa/kernel/syscalls/syscall.tbl | 1
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h | 5
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_arcturus.c | 1
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v10.c | 1
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v7.c | 2
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v8.c | 2
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v9.c | 2
drivers/gpu/drm/i915/gvt/kvmgt.c | 2
drivers/usb/gadget/function/f_fs.c | 10
drivers/usb/gadget/legacy/inode.c | 6
drivers/vfio/vfio_iommu_type1.c | 6
drivers/vhost/vhost.c | 8
fs/aio.c | 1
fs/io-wq.c | 15 -
fs/io_uring.c | 11
fs/nilfs2/segment.c | 2
fs/ocfs2/mmap.c | 2
include/linux/compat.h | 10
include/linux/kthread.h | 9
include/linux/mm.h | 3
include/linux/mmu_context.h | 5
include/linux/mmzone.h | 14
include/linux/pid.h | 1
include/linux/stacktrace.h | 2
include/linux/syscalls.h | 16 -
include/uapi/asm-generic/unistd.h | 7
kernel/exit.c | 17 -
kernel/kcov.c | 26 +
kernel/kthread.c | 95 +++++-
kernel/pid.c | 17 +
kernel/sys_ni.c | 2
lib/Kconfig.debug | 10
lib/bitmap.c | 9
lib/lz4/lz4_decompress.c | 3
lib/test_bitops.c | 53 +++
mm/Makefile | 2
mm/debug_vm_pgtable.c | 6
mm/madvise.c | 295 ++++++++++++++------
mm/mmu_context.c | 64 ----
mm/oom_kill.c | 6
mm/vmacache.c | 4
scripts/checkpatch.pl | 4
scripts/spelling.txt | 9
tools/testing/selftests/vm/khugepaged.c | 2
62 files changed, 526 insertions(+), 285 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2020-06-09 4:29 incoming Andrew Morton
@ 2020-06-09 16:58 ` Linus Torvalds
0 siblings, 0 replies; 409+ messages in thread
From: Linus Torvalds @ 2020-06-09 16:58 UTC (permalink / raw)
To: Andrew Morton; +Cc: mm-commits, Linux-MM
On Mon, Jun 8, 2020 at 9:29 PM Andrew Morton <akpm@linux-foundation.org> wrote:
>
> 942 files changed, 4580 insertions(+), 5662 deletions(-)
If you use proper tools, add a "-M" to your diff script, so that you see
941 files changed, 2614 insertions(+), 3696 deletions(-)
because a big portion of the lines were due to a rename:
rename include/{asm-generic => linux}/pgtable.h (91%)
but at some earlier point you mentioned "diffstat", so I guess "proper
tools" isn't an option ;(
Linus
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-06-09 4:29 Andrew Morton
2020-06-09 16:58 ` incoming Linus Torvalds
0 siblings, 1 reply; 409+ messages in thread
From: Andrew Morton @ 2020-06-09 4:29 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
- a kernel-wide sweep of show_stack()
- pagetable cleanups
- abstract out accesses to mmap_sem - prep for mmap_sem scalability work
- hch's user acess work
93 patches, based on abfbb29297c27e3f101f348dc9e467b0fe70f919:
Subsystems affected by this patch series:
debug
mm/pagemap
mm/maccess
mm/documentation
Subsystem: debug
Dmitry Safonov <dima@arista.com>:
Patch series "Add log level to show_stack()", v3:
kallsyms/printk: add loglvl to print_ip_sym()
alpha: add show_stack_loglvl()
arc: add show_stack_loglvl()
arm/asm: add loglvl to c_backtrace()
arm: add loglvl to unwind_backtrace()
arm: add loglvl to dump_backtrace()
arm: wire up dump_backtrace_{entry,stm}
arm: add show_stack_loglvl()
arm64: add loglvl to dump_backtrace()
arm64: add show_stack_loglvl()
c6x: add show_stack_loglvl()
csky: add show_stack_loglvl()
h8300: add show_stack_loglvl()
hexagon: add show_stack_loglvl()
ia64: pass log level as arg into ia64_do_show_stack()
ia64: add show_stack_loglvl()
m68k: add show_stack_loglvl()
microblaze: add loglvl to microblaze_unwind_inner()
microblaze: add loglvl to microblaze_unwind()
microblaze: add show_stack_loglvl()
mips: add show_stack_loglvl()
nds32: add show_stack_loglvl()
nios2: add show_stack_loglvl()
openrisc: add show_stack_loglvl()
parisc: add show_stack_loglvl()
powerpc: add show_stack_loglvl()
riscv: add show_stack_loglvl()
s390: add show_stack_loglvl()
sh: add loglvl to dump_mem()
sh: remove needless printk()
sh: add loglvl to printk_address()
sh: add loglvl to show_trace()
sh: add show_stack_loglvl()
sparc: add show_stack_loglvl()
um/sysrq: remove needless variable sp
um: add show_stack_loglvl()
unicore32: remove unused pmode argument in c_backtrace()
unicore32: add loglvl to c_backtrace()
unicore32: add show_stack_loglvl()
x86: add missing const qualifiers for log_lvl
x86: add show_stack_loglvl()
xtensa: add loglvl to show_trace()
xtensa: add show_stack_loglvl()
sysrq: use show_stack_loglvl()
x86/amd_gart: print stacktrace for a leak with KERN_ERR
power: use show_stack_loglvl()
kdb: don't play with console_loglevel
sched: print stack trace with KERN_INFO
kernel: use show_stack_loglvl()
kernel: rename show_stack_loglvl() => show_stack()
Subsystem: mm/pagemap
Mike Rapoport <rppt@linux.ibm.com>:
Patch series "mm: consolidate definitions of page table accessors", v2:
mm: don't include asm/pgtable.h if linux/mm.h is already included
mm: introduce include/linux/pgtable.h
mm: reorder includes after introduction of linux/pgtable.h
csky: replace definitions of __pXd_offset() with pXd_index()
m68k/mm/motorola: move comment about page table allocation funcitons
m68k/mm: move {cache,nocahe}_page() definitions close to their user
x86/mm: simplify init_trampoline() and surrounding logic
mm: pgtable: add shortcuts for accessing kernel PMD and PTE
mm: consolidate pte_index() and pte_offset_*() definitions
Michel Lespinasse <walken@google.com>:
mmap locking API: initial implementation as rwsem wrappers
MMU notifier: use the new mmap locking API
DMA reservations: use the new mmap locking API
mmap locking API: use coccinelle to convert mmap_sem rwsem call sites
mmap locking API: convert mmap_sem call sites missed by coccinelle
mmap locking API: convert nested write lock sites
mmap locking API: add mmap_read_trylock_non_owner()
mmap locking API: add MMAP_LOCK_INITIALIZER
mmap locking API: add mmap_assert_locked() and mmap_assert_write_locked()
mmap locking API: rename mmap_sem to mmap_lock
mmap locking API: convert mmap_sem API comments
mmap locking API: convert mmap_sem comments
Subsystem: mm/maccess
Christoph Hellwig <hch@lst.de>:
Patch series "clean up and streamline probe_kernel_* and friends", v4:
maccess: unexport probe_kernel_write()
maccess: remove various unused weak aliases
maccess: remove duplicate kerneldoc comments
maccess: clarify kerneldoc comments
maccess: update the top of file comment
maccess: rename strncpy_from_unsafe_user to strncpy_from_user_nofault
maccess: rename strncpy_from_unsafe_strict to strncpy_from_kernel_nofault
maccess: rename strnlen_unsafe_user to strnlen_user_nofault
maccess: remove probe_read_common and probe_write_common
maccess: unify the probe kernel arch hooks
bpf: factor out a bpf_trace_copy_string helper
bpf: handle the compat string in bpf_trace_copy_string better
Andrew Morton <akpm@linux-foundation.org>:
bpf:bpf_seq_printf(): handle potentially unsafe format string better
Christoph Hellwig <hch@lst.de>:
bpf: rework the compat kernel probe handling
tracing/kprobes: handle mixed kernel/userspace probes better
maccess: remove strncpy_from_unsafe
maccess: always use strict semantics for probe_kernel_read
maccess: move user access routines together
maccess: allow architectures to provide kernel probing directly
x86: use non-set_fs based maccess routines
maccess: return -ERANGE when probe_kernel_read() fails
Subsystem: mm/documentation
Luis Chamberlain <mcgrof@kernel.org>:
include/linux/cache.h: expand documentation over __read_mostly
Documentation/admin-guide/mm/numa_memory_policy.rst | 10
Documentation/admin-guide/mm/userfaultfd.rst | 2
Documentation/filesystems/locking.rst | 2
Documentation/vm/hmm.rst | 6
Documentation/vm/transhuge.rst | 4
arch/alpha/boot/bootp.c | 1
arch/alpha/boot/bootpz.c | 1
arch/alpha/boot/main.c | 1
arch/alpha/include/asm/io.h | 1
arch/alpha/include/asm/pgtable.h | 16
arch/alpha/kernel/process.c | 1
arch/alpha/kernel/proto.h | 4
arch/alpha/kernel/ptrace.c | 1
arch/alpha/kernel/setup.c | 1
arch/alpha/kernel/smp.c | 1
arch/alpha/kernel/sys_alcor.c | 1
arch/alpha/kernel/sys_cabriolet.c | 1
arch/alpha/kernel/sys_dp264.c | 1
arch/alpha/kernel/sys_eb64p.c | 1
arch/alpha/kernel/sys_eiger.c | 1
arch/alpha/kernel/sys_jensen.c | 1
arch/alpha/kernel/sys_marvel.c | 1
arch/alpha/kernel/sys_miata.c | 1
arch/alpha/kernel/sys_mikasa.c | 1
arch/alpha/kernel/sys_nautilus.c | 1
arch/alpha/kernel/sys_noritake.c | 1
arch/alpha/kernel/sys_rawhide.c | 1
arch/alpha/kernel/sys_ruffian.c | 1
arch/alpha/kernel/sys_rx164.c | 1
arch/alpha/kernel/sys_sable.c | 1
arch/alpha/kernel/sys_sio.c | 1
arch/alpha/kernel/sys_sx164.c | 1
arch/alpha/kernel/sys_takara.c | 1
arch/alpha/kernel/sys_titan.c | 1
arch/alpha/kernel/sys_wildfire.c | 1
arch/alpha/kernel/traps.c | 40
arch/alpha/mm/fault.c | 12
arch/alpha/mm/init.c | 1
arch/arc/include/asm/bug.h | 3
arch/arc/include/asm/pgtable.h | 24
arch/arc/kernel/process.c | 4
arch/arc/kernel/stacktrace.c | 29
arch/arc/kernel/troubleshoot.c | 6
arch/arc/mm/fault.c | 6
arch/arc/mm/highmem.c | 14
arch/arc/mm/tlbex.S | 4
arch/arm/include/asm/bug.h | 3
arch/arm/include/asm/efi.h | 3
arch/arm/include/asm/fixmap.h | 4
arch/arm/include/asm/idmap.h | 2
arch/arm/include/asm/pgtable-2level.h | 1
arch/arm/include/asm/pgtable-3level.h | 7
arch/arm/include/asm/pgtable-nommu.h | 3
arch/arm/include/asm/pgtable.h | 25
arch/arm/include/asm/traps.h | 3
arch/arm/include/asm/unwind.h | 3
arch/arm/kernel/head.S | 4
arch/arm/kernel/machine_kexec.c | 1
arch/arm/kernel/module.c | 1
arch/arm/kernel/process.c | 4
arch/arm/kernel/ptrace.c | 1
arch/arm/kernel/smp.c | 1
arch/arm/kernel/suspend.c | 4
arch/arm/kernel/swp_emulate.c | 4
arch/arm/kernel/traps.c | 61
arch/arm/kernel/unwind.c | 7
arch/arm/kernel/vdso.c | 2
arch/arm/kernel/vmlinux.lds.S | 4
arch/arm/lib/backtrace-clang.S | 9
arch/arm/lib/backtrace.S | 14
arch/arm/lib/uaccess_with_memcpy.c | 16
arch/arm/mach-ebsa110/core.c | 1
arch/arm/mach-footbridge/common.c | 1
arch/arm/mach-imx/mm-imx21.c | 1
arch/arm/mach-imx/mm-imx27.c | 1
arch/arm/mach-imx/mm-imx3.c | 1
arch/arm/mach-integrator/core.c | 4
arch/arm/mach-iop32x/i2c.c | 1
arch/arm/mach-iop32x/iq31244.c | 1
arch/arm/mach-iop32x/iq80321.c | 1
arch/arm/mach-iop32x/n2100.c | 1
arch/arm/mach-ixp4xx/common.c | 1
arch/arm/mach-keystone/platsmp.c | 4
arch/arm/mach-sa1100/assabet.c | 3
arch/arm/mach-sa1100/hackkit.c | 4
arch/arm/mach-tegra/iomap.h | 2
arch/arm/mach-zynq/common.c | 4
arch/arm/mm/copypage-v4mc.c | 1
arch/arm/mm/copypage-v6.c | 1
arch/arm/mm/copypage-xscale.c | 1
arch/arm/mm/dump.c | 1
arch/arm/mm/fault-armv.c | 1
arch/arm/mm/fault.c | 9
arch/arm/mm/highmem.c | 4
arch/arm/mm/idmap.c | 4
arch/arm/mm/ioremap.c | 31
arch/arm/mm/mm.h | 8
arch/arm/mm/mmu.c | 7
arch/arm/mm/pageattr.c | 1
arch/arm/mm/proc-arm1020.S | 4
arch/arm/mm/proc-arm1020e.S | 4
arch/arm/mm/proc-arm1022.S | 4
arch/arm/mm/proc-arm1026.S | 4
arch/arm/mm/proc-arm720.S | 4
arch/arm/mm/proc-arm740.S | 4
arch/arm/mm/proc-arm7tdmi.S | 4
arch/arm/mm/proc-arm920.S | 4
arch/arm/mm/proc-arm922.S | 4
arch/arm/mm/proc-arm925.S | 4
arch/arm/mm/proc-arm926.S | 4
arch/arm/mm/proc-arm940.S | 4
arch/arm/mm/proc-arm946.S | 4
arch/arm/mm/proc-arm9tdmi.S | 4
arch/arm/mm/proc-fa526.S | 4
arch/arm/mm/proc-feroceon.S | 4
arch/arm/mm/proc-mohawk.S | 4
arch/arm/mm/proc-sa110.S | 4
arch/arm/mm/proc-sa1100.S | 4
arch/arm/mm/proc-v6.S | 4
arch/arm/mm/proc-v7.S | 4
arch/arm/mm/proc-xsc3.S | 4
arch/arm/mm/proc-xscale.S | 4
arch/arm/mm/pv-fixup-asm.S | 4
arch/arm64/include/asm/io.h | 4
arch/arm64/include/asm/kernel-pgtable.h | 2
arch/arm64/include/asm/kvm_mmu.h | 4
arch/arm64/include/asm/mmu_context.h | 4
arch/arm64/include/asm/pgtable.h | 40
arch/arm64/include/asm/stacktrace.h | 3
arch/arm64/include/asm/stage2_pgtable.h | 2
arch/arm64/include/asm/vmap_stack.h | 4
arch/arm64/kernel/acpi.c | 4
arch/arm64/kernel/head.S | 4
arch/arm64/kernel/hibernate.c | 5
arch/arm64/kernel/kaslr.c | 4
arch/arm64/kernel/process.c | 2
arch/arm64/kernel/ptrace.c | 1
arch/arm64/kernel/smp.c | 1
arch/arm64/kernel/suspend.c | 4
arch/arm64/kernel/traps.c | 37
arch/arm64/kernel/vdso.c | 8
arch/arm64/kernel/vmlinux.lds.S | 3
arch/arm64/kvm/mmu.c | 14
arch/arm64/mm/dump.c | 1
arch/arm64/mm/fault.c | 9
arch/arm64/mm/kasan_init.c | 3
arch/arm64/mm/mmu.c | 8
arch/arm64/mm/pageattr.c | 1
arch/arm64/mm/proc.S | 4
arch/c6x/include/asm/pgtable.h | 3
arch/c6x/kernel/traps.c | 28
arch/csky/include/asm/io.h | 2
arch/csky/include/asm/pgtable.h | 37
arch/csky/kernel/module.c | 1
arch/csky/kernel/ptrace.c | 5
arch/csky/kernel/stacktrace.c | 20
arch/csky/kernel/vdso.c | 4
arch/csky/mm/fault.c | 10
arch/csky/mm/highmem.c | 2
arch/csky/mm/init.c | 7
arch/csky/mm/tlb.c | 1
arch/h8300/include/asm/pgtable.h | 1
arch/h8300/kernel/process.c | 1
arch/h8300/kernel/setup.c | 1
arch/h8300/kernel/signal.c | 1
arch/h8300/kernel/traps.c | 26
arch/h8300/mm/fault.c | 1
arch/h8300/mm/init.c | 1
arch/h8300/mm/memory.c | 1
arch/hexagon/include/asm/fixmap.h | 4
arch/hexagon/include/asm/pgtable.h | 55
arch/hexagon/kernel/traps.c | 39
arch/hexagon/kernel/vdso.c | 4
arch/hexagon/mm/uaccess.c | 2
arch/hexagon/mm/vm_fault.c | 9
arch/ia64/include/asm/pgtable.h | 34
arch/ia64/include/asm/ptrace.h | 1
arch/ia64/include/asm/uaccess.h | 2
arch/ia64/kernel/efi.c | 1
arch/ia64/kernel/entry.S | 4
arch/ia64/kernel/head.S | 5
arch/ia64/kernel/irq_ia64.c | 4
arch/ia64/kernel/ivt.S | 4
arch/ia64/kernel/kprobes.c | 4
arch/ia64/kernel/mca.c | 2
arch/ia64/kernel/mca_asm.S | 4
arch/ia64/kernel/perfmon.c | 8
arch/ia64/kernel/process.c | 37
arch/ia64/kernel/ptrace.c | 1
arch/ia64/kernel/relocate_kernel.S | 6
arch/ia64/kernel/setup.c | 4
arch/ia64/kernel/smp.c | 1
arch/ia64/kernel/smpboot.c | 1
arch/ia64/kernel/uncached.c | 4
arch/ia64/kernel/vmlinux.lds.S | 4
arch/ia64/mm/contig.c | 1
arch/ia64/mm/fault.c | 17
arch/ia64/mm/init.c | 12
arch/m68k/68000/m68EZ328.c | 2
arch/m68k/68000/m68VZ328.c | 4
arch/m68k/68000/timers.c | 1
arch/m68k/amiga/config.c | 1
arch/m68k/apollo/config.c | 1
arch/m68k/atari/atasound.c | 1
arch/m68k/atari/stram.c | 1
arch/m68k/bvme6000/config.c | 1
arch/m68k/include/asm/mcf_pgtable.h | 63
arch/m68k/include/asm/motorola_pgalloc.h | 8
arch/m68k/include/asm/motorola_pgtable.h | 84 -
arch/m68k/include/asm/pgtable_mm.h | 1
arch/m68k/include/asm/pgtable_no.h | 2
arch/m68k/include/asm/sun3_pgtable.h | 24
arch/m68k/include/asm/sun3xflop.h | 4
arch/m68k/kernel/head.S | 4
arch/m68k/kernel/process.c | 1
arch/m68k/kernel/ptrace.c | 1
arch/m68k/kernel/setup_no.c | 1
arch/m68k/kernel/signal.c | 1
arch/m68k/kernel/sys_m68k.c | 14
arch/m68k/kernel/traps.c | 27
arch/m68k/kernel/uboot.c | 1
arch/m68k/mac/config.c | 1
arch/m68k/mm/fault.c | 10
arch/m68k/mm/init.c | 2
arch/m68k/mm/mcfmmu.c | 1
arch/m68k/mm/motorola.c | 65
arch/m68k/mm/sun3kmap.c | 1
arch/m68k/mm/sun3mmu.c | 1
arch/m68k/mvme147/config.c | 1
arch/m68k/mvme16x/config.c | 1
arch/m68k/q40/config.c | 1
arch/m68k/sun3/config.c | 1
arch/m68k/sun3/dvma.c | 1
arch/m68k/sun3/mmu_emu.c | 1
arch/m68k/sun3/sun3dvma.c | 1
arch/m68k/sun3x/dvma.c | 1
arch/m68k/sun3x/prom.c | 1
arch/microblaze/include/asm/pgalloc.h | 4
arch/microblaze/include/asm/pgtable.h | 23
arch/microblaze/include/asm/uaccess.h | 2
arch/microblaze/include/asm/unwind.h | 3
arch/microblaze/kernel/hw_exception_handler.S | 4
arch/microblaze/kernel/module.c | 4
arch/microblaze/kernel/setup.c | 4
arch/microblaze/kernel/signal.c | 9
arch/microblaze/kernel/stacktrace.c | 4
arch/microblaze/kernel/traps.c | 28
arch/microblaze/kernel/unwind.c | 46
arch/microblaze/mm/fault.c | 17
arch/microblaze/mm/init.c | 9
arch/microblaze/mm/pgtable.c | 4
arch/mips/fw/arc/memory.c | 1
arch/mips/include/asm/fixmap.h | 3
arch/mips/include/asm/mach-generic/floppy.h | 1
arch/mips/include/asm/mach-jazz/floppy.h | 1
arch/mips/include/asm/pgtable-32.h | 22
arch/mips/include/asm/pgtable-64.h | 32
arch/mips/include/asm/pgtable.h | 2
arch/mips/jazz/irq.c | 4
arch/mips/jazz/jazzdma.c | 1
arch/mips/jazz/setup.c | 4
arch/mips/kernel/module.c | 1
arch/mips/kernel/process.c | 1
arch/mips/kernel/ptrace.c | 1
arch/mips/kernel/ptrace32.c | 1
arch/mips/kernel/smp-bmips.c | 1
arch/mips/kernel/traps.c | 58
arch/mips/kernel/vdso.c | 4
arch/mips/kvm/mips.c | 4
arch/mips/kvm/mmu.c | 20
arch/mips/kvm/tlb.c | 1
arch/mips/kvm/trap_emul.c | 2
arch/mips/lib/dump_tlb.c | 1
arch/mips/lib/r3k_dump_tlb.c | 1
arch/mips/mm/c-octeon.c | 1
arch/mips/mm/c-r3k.c | 11
arch/mips/mm/c-r4k.c | 11
arch/mips/mm/c-tx39.c | 11
arch/mips/mm/fault.c | 12
arch/mips/mm/highmem.c | 2
arch/mips/mm/init.c | 1
arch/mips/mm/page.c | 1
arch/mips/mm/pgtable-32.c | 1
arch/mips/mm/pgtable-64.c | 1
arch/mips/mm/sc-ip22.c | 1
arch/mips/mm/sc-mips.c | 1
arch/mips/mm/sc-r5k.c | 1
arch/mips/mm/tlb-r3k.c | 1
arch/mips/mm/tlb-r4k.c | 1
arch/mips/mm/tlbex.c | 4
arch/mips/sgi-ip27/ip27-init.c | 1
arch/mips/sgi-ip27/ip27-timer.c | 1
arch/mips/sgi-ip32/ip32-memory.c | 1
arch/nds32/include/asm/highmem.h | 3
arch/nds32/include/asm/pgtable.h | 22
arch/nds32/kernel/head.S | 4
arch/nds32/kernel/module.c | 2
arch/nds32/kernel/traps.c | 33
arch/nds32/kernel/vdso.c | 6
arch/nds32/mm/fault.c | 17
arch/nds32/mm/init.c | 13
arch/nds32/mm/proc.c | 7
arch/nios2/include/asm/pgtable.h | 24
arch/nios2/kernel/module.c | 1
arch/nios2/kernel/nios2_ksyms.c | 4
arch/nios2/kernel/traps.c | 35
arch/nios2/mm/fault.c | 14
arch/nios2/mm/init.c | 5
arch/nios2/mm/pgtable.c | 1
arch/nios2/mm/tlb.c | 1
arch/openrisc/include/asm/io.h | 3
arch/openrisc/include/asm/pgtable.h | 33
arch/openrisc/include/asm/tlbflush.h | 1
arch/openrisc/kernel/asm-offsets.c | 1
arch/openrisc/kernel/entry.S | 4
arch/openrisc/kernel/head.S | 4
arch/openrisc/kernel/or32_ksyms.c | 4
arch/openrisc/kernel/process.c | 1
arch/openrisc/kernel/ptrace.c | 1
arch/openrisc/kernel/setup.c | 1
arch/openrisc/kernel/traps.c | 27
arch/openrisc/mm/fault.c | 12
arch/openrisc/mm/init.c | 1
arch/openrisc/mm/ioremap.c | 4
arch/openrisc/mm/tlb.c | 1
arch/parisc/include/asm/io.h | 2
arch/parisc/include/asm/mmu_context.h | 1
arch/parisc/include/asm/pgtable.h | 33
arch/parisc/kernel/asm-offsets.c | 4
arch/parisc/kernel/entry.S | 4
arch/parisc/kernel/head.S | 4
arch/parisc/kernel/module.c | 1
arch/parisc/kernel/pacache.S | 4
arch/parisc/kernel/pci-dma.c | 2
arch/parisc/kernel/pdt.c | 4
arch/parisc/kernel/ptrace.c | 1
arch/parisc/kernel/smp.c | 1
arch/parisc/kernel/traps.c | 42
arch/parisc/lib/memcpy.c | 14
arch/parisc/mm/fault.c | 10
arch/parisc/mm/fixmap.c | 6
arch/parisc/mm/init.c | 1
arch/powerpc/include/asm/book3s/32/pgtable.h | 20
arch/powerpc/include/asm/book3s/64/pgtable.h | 43
arch/powerpc/include/asm/fixmap.h | 4
arch/powerpc/include/asm/io.h | 1
arch/powerpc/include/asm/kup.h | 2
arch/powerpc/include/asm/nohash/32/pgtable.h | 17
arch/powerpc/include/asm/nohash/64/pgtable-4k.h | 4
arch/powerpc/include/asm/nohash/64/pgtable.h | 22
arch/powerpc/include/asm/nohash/pgtable.h | 2
arch/powerpc/include/asm/pgtable.h | 28
arch/powerpc/include/asm/pkeys.h | 2
arch/powerpc/include/asm/tlb.h | 2
arch/powerpc/kernel/asm-offsets.c | 1
arch/powerpc/kernel/btext.c | 4
arch/powerpc/kernel/fpu.S | 3
arch/powerpc/kernel/head_32.S | 4
arch/powerpc/kernel/head_40x.S | 4
arch/powerpc/kernel/head_44x.S | 4
arch/powerpc/kernel/head_8xx.S | 4
arch/powerpc/kernel/head_fsl_booke.S | 4
arch/powerpc/kernel/io-workarounds.c | 4
arch/powerpc/kernel/irq.c | 4
arch/powerpc/kernel/mce_power.c | 4
arch/powerpc/kernel/paca.c | 4
arch/powerpc/kernel/process.c | 30
arch/powerpc/kernel/prom.c | 4
arch/powerpc/kernel/prom_init.c | 4
arch/powerpc/kernel/rtas_pci.c | 4
arch/powerpc/kernel/setup-common.c | 4
arch/powerpc/kernel/setup_32.c | 4
arch/powerpc/kernel/setup_64.c | 4
arch/powerpc/kernel/signal_32.c | 1
arch/powerpc/kernel/signal_64.c | 1
arch/powerpc/kernel/smp.c | 4
arch/powerpc/kernel/stacktrace.c | 2
arch/powerpc/kernel/traps.c | 1
arch/powerpc/kernel/vdso.c | 7
arch/powerpc/kvm/book3s_64_mmu_radix.c | 4
arch/powerpc/kvm/book3s_hv.c | 6
arch/powerpc/kvm/book3s_hv_nested.c | 4
arch/powerpc/kvm/book3s_hv_rm_xics.c | 4
arch/powerpc/kvm/book3s_hv_rm_xive.c | 4
arch/powerpc/kvm/book3s_hv_uvmem.c | 18
arch/powerpc/kvm/e500_mmu_host.c | 4
arch/powerpc/kvm/fpu.S | 4
arch/powerpc/lib/code-patching.c | 1
arch/powerpc/mm/book3s32/hash_low.S | 4
arch/powerpc/mm/book3s32/mmu.c | 2
arch/powerpc/mm/book3s32/tlb.c | 6
arch/powerpc/mm/book3s64/hash_hugetlbpage.c | 1
arch/powerpc/mm/book3s64/hash_native.c | 4
arch/powerpc/mm/book3s64/hash_pgtable.c | 5
arch/powerpc/mm/book3s64/hash_utils.c | 4
arch/powerpc/mm/book3s64/iommu_api.c | 4
arch/powerpc/mm/book3s64/radix_hugetlbpage.c | 1
arch/powerpc/mm/book3s64/radix_pgtable.c | 1
arch/powerpc/mm/book3s64/slb.c | 4
arch/powerpc/mm/book3s64/subpage_prot.c | 16
arch/powerpc/mm/copro_fault.c | 4
arch/powerpc/mm/fault.c | 23
arch/powerpc/mm/hugetlbpage.c | 1
arch/powerpc/mm/init-common.c | 4
arch/powerpc/mm/init_32.c | 1
arch/powerpc/mm/init_64.c | 1
arch/powerpc/mm/kasan/8xx.c | 4
arch/powerpc/mm/kasan/book3s_32.c | 2
arch/powerpc/mm/kasan/kasan_init_32.c | 8
arch/powerpc/mm/mem.c | 1
arch/powerpc/mm/nohash/40x.c | 5
arch/powerpc/mm/nohash/8xx.c | 2
arch/powerpc/mm/nohash/fsl_booke.c | 1
arch/powerpc/mm/nohash/tlb_low_64e.S | 4
arch/powerpc/mm/pgtable.c | 2
arch/powerpc/mm/pgtable_32.c | 5
arch/powerpc/mm/pgtable_64.c | 1
arch/powerpc/mm/ptdump/8xx.c | 2
arch/powerpc/mm/ptdump/bats.c | 4
arch/powerpc/mm/ptdump/book3s64.c | 2
arch/powerpc/mm/ptdump/hashpagetable.c | 1
arch/powerpc/mm/ptdump/ptdump.c | 1
arch/powerpc/mm/ptdump/shared.c | 2
arch/powerpc/oprofile/cell/spu_task_sync.c | 6
arch/powerpc/perf/callchain.c | 1
arch/powerpc/perf/callchain_32.c | 1
arch/powerpc/perf/callchain_64.c | 1
arch/powerpc/platforms/85xx/corenet_generic.c | 4
arch/powerpc/platforms/85xx/mpc85xx_cds.c | 4
arch/powerpc/platforms/85xx/qemu_e500.c | 4
arch/powerpc/platforms/85xx/sbc8548.c | 4
arch/powerpc/platforms/85xx/smp.c | 4
arch/powerpc/platforms/86xx/mpc86xx_smp.c | 4
arch/powerpc/platforms/8xx/cpm1.c | 1
arch/powerpc/platforms/8xx/micropatch.c | 1
arch/powerpc/platforms/cell/cbe_regs.c | 4
arch/powerpc/platforms/cell/interrupt.c | 4
arch/powerpc/platforms/cell/pervasive.c | 4
arch/powerpc/platforms/cell/setup.c | 1
arch/powerpc/platforms/cell/smp.c | 4
arch/powerpc/platforms/cell/spider-pic.c | 4
arch/powerpc/platforms/cell/spufs/file.c | 10
arch/powerpc/platforms/chrp/pci.c | 4
arch/powerpc/platforms/chrp/setup.c | 1
arch/powerpc/platforms/chrp/smp.c | 4
arch/powerpc/platforms/maple/setup.c | 1
arch/powerpc/platforms/maple/time.c | 1
arch/powerpc/platforms/powermac/setup.c | 1
arch/powerpc/platforms/powermac/smp.c | 4
arch/powerpc/platforms/powermac/time.c | 1
arch/powerpc/platforms/pseries/lpar.c | 4
arch/powerpc/platforms/pseries/setup.c | 1
arch/powerpc/platforms/pseries/smp.c | 4
arch/powerpc/sysdev/cpm2.c | 1
arch/powerpc/sysdev/fsl_85xx_cache_sram.c | 2
arch/powerpc/sysdev/mpic.c | 4
arch/powerpc/xmon/xmon.c | 1
arch/riscv/include/asm/fixmap.h | 4
arch/riscv/include/asm/io.h | 4
arch/riscv/include/asm/kasan.h | 4
arch/riscv/include/asm/pgtable-64.h | 7
arch/riscv/include/asm/pgtable.h | 22
arch/riscv/kernel/module.c | 2
arch/riscv/kernel/setup.c | 1
arch/riscv/kernel/soc.c | 2
arch/riscv/kernel/stacktrace.c | 23
arch/riscv/kernel/vdso.c | 4
arch/riscv/mm/cacheflush.c | 3
arch/riscv/mm/fault.c | 14
arch/riscv/mm/init.c | 31
arch/riscv/mm/kasan_init.c | 4
arch/riscv/mm/pageattr.c | 6
arch/riscv/mm/ptdump.c | 2
arch/s390/boot/ipl_parm.c | 4
arch/s390/boot/kaslr.c | 4
arch/s390/include/asm/hugetlb.h | 4
arch/s390/include/asm/kasan.h | 4
arch/s390/include/asm/pgtable.h | 15
arch/s390/include/asm/tlbflush.h | 1
arch/s390/kernel/asm-offsets.c | 4
arch/s390/kernel/dumpstack.c | 25
arch/s390/kernel/machine_kexec.c | 1
arch/s390/kernel/ptrace.c | 1
arch/s390/kernel/uv.c | 4
arch/s390/kernel/vdso.c | 5
arch/s390/kvm/gaccess.c | 8
arch/s390/kvm/interrupt.c | 4
arch/s390/kvm/kvm-s390.c | 32
arch/s390/kvm/priv.c | 38
arch/s390/mm/dump_pagetables.c | 1
arch/s390/mm/extmem.c | 4
arch/s390/mm/fault.c | 17
arch/s390/mm/gmap.c | 80
arch/s390/mm/init.c | 1
arch/s390/mm/kasan_init.c | 4
arch/s390/mm/pageattr.c | 13
arch/s390/mm/pgalloc.c | 2
arch/s390/mm/pgtable.c | 1
arch/s390/mm/vmem.c | 1
arch/s390/pci/pci_mmio.c | 4
arch/sh/include/asm/io.h | 2
arch/sh/include/asm/kdebug.h | 6
arch/sh/include/asm/pgtable-3level.h | 7
arch/sh/include/asm/pgtable.h | 2
arch/sh/include/asm/pgtable_32.h | 25
arch/sh/include/asm/processor_32.h | 2
arch/sh/kernel/dumpstack.c | 54
arch/sh/kernel/machine_kexec.c | 1
arch/sh/kernel/process_32.c | 2
arch/sh/kernel/ptrace_32.c | 1
arch/sh/kernel/signal_32.c | 1
arch/sh/kernel/sys_sh.c | 6
arch/sh/kernel/traps.c | 4
arch/sh/kernel/vsyscall/vsyscall.c | 4
arch/sh/mm/cache-sh3.c | 1
arch/sh/mm/cache-sh4.c | 11
arch/sh/mm/cache-sh7705.c | 1
arch/sh/mm/fault.c | 16
arch/sh/mm/kmap.c | 5
arch/sh/mm/nommu.c | 1
arch/sh/mm/pmb.c | 4
arch/sparc/include/asm/floppy_32.h | 4
arch/sparc/include/asm/highmem.h | 4
arch/sparc/include/asm/ide.h | 2
arch/sparc/include/asm/io-unit.h | 4
arch/sparc/include/asm/pgalloc_32.h | 4
arch/sparc/include/asm/pgalloc_64.h | 2
arch/sparc/include/asm/pgtable_32.h | 34
arch/sparc/include/asm/pgtable_64.h | 32
arch/sparc/kernel/cpu.c | 4
arch/sparc/kernel/entry.S | 4
arch/sparc/kernel/head_64.S | 4
arch/sparc/kernel/ktlb.S | 4
arch/sparc/kernel/leon_smp.c | 1
arch/sparc/kernel/pci.c | 4
arch/sparc/kernel/process_32.c | 29
arch/sparc/kernel/process_64.c | 3
arch/sparc/kernel/ptrace_32.c | 1
arch/sparc/kernel/ptrace_64.c | 1
arch/sparc/kernel/setup_32.c | 1
arch/sparc/kernel/setup_64.c | 1
arch/sparc/kernel/signal32.c | 1
arch/sparc/kernel/signal_32.c | 1
arch/sparc/kernel/signal_64.c | 1
arch/sparc/kernel/smp_32.c | 1
arch/sparc/kernel/smp_64.c | 1
arch/sparc/kernel/sun4m_irq.c | 4
arch/sparc/kernel/trampoline_64.S | 4
arch/sparc/kernel/traps_32.c | 4
arch/sparc/kernel/traps_64.c | 24
arch/sparc/lib/clear_page.S | 4
arch/sparc/lib/copy_page.S | 2
arch/sparc/mm/fault_32.c | 21
arch/sparc/mm/fault_64.c | 17
arch/sparc/mm/highmem.c | 12
arch/sparc/mm/hugetlbpage.c | 1
arch/sparc/mm/init_32.c | 1
arch/sparc/mm/init_64.c | 7
arch/sparc/mm/io-unit.c | 11
arch/sparc/mm/iommu.c | 9
arch/sparc/mm/tlb.c | 1
arch/sparc/mm/tsb.c | 4
arch/sparc/mm/ultra.S | 4
arch/sparc/vdso/vma.c | 4
arch/um/drivers/mconsole_kern.c | 2
arch/um/include/asm/mmu_context.h | 5
arch/um/include/asm/pgtable-3level.h | 4
arch/um/include/asm/pgtable.h | 69
arch/um/kernel/maccess.c | 12
arch/um/kernel/mem.c | 10
arch/um/kernel/process.c | 1
arch/um/kernel/skas/mmu.c | 3
arch/um/kernel/skas/uaccess.c | 1
arch/um/kernel/sysrq.c | 35
arch/um/kernel/tlb.c | 5
arch/um/kernel/trap.c | 15
arch/um/kernel/um_arch.c | 1
arch/unicore32/include/asm/pgtable.h | 19
arch/unicore32/kernel/hibernate.c | 4
arch/unicore32/kernel/hibernate_asm.S | 4
arch/unicore32/kernel/module.c | 1
arch/unicore32/kernel/setup.h | 4
arch/unicore32/kernel/traps.c | 50
arch/unicore32/lib/backtrace.S | 24
arch/unicore32/mm/alignment.c | 4
arch/unicore32/mm/fault.c | 9
arch/unicore32/mm/mm.h | 10
arch/unicore32/mm/proc-ucv2.S | 4
arch/x86/boot/compressed/kaslr_64.c | 4
arch/x86/entry/vdso/vma.c | 14
arch/x86/events/core.c | 4
arch/x86/include/asm/agp.h | 2
arch/x86/include/asm/asm-prototypes.h | 4
arch/x86/include/asm/efi.h | 4
arch/x86/include/asm/iomap.h | 1
arch/x86/include/asm/kaslr.h | 2
arch/x86/include/asm/mmu.h | 2
arch/x86/include/asm/pgtable-3level.h | 8
arch/x86/include/asm/pgtable.h | 89 -
arch/x86/include/asm/pgtable_32.h | 11
arch/x86/include/asm/pgtable_64.h | 4
arch/x86/include/asm/setup.h | 12
arch/x86/include/asm/stacktrace.h | 2
arch/x86/include/asm/uaccess.h | 16
arch/x86/include/asm/xen/hypercall.h | 4
arch/x86/include/asm/xen/page.h | 1
arch/x86/kernel/acpi/boot.c | 4
arch/x86/kernel/acpi/sleep.c | 4
arch/x86/kernel/alternative.c | 1
arch/x86/kernel/amd_gart_64.c | 5
arch/x86/kernel/apic/apic_numachip.c | 4
arch/x86/kernel/cpu/bugs.c | 4
arch/x86/kernel/cpu/common.c | 4
arch/x86/kernel/cpu/intel.c | 4
arch/x86/kernel/cpu/resctrl/pseudo_lock.c | 6
arch/x86/kernel/cpu/resctrl/rdtgroup.c | 6
arch/x86/kernel/crash_core_32.c | 4
arch/x86/kernel/crash_core_64.c | 4
arch/x86/kernel/doublefault_32.c | 1
arch/x86/kernel/dumpstack.c | 21
arch/x86/kernel/early_printk.c | 4
arch/x86/kernel/espfix_64.c | 2
arch/x86/kernel/head64.c | 4
arch/x86/kernel/head_64.S | 4
arch/x86/kernel/i8259.c | 4
arch/x86/kernel/irqinit.c | 4
arch/x86/kernel/kprobes/core.c | 4
arch/x86/kernel/kprobes/opt.c | 4
arch/x86/kernel/ldt.c | 2
arch/x86/kernel/machine_kexec_32.c | 1
arch/x86/kernel/machine_kexec_64.c | 1
arch/x86/kernel/module.c | 1
arch/x86/kernel/paravirt.c | 4
arch/x86/kernel/process_32.c | 1
arch/x86/kernel/process_64.c | 1
arch/x86/kernel/ptrace.c | 1
arch/x86/kernel/reboot.c | 4
arch/x86/kernel/smpboot.c | 4
arch/x86/kernel/tboot.c | 3
arch/x86/kernel/vm86_32.c | 4
arch/x86/kvm/mmu/paging_tmpl.h | 8
arch/x86/mm/cpu_entry_area.c | 4
arch/x86/mm/debug_pagetables.c | 2
arch/x86/mm/dump_pagetables.c | 1
arch/x86/mm/fault.c | 22
arch/x86/mm/init.c | 22
arch/x86/mm/init_32.c | 27
arch/x86/mm/init_64.c | 1
arch/x86/mm/ioremap.c | 4
arch/x86/mm/kasan_init_64.c | 1
arch/x86/mm/kaslr.c | 37
arch/x86/mm/maccess.c | 44
arch/x86/mm/mem_encrypt_boot.S | 2
arch/x86/mm/mmio-mod.c | 4
arch/x86/mm/pat/cpa-test.c | 1
arch/x86/mm/pat/memtype.c | 1
arch/x86/mm/pat/memtype_interval.c | 4
arch/x86/mm/pgtable.c | 1
arch/x86/mm/pgtable_32.c | 1
arch/x86/mm/pti.c | 1
arch/x86/mm/setup_nx.c | 4
arch/x86/platform/efi/efi_32.c | 4
arch/x86/platform/efi/efi_64.c | 1
arch/x86/platform/olpc/olpc_ofw.c | 4
arch/x86/power/cpu.c | 4
arch/x86/power/hibernate.c | 4
arch/x86/power/hibernate_32.c | 4
arch/x86/power/hibernate_64.c | 4
arch/x86/realmode/init.c | 4
arch/x86/um/vdso/vma.c | 4
arch/x86/xen/enlighten_pv.c | 1
arch/x86/xen/grant-table.c | 1
arch/x86/xen/mmu_pv.c | 4
arch/x86/xen/smp_pv.c | 2
arch/xtensa/include/asm/fixmap.h | 12
arch/xtensa/include/asm/highmem.h | 4
arch/xtensa/include/asm/initialize_mmu.h | 2
arch/xtensa/include/asm/mmu_context.h | 4
arch/xtensa/include/asm/pgtable.h | 20
arch/xtensa/kernel/entry.S | 4
arch/xtensa/kernel/process.c | 1
arch/xtensa/kernel/ptrace.c | 1
arch/xtensa/kernel/setup.c | 1
arch/xtensa/kernel/traps.c | 42
arch/xtensa/kernel/vectors.S | 4
arch/xtensa/mm/cache.c | 4
arch/xtensa/mm/fault.c | 12
arch/xtensa/mm/highmem.c | 2
arch/xtensa/mm/ioremap.c | 4
arch/xtensa/mm/kasan_init.c | 10
arch/xtensa/mm/misc.S | 4
arch/xtensa/mm/mmu.c | 5
drivers/acpi/scan.c | 3
drivers/android/binder_alloc.c | 14
drivers/atm/fore200e.c | 4
drivers/base/power/main.c | 4
drivers/block/z2ram.c | 4
drivers/char/agp/frontend.c | 1
drivers/char/agp/generic.c | 1
drivers/char/bsr.c | 1
drivers/char/mspec.c | 3
drivers/dma-buf/dma-resv.c | 5
drivers/firmware/efi/arm-runtime.c | 4
drivers/firmware/efi/efi.c | 2
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h | 2
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v7.c | 2
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v8.c | 2
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c | 4
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c | 10
drivers/gpu/drm/amd/amdkfd/kfd_events.c | 4
drivers/gpu/drm/drm_vm.c | 4
drivers/gpu/drm/etnaviv/etnaviv_gem.c | 2
drivers/gpu/drm/i915/gem/i915_gem_mman.c | 4
drivers/gpu/drm/i915/gem/i915_gem_userptr.c | 14
drivers/gpu/drm/i915/i915_mm.c | 1
drivers/gpu/drm/i915/i915_perf.c | 2
drivers/gpu/drm/nouveau/nouveau_svm.c | 22
drivers/gpu/drm/radeon/radeon_cs.c | 4
drivers/gpu/drm/radeon/radeon_gem.c | 6
drivers/gpu/drm/ttm/ttm_bo_vm.c | 10
drivers/infiniband/core/umem_odp.c | 4
drivers/infiniband/core/uverbs_main.c | 6
drivers/infiniband/hw/hfi1/mmu_rb.c | 2
drivers/infiniband/hw/mlx4/mr.c | 4
drivers/infiniband/hw/qib/qib_file_ops.c | 4
drivers/infiniband/hw/qib/qib_user_pages.c | 6
drivers/infiniband/hw/usnic/usnic_uiom.c | 4
drivers/infiniband/sw/rdmavt/mmap.c | 1
drivers/infiniband/sw/rxe/rxe_mmap.c | 1
drivers/infiniband/sw/siw/siw_mem.c | 4
drivers/iommu/amd_iommu_v2.c | 4
drivers/iommu/intel-svm.c | 4
drivers/macintosh/macio-adb.c | 4
drivers/macintosh/mediabay.c | 4
drivers/macintosh/via-pmu.c | 4
drivers/media/pci/bt8xx/bt878.c | 4
drivers/media/pci/bt8xx/btcx-risc.c | 4
drivers/media/pci/bt8xx/bttv-risc.c | 4
drivers/media/platform/davinci/vpbe_display.c | 1
drivers/media/v4l2-core/v4l2-common.c | 1
drivers/media/v4l2-core/videobuf-core.c | 4
drivers/media/v4l2-core/videobuf-dma-contig.c | 4
drivers/media/v4l2-core/videobuf-dma-sg.c | 10
drivers/media/v4l2-core/videobuf-vmalloc.c | 4
drivers/misc/cxl/cxllib.c | 9
drivers/misc/cxl/fault.c | 4
drivers/misc/genwqe/card_utils.c | 2
drivers/misc/sgi-gru/grufault.c | 25
drivers/misc/sgi-gru/grufile.c | 4
drivers/mtd/ubi/ubi.h | 2
drivers/net/ethernet/amd/7990.c | 4
drivers/net/ethernet/amd/hplance.c | 4
drivers/net/ethernet/amd/mvme147.c | 4
drivers/net/ethernet/amd/sun3lance.c | 4
drivers/net/ethernet/amd/sunlance.c | 4
drivers/net/ethernet/apple/bmac.c | 4
drivers/net/ethernet/apple/mace.c | 4
drivers/net/ethernet/freescale/fs_enet/fs_enet-main.c | 4
drivers/net/ethernet/freescale/fs_enet/mac-fcc.c | 4
drivers/net/ethernet/freescale/fs_enet/mii-fec.c | 4
drivers/net/ethernet/i825xx/82596.c | 4
drivers/net/ethernet/korina.c | 4
drivers/net/ethernet/marvell/pxa168_eth.c | 4
drivers/net/ethernet/natsemi/jazzsonic.c | 4
drivers/net/ethernet/natsemi/macsonic.c | 4
drivers/net/ethernet/natsemi/xtsonic.c | 4
drivers/net/ethernet/sun/sunbmac.c | 4
drivers/net/ethernet/sun/sunhme.c | 1
drivers/net/ethernet/sun/sunqe.c | 4
drivers/oprofile/buffer_sync.c | 12
drivers/sbus/char/flash.c | 1
drivers/sbus/char/uctrl.c | 1
drivers/scsi/53c700.c | 4
drivers/scsi/a2091.c | 1
drivers/scsi/a3000.c | 1
drivers/scsi/arm/cumana_2.c | 4
drivers/scsi/arm/eesox.c | 4
drivers/scsi/arm/powertec.c | 4
drivers/scsi/dpt_i2o.c | 4
drivers/scsi/gvp11.c | 1
drivers/scsi/lasi700.c | 1
drivers/scsi/mac53c94.c | 4
drivers/scsi/mesh.c | 4
drivers/scsi/mvme147.c | 1
drivers/scsi/qlogicpti.c | 4
drivers/scsi/sni_53c710.c | 1
drivers/scsi/zorro_esp.c | 4
drivers/staging/android/ashmem.c | 4
drivers/staging/comedi/comedi_fops.c | 2
drivers/staging/kpc2000/kpc_dma/fileops.c | 4
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c | 4
drivers/tee/optee/call.c | 4
drivers/tty/sysrq.c | 4
drivers/tty/vt/consolemap.c | 2
drivers/vfio/pci/vfio_pci.c | 22
drivers/vfio/vfio_iommu_type1.c | 8
drivers/vhost/vdpa.c | 4
drivers/video/console/newport_con.c | 1
drivers/video/fbdev/acornfb.c | 1
drivers/video/fbdev/atafb.c | 1
drivers/video/fbdev/cirrusfb.c | 1
drivers/video/fbdev/cyber2000fb.c | 1
drivers/video/fbdev/fb-puv3.c | 1
drivers/video/fbdev/hitfb.c | 1
drivers/video/fbdev/neofb.c | 1
drivers/video/fbdev/q40fb.c | 1
drivers/video/fbdev/savage/savagefb_driver.c | 1
drivers/xen/balloon.c | 1
drivers/xen/gntdev.c | 6
drivers/xen/grant-table.c | 1
drivers/xen/privcmd.c | 15
drivers/xen/xenbus/xenbus_probe.c | 1
drivers/xen/xenbus/xenbus_probe_backend.c | 1
drivers/xen/xenbus/xenbus_probe_frontend.c | 1
fs/aio.c | 4
fs/coredump.c | 8
fs/exec.c | 18
fs/ext2/file.c | 2
fs/ext4/super.c | 6
fs/hugetlbfs/inode.c | 2
fs/io_uring.c | 4
fs/kernfs/file.c | 4
fs/proc/array.c | 1
fs/proc/base.c | 24
fs/proc/meminfo.c | 1
fs/proc/nommu.c | 1
fs/proc/task_mmu.c | 34
fs/proc/task_nommu.c | 18
fs/proc/vmcore.c | 1
fs/userfaultfd.c | 46
fs/xfs/xfs_file.c | 2
fs/xfs/xfs_inode.c | 14
fs/xfs/xfs_iops.c | 4
include/asm-generic/io.h | 2
include/asm-generic/pgtable-nopmd.h | 1
include/asm-generic/pgtable-nopud.h | 1
include/asm-generic/pgtable.h | 1322 ----------------
include/linux/cache.h | 10
include/linux/crash_dump.h | 3
include/linux/dax.h | 1
include/linux/dma-noncoherent.h | 2
include/linux/fs.h | 4
include/linux/hmm.h | 2
include/linux/huge_mm.h | 2
include/linux/hugetlb.h | 2
include/linux/io-mapping.h | 4
include/linux/kallsyms.h | 4
include/linux/kasan.h | 4
include/linux/mempolicy.h | 2
include/linux/mm.h | 15
include/linux/mm_types.h | 4
include/linux/mmap_lock.h | 128 +
include/linux/mmu_notifier.h | 13
include/linux/pagemap.h | 2
include/linux/pgtable.h | 1444 +++++++++++++++++-
include/linux/rmap.h | 2
include/linux/sched/debug.h | 7
include/linux/sched/mm.h | 10
include/linux/uaccess.h | 62
include/xen/arm/page.h | 4
init/init_task.c | 1
ipc/shm.c | 8
kernel/acct.c | 6
kernel/bpf/stackmap.c | 21
kernel/bpf/syscall.c | 2
kernel/cgroup/cpuset.c | 4
kernel/debug/kdb/kdb_bt.c | 17
kernel/events/core.c | 10
kernel/events/uprobes.c | 20
kernel/exit.c | 11
kernel/fork.c | 15
kernel/futex.c | 4
kernel/locking/lockdep.c | 4
kernel/locking/rtmutex-debug.c | 4
kernel/power/snapshot.c | 1
kernel/relay.c | 2
kernel/sched/core.c | 10
kernel/sched/fair.c | 4
kernel/sys.c | 22
kernel/trace/bpf_trace.c | 176 +-
kernel/trace/ftrace.c | 8
kernel/trace/trace_kprobe.c | 80
kernel/trace/trace_output.c | 4
lib/dump_stack.c | 4
lib/ioremap.c | 1
lib/test_hmm.c | 14
lib/test_lockup.c | 16
mm/debug.c | 10
mm/debug_vm_pgtable.c | 1
mm/filemap.c | 46
mm/frame_vector.c | 6
mm/gup.c | 73
mm/hmm.c | 2
mm/huge_memory.c | 8
mm/hugetlb.c | 3
mm/init-mm.c | 6
mm/internal.h | 6
mm/khugepaged.c | 72
mm/ksm.c | 48
mm/maccess.c | 496 +++---
mm/madvise.c | 40
mm/memcontrol.c | 10
mm/memory.c | 61
mm/mempolicy.c | 36
mm/migrate.c | 16
mm/mincore.c | 8
mm/mlock.c | 22
mm/mmap.c | 74
mm/mmu_gather.c | 2
mm/mmu_notifier.c | 22
mm/mprotect.c | 22
mm/mremap.c | 14
mm/msync.c | 8
mm/nommu.c | 22
mm/oom_kill.c | 14
mm/page_io.c | 1
mm/page_reporting.h | 2
mm/pagewalk.c | 12
mm/pgtable-generic.c | 6
mm/process_vm_access.c | 4
mm/ptdump.c | 4
mm/rmap.c | 12
mm/shmem.c | 5
mm/sparse-vmemmap.c | 1
mm/sparse.c | 1
mm/swap_state.c | 5
mm/swapfile.c | 5
mm/userfaultfd.c | 26
mm/util.c | 12
mm/vmacache.c | 1
mm/zsmalloc.c | 4
net/ipv4/tcp.c | 8
net/xdp/xdp_umem.c | 4
security/keys/keyctl.c | 2
sound/core/oss/pcm_oss.c | 2
sound/core/sgbuf.c | 1
sound/pci/hda/hda_intel.c | 4
sound/soc/intel/common/sst-firmware.c | 4
sound/soc/intel/haswell/sst-haswell-pcm.c | 4
tools/include/linux/kallsyms.h | 2
virt/kvm/async_pf.c | 4
virt/kvm/kvm_main.c | 9
942 files changed, 4580 insertions(+), 5662 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-06-08 4:35 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-06-08 4:35 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
Various trees. Mainly those parts of MM whose linux-next dependents
are now merged. I'm still sitting on ~160 patches which await merges
from -next.
54 patches, based on 9aa900c8094dba7a60dc805ecec1e9f720744ba1.
Subsystems affected by this patch series:
mm/proc
ipc
dynamic-debug
panic
lib
sysctl
mm/gup
mm/pagemap
Subsystem: mm/proc
SeongJae Park <sjpark@amazon.de>:
mm/page_idle.c: skip offline pages
Subsystem: ipc
Jules Irenge <jbi.octave@gmail.com>:
ipc/msg: add missing annotation for freeque()
Giuseppe Scrivano <gscrivan@redhat.com>:
ipc/namespace.c: use a work queue to free_ipc
Subsystem: dynamic-debug
Orson Zhai <orson.zhai@unisoc.com>:
dynamic_debug: add an option to enable dynamic debug for modules only
Subsystem: panic
Rafael Aquini <aquini@redhat.com>:
kernel: add panic_on_taint
Subsystem: lib
Manfred Spraul <manfred@colorfullife.com>:
xarray.h: correct return code documentation for xa_store_{bh,irq}()
Subsystem: sysctl
Vlastimil Babka <vbabka@suse.cz>:
Patch series "support setting sysctl parameters from kernel command line", v3:
kernel/sysctl: support setting sysctl parameters from kernel command line
kernel/sysctl: support handling command line aliases
kernel/hung_task convert hung_task_panic boot parameter to sysctl
tools/testing/selftests/sysctl/sysctl.sh: support CONFIG_TEST_SYSCTL=y
lib/test_sysctl: support testing of sysctl. boot parameter
"Guilherme G. Piccoli" <gpiccoli@canonical.com>:
kernel/watchdog.c: convert {soft/hard}lockup boot parameters to sysctl aliases
kernel/hung_task.c: introduce sysctl to print all traces when a hung task is detected
panic: add sysctl to dump all CPUs backtraces on oops event
Rafael Aquini <aquini@redhat.com>:
kernel/sysctl.c: ignore out-of-range taint bits introduced via kernel.tainted
Subsystem: mm/gup
Souptick Joarder <jrdr.linux@gmail.com>:
mm/gup.c: convert to use get_user_{page|pages}_fast_only()
John Hubbard <jhubbard@nvidia.com>:
mm/gup: update pin_user_pages.rst for "case 3" (mmu notifiers)
Patch series "mm/gup: introduce pin_user_pages_locked(), use it in frame_vector.c", v2:
mm/gup: introduce pin_user_pages_locked()
mm/gup: frame_vector: convert get_user_pages() --> pin_user_pages()
mm/gup: documentation fix for pin_user_pages*() APIs
Patch series "vhost, docs: convert to pin_user_pages(), new "case 5"":
docs: mm/gup: pin_user_pages.rst: add a "case 5"
vhost: convert get_user_pages() --> pin_user_pages()
Subsystem: mm/pagemap
Alexander Gordeev <agordeev@linux.ibm.com>:
mm/mmap.c: add more sanity checks to get_unmapped_area()
mm/mmap.c: do not allow mappings outside of allowed limits
Christoph Hellwig <hch@lst.de>:
Patch series "sort out the flush_icache_range mess", v2:
arm: fix the flush_icache_range arguments in set_fiq_handler
nds32: unexport flush_icache_page
powerpc: unexport flush_icache_user_range
unicore32: remove flush_cache_user_range
asm-generic: fix the inclusion guards for cacheflush.h
asm-generic: don't include <linux/mm.h> in cacheflush.h
asm-generic: improve the flush_dcache_page stub
alpha: use asm-generic/cacheflush.h
arm64: use asm-generic/cacheflush.h
c6x: use asm-generic/cacheflush.h
hexagon: use asm-generic/cacheflush.h
ia64: use asm-generic/cacheflush.h
microblaze: use asm-generic/cacheflush.h
m68knommu: use asm-generic/cacheflush.h
openrisc: use asm-generic/cacheflush.h
powerpc: use asm-generic/cacheflush.h
riscv: use asm-generic/cacheflush.h
arm,sparc,unicore32: remove flush_icache_user_range
mm: rename flush_icache_user_range to flush_icache_user_page
asm-generic: add a flush_icache_user_range stub
sh: implement flush_icache_user_range
xtensa: implement flush_icache_user_range
arm: rename flush_cache_user_range to flush_icache_user_range
m68k: implement flush_icache_user_range
exec: only build read_code when needed
exec: use flush_icache_user_range in read_code
binfmt_flat: use flush_icache_user_range
nommu: use flush_icache_user_range in brk and mmap
module: move the set_fs hack for flush_icache_range to m68k
Konstantin Khlebnikov <khlebnikov@yandex-team.ru>:
doc: cgroup: update note about conditions when oom killer is invoked
Documentation/admin-guide/cgroup-v2.rst | 17 +-
Documentation/admin-guide/dynamic-debug-howto.rst | 5
Documentation/admin-guide/kdump/kdump.rst | 8 +
Documentation/admin-guide/kernel-parameters.txt | 34 +++-
Documentation/admin-guide/sysctl/kernel.rst | 37 ++++
Documentation/core-api/pin_user_pages.rst | 47 ++++--
arch/alpha/include/asm/cacheflush.h | 38 +----
arch/alpha/kernel/smp.c | 2
arch/arm/include/asm/cacheflush.h | 7
arch/arm/kernel/fiq.c | 4
arch/arm/kernel/traps.c | 2
arch/arm64/include/asm/cacheflush.h | 46 ------
arch/c6x/include/asm/cacheflush.h | 19 --
arch/hexagon/include/asm/cacheflush.h | 19 --
arch/ia64/include/asm/cacheflush.h | 30 ----
arch/m68k/include/asm/cacheflush_mm.h | 6
arch/m68k/include/asm/cacheflush_no.h | 19 --
arch/m68k/mm/cache.c | 13 +
arch/microblaze/include/asm/cacheflush.h | 29 ---
arch/nds32/include/asm/cacheflush.h | 4
arch/nds32/mm/cacheflush.c | 3
arch/openrisc/include/asm/cacheflush.h | 33 ----
arch/powerpc/include/asm/cacheflush.h | 46 +-----
arch/powerpc/kvm/book3s_64_mmu_hv.c | 2
arch/powerpc/kvm/book3s_64_mmu_radix.c | 2
arch/powerpc/mm/mem.c | 3
arch/powerpc/perf/callchain_64.c | 4
arch/riscv/include/asm/cacheflush.h | 65 --------
arch/sh/include/asm/cacheflush.h | 1
arch/sparc/include/asm/cacheflush_32.h | 2
arch/sparc/include/asm/cacheflush_64.h | 1
arch/um/include/asm/tlb.h | 2
arch/unicore32/include/asm/cacheflush.h | 11 -
arch/x86/include/asm/cacheflush.h | 2
arch/xtensa/include/asm/cacheflush.h | 2
drivers/media/platform/omap3isp/ispvideo.c | 2
drivers/nvdimm/pmem.c | 3
drivers/vhost/vhost.c | 5
fs/binfmt_flat.c | 2
fs/exec.c | 5
fs/proc/proc_sysctl.c | 163 ++++++++++++++++++++--
include/asm-generic/cacheflush.h | 25 +--
include/linux/dev_printk.h | 6
include/linux/dynamic_debug.h | 2
include/linux/ipc_namespace.h | 2
include/linux/kernel.h | 9 +
include/linux/mm.h | 12 +
include/linux/net.h | 3
include/linux/netdevice.h | 6
include/linux/printk.h | 9 -
include/linux/sched/sysctl.h | 7
include/linux/sysctl.h | 4
include/linux/xarray.h | 4
include/rdma/ib_verbs.h | 6
init/main.c | 2
ipc/msg.c | 2
ipc/namespace.c | 24 ++-
kernel/events/core.c | 4
kernel/events/uprobes.c | 2
kernel/hung_task.c | 30 ++--
kernel/module.c | 8 -
kernel/panic.c | 45 ++++++
kernel/sysctl.c | 38 ++++-
kernel/watchdog.c | 37 +---
lib/Kconfig.debug | 12 +
lib/Makefile | 2
lib/dynamic_debug.c | 9 -
lib/test_sysctl.c | 13 +
mm/frame_vector.c | 7
mm/gup.c | 74 +++++++--
mm/mmap.c | 28 ++-
mm/nommu.c | 4
mm/page_alloc.c | 9 -
mm/page_idle.c | 7
tools/testing/selftests/sysctl/sysctl.sh | 44 +++++
virt/kvm/kvm_main.c | 8 -
76 files changed, 732 insertions(+), 517 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-06-04 23:45 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-06-04 23:45 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
- More MM work. 100ish more to go. Mike's "mm: remove
__ARCH_HAS_5LEVEL_HACK" series should fix the current ppc issue.
- Various other little subsystems
127 patches, based on 6929f71e46bdddbf1c4d67c2728648176c67c555.
Subsystems affected by this patch series:
kcov
mm/pagemap
mm/vmalloc
mm/kmap
mm/util
mm/memory-hotplug
mm/cleanups
mm/zram
procfs
core-kernel
get_maintainer
lib
bitops
checkpatch
binfmt
init
fat
seq_file
exec
rapidio
relay
selftests
ubsan
Subsystem: kcov
Andrey Konovalov <andreyknvl@google.com>:
Patch series "kcov: collect coverage from usb soft interrupts", v4:
kcov: cleanup debug messages
kcov: fix potential use-after-free in kcov_remote_start
kcov: move t->kcov assignments into kcov_start/stop
kcov: move t->kcov_sequence assignment
kcov: use t->kcov_mode as enabled indicator
kcov: collect coverage from interrupts
usb: core: kcov: collect coverage from usb complete callback
Subsystem: mm/pagemap
Feng Tang <feng.tang@intel.com>:
mm/util.c: remove the VM_WARN_ONCE for vm_committed_as underflow check
Mike Rapoport <rppt@linux.ibm.com>:
Patch series "mm: remove __ARCH_HAS_5LEVEL_HACK", v4:
h8300: remove usage of __ARCH_USE_5LEVEL_HACK
arm: add support for folded p4d page tables
arm64: add support for folded p4d page tables
hexagon: remove __ARCH_USE_5LEVEL_HACK
ia64: add support for folded p4d page tables
nios2: add support for folded p4d page tables
openrisc: add support for folded p4d page tables
powerpc: add support for folded p4d page tables
Geert Uytterhoeven <geert+renesas@glider.be>:
sh: fault: modernize printing of kernel messages
Mike Rapoport <rppt@linux.ibm.com>:
sh: drop __pXd_offset() macros that duplicate pXd_index() ones
sh: add support for folded p4d page tables
unicore32: remove __ARCH_USE_5LEVEL_HACK
asm-generic: remove pgtable-nop4d-hack.h
mm: remove __ARCH_HAS_5LEVEL_HACK and include/asm-generic/5level-fixup.h
Anshuman Khandual <anshuman.khandual@arm.com>:
Patch series "mm/debug: Add tests validating architecture page table:
x86/mm: define mm_p4d_folded()
mm/debug: add tests validating architecture page table helpers
Subsystem: mm/vmalloc
Jeongtae Park <jtp.park@samsung.com>:
mm/vmalloc: fix a typo in comment
Subsystem: mm/kmap
Ira Weiny <ira.weiny@intel.com>:
Patch series "Remove duplicated kmap code", v3:
arch/kmap: remove BUG_ON()
arch/xtensa: move kmap build bug out of the way
arch/kmap: remove redundant arch specific kmaps
arch/kunmap: remove duplicate kunmap implementations
{x86,powerpc,microblaze}/kmap: move preempt disable
arch/kmap_atomic: consolidate duplicate code
arch/kunmap_atomic: consolidate duplicate code
arch/kmap: ensure kmap_prot visibility
arch/kmap: don't hard code kmap_prot values
arch/kmap: define kmap_atomic_prot() for all arch's
drm: remove drm specific kmap_atomic code
kmap: remove kmap_atomic_to_page()
parisc/kmap: remove duplicate kmap code
sparc: remove unnecessary includes
kmap: consolidate kmap_prot definitions
Subsystem: mm/util
Waiman Long <longman@redhat.com>:
mm: add kvfree_sensitive() for freeing sensitive data objects
Subsystem: mm/memory-hotplug
Vishal Verma <vishal.l.verma@intel.com>:
mm/memory_hotplug: refrain from adding memory into an impossible node
David Hildenbrand <david@redhat.com>:
powerpc/pseries/hotplug-memory: stop checking is_mem_section_removable()
mm/memory_hotplug: remove is_mem_section_removable()
Patch series "mm/memory_hotplug: handle memblocks only with:
mm/memory_hotplug: set node_start_pfn of hotadded pgdat to 0
mm/memory_hotplug: handle memblocks only with CONFIG_ARCH_KEEP_MEMBLOCK
Patch series "mm/memory_hotplug: Interface to add driver-managed system:
mm/memory_hotplug: introduce add_memory_driver_managed()
kexec_file: don't place kexec images on IORESOURCE_MEM_DRIVER_MANAGED
device-dax: add memory via add_memory_driver_managed()
Michal Hocko <mhocko@kernel.org>:
mm/memory_hotplug: disable the functionality for 32b
Subsystem: mm/cleanups
chenqiwu <chenqiwu@xiaomi.com>:
mm: replace zero-length array with flexible-array member
Ethon Paul <ethp@qq.com>:
mm/memory_hotplug: fix a typo in comment "recoreded"->"recorded"
mm: ksm: fix a typo in comment "alreaady"->"already"
mm: mmap: fix a typo in comment "compatbility"->"compatibility"
mm/hugetlb: fix a typos in comments
mm/vmsan: fix some typos in comment
mm/compaction: fix a typo in comment "pessemistic"->"pessimistic"
mm/memblock: fix a typo in comment "implict"->"implicit"
mm/list_lru: fix a typo in comment "numbesr"->"numbers"
mm/filemap: fix a typo in comment "unneccssary"->"unnecessary"
mm/frontswap: fix some typos in frontswap.c
mm, memcg: fix some typos in memcontrol.c
mm: fix a typo in comment "strucure"->"structure"
mm/slub: fix a typo in comment "disambiguiation"->"disambiguation"
mm/sparse: fix a typo in comment "convienence"->"convenience"
mm/page-writeback: fix a typo in comment "effictive"->"effective"
mm/memory: fix a typo in comment "attampt"->"attempt"
Zou Wei <zou_wei@huawei.com>:
mm: use false for bool variable
Jason Yan <yanaijie@huawei.com>:
include/linux/mm.h: return true in cpupid_pid_unset()
Subsystem: mm/zram
Andy Shevchenko <andriy.shevchenko@linux.intel.com>:
zcomp: Use ARRAY_SIZE() for backends list
Subsystem: procfs
Alexey Dobriyan <adobriyan@gmail.com>:
proc: rename "catch" function argument
Subsystem: core-kernel
Jason Yan <yanaijie@huawei.com>:
user.c: make uidhash_table static
Subsystem: get_maintainer
Joe Perches <joe@perches.com>:
get_maintainer: add email addresses from .yaml files
get_maintainer: fix unexpected behavior for path/to//file (double slashes)
Subsystem: lib
Christophe JAILLET <christophe.jaillet@wanadoo.fr>:
lib/math: avoid trailing newline hidden in pr_fmt()
KP Singh <kpsingh@chromium.org>:
lib: Add might_fault() to strncpy_from_user.
Jason Yan <yanaijie@huawei.com>:
lib/test_lockup.c: make test_inode static
Jann Horn <jannh@google.com>:
lib/zlib: remove outdated and incorrect pre-increment optimization
Joe Perches <joe@perches.com>:
lib/percpu-refcount.c: use a more common logging style
Tan Hu <tan.hu@zte.com.cn>:
lib/flex_proportions.c: cleanup __fprop_inc_percpu_max
Jesse Brandeburg <jesse.brandeburg@intel.com>:
lib: make a test module with set/clear bit
Subsystem: bitops
Arnd Bergmann <arnd@arndb.de>:
include/linux/bitops.h: avoid clang shift-count-overflow warnings
Subsystem: checkpatch
Joe Perches <joe@perches.com>:
checkpatch: additional MAINTAINER section entry ordering checks
checkpatch: look for c99 comments in ctx_locate_comment
checkpatch: disallow --git and --file/--fix
Geert Uytterhoeven <geert+renesas@glider.be>:
checkpatch: use patch subject when reading from stdin
Subsystem: binfmt
Anthony Iliopoulos <ailiop@suse.com>:
fs/binfmt_elf: remove redundant elf_map ifndef
Nick Desaulniers <ndesaulniers@google.com>:
elfnote: mark all .note sections SHF_ALLOC
Subsystem: init
Chris Down <chris@chrisdown.name>:
init: allow distribution configuration of default init
Subsystem: fat
OGAWA Hirofumi <hirofumi@mail.parknet.co.jp>:
fat: don't allow to mount if the FAT length == 0
fat: improve the readahead for FAT entries
Subsystem: seq_file
Joe Perches <joe@perches.com>:
fs/seq_file.c: seq_read: Update pr_info_ratelimited
Kefeng Wang <wangkefeng.wang@huawei.com>:
Patch series "seq_file: Introduce DEFINE_SEQ_ATTRIBUTE() helper macro":
include/linux/seq_file.h: introduce DEFINE_SEQ_ATTRIBUTE() helper macro
mm/vmstat.c: convert to use DEFINE_SEQ_ATTRIBUTE macro
kernel/kprobes.c: convert to use DEFINE_SEQ_ATTRIBUTE macro
Subsystem: exec
Christoph Hellwig <hch@lst.de>:
exec: simplify the copy_strings_kernel calling convention
exec: open code copy_string_kernel
Subsystem: rapidio
Madhuparna Bhowmik <madhuparnabhowmik10@gmail.com>:
rapidio: avoid data race between file operation callbacks and mport_cdev_add().
John Hubbard <jhubbard@nvidia.com>:
rapidio: convert get_user_pages() --> pin_user_pages()
Subsystem: relay
Daniel Axtens <dja@axtens.net>:
kernel/relay.c: handle alloc_percpu returning NULL in relay_open
Pengcheng Yang <yangpc@wangsu.com>:
kernel/relay.c: fix read_pos error when multiple readers
Subsystem: selftests
Ram Pai <linuxram@us.ibm.com>:
Patch series "selftests, powerpc, x86: Memory Protection Keys", v19:
selftests/x86/pkeys: move selftests to arch-neutral directory
selftests/vm/pkeys: rename all references to pkru to a generic name
selftests/vm/pkeys: move generic definitions to header file
Thiago Jung Bauermann <bauerman@linux.ibm.com>:
selftests/vm/pkeys: move some definitions to arch-specific header
selftests/vm/pkeys: make gcc check arguments of sigsafe_printf()
Sandipan Das <sandipan@linux.ibm.com>:
selftests: vm: pkeys: Use sane types for pkey register
selftests: vm: pkeys: add helpers for pkey bits
Ram Pai <linuxram@us.ibm.com>:
selftests/vm/pkeys: fix pkey_disable_clear()
selftests/vm/pkeys: fix assertion in pkey_disable_set/clear()
selftests/vm/pkeys: fix alloc_random_pkey() to make it really random
Sandipan Das <sandipan@linux.ibm.com>:
selftests: vm: pkeys: use the correct huge page size
Ram Pai <linuxram@us.ibm.com>:
selftests/vm/pkeys: introduce generic pkey abstractions
selftests/vm/pkeys: introduce powerpc support
"Desnes A. Nunes do Rosario" <desnesn@linux.vnet.ibm.com>:
selftests/vm/pkeys: fix number of reserved powerpc pkeys
Ram Pai <linuxram@us.ibm.com>:
selftests/vm/pkeys: fix assertion in test_pkey_alloc_exhaust()
selftests/vm/pkeys: improve checks to determine pkey support
selftests/vm/pkeys: associate key on a mapped page and detect access violation
selftests/vm/pkeys: associate key on a mapped page and detect write violation
selftests/vm/pkeys: detect write violation on a mapped access-denied-key page
selftests/vm/pkeys: introduce a sub-page allocator
selftests/vm/pkeys: test correct behaviour of pkey-0
selftests/vm/pkeys: override access right definitions on powerpc
Sandipan Das <sandipan@linux.ibm.com>:
selftests: vm: pkeys: use the correct page size on powerpc
selftests: vm: pkeys: fix multilib builds for x86
Jagadeesh Pagadala <jagdsh.linux@gmail.com>:
tools/testing/selftests/vm: remove duplicate headers
Subsystem: ubsan
Arnd Bergmann <arnd@arndb.de>:
lib/ubsan.c: fix gcc-10 warnings
Documentation/dev-tools/kcov.rst | 17
Documentation/features/debug/debug-vm-pgtable/arch-support.txt | 34
arch/arc/Kconfig | 1
arch/arc/include/asm/highmem.h | 20
arch/arc/mm/highmem.c | 34
arch/arm/include/asm/highmem.h | 9
arch/arm/include/asm/pgtable.h | 1
arch/arm/lib/uaccess_with_memcpy.c | 7
arch/arm/mach-sa1100/assabet.c | 2
arch/arm/mm/dump.c | 29
arch/arm/mm/fault-armv.c | 7
arch/arm/mm/fault.c | 22
arch/arm/mm/highmem.c | 41
arch/arm/mm/idmap.c | 3
arch/arm/mm/init.c | 2
arch/arm/mm/ioremap.c | 12
arch/arm/mm/mm.h | 2
arch/arm/mm/mmu.c | 35
arch/arm/mm/pgd.c | 40
arch/arm64/Kconfig | 1
arch/arm64/include/asm/kvm_mmu.h | 10
arch/arm64/include/asm/pgalloc.h | 10
arch/arm64/include/asm/pgtable-types.h | 5
arch/arm64/include/asm/pgtable.h | 37
arch/arm64/include/asm/stage2_pgtable.h | 48
arch/arm64/kernel/hibernate.c | 44
arch/arm64/kvm/mmu.c | 209
arch/arm64/mm/fault.c | 9
arch/arm64/mm/hugetlbpage.c | 15
arch/arm64/mm/kasan_init.c | 26
arch/arm64/mm/mmu.c | 52
arch/arm64/mm/pageattr.c | 7
arch/csky/include/asm/highmem.h | 12
arch/csky/mm/highmem.c | 64
arch/h8300/include/asm/pgtable.h | 1
arch/hexagon/include/asm/fixmap.h | 4
arch/hexagon/include/asm/pgtable.h | 1
arch/ia64/include/asm/pgalloc.h | 4
arch/ia64/include/asm/pgtable.h | 17
arch/ia64/mm/fault.c | 7
arch/ia64/mm/hugetlbpage.c | 18
arch/ia64/mm/init.c | 28
arch/microblaze/include/asm/highmem.h | 55
arch/microblaze/mm/highmem.c | 21
arch/microblaze/mm/init.c | 3
arch/mips/include/asm/highmem.h | 11
arch/mips/mm/cache.c | 6
arch/mips/mm/highmem.c | 62
arch/nds32/include/asm/highmem.h | 9
arch/nds32/mm/highmem.c | 49
arch/nios2/include/asm/pgtable.h | 3
arch/nios2/mm/fault.c | 9
arch/nios2/mm/ioremap.c | 6
arch/openrisc/include/asm/pgtable.h | 1
arch/openrisc/mm/fault.c | 10
arch/openrisc/mm/init.c | 4
arch/parisc/include/asm/cacheflush.h | 32
arch/powerpc/Kconfig | 1
arch/powerpc/include/asm/book3s/32/pgtable.h | 1
arch/powerpc/include/asm/book3s/64/hash.h | 4
arch/powerpc/include/asm/book3s/64/pgalloc.h | 4
arch/powerpc/include/asm/book3s/64/pgtable.h | 60
arch/powerpc/include/asm/book3s/64/radix.h | 6
arch/powerpc/include/asm/highmem.h | 56
arch/powerpc/include/asm/nohash/32/pgtable.h | 1
arch/powerpc/include/asm/nohash/64/pgalloc.h | 2
arch/powerpc/include/asm/nohash/64/pgtable-4k.h | 32
arch/powerpc/include/asm/nohash/64/pgtable.h | 6
arch/powerpc/include/asm/pgtable.h | 10
arch/powerpc/kvm/book3s_64_mmu_radix.c | 32
arch/powerpc/lib/code-patching.c | 7
arch/powerpc/mm/book3s64/hash_pgtable.c | 4
arch/powerpc/mm/book3s64/radix_pgtable.c | 26
arch/powerpc/mm/book3s64/subpage_prot.c | 6
arch/powerpc/mm/highmem.c | 26
arch/powerpc/mm/hugetlbpage.c | 28
arch/powerpc/mm/kasan/kasan_init_32.c | 2
arch/powerpc/mm/mem.c | 3
arch/powerpc/mm/nohash/book3e_pgtable.c | 15
arch/powerpc/mm/pgtable.c | 30
arch/powerpc/mm/pgtable_64.c | 10
arch/powerpc/mm/ptdump/hashpagetable.c | 20
arch/powerpc/mm/ptdump/ptdump.c | 12
arch/powerpc/platforms/pseries/hotplug-memory.c | 26
arch/powerpc/xmon/xmon.c | 27
arch/s390/Kconfig | 1
arch/sh/include/asm/pgtable-2level.h | 1
arch/sh/include/asm/pgtable-3level.h | 1
arch/sh/include/asm/pgtable_32.h | 5
arch/sh/include/asm/pgtable_64.h | 5
arch/sh/kernel/io_trapped.c | 7
arch/sh/mm/cache-sh4.c | 4
arch/sh/mm/cache-sh5.c | 7
arch/sh/mm/fault.c | 64
arch/sh/mm/hugetlbpage.c | 28
arch/sh/mm/init.c | 15
arch/sh/mm/kmap.c | 2
arch/sh/mm/tlbex_32.c | 6
arch/sh/mm/tlbex_64.c | 7
arch/sparc/include/asm/highmem.h | 29
arch/sparc/mm/highmem.c | 31
arch/sparc/mm/io-unit.c | 1
arch/sparc/mm/iommu.c | 1
arch/unicore32/include/asm/pgtable.h | 1
arch/unicore32/kernel/hibernate.c | 4
arch/x86/Kconfig | 1
arch/x86/include/asm/fixmap.h | 1
arch/x86/include/asm/highmem.h | 37
arch/x86/include/asm/pgtable_64.h | 6
arch/x86/mm/highmem_32.c | 52
arch/xtensa/include/asm/highmem.h | 31
arch/xtensa/mm/highmem.c | 28
drivers/block/zram/zcomp.c | 7
drivers/dax/dax-private.h | 1
drivers/dax/kmem.c | 28
drivers/gpu/drm/ttm/ttm_bo_util.c | 56
drivers/gpu/drm/vmwgfx/vmwgfx_blit.c | 17
drivers/rapidio/devices/rio_mport_cdev.c | 27
drivers/usb/core/hcd.c | 3
fs/binfmt_elf.c | 4
fs/binfmt_em86.c | 6
fs/binfmt_misc.c | 4
fs/binfmt_script.c | 6
fs/exec.c | 58
fs/fat/fatent.c | 103
fs/fat/inode.c | 6
fs/proc/array.c | 8
fs/seq_file.c | 7
include/asm-generic/5level-fixup.h | 59
include/asm-generic/pgtable-nop4d-hack.h | 64
include/asm-generic/pgtable-nopud.h | 4
include/drm/ttm/ttm_bo_api.h | 4
include/linux/binfmts.h | 3
include/linux/bitops.h | 2
include/linux/elfnote.h | 2
include/linux/highmem.h | 89
include/linux/ioport.h | 1
include/linux/memory_hotplug.h | 9
include/linux/mm.h | 12
include/linux/sched.h | 3
include/linux/seq_file.h | 19
init/Kconfig | 10
init/main.c | 10
kernel/kcov.c | 282 -
kernel/kexec_file.c | 5
kernel/kprobes.c | 34
kernel/relay.c | 22
kernel/user.c | 2
lib/Kconfig.debug | 44
lib/Makefile | 2
lib/flex_proportions.c | 7
lib/math/prime_numbers.c | 10
lib/percpu-refcount.c | 6
lib/strncpy_from_user.c | 1
lib/test_bitops.c | 60
lib/test_lockup.c | 2
lib/ubsan.c | 33
lib/zlib_inflate/inffast.c | 91
mm/Kconfig | 4
mm/Makefile | 1
mm/compaction.c | 2
mm/debug_vm_pgtable.c | 382 +
mm/filemap.c | 2
mm/frontswap.c | 6
mm/huge_memory.c | 2
mm/hugetlb.c | 16
mm/internal.h | 2
mm/kasan/init.c | 11
mm/ksm.c | 10
mm/list_lru.c | 2
mm/memblock.c | 2
mm/memcontrol.c | 4
mm/memory.c | 10
mm/memory_hotplug.c | 179
mm/mmap.c | 2
mm/mremap.c | 2
mm/page-writeback.c | 2
mm/slub.c | 2
mm/sparse.c | 2
mm/util.c | 22
mm/vmalloc.c | 2
mm/vmscan.c | 6
mm/vmstat.c | 32
mm/zbud.c | 2
scripts/checkpatch.pl | 62
scripts/get_maintainer.pl | 46
security/keys/internal.h | 11
security/keys/keyctl.c | 16
tools/testing/selftests/lib/config | 1
tools/testing/selftests/vm/.gitignore | 1
tools/testing/selftests/vm/Makefile | 75
tools/testing/selftests/vm/mremap_dontunmap.c | 1
tools/testing/selftests/vm/pkey-helpers.h | 557 +-
tools/testing/selftests/vm/pkey-powerpc.h | 153
tools/testing/selftests/vm/pkey-x86.h | 191
tools/testing/selftests/vm/protection_keys.c | 2370 ++++++++--
tools/testing/selftests/x86/.gitignore | 1
tools/testing/selftests/x86/Makefile | 2
tools/testing/selftests/x86/pkey-helpers.h | 219
tools/testing/selftests/x86/protection_keys.c | 1506 ------
200 files changed, 5182 insertions(+), 4033 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-06-03 22:55 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-06-03 22:55 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
More mm/ work, plenty more to come.
131 patches, based on d6f9469a03d832dcd17041ed67774ffb5f3e73b3.
Subsystems affected by this patch series:
mm/slub
mm/memcg
mm/gup
mm/kasan
mm/pagealloc
mm/hugetlb
mm/vmscan
mm/tools
mm/mempolicy
mm/memblock
mm/hugetlbfs
mm/thp
mm/mmap
mm/kconfig
Subsystem: mm/slub
Wang Hai <wanghai38@huawei.com>:
mm/slub: fix a memory leak in sysfs_slab_add()
Subsystem: mm/memcg
Shakeel Butt <shakeelb@google.com>:
mm/memcg: optimize memory.numa_stat like memory.stat
Subsystem: mm/gup
John Hubbard <jhubbard@nvidia.com>:
Patch series "mm/gup, drm/i915: refactor gup_fast, convert to pin_user_pages()", v2:
mm/gup: move __get_user_pages_fast() down a few lines in gup.c
mm/gup: refactor and de-duplicate gup_fast() code
mm/gup: introduce pin_user_pages_fast_only()
drm/i915: convert get_user_pages() --> pin_user_pages()
mm/gup: might_lock_read(mmap_sem) in get_user_pages_fast()
Subsystem: mm/kasan
Daniel Axtens <dja@axtens.net>:
Patch series "Fix some incompatibilites between KASAN and FORTIFY_SOURCE", v4:
kasan: stop tests being eliminated as dead code with FORTIFY_SOURCE
string.h: fix incompatibility between FORTIFY_SOURCE and KASAN
Subsystem: mm/pagealloc
Michal Hocko <mhocko@suse.com>:
mm: clarify __GFP_MEMALLOC usage
Mike Rapoport <rppt@linux.ibm.com>:
Patch series "mm: rework free_area_init*() funcitons":
mm: memblock: replace dereferences of memblock_region.nid with API calls
mm: make early_pfn_to_nid() and related defintions close to each other
mm: remove CONFIG_HAVE_MEMBLOCK_NODE_MAP option
mm: free_area_init: use maximal zone PFNs rather than zone sizes
mm: use free_area_init() instead of free_area_init_nodes()
alpha: simplify detection of memory zone boundaries
arm: simplify detection of memory zone boundaries
arm64: simplify detection of memory zone boundaries for UMA configs
csky: simplify detection of memory zone boundaries
m68k: mm: simplify detection of memory zone boundaries
parisc: simplify detection of memory zone boundaries
sparc32: simplify detection of memory zone boundaries
unicore32: simplify detection of memory zone boundaries
xtensa: simplify detection of memory zone boundaries
Baoquan He <bhe@redhat.com>:
mm: memmap_init: iterate over memblock regions rather that check each PFN
Mike Rapoport <rppt@linux.ibm.com>:
mm: remove early_pfn_in_nid() and CONFIG_NODES_SPAN_OTHER_NODES
mm: free_area_init: allow defining max_zone_pfn in descending order
mm: rename free_area_init_node() to free_area_init_memoryless_node()
mm: clean up free_area_init_node() and its helpers
mm: simplify find_min_pfn_with_active_regions()
docs/vm: update memory-models documentation
Wei Yang <richard.weiyang@gmail.com>:
Patch series "mm/page_alloc.c: cleanup on check page", v3:
mm/page_alloc.c: bad_[reason|flags] is not necessary when PageHWPoison
mm/page_alloc.c: bad_flags is not necessary for bad_page()
mm/page_alloc.c: rename free_pages_check_bad() to check_free_page_bad()
mm/page_alloc.c: rename free_pages_check() to check_free_page()
mm/page_alloc.c: extract check_[new|free]_page_bad() common part to page_bad_reason()
Roman Gushchin <guro@fb.com>:
mm,page_alloc,cma: conditionally prefer cma pageblocks for movable allocations
Baoquan He <bhe@redhat.com>:
mm/page_alloc.c: remove unused free_bootmem_with_active_regions
Patch series "improvements about lowmem_reserve and /proc/zoneinfo", v2:
mm/page_alloc.c: only tune sysctl_lowmem_reserve_ratio value once when changing it
mm/page_alloc.c: clear out zone->lowmem_reserve[] if the zone is empty
mm/vmstat.c: do not show lowmem reserve protection information of empty zone
Joonsoo Kim <iamjoonsoo.kim@lge.com>:
Patch series "integrate classzone_idx and high_zoneidx", v5:
mm/page_alloc: use ac->high_zoneidx for classzone_idx
mm/page_alloc: integrate classzone_idx and high_zoneidx
Wei Yang <richard.weiyang@gmail.com>:
mm/page_alloc.c: use NODE_MASK_NONE in build_zonelists()
mm: rename gfpflags_to_migratetype to gfp_migratetype for same convention
Sandipan Das <sandipan@linux.ibm.com>:
mm/page_alloc.c: reset numa stats for boot pagesets
Charan Teja Reddy <charante@codeaurora.org>:
mm, page_alloc: reset the zone->watermark_boost early
Anshuman Khandual <anshuman.khandual@arm.com>:
mm/page_alloc: restrict and formalize compound_page_dtors[]
Daniel Jordan <daniel.m.jordan@oracle.com>:
Patch series "initialize deferred pages with interrupts enabled", v4:
mm/pagealloc.c: call touch_nmi_watchdog() on max order boundaries in deferred init
Pavel Tatashin <pasha.tatashin@soleen.com>:
mm: initialize deferred pages with interrupts enabled
mm: call cond_resched() from deferred_init_memmap()
Daniel Jordan <daniel.m.jordan@oracle.com>:
Patch series "padata: parallelize deferred page init", v3:
padata: remove exit routine
padata: initialize earlier
padata: allocate work structures for parallel jobs from a pool
padata: add basic support for multithreaded jobs
mm: don't track number of pages during deferred initialization
mm: parallelize deferred_init_memmap()
mm: make deferred init's max threads arch-specific
padata: document multithreaded jobs
Chen Tao <chentao107@huawei.com>:
mm/page_alloc.c: add missing newline
Subsystem: mm/hugetlb
"Kirill A. Shutemov" <kirill.shutemov@linux.intel.com>:
Patch series "thp/khugepaged improvements and CoW semantics", v4:
khugepaged: add self test
khugepaged: do not stop collapse if less than half PTEs are referenced
khugepaged: drain all LRU caches before scanning pages
khugepaged: drain LRU add pagevec after swapin
khugepaged: allow to collapse a page shared across fork
khugepaged: allow to collapse PTE-mapped compound pages
thp: change CoW semantics for anon-THP
khugepaged: introduce 'max_ptes_shared' tunable
Mike Kravetz <mike.kravetz@oracle.com>:
Patch series "Clean up hugetlb boot command line processing", v4:
hugetlbfs: add arch_hugetlb_valid_size
hugetlbfs: move hugepagesz= parsing to arch independent code
hugetlbfs: remove hugetlb_add_hstate() warning for existing hstate
hugetlbfs: clean up command line processing
hugetlbfs: fix changes to command line processing
Li Xinhai <lixinhai.lxh@gmail.com>:
mm/hugetlb: avoid unnecessary check on pud and pmd entry in huge_pte_offset
Anshuman Khandual <anshuman.khandual@arm.com>:
Patch series "mm/hugetlb: Add some new generic fallbacks", v3:
arm64/mm: drop __HAVE_ARCH_HUGE_PTEP_GET
mm/hugetlb: define a generic fallback for is_hugepage_only_range()
mm/hugetlb: define a generic fallback for arch_clear_hugepage_flags()
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
mm: simplify calling a compound page destructor
Subsystem: mm/vmscan
Wei Yang <richard.weiyang@gmail.com>:
mm/vmscan.c: use update_lru_size() in update_lru_sizes()
Jaewon Kim <jaewon31.kim@samsung.com>:
mm/vmscan: count layzfree pages and fix nr_isolated_* mismatch
Maninder Singh <maninder1.s@samsung.com>:
mm/vmscan.c: change prototype for shrink_page_list
Qiwu Chen <qiwuchen55@gmail.com>:
mm/vmscan: update the comment of should_continue_reclaim()
Johannes Weiner <hannes@cmpxchg.org>:
Patch series "mm: memcontrol: charge swapin pages on instantiation", v2:
mm: fix NUMA node file count error in replace_page_cache()
mm: memcontrol: fix stat-corrupting race in charge moving
mm: memcontrol: drop @compound parameter from memcg charging API
mm: shmem: remove rare optimization when swapin races with hole punching
mm: memcontrol: move out cgroup swaprate throttling
mm: memcontrol: convert page cache to a new mem_cgroup_charge() API
mm: memcontrol: prepare uncharging for removal of private page type counters
mm: memcontrol: prepare move_account for removal of private page type counters
mm: memcontrol: prepare cgroup vmstat infrastructure for native anon counters
mm: memcontrol: switch to native NR_FILE_PAGES and NR_SHMEM counters
mm: memcontrol: switch to native NR_ANON_MAPPED counter
mm: memcontrol: switch to native NR_ANON_THPS counter
mm: memcontrol: convert anon and file-thp to new mem_cgroup_charge() API
mm: memcontrol: drop unused try/commit/cancel charge API
mm: memcontrol: prepare swap controller setup for integration
mm: memcontrol: make swap tracking an integral part of memory control
mm: memcontrol: charge swapin pages on instantiation
Alex Shi <alex.shi@linux.alibaba.com>:
mm: memcontrol: document the new swap control behavior
Johannes Weiner <hannes@cmpxchg.org>:
mm: memcontrol: delete unused lrucare handling
mm: memcontrol: update page->mem_cgroup stability rules
mm: fix LRU balancing effect of new transparent huge pages
mm: keep separate anon and file statistics on page reclaim activity
mm: allow swappiness that prefers reclaiming anon over the file workingset
mm: fold and remove lru_cache_add_anon() and lru_cache_add_file()
mm: workingset: let cache workingset challenge anon
mm: remove use-once cache bias from LRU balancing
mm: vmscan: drop unnecessary div0 avoidance rounding in get_scan_count()
mm: base LRU balancing on an explicit cost model
mm: deactivations shouldn't bias the LRU balance
mm: only count actual rotations as LRU reclaim cost
mm: balance LRU lists based on relative thrashing
mm: vmscan: determine anon/file pressure balance at the reclaim root
mm: vmscan: reclaim writepage is IO cost
mm: vmscan: limit the range of LRU type balancing
Shakeel Butt <shakeelb@google.com>:
mm: swap: fix vmstats for huge pages
mm: swap: memcg: fix memcg stats for huge pages
Subsystem: mm/tools
Changhee Han <ch0.han@lge.com>:
tools/vm/page_owner_sort.c: filter out unneeded line
Subsystem: mm/mempolicy
Michal Hocko <mhocko@suse.com>:
mm, mempolicy: fix up gup usage in lookup_node
Subsystem: mm/memblock
chenqiwu <chenqiwu@xiaomi.com>:
include/linux/memblock.h: fix minor typo and unclear comment
Mike Rapoport <rppt@linux.ibm.com>:
sparc32: register memory occupied by kernel as memblock.memory
Subsystem: mm/hugetlbfs
Shijie Hu <hushijie3@huawei.com>:
hugetlbfs: get unmapped area below TASK_UNMAPPED_BASE for hugetlbfs
Subsystem: mm/thp
Yang Shi <yang.shi@linux.alibaba.com>:
mm: thp: don't need to drain lru cache when splitting and mlocking THP
Anshuman Khandual <anshuman.khandual@arm.com>:
Patch series "mm/thp: Rename pmd_mknotpresent() as pmd_mknotvalid()", v2:
powerpc/mm: drop platform defined pmd_mknotpresent()
mm/thp: rename pmd_mknotpresent() as pmd_mkinvalid()
Subsystem: mm/mmap
Scott Cheloha <cheloha@linux.vnet.ibm.com>:
drivers/base/memory.c: cache memory blocks in xarray to accelerate lookup
Subsystem: mm/kconfig
Zong Li <zong.li@sifive.com>:
Patch series "Extract DEBUG_WX to shared use":
mm: add DEBUG_WX support
riscv: support DEBUG_WX
x86: mm: use ARCH_HAS_DEBUG_WX instead of arch defined
arm64: mm: use ARCH_HAS_DEBUG_WX instead of arch defined
Documentation/admin-guide/cgroup-v1/memory.rst | 19
Documentation/admin-guide/kernel-parameters.txt | 40
Documentation/admin-guide/mm/hugetlbpage.rst | 35
Documentation/admin-guide/mm/transhuge.rst | 7
Documentation/admin-guide/sysctl/vm.rst | 23
Documentation/core-api/padata.rst | 41
Documentation/features/vm/numa-memblock/arch-support.txt | 34
Documentation/vm/memory-model.rst | 9
Documentation/vm/page_owner.rst | 3
arch/alpha/mm/init.c | 16
arch/alpha/mm/numa.c | 22
arch/arc/include/asm/hugepage.h | 2
arch/arc/mm/init.c | 41
arch/arm/include/asm/hugetlb.h | 7
arch/arm/include/asm/pgtable-3level.h | 2
arch/arm/mm/init.c | 66
arch/arm64/Kconfig | 2
arch/arm64/Kconfig.debug | 29
arch/arm64/include/asm/hugetlb.h | 13
arch/arm64/include/asm/pgtable.h | 2
arch/arm64/mm/hugetlbpage.c | 48
arch/arm64/mm/init.c | 56
arch/arm64/mm/numa.c | 9
arch/c6x/mm/init.c | 8
arch/csky/kernel/setup.c | 26
arch/h8300/mm/init.c | 6
arch/hexagon/mm/init.c | 6
arch/ia64/Kconfig | 1
arch/ia64/include/asm/hugetlb.h | 5
arch/ia64/mm/contig.c | 2
arch/ia64/mm/discontig.c | 2
arch/m68k/mm/init.c | 6
arch/m68k/mm/mcfmmu.c | 9
arch/m68k/mm/motorola.c | 15
arch/m68k/mm/sun3mmu.c | 10
arch/microblaze/Kconfig | 1
arch/microblaze/mm/init.c | 2
arch/mips/Kconfig | 1
arch/mips/include/asm/hugetlb.h | 11
arch/mips/include/asm/pgtable.h | 2
arch/mips/loongson64/numa.c | 2
arch/mips/mm/init.c | 2
arch/mips/sgi-ip27/ip27-memory.c | 2
arch/nds32/mm/init.c | 11
arch/nios2/mm/init.c | 8
arch/openrisc/mm/init.c | 9
arch/parisc/include/asm/hugetlb.h | 10
arch/parisc/mm/init.c | 22
arch/powerpc/Kconfig | 10
arch/powerpc/include/asm/book3s/64/pgtable.h | 4
arch/powerpc/include/asm/hugetlb.h | 5
arch/powerpc/mm/hugetlbpage.c | 38
arch/powerpc/mm/mem.c | 2
arch/riscv/Kconfig | 2
arch/riscv/include/asm/hugetlb.h | 10
arch/riscv/include/asm/ptdump.h | 11
arch/riscv/mm/hugetlbpage.c | 44
arch/riscv/mm/init.c | 5
arch/s390/Kconfig | 1
arch/s390/include/asm/hugetlb.h | 8
arch/s390/mm/hugetlbpage.c | 34
arch/s390/mm/init.c | 2
arch/sh/Kconfig | 1
arch/sh/include/asm/hugetlb.h | 7
arch/sh/mm/init.c | 2
arch/sparc/Kconfig | 10
arch/sparc/include/asm/hugetlb.h | 10
arch/sparc/mm/init_32.c | 1
arch/sparc/mm/init_64.c | 67
arch/sparc/mm/srmmu.c | 21
arch/um/kernel/mem.c | 12
arch/unicore32/include/asm/memory.h | 2
arch/unicore32/include/mach/memory.h | 6
arch/unicore32/kernel/pci.c | 14
arch/unicore32/mm/init.c | 43
arch/x86/Kconfig | 11
arch/x86/Kconfig.debug | 27
arch/x86/include/asm/hugetlb.h | 10
arch/x86/include/asm/pgtable.h | 2
arch/x86/mm/hugetlbpage.c | 35
arch/x86/mm/init.c | 2
arch/x86/mm/init_64.c | 12
arch/x86/mm/kmmio.c | 2
arch/x86/mm/numa.c | 11
arch/xtensa/mm/init.c | 8
drivers/base/memory.c | 44
drivers/gpu/drm/i915/gem/i915_gem_userptr.c | 22
fs/cifs/file.c | 10
fs/fuse/dev.c | 2
fs/hugetlbfs/inode.c | 67
include/asm-generic/hugetlb.h | 2
include/linux/compaction.h | 9
include/linux/gfp.h | 7
include/linux/hugetlb.h | 16
include/linux/memblock.h | 15
include/linux/memcontrol.h | 102 -
include/linux/mm.h | 52
include/linux/mmzone.h | 46
include/linux/padata.h | 43
include/linux/string.h | 60
include/linux/swap.h | 17
include/linux/vm_event_item.h | 4
include/linux/vmstat.h | 2
include/trace/events/compaction.h | 22
include/trace/events/huge_memory.h | 3
include/trace/events/vmscan.h | 14
init/Kconfig | 17
init/main.c | 2
kernel/events/uprobes.c | 22
kernel/padata.c | 293 +++-
kernel/sysctl.c | 3
lib/test_kasan.c | 29
mm/Kconfig | 9
mm/Kconfig.debug | 32
mm/compaction.c | 70 -
mm/filemap.c | 55
mm/gup.c | 237 ++-
mm/huge_memory.c | 282 ----
mm/hugetlb.c | 260 ++-
mm/internal.h | 25
mm/khugepaged.c | 316 ++--
mm/memblock.c | 19
mm/memcontrol.c | 642 +++------
mm/memory.c | 103 -
mm/memory_hotplug.c | 10
mm/mempolicy.c | 5
mm/migrate.c | 30
mm/oom_kill.c | 4
mm/page_alloc.c | 735 ++++------
mm/page_owner.c | 7
mm/pgtable-generic.c | 2
mm/rmap.c | 53
mm/shmem.c | 156 --
mm/slab.c | 4
mm/slub.c | 8
mm/swap.c | 199 +-
mm/swap_cgroup.c | 10
mm/swap_state.c | 110 -
mm/swapfile.c | 39
mm/userfaultfd.c | 15
mm/vmscan.c | 344 ++--
mm/vmstat.c | 16
mm/workingset.c | 23
tools/testing/selftests/vm/.gitignore | 1
tools/testing/selftests/vm/Makefile | 1
tools/testing/selftests/vm/khugepaged.c | 1035 +++++++++++++++
tools/vm/page_owner_sort.c | 5
147 files changed, 3876 insertions(+), 3108 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2020-06-02 21:38 ` incoming Andrew Morton
@ 2020-06-02 22:18 ` Linus Torvalds
0 siblings, 0 replies; 409+ messages in thread
From: Linus Torvalds @ 2020-06-02 22:18 UTC (permalink / raw)
To: Andrew Morton; +Cc: mm-commits, Linux-MM
On Tue, Jun 2, 2020 at 2:38 PM Andrew Morton <akpm@linux-foundation.org> wrote:
>
> On Tue, 2 Jun 2020 13:45:49 -0700 Linus Torvalds <torvalds@linux-foundation.org> wrote:
> >
> > Hmm. I have no issues with conflicts, and already took your previous series.
>
> Well that's odd.
I meant "I saw the conflicts and had no issue with them". Nothing odd.
And I actually much prefer seeing conflicts from your series (against
other pulls I've done) over having you delay your patch bombs because
of any fear for them.
Linus
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2020-06-02 20:45 ` incoming Linus Torvalds
@ 2020-06-02 21:38 ` Andrew Morton
2020-06-02 22:18 ` incoming Linus Torvalds
0 siblings, 1 reply; 409+ messages in thread
From: Andrew Morton @ 2020-06-02 21:38 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, Linux-MM
On Tue, 2 Jun 2020 13:45:49 -0700 Linus Torvalds <torvalds@linux-foundation.org> wrote:
> On Tue, Jun 2, 2020 at 1:08 PM Andrew Morton <akpm@linux-foundation.org> wrote:
> >
> > The local_lock merge made rather a mess of all of this. I'm
> > cooking up a full resend of the same material.
>
> Hmm. I have no issues with conflicts, and already took your previous series.
Well that's odd.
> I've pushed it out now - does my tree match what you expect?
Yup, thanks.
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2020-06-02 20:08 ` incoming Andrew Morton
@ 2020-06-02 20:45 ` Linus Torvalds
2020-06-02 21:38 ` incoming Andrew Morton
0 siblings, 1 reply; 409+ messages in thread
From: Linus Torvalds @ 2020-06-02 20:45 UTC (permalink / raw)
To: Andrew Morton; +Cc: mm-commits, Linux-MM
On Tue, Jun 2, 2020 at 1:08 PM Andrew Morton <akpm@linux-foundation.org> wrote:
>
> The local_lock merge made rather a mess of all of this. I'm
> cooking up a full resend of the same material.
Hmm. I have no issues with conflicts, and already took your previous series.
I've pushed it out now - does my tree match what you expect?
Linus
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-06-02 20:09 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-06-02 20:09 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
A few little subsystems and a start of a lot of MM patches.
128 patches, based on f359287765c04711ff54fbd11645271d8e5ff763:
Subsystems affected by this patch series:
squashfs
ocfs2
parisc
vfs
mm/slab-generic
mm/slub
mm/debug
mm/pagecache
mm/gup
mm/swap
mm/memcg
mm/pagemap
mm/memory-failure
mm/vmalloc
mm/kasan
Subsystem: squashfs
Philippe Liard <pliard@google.com>:
squashfs: migrate from ll_rw_block usage to BIO
Subsystem: ocfs2
Jules Irenge <jbi.octave@gmail.com>:
ocfs2: add missing annotation for dlm_empty_lockres()
Gang He <ghe@suse.com>:
ocfs2: mount shared volume without ha stack
Subsystem: parisc
Andrew Morton <akpm@linux-foundation.org>:
arch/parisc/include/asm/pgtable.h: remove unused `old_pte'
Subsystem: vfs
Jeff Layton <jlayton@redhat.com>:
Patch series "vfs: have syncfs() return error when there are writeback:
vfs: track per-sb writeback errors and report them to syncfs
fs/buffer.c: record blockdev write errors in super_block that it backs
Subsystem: mm/slab-generic
Vlastimil Babka <vbabka@suse.cz>:
usercopy: mark dma-kmalloc caches as usercopy caches
Subsystem: mm/slub
Dongli Zhang <dongli.zhang@oracle.com>:
mm/slub.c: fix corrupted freechain in deactivate_slab()
Christoph Lameter <cl@linux.com>:
slub: Remove userspace notifier for cache add/remove
Christopher Lameter <cl@linux.com>:
slub: remove kmalloc under list_lock from list_slab_objects() V2
Qian Cai <cai@lca.pw>:
mm/slub: fix stack overruns with SLUB_STATS
Andrew Morton <akpm@linux-foundation.org>:
Documentation/vm/slub.rst: s/Toggle/Enable/
Subsystem: mm/debug
Vlastimil Babka <vbabka@suse.cz>:
mm, dump_page(): do not crash with invalid mapping pointer
Subsystem: mm/pagecache
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
Patch series "Change readahead API", v11:
mm: move readahead prototypes from mm.h
mm: return void from various readahead functions
mm: ignore return value of ->readpages
mm: move readahead nr_pages check into read_pages
mm: add new readahead_control API
mm: use readahead_control to pass arguments
mm: rename various 'offset' parameters to 'index'
mm: rename readahead loop variable to 'i'
mm: remove 'page_offset' from readahead loop
mm: put readahead pages in cache earlier
mm: add readahead address space operation
mm: move end_index check out of readahead loop
mm: add page_cache_readahead_unbounded
mm: document why we don't set PageReadahead
mm: use memalloc_nofs_save in readahead path
fs: convert mpage_readpages to mpage_readahead
btrfs: convert from readpages to readahead
erofs: convert uncompressed files from readpages to readahead
erofs: convert compressed files from readpages to readahead
ext4: convert from readpages to readahead
ext4: pass the inode to ext4_mpage_readpages
f2fs: convert from readpages to readahead
f2fs: pass the inode to f2fs_mpage_readpages
fuse: convert from readpages to readahead
iomap: convert from readpages to readahead
Guoqing Jiang <guoqing.jiang@cloud.ionos.com>:
Patch series "Introduce attach/detach_page_private to cleanup code":
include/linux/pagemap.h: introduce attach/detach_page_private
md: remove __clear_page_buffers and use attach/detach_page_private
btrfs: use attach/detach_page_private
fs/buffer.c: use attach/detach_page_private
f2fs: use attach/detach_page_private
iomap: use attach/detach_page_private
ntfs: replace attach_page_buffers with attach_page_private
orangefs: use attach/detach_page_private
buffer_head.h: remove attach_page_buffers
mm/migrate.c: call detach_page_private to cleanup code
mm_types.h: change set_page_private to inline function
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
mm/filemap.c: remove misleading comment
Chao Yu <yuchao0@huawei.com>:
mm/page-writeback.c: remove unused variable
NeilBrown <neilb@suse.de>:
mm/writeback: replace PF_LESS_THROTTLE with PF_LOCAL_THROTTLE
mm/writeback: discard NR_UNSTABLE_NFS, use NR_WRITEBACK instead
Subsystem: mm/gup
Souptick Joarder <jrdr.linux@gmail.com>:
mm/gup.c: update the documentation
John Hubbard <jhubbard@nvidia.com>:
mm/gup: introduce pin_user_pages_unlocked
ivtv: convert get_user_pages() --> pin_user_pages()
Miles Chen <miles.chen@mediatek.com>:
mm/gup.c: further document vma_permits_fault()
Subsystem: mm/swap
chenqiwu <chenqiwu@xiaomi.com>:
mm/swapfile: use list_{prev,next}_entry() instead of open-coding
Qian Cai <cai@lca.pw>:
mm/swap_state: fix a data race in swapin_nr_pages
Andrea Righi <andrea.righi@canonical.com>:
mm: swap: properly update readahead statistics in unuse_pte_range()
Wei Yang <richard.weiyang@gmail.com>:
mm/swapfile.c: offset is only used when there is more slots
mm/swapfile.c: explicitly show ssd/non-ssd is handled mutually exclusive
mm/swapfile.c: remove the unnecessary goto for SSD case
mm/swapfile.c: simplify the calculation of n_goal
mm/swapfile.c: remove the extra check in scan_swap_map_slots()
mm/swapfile.c: found_free could be represented by (tmp < max)
mm/swapfile.c: tmp is always smaller than max
mm/swapfile.c: omit a duplicate code by compare tmp and max first
Huang Ying <ying.huang@intel.com>:
swap: try to scan more free slots even when fragmented
Wei Yang <richard.weiyang@gmail.com>:
mm/swapfile.c: classify SWAP_MAP_XXX to make it more readable
mm/swapfile.c: __swap_entry_free() always free 1 entry
Huang Ying <ying.huang@intel.com>:
mm/swapfile.c: use prandom_u32_max()
swap: reduce lock contention on swap cache from swap slots allocation
Randy Dunlap <rdunlap@infradead.org>:
mm: swapfile: fix /proc/swaps heading and Size/Used/Priority alignment
Miaohe Lin <linmiaohe@huawei.com>:
include/linux/swap.h: delete meaningless __add_to_swap_cache() declaration
Subsystem: mm/memcg
Yafang Shao <laoar.shao@gmail.com>:
mm, memcg: add workingset_restore in memory.stat
Kaixu Xia <kaixuxia@tencent.com>:
mm: memcontrol: simplify value comparison between count and limit
Shakeel Butt <shakeelb@google.com>:
memcg: expose root cgroup's memory.stat
Jakub Kicinski <kuba@kernel.org>:
Patch series "memcg: Slow down swap allocation as the available space gets:
mm/memcg: prepare for swap over-high accounting and penalty calculation
mm/memcg: move penalty delay clamping out of calculate_high_delay()
mm/memcg: move cgroup high memory limit setting into struct page_counter
mm/memcg: automatically penalize tasks with high swap use
Zefan Li <lizefan@huawei.com>:
memcg: fix memcg_kmem_bypass() for remote memcg charging
Subsystem: mm/pagemap
Steven Price <steven.price@arm.com>:
Patch series "Fix W+X debug feature on x86":
x86: mm: ptdump: calculate effective permissions correctly
mm: ptdump: expand type of 'val' in note_page()
Huang Ying <ying.huang@intel.com>:
/proc/PID/smaps: Add PMD migration entry parsing
chenqiwu <chenqiwu@xiaomi.com>:
mm/memory: remove unnecessary pte_devmap case in copy_one_pte()
Subsystem: mm/memory-failure
Wetp Zhang <wetp.zy@linux.alibaba.com>:
mm, memory_failure: don't send BUS_MCEERR_AO for action required error
Subsystem: mm/vmalloc
Christoph Hellwig <hch@lst.de>:
Patch series "decruft the vmalloc API", v2:
x86/hyperv: use vmalloc_exec for the hypercall page
x86: fix vmap arguments in map_irq_stack
staging: android: ion: use vmap instead of vm_map_ram
staging: media: ipu3: use vmap instead of reimplementing it
dma-mapping: use vmap insted of reimplementing it
powerpc: add an ioremap_phb helper
powerpc: remove __ioremap_at and __iounmap_at
mm: remove __get_vm_area
mm: unexport unmap_kernel_range_noflush
mm: rename CONFIG_PGTABLE_MAPPING to CONFIG_ZSMALLOC_PGTABLE_MAPPING
mm: only allow page table mappings for built-in zsmalloc
mm: pass addr as unsigned long to vb_free
mm: remove vmap_page_range_noflush and vunmap_page_range
mm: rename vmap_page_range to map_kernel_range
mm: don't return the number of pages from map_kernel_range{,_noflush}
mm: remove map_vm_range
mm: remove unmap_vmap_area
mm: remove the prot argument from vm_map_ram
mm: enforce that vmap can't map pages executable
gpu/drm: remove the powerpc hack in drm_legacy_sg_alloc
mm: remove the pgprot argument to __vmalloc
mm: remove the prot argument to __vmalloc_node
mm: remove both instances of __vmalloc_node_flags
mm: remove __vmalloc_node_flags_caller
mm: switch the test_vmalloc module to use __vmalloc_node
mm: remove vmalloc_user_node_flags
arm64: use __vmalloc_node in arch_alloc_vmap_stack
powerpc: use __vmalloc_node in alloc_vm_stack
s390: use __vmalloc_node in stack_alloc
Joerg Roedel <jroedel@suse.de>:
Patch series "mm: Get rid of vmalloc_sync_(un)mappings()", v3:
mm: add functions to track page directory modifications
mm/vmalloc: track which page-table levels were modified
mm/ioremap: track which page-table levels were modified
x86/mm/64: implement arch_sync_kernel_mappings()
x86/mm/32: implement arch_sync_kernel_mappings()
mm: remove vmalloc_sync_(un)mappings()
x86/mm: remove vmalloc faulting
Subsystem: mm/kasan
Andrey Konovalov <andreyknvl@google.com>:
kasan: fix clang compilation warning due to stack protector
Kees Cook <keescook@chromium.org>:
ubsan: entirely disable alignment checks under UBSAN_TRAP
Jing Xia <jing.xia@unisoc.com>:
mm/mm_init.c: report kasan-tag information stored in page->flags
Andrey Konovalov <andreyknvl@google.com>:
kasan: move kasan_report() into report.c
Documentation/admin-guide/cgroup-v2.rst | 24 +
Documentation/core-api/cachetlb.rst | 2
Documentation/filesystems/locking.rst | 6
Documentation/filesystems/proc.rst | 4
Documentation/filesystems/vfs.rst | 15
Documentation/vm/slub.rst | 2
arch/arm/configs/omap2plus_defconfig | 2
arch/arm64/include/asm/pgtable.h | 3
arch/arm64/include/asm/vmap_stack.h | 6
arch/arm64/mm/dump.c | 2
arch/parisc/include/asm/pgtable.h | 2
arch/powerpc/include/asm/io.h | 10
arch/powerpc/include/asm/pci-bridge.h | 2
arch/powerpc/kernel/irq.c | 5
arch/powerpc/kernel/isa-bridge.c | 28 +
arch/powerpc/kernel/pci_64.c | 56 +-
arch/powerpc/mm/ioremap_64.c | 50 --
arch/riscv/include/asm/pgtable.h | 4
arch/riscv/mm/ptdump.c | 2
arch/s390/kernel/setup.c | 9
arch/sh/kernel/cpu/sh4/sq.c | 3
arch/x86/hyperv/hv_init.c | 5
arch/x86/include/asm/kvm_host.h | 3
arch/x86/include/asm/pgtable-2level_types.h | 2
arch/x86/include/asm/pgtable-3level_types.h | 2
arch/x86/include/asm/pgtable_64_types.h | 2
arch/x86/include/asm/pgtable_types.h | 8
arch/x86/include/asm/switch_to.h | 23 -
arch/x86/kernel/irq_64.c | 2
arch/x86/kernel/setup_percpu.c | 6
arch/x86/kvm/svm/sev.c | 3
arch/x86/mm/dump_pagetables.c | 35 +
arch/x86/mm/fault.c | 196 ----------
arch/x86/mm/init_64.c | 5
arch/x86/mm/pti.c | 8
arch/x86/mm/tlb.c | 37 -
block/blk-core.c | 1
drivers/acpi/apei/ghes.c | 6
drivers/base/node.c | 2
drivers/block/drbd/drbd_bitmap.c | 4
drivers/block/loop.c | 2
drivers/dax/device.c | 1
drivers/gpu/drm/drm_scatter.c | 11
drivers/gpu/drm/etnaviv/etnaviv_dump.c | 4
drivers/gpu/drm/i915/gem/selftests/mock_dmabuf.c | 2
drivers/lightnvm/pblk-init.c | 5
drivers/md/dm-bufio.c | 4
drivers/md/md-bitmap.c | 12
drivers/media/common/videobuf2/videobuf2-dma-sg.c | 3
drivers/media/common/videobuf2/videobuf2-vmalloc.c | 3
drivers/media/pci/ivtv/ivtv-udma.c | 19 -
drivers/media/pci/ivtv/ivtv-yuv.c | 17
drivers/media/pci/ivtv/ivtvfb.c | 4
drivers/mtd/ubi/io.c | 4
drivers/pcmcia/electra_cf.c | 45 --
drivers/scsi/sd_zbc.c | 3
drivers/staging/android/ion/ion_heap.c | 4
drivers/staging/media/ipu3/ipu3-css-pool.h | 4
drivers/staging/media/ipu3/ipu3-dmamap.c | 30 -
fs/block_dev.c | 7
fs/btrfs/disk-io.c | 4
fs/btrfs/extent_io.c | 64 ---
fs/btrfs/extent_io.h | 3
fs/btrfs/inode.c | 39 --
fs/buffer.c | 23 -
fs/erofs/data.c | 41 --
fs/erofs/decompressor.c | 2
fs/erofs/zdata.c | 31 -
fs/exfat/inode.c | 7
fs/ext2/inode.c | 10
fs/ext4/ext4.h | 5
fs/ext4/inode.c | 25 -
fs/ext4/readpage.c | 25 -
fs/ext4/verity.c | 35 -
fs/f2fs/data.c | 56 +-
fs/f2fs/f2fs.h | 14
fs/f2fs/verity.c | 35 -
fs/fat/inode.c | 7
fs/file_table.c | 1
fs/fs-writeback.c | 1
fs/fuse/file.c | 100 +----
fs/gfs2/aops.c | 23 -
fs/gfs2/dir.c | 9
fs/gfs2/quota.c | 2
fs/hpfs/file.c | 7
fs/iomap/buffered-io.c | 113 +----
fs/iomap/trace.h | 2
fs/isofs/inode.c | 7
fs/jfs/inode.c | 7
fs/mpage.c | 38 --
fs/nfs/blocklayout/extent_tree.c | 2
fs/nfs/internal.h | 10
fs/nfs/write.c | 4
fs/nfsd/vfs.c | 9
fs/nilfs2/inode.c | 15
fs/ntfs/aops.c | 2
fs/ntfs/malloc.h | 2
fs/ntfs/mft.c | 2
fs/ocfs2/aops.c | 34 -
fs/ocfs2/dlm/dlmmaster.c | 1
fs/ocfs2/ocfs2.h | 4
fs/ocfs2/slot_map.c | 46 +-
fs/ocfs2/super.c | 21 +
fs/omfs/file.c | 7
fs/open.c | 3
fs/orangefs/inode.c | 32 -
fs/proc/meminfo.c | 3
fs/proc/task_mmu.c | 16
fs/qnx6/inode.c | 7
fs/reiserfs/inode.c | 8
fs/squashfs/block.c | 273 +++++++-------
fs/squashfs/decompressor.h | 5
fs/squashfs/decompressor_multi.c | 9
fs/squashfs/decompressor_multi_percpu.c | 17
fs/squashfs/decompressor_single.c | 9
fs/squashfs/lz4_wrapper.c | 17
fs/squashfs/lzo_wrapper.c | 17
fs/squashfs/squashfs.h | 4
fs/squashfs/xz_wrapper.c | 51 +-
fs/squashfs/zlib_wrapper.c | 63 +--
fs/squashfs/zstd_wrapper.c | 62 +--
fs/sync.c | 6
fs/ubifs/debug.c | 2
fs/ubifs/lprops.c | 2
fs/ubifs/lpt_commit.c | 4
fs/ubifs/orphan.c | 2
fs/udf/inode.c | 7
fs/xfs/kmem.c | 2
fs/xfs/xfs_aops.c | 13
fs/xfs/xfs_buf.c | 2
fs/zonefs/super.c | 7
include/asm-generic/5level-fixup.h | 5
include/asm-generic/pgtable.h | 27 +
include/linux/buffer_head.h | 8
include/linux/fs.h | 18
include/linux/iomap.h | 3
include/linux/memcontrol.h | 4
include/linux/mm.h | 67 ++-
include/linux/mm_types.h | 6
include/linux/mmzone.h | 1
include/linux/mpage.h | 4
include/linux/page_counter.h | 8
include/linux/pagemap.h | 193 ++++++++++
include/linux/ptdump.h | 3
include/linux/sched.h | 3
include/linux/swap.h | 17
include/linux/vmalloc.h | 49 +-
include/linux/zsmalloc.h | 2
include/trace/events/erofs.h | 6
include/trace/events/f2fs.h | 6
include/trace/events/writeback.h | 5
kernel/bpf/core.c | 6
kernel/bpf/syscall.c | 29 -
kernel/dma/remap.c | 48 --
kernel/groups.c | 2
kernel/module.c | 3
kernel/notifier.c | 1
kernel/sys.c | 2
kernel/trace/trace.c | 12
lib/Kconfig.ubsan | 2
lib/ioremap.c | 46 +-
lib/test_vmalloc.c | 26 -
mm/Kconfig | 4
mm/debug.c | 56 ++
mm/fadvise.c | 6
mm/filemap.c | 1
mm/gup.c | 77 +++-
mm/internal.h | 14
mm/kasan/Makefile | 21 -
mm/kasan/common.c | 19 -
mm/kasan/report.c | 22 +
mm/memcontrol.c | 198 +++++++---
mm/memory-failure.c | 15
mm/memory.c | 2
mm/migrate.c | 9
mm/mm_init.c | 16
mm/nommu.c | 52 +-
mm/page-writeback.c | 62 ++-
mm/page_alloc.c | 7
mm/percpu.c | 2
mm/ptdump.c | 17
mm/readahead.c | 349 ++++++++++--------
mm/slab_common.c | 3
mm/slub.c | 67 ++-
mm/swap_state.c | 5
mm/swapfile.c | 194 ++++++----
mm/util.c | 2
mm/vmalloc.c | 399 ++++++++-------------
mm/vmscan.c | 4
mm/vmstat.c | 11
mm/zsmalloc.c | 12
net/bridge/netfilter/ebtables.c | 6
net/ceph/ceph_common.c | 3
sound/core/memalloc.c | 2
sound/core/pcm_memory.c | 2
195 files changed, 2292 insertions(+), 2288 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2020-06-02 4:44 incoming Andrew Morton
@ 2020-06-02 20:08 ` Andrew Morton
2020-06-02 20:45 ` incoming Linus Torvalds
0 siblings, 1 reply; 409+ messages in thread
From: Andrew Morton @ 2020-06-02 20:08 UTC (permalink / raw)
To: Linus Torvalds, mm-commits, linux-mm
The local_lock merge made rather a mess of all of this. I'm
cooking up a full resend of the same material.
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-06-02 4:44 Andrew Morton
2020-06-02 20:08 ` incoming Andrew Morton
0 siblings, 1 reply; 409+ messages in thread
From: Andrew Morton @ 2020-06-02 4:44 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
A few little subsystems and a start of a lot of MM patches.
128 patches, based on 9bf9511e3d9f328c03f6f79bfb741c3d18f2f2c0:
Subsystems affected by this patch series:
squashfs
ocfs2
parisc
vfs
mm/slab-generic
mm/slub
mm/debug
mm/pagecache
mm/gup
mm/swap
mm/memcg
mm/pagemap
mm/memory-failure
mm/vmalloc
mm/kasan
Subsystem: squashfs
Philippe Liard <pliard@google.com>:
squashfs: migrate from ll_rw_block usage to BIO
Subsystem: ocfs2
Jules Irenge <jbi.octave@gmail.com>:
ocfs2: add missing annotation for dlm_empty_lockres()
Gang He <ghe@suse.com>:
ocfs2: mount shared volume without ha stack
Subsystem: parisc
Andrew Morton <akpm@linux-foundation.org>:
arch/parisc/include/asm/pgtable.h: remove unused `old_pte'
Subsystem: vfs
Jeff Layton <jlayton@redhat.com>:
Patch series "vfs: have syncfs() return error when there are writeback:
vfs: track per-sb writeback errors and report them to syncfs
fs/buffer.c: record blockdev write errors in super_block that it backs
Subsystem: mm/slab-generic
Vlastimil Babka <vbabka@suse.cz>:
usercopy: mark dma-kmalloc caches as usercopy caches
Subsystem: mm/slub
Dongli Zhang <dongli.zhang@oracle.com>:
mm/slub.c: fix corrupted freechain in deactivate_slab()
Christoph Lameter <cl@linux.com>:
slub: Remove userspace notifier for cache add/remove
Christopher Lameter <cl@linux.com>:
slub: remove kmalloc under list_lock from list_slab_objects() V2
Qian Cai <cai@lca.pw>:
mm/slub: fix stack overruns with SLUB_STATS
Andrew Morton <akpm@linux-foundation.org>:
Documentation/vm/slub.rst: s/Toggle/Enable/
Subsystem: mm/debug
Vlastimil Babka <vbabka@suse.cz>:
mm, dump_page(): do not crash with invalid mapping pointer
Subsystem: mm/pagecache
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
Patch series "Change readahead API", v11:
mm: move readahead prototypes from mm.h
mm: return void from various readahead functions
mm: ignore return value of ->readpages
mm: move readahead nr_pages check into read_pages
mm: add new readahead_control API
mm: use readahead_control to pass arguments
mm: rename various 'offset' parameters to 'index'
mm: rename readahead loop variable to 'i'
mm: remove 'page_offset' from readahead loop
mm: put readahead pages in cache earlier
mm: add readahead address space operation
mm: move end_index check out of readahead loop
mm: add page_cache_readahead_unbounded
mm: document why we don't set PageReadahead
mm: use memalloc_nofs_save in readahead path
fs: convert mpage_readpages to mpage_readahead
btrfs: convert from readpages to readahead
erofs: convert uncompressed files from readpages to readahead
erofs: convert compressed files from readpages to readahead
ext4: convert from readpages to readahead
ext4: pass the inode to ext4_mpage_readpages
f2fs: convert from readpages to readahead
f2fs: pass the inode to f2fs_mpage_readpages
fuse: convert from readpages to readahead
iomap: convert from readpages to readahead
Guoqing Jiang <guoqing.jiang@cloud.ionos.com>:
Patch series "Introduce attach/detach_page_private to cleanup code":
include/linux/pagemap.h: introduce attach/detach_page_private
md: remove __clear_page_buffers and use attach/detach_page_private
btrfs: use attach/detach_page_private
fs/buffer.c: use attach/detach_page_private
f2fs: use attach/detach_page_private
iomap: use attach/detach_page_private
ntfs: replace attach_page_buffers with attach_page_private
orangefs: use attach/detach_page_private
buffer_head.h: remove attach_page_buffers
mm/migrate.c: call detach_page_private to cleanup code
mm_types.h: change set_page_private to inline function
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
mm/filemap.c: remove misleading comment
Chao Yu <yuchao0@huawei.com>:
mm/page-writeback.c: remove unused variable
NeilBrown <neilb@suse.de>:
mm/writeback: replace PF_LESS_THROTTLE with PF_LOCAL_THROTTLE
mm/writeback: discard NR_UNSTABLE_NFS, use NR_WRITEBACK instead
Subsystem: mm/gup
Souptick Joarder <jrdr.linux@gmail.com>:
mm/gup.c: update the documentation
John Hubbard <jhubbard@nvidia.com>:
mm/gup: introduce pin_user_pages_unlocked
ivtv: convert get_user_pages() --> pin_user_pages()
Miles Chen <miles.chen@mediatek.com>:
mm/gup.c: further document vma_permits_fault()
Subsystem: mm/swap
chenqiwu <chenqiwu@xiaomi.com>:
mm/swapfile: use list_{prev,next}_entry() instead of open-coding
Qian Cai <cai@lca.pw>:
mm/swap_state: fix a data race in swapin_nr_pages
Andrea Righi <andrea.righi@canonical.com>:
mm: swap: properly update readahead statistics in unuse_pte_range()
Wei Yang <richard.weiyang@gmail.com>:
mm/swapfile.c: offset is only used when there is more slots
mm/swapfile.c: explicitly show ssd/non-ssd is handled mutually exclusive
mm/swapfile.c: remove the unnecessary goto for SSD case
mm/swapfile.c: simplify the calculation of n_goal
mm/swapfile.c: remove the extra check in scan_swap_map_slots()
mm/swapfile.c: found_free could be represented by (tmp < max)
mm/swapfile.c: tmp is always smaller than max
mm/swapfile.c: omit a duplicate code by compare tmp and max first
Huang Ying <ying.huang@intel.com>:
swap: try to scan more free slots even when fragmented
Wei Yang <richard.weiyang@gmail.com>:
mm/swapfile.c: classify SWAP_MAP_XXX to make it more readable
mm/swapfile.c: __swap_entry_free() always free 1 entry
Huang Ying <ying.huang@intel.com>:
mm/swapfile.c: use prandom_u32_max()
swap: reduce lock contention on swap cache from swap slots allocation
Randy Dunlap <rdunlap@infradead.org>:
mm: swapfile: fix /proc/swaps heading and Size/Used/Priority alignment
Miaohe Lin <linmiaohe@huawei.com>:
include/linux/swap.h: delete meaningless __add_to_swap_cache() declaration
Subsystem: mm/memcg
Yafang Shao <laoar.shao@gmail.com>:
mm, memcg: add workingset_restore in memory.stat
Kaixu Xia <kaixuxia@tencent.com>:
mm: memcontrol: simplify value comparison between count and limit
Shakeel Butt <shakeelb@google.com>:
memcg: expose root cgroup's memory.stat
Jakub Kicinski <kuba@kernel.org>:
Patch series "memcg: Slow down swap allocation as the available space gets:
mm/memcg: prepare for swap over-high accounting and penalty calculation
mm/memcg: move penalty delay clamping out of calculate_high_delay()
mm/memcg: move cgroup high memory limit setting into struct page_counter
mm/memcg: automatically penalize tasks with high swap use
Zefan Li <lizefan@huawei.com>:
memcg: fix memcg_kmem_bypass() for remote memcg charging
Subsystem: mm/pagemap
Steven Price <steven.price@arm.com>:
Patch series "Fix W+X debug feature on x86":
x86: mm: ptdump: calculate effective permissions correctly
mm: ptdump: expand type of 'val' in note_page()
Huang Ying <ying.huang@intel.com>:
/proc/PID/smaps: Add PMD migration entry parsing
chenqiwu <chenqiwu@xiaomi.com>:
mm/memory: remove unnecessary pte_devmap case in copy_one_pte()
Subsystem: mm/memory-failure
Wetp Zhang <wetp.zy@linux.alibaba.com>:
mm, memory_failure: don't send BUS_MCEERR_AO for action required error
Subsystem: mm/vmalloc
Christoph Hellwig <hch@lst.de>:
Patch series "decruft the vmalloc API", v2:
x86/hyperv: use vmalloc_exec for the hypercall page
x86: fix vmap arguments in map_irq_stack
staging: android: ion: use vmap instead of vm_map_ram
staging: media: ipu3: use vmap instead of reimplementing it
dma-mapping: use vmap insted of reimplementing it
powerpc: add an ioremap_phb helper
powerpc: remove __ioremap_at and __iounmap_at
mm: remove __get_vm_area
mm: unexport unmap_kernel_range_noflush
mm: rename CONFIG_PGTABLE_MAPPING to CONFIG_ZSMALLOC_PGTABLE_MAPPING
mm: only allow page table mappings for built-in zsmalloc
mm: pass addr as unsigned long to vb_free
mm: remove vmap_page_range_noflush and vunmap_page_range
mm: rename vmap_page_range to map_kernel_range
mm: don't return the number of pages from map_kernel_range{,_noflush}
mm: remove map_vm_range
mm: remove unmap_vmap_area
mm: remove the prot argument from vm_map_ram
mm: enforce that vmap can't map pages executable
gpu/drm: remove the powerpc hack in drm_legacy_sg_alloc
mm: remove the pgprot argument to __vmalloc
mm: remove the prot argument to __vmalloc_node
mm: remove both instances of __vmalloc_node_flags
mm: remove __vmalloc_node_flags_caller
mm: switch the test_vmalloc module to use __vmalloc_node
mm: remove vmalloc_user_node_flags
arm64: use __vmalloc_node in arch_alloc_vmap_stack
powerpc: use __vmalloc_node in alloc_vm_stack
s390: use __vmalloc_node in stack_alloc
Joerg Roedel <jroedel@suse.de>:
Patch series "mm: Get rid of vmalloc_sync_(un)mappings()", v3:
mm: add functions to track page directory modifications
mm/vmalloc: track which page-table levels were modified
mm/ioremap: track which page-table levels were modified
x86/mm/64: implement arch_sync_kernel_mappings()
x86/mm/32: implement arch_sync_kernel_mappings()
mm: remove vmalloc_sync_(un)mappings()
x86/mm: remove vmalloc faulting
Subsystem: mm/kasan
Andrey Konovalov <andreyknvl@google.com>:
kasan: fix clang compilation warning due to stack protector
Kees Cook <keescook@chromium.org>:
ubsan: entirely disable alignment checks under UBSAN_TRAP
Jing Xia <jing.xia@unisoc.com>:
mm/mm_init.c: report kasan-tag information stored in page->flags
Andrey Konovalov <andreyknvl@google.com>:
kasan: move kasan_report() into report.c
Documentation/admin-guide/cgroup-v2.rst | 24 +
Documentation/core-api/cachetlb.rst | 2
Documentation/filesystems/locking.rst | 6
Documentation/filesystems/proc.rst | 4
Documentation/filesystems/vfs.rst | 15
Documentation/vm/slub.rst | 2
arch/arm/configs/omap2plus_defconfig | 2
arch/arm64/include/asm/pgtable.h | 3
arch/arm64/include/asm/vmap_stack.h | 6
arch/arm64/mm/dump.c | 2
arch/parisc/include/asm/pgtable.h | 2
arch/powerpc/include/asm/io.h | 10
arch/powerpc/include/asm/pci-bridge.h | 2
arch/powerpc/kernel/irq.c | 5
arch/powerpc/kernel/isa-bridge.c | 28 +
arch/powerpc/kernel/pci_64.c | 56 +-
arch/powerpc/mm/ioremap_64.c | 50 --
arch/riscv/include/asm/pgtable.h | 4
arch/riscv/mm/ptdump.c | 2
arch/s390/kernel/setup.c | 9
arch/sh/kernel/cpu/sh4/sq.c | 3
arch/x86/hyperv/hv_init.c | 5
arch/x86/include/asm/kvm_host.h | 3
arch/x86/include/asm/pgtable-2level_types.h | 2
arch/x86/include/asm/pgtable-3level_types.h | 2
arch/x86/include/asm/pgtable_64_types.h | 2
arch/x86/include/asm/pgtable_types.h | 8
arch/x86/include/asm/switch_to.h | 23 -
arch/x86/kernel/irq_64.c | 2
arch/x86/kernel/setup_percpu.c | 6
arch/x86/kvm/svm/sev.c | 3
arch/x86/mm/dump_pagetables.c | 35 +
arch/x86/mm/fault.c | 196 ----------
arch/x86/mm/init_64.c | 5
arch/x86/mm/pti.c | 8
arch/x86/mm/tlb.c | 37 -
block/blk-core.c | 1
drivers/acpi/apei/ghes.c | 6
drivers/base/node.c | 2
drivers/block/drbd/drbd_bitmap.c | 4
drivers/block/loop.c | 2
drivers/dax/device.c | 1
drivers/gpu/drm/drm_scatter.c | 11
drivers/gpu/drm/etnaviv/etnaviv_dump.c | 4
drivers/gpu/drm/i915/gem/selftests/mock_dmabuf.c | 2
drivers/lightnvm/pblk-init.c | 5
drivers/md/dm-bufio.c | 4
drivers/md/md-bitmap.c | 12
drivers/media/common/videobuf2/videobuf2-dma-sg.c | 3
drivers/media/common/videobuf2/videobuf2-vmalloc.c | 3
drivers/media/pci/ivtv/ivtv-udma.c | 19 -
drivers/media/pci/ivtv/ivtv-yuv.c | 17
drivers/media/pci/ivtv/ivtvfb.c | 4
drivers/mtd/ubi/io.c | 4
drivers/pcmcia/electra_cf.c | 45 --
drivers/scsi/sd_zbc.c | 3
drivers/staging/android/ion/ion_heap.c | 4
drivers/staging/media/ipu3/ipu3-css-pool.h | 4
drivers/staging/media/ipu3/ipu3-dmamap.c | 30 -
fs/block_dev.c | 7
fs/btrfs/disk-io.c | 4
fs/btrfs/extent_io.c | 64 ---
fs/btrfs/extent_io.h | 3
fs/btrfs/inode.c | 39 --
fs/buffer.c | 23 -
fs/erofs/data.c | 41 --
fs/erofs/decompressor.c | 2
fs/erofs/zdata.c | 31 -
fs/exfat/inode.c | 7
fs/ext2/inode.c | 10
fs/ext4/ext4.h | 5
fs/ext4/inode.c | 25 -
fs/ext4/readpage.c | 25 -
fs/ext4/verity.c | 35 -
fs/f2fs/data.c | 56 +-
fs/f2fs/f2fs.h | 14
fs/f2fs/verity.c | 35 -
fs/fat/inode.c | 7
fs/file_table.c | 1
fs/fs-writeback.c | 1
fs/fuse/file.c | 100 +----
fs/gfs2/aops.c | 23 -
fs/gfs2/dir.c | 9
fs/gfs2/quota.c | 2
fs/hpfs/file.c | 7
fs/iomap/buffered-io.c | 113 +----
fs/iomap/trace.h | 2
fs/isofs/inode.c | 7
fs/jfs/inode.c | 7
fs/mpage.c | 38 --
fs/nfs/blocklayout/extent_tree.c | 2
fs/nfs/internal.h | 10
fs/nfs/write.c | 4
fs/nfsd/vfs.c | 9
fs/nilfs2/inode.c | 15
fs/ntfs/aops.c | 2
fs/ntfs/malloc.h | 2
fs/ntfs/mft.c | 2
fs/ocfs2/aops.c | 34 -
fs/ocfs2/dlm/dlmmaster.c | 1
fs/ocfs2/ocfs2.h | 4
fs/ocfs2/slot_map.c | 46 +-
fs/ocfs2/super.c | 21 +
fs/omfs/file.c | 7
fs/open.c | 3
fs/orangefs/inode.c | 32 -
fs/proc/meminfo.c | 3
fs/proc/task_mmu.c | 16
fs/qnx6/inode.c | 7
fs/reiserfs/inode.c | 8
fs/squashfs/block.c | 273 +++++++-------
fs/squashfs/decompressor.h | 5
fs/squashfs/decompressor_multi.c | 9
fs/squashfs/decompressor_multi_percpu.c | 17
fs/squashfs/decompressor_single.c | 9
fs/squashfs/lz4_wrapper.c | 17
fs/squashfs/lzo_wrapper.c | 17
fs/squashfs/squashfs.h | 4
fs/squashfs/xz_wrapper.c | 51 +-
fs/squashfs/zlib_wrapper.c | 63 +--
fs/squashfs/zstd_wrapper.c | 62 +--
fs/sync.c | 6
fs/ubifs/debug.c | 2
fs/ubifs/lprops.c | 2
fs/ubifs/lpt_commit.c | 4
fs/ubifs/orphan.c | 2
fs/udf/inode.c | 7
fs/xfs/kmem.c | 2
fs/xfs/xfs_aops.c | 13
fs/xfs/xfs_buf.c | 2
fs/zonefs/super.c | 7
include/asm-generic/5level-fixup.h | 5
include/asm-generic/pgtable.h | 27 +
include/linux/buffer_head.h | 8
include/linux/fs.h | 18
include/linux/iomap.h | 3
include/linux/memcontrol.h | 4
include/linux/mm.h | 67 ++-
include/linux/mm_types.h | 6
include/linux/mmzone.h | 1
include/linux/mpage.h | 4
include/linux/page_counter.h | 8
include/linux/pagemap.h | 193 ++++++++++
include/linux/ptdump.h | 3
include/linux/sched.h | 3
include/linux/swap.h | 17
include/linux/vmalloc.h | 49 +-
include/linux/zsmalloc.h | 2
include/trace/events/erofs.h | 6
include/trace/events/f2fs.h | 6
include/trace/events/writeback.h | 5
kernel/bpf/core.c | 6
kernel/bpf/syscall.c | 29 -
kernel/dma/remap.c | 48 --
kernel/groups.c | 2
kernel/module.c | 3
kernel/notifier.c | 1
kernel/sys.c | 2
kernel/trace/trace.c | 12
lib/Kconfig.ubsan | 2
lib/ioremap.c | 46 +-
lib/test_vmalloc.c | 26 -
mm/Kconfig | 4
mm/debug.c | 56 ++
mm/fadvise.c | 6
mm/filemap.c | 1
mm/gup.c | 77 +++-
mm/internal.h | 14
mm/kasan/Makefile | 21 -
mm/kasan/common.c | 19 -
mm/kasan/report.c | 22 +
mm/memcontrol.c | 198 +++++++---
mm/memory-failure.c | 15
mm/memory.c | 2
mm/migrate.c | 9
mm/mm_init.c | 16
mm/nommu.c | 52 +-
mm/page-writeback.c | 62 ++-
mm/page_alloc.c | 7
mm/percpu.c | 2
mm/ptdump.c | 17
mm/readahead.c | 349 ++++++++++--------
mm/slab_common.c | 3
mm/slub.c | 67 ++-
mm/swap_state.c | 5
mm/swapfile.c | 194 ++++++----
mm/util.c | 2
mm/vmalloc.c | 399 ++++++++-------------
mm/vmscan.c | 4
mm/vmstat.c | 11
mm/zsmalloc.c | 12
net/bridge/netfilter/ebtables.c | 6
net/ceph/ceph_common.c | 3
sound/core/memalloc.c | 2
sound/core/pcm_memory.c | 2
195 files changed, 2292 insertions(+), 2288 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2020-05-29 21:12 ` incoming Andrew Morton
@ 2020-05-29 21:20 ` Linus Torvalds
0 siblings, 0 replies; 409+ messages in thread
From: Linus Torvalds @ 2020-05-29 21:20 UTC (permalink / raw)
To: Andrew Morton; +Cc: mm-commits, Linux-MM
On Fri, May 29, 2020 at 2:12 PM Andrew Morton <akpm@linux-foundation.org> wrote:
>
> Stupid diffstat. Means that basically all my diffstats are very wrong.
I'm actually used to diffstats not matching 100%/
Usually it's not due to this issue - a "git diff --stat" *will* give
the stat from the actual combined diff result - but with git diffstats
the issue is that I might have gotten a patch from another source.
So the diffstat I see after-the-merge is possibly different from the
pre-merge diffstat simply due to merge issues.
So then I usually take a look at "ok, why did that diffstat differ"
and go "Ahh".
In your case, when I looked at the diffstat, I couldn't for the life
of me see how you would have gotten the diffstat you did, since I only
saw a single patch with no merge issues.
> Thanks for spotting it.
>
> I can fix that...
I can also just live with it, knowing what your workflow is. The
diffstat matching exactly just isn't that important - in fact,
different versions of "diff" can give slightly different output anyway
depending on diff algorithms even when they are looking at the exact
same before/after state. There's not necessarily always only one way
to generate a valid diff.
So to me, the diffstat is more of a guide than a hard thing, and I
want to see the rough outline,
In fact, one reason I want to see it in pull requests is actually just
that I want to get a feel for what changes even before I do the pull
or merge, so it's not just a "match against what I get" thing.
Linus
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2020-05-29 20:38 ` incoming Linus Torvalds
@ 2020-05-29 21:12 ` Andrew Morton
2020-05-29 21:20 ` incoming Linus Torvalds
0 siblings, 1 reply; 409+ messages in thread
From: Andrew Morton @ 2020-05-29 21:12 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, Linux-MM
On Fri, 29 May 2020 13:38:35 -0700 Linus Torvalds <torvalds@linux-foundation.org> wrote:
> On Fri, May 29, 2020 at 1:31 PM Andrew Morton <akpm@linux-foundation.org> wrote:
> >
> > Bah. I got lazy (didn't want to interrupt an ongoing build) so I
> > generated the diffstat prior to folding two patches into a single one.
> > Evidently diffstat isn't as smart as I had assumed!
>
> Ahh. Yes - given two patches, diffstat just adds up the line number
> counts for the individual diffs, it doesn't count some kind of
> "combined diff result" line counts.
Stupid diffstat. Means that basically all my diffstats are very wrong.
Thanks for spotting it.
I can fix that...
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2020-05-29 20:31 ` incoming Andrew Morton
@ 2020-05-29 20:38 ` Linus Torvalds
2020-05-29 21:12 ` incoming Andrew Morton
0 siblings, 1 reply; 409+ messages in thread
From: Linus Torvalds @ 2020-05-29 20:38 UTC (permalink / raw)
To: Andrew Morton; +Cc: mm-commits, Linux-MM
On Fri, May 29, 2020 at 1:31 PM Andrew Morton <akpm@linux-foundation.org> wrote:
>
> Bah. I got lazy (didn't want to interrupt an ongoing build) so I
> generated the diffstat prior to folding two patches into a single one.
> Evidently diffstat isn't as smart as I had assumed!
Ahh. Yes - given two patches, diffstat just adds up the line number
counts for the individual diffs, it doesn't count some kind of
"combined diff result" line counts.
Linus
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2020-05-28 20:10 ` incoming Linus Torvalds
@ 2020-05-29 20:31 ` Andrew Morton
2020-05-29 20:38 ` incoming Linus Torvalds
0 siblings, 1 reply; 409+ messages in thread
From: Andrew Morton @ 2020-05-29 20:31 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, Linux-MM
On Thu, 28 May 2020 13:10:18 -0700 Linus Torvalds <torvalds@linux-foundation.org> wrote:
> Hmm..
>
> On Wed, May 27, 2020 at 10:20 PM Andrew Morton
> <akpm@linux-foundation.org> wrote:
> >
> > fs/binfmt_elf.c | 2 +-
> > include/asm-generic/topology.h | 2 +-
> > include/linux/mm.h | 19 +++++++++++++++----
> > mm/khugepaged.c | 1 +
> > mm/z3fold.c | 3 +++
> > 5 files changed, 21 insertions(+), 6 deletions(-)
>
> I wonder how you generate that diffstat.
>
> The change to <linux/mm.h> simply doesn't match what you sent me. The
> patch you sent me that changed mm.h had this:
>
> include/linux/mm.h | 15 +++++++++++++--
> 1 file changed, 13 insertions(+), 2 deletions(-)
>
> (note 15 lines changed: it's +13 and -2) but now suddenly in your
> overall diffstat you have that
>
> include/linux/mm.h | 19 +++++++++++++++----
>
> with +15/-4.
>
> So your diffstat simply doesn't match what you are sending. What's going on?
>
Bah. I got lazy (didn't want to interrupt an ongoing build) so I
generated the diffstat prior to folding two patches into a single one.
Evidently diffstat isn't as smart as I had assumed!
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2020-05-28 5:20 incoming Andrew Morton
@ 2020-05-28 20:10 ` Linus Torvalds
2020-05-29 20:31 ` incoming Andrew Morton
0 siblings, 1 reply; 409+ messages in thread
From: Linus Torvalds @ 2020-05-28 20:10 UTC (permalink / raw)
To: Andrew Morton; +Cc: mm-commits, Linux-MM
Hmm..
On Wed, May 27, 2020 at 10:20 PM Andrew Morton
<akpm@linux-foundation.org> wrote:
>
> fs/binfmt_elf.c | 2 +-
> include/asm-generic/topology.h | 2 +-
> include/linux/mm.h | 19 +++++++++++++++----
> mm/khugepaged.c | 1 +
> mm/z3fold.c | 3 +++
> 5 files changed, 21 insertions(+), 6 deletions(-)
I wonder how you generate that diffstat.
The change to <linux/mm.h> simply doesn't match what you sent me. The
patch you sent me that changed mm.h had this:
include/linux/mm.h | 15 +++++++++++++--
1 file changed, 13 insertions(+), 2 deletions(-)
(note 15 lines changed: it's +13 and -2) but now suddenly in your
overall diffstat you have that
include/linux/mm.h | 19 +++++++++++++++----
with +15/-4.
So your diffstat simply doesn't match what you are sending. What's going on?
Linus
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-05-28 5:20 Andrew Morton
2020-05-28 20:10 ` incoming Linus Torvalds
0 siblings, 1 reply; 409+ messages in thread
From: Andrew Morton @ 2020-05-28 5:20 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
5 fixes, based on 444fc5cde64330661bf59944c43844e7d4c2ccd8:
Qian Cai <cai@lca.pw>:
mm/z3fold: silence kmemleak false positives of slots
Hugh Dickins <hughd@google.com>:
mm,thp: stop leaking unreleased file pages
Konstantin Khlebnikov <khlebnikov@yandex-team.ru>:
mm: remove VM_BUG_ON(PageSlab()) from page_mapcount()
Alexander Potapenko <glider@google.com>:
fs/binfmt_elf.c: allocate initialized memory in fill_thread_core_info()
Arnd Bergmann <arnd@arndb.de>:
include/asm-generic/topology.h: guard cpumask_of_node() macro argument
fs/binfmt_elf.c | 2 +-
include/asm-generic/topology.h | 2 +-
include/linux/mm.h | 19 +++++++++++++++----
mm/khugepaged.c | 1 +
mm/z3fold.c | 3 +++
5 files changed, 21 insertions(+), 6 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-05-23 5:22 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-05-23 5:22 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
11 fixes, based on 444565650a5fe9c63ddf153e6198e31705dedeb2:
David Hildenbrand <david@redhat.com>:
device-dax: don't leak kernel memory to user space after unloading kmem
Nick Desaulniers <ndesaulniers@google.com>:
x86: bitops: fix build regression
John Hubbard <jhubbard@nvidia.com>:
rapidio: fix an error in get_user_pages_fast() error handling
selftests/vm/.gitignore: add mremap_dontunmap
selftests/vm/write_to_hugetlbfs.c: fix unused variable warning
Marco Elver <elver@google.com>:
kasan: disable branch tracing for core runtime
Arnd Bergmann <arnd@arndb.de>:
sh: include linux/time_types.h for sockios
Naoya Horiguchi <n-horiguchi@ah.jp.nec.com>:
MAINTAINERS: update email address for Naoya Horiguchi
Mike Rapoport <rppt@linux.ibm.com>:
sparc32: use PUD rather than PGD to get PMD in srmmu_nocache_init()
Uladzislau Rezki <uladzislau.rezki@sony.com>:
z3fold: fix use-after-free when freeing handles
Baoquan He <bhe@redhat.com>:
MAINTAINERS: add files related to kdump
MAINTAINERS | 7 ++++++-
arch/sh/include/uapi/asm/sockios.h | 2 ++
arch/sparc/mm/srmmu.c | 2 +-
arch/x86/include/asm/bitops.h | 12 ++++++------
drivers/dax/kmem.c | 14 +++++++++++---
drivers/rapidio/devices/rio_mport_cdev.c | 5 +++++
mm/kasan/Makefile | 16 ++++++++--------
mm/kasan/generic.c | 1 -
mm/kasan/tags.c | 1 -
mm/z3fold.c | 11 ++++++-----
tools/testing/selftests/vm/.gitignore | 1 +
tools/testing/selftests/vm/write_to_hugetlbfs.c | 2 --
12 files changed, 46 insertions(+), 28 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-05-14 0:50 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-05-14 0:50 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
7 fixes, based on 24085f70a6e1b0cb647ec92623284641d8270637:
Yafang Shao <laoar.shao@gmail.com>:
mm, memcg: fix inconsistent oom event behavior
Roman Penyaev <rpenyaev@suse.de>:
epoll: call final ep_events_available() check under the lock
Peter Xu <peterx@redhat.com>:
mm/gup: fix fixup_user_fault() on multiple retries
Brian Geffon <bgeffon@google.com>:
userfaultfd: fix remap event with MREMAP_DONTUNMAP
Vasily Averin <vvs@virtuozzo.com>:
ipc/util.c: sysvipc_find_ipc() incorrectly updates position index
Andrey Konovalov <andreyknvl@google.com>:
kasan: consistently disable debugging features
kasan: add missing functions declarations to kasan.h
fs/eventpoll.c | 48 ++++++++++++++++++++++++++-------------------
include/linux/memcontrol.h | 2 +
ipc/util.c | 12 +++++------
mm/gup.c | 12 ++++++-----
mm/kasan/Makefile | 15 +++++++++-----
mm/kasan/kasan.h | 34 ++++++++++++++++++++++++++++++-
mm/mremap.c | 2 -
7 files changed, 86 insertions(+), 39 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-05-08 1:35 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-05-08 1:35 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
14 fixes and one selftest to verify the ipc fixes herein.
15 patches, based on a811c1fa0a02c062555b54651065899437bacdbe:
Oleg Nesterov <oleg@redhat.com>:
ipc/mqueue.c: change __do_notify() to bypass check_kill_permission()
Yafang Shao <laoar.shao@gmail.com>:
mm, memcg: fix error return value of mem_cgroup_css_alloc()
David Hildenbrand <david@redhat.com>:
mm/page_alloc: fix watchdog soft lockups during set_zone_contiguous()
Maciej Grochowski <maciej.grochowski@pm.me>:
kernel/kcov.c: fix typos in kcov_remote_start documentation
Ivan Delalande <colona@arista.com>:
scripts/decodecode: fix trapping instruction formatting
Janakarajan Natarajan <Janakarajan.Natarajan@amd.com>:
arch/x86/kvm/svm/sev.c: change flag passed to GUP fast in sev_pin_memory()
Khazhismel Kumykov <khazhy@google.com>:
eventpoll: fix missing wakeup for ovflist in ep_poll_callback
Aymeric Agon-Rambosson <aymeric.agon@yandex.com>:
scripts/gdb: repair rb_first() and rb_last()
Waiman Long <longman@redhat.com>:
mm/slub: fix incorrect interpretation of s->offset
Filipe Manana <fdmanana@suse.com>:
percpu: make pcpu_alloc() aware of current gfp context
Roman Penyaev <rpenyaev@suse.de>:
kselftests: introduce new epoll60 testcase for catching lost wakeups
epoll: atomically remove wait entry on wake up
Qiwu Chen <qiwuchen55@gmail.com>:
mm/vmscan: remove unnecessary argument description of isolate_lru_pages()
Kees Cook <keescook@chromium.org>:
ubsan: disable UBSAN_ALIGNMENT under COMPILE_TEST
Henry Willard <henry.willard@oracle.com>:
mm: limit boost_watermark on small zones
arch/x86/kvm/svm/sev.c | 2
fs/eventpoll.c | 61 ++--
ipc/mqueue.c | 34 +-
kernel/kcov.c | 4
lib/Kconfig.ubsan | 15 -
mm/memcontrol.c | 15 -
mm/page_alloc.c | 9
mm/percpu.c | 14
mm/slub.c | 45 ++-
mm/vmscan.c | 1
scripts/decodecode | 2
scripts/gdb/linux/rbtree.py | 4
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c | 146 ++++++++++
tools/testing/selftests/wireguard/qemu/debug.config | 1
14 files changed, 275 insertions(+), 78 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-04-21 1:13 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-04-21 1:13 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
15 fixes, based on ae83d0b416db002fe95601e7f97f64b59514d936:
Masahiro Yamada <masahiroy@kernel.org>:
sh: fix build error in mm/init.c
Kees Cook <keescook@chromium.org>:
slub: avoid redzone when choosing freepointer location
Peter Xu <peterx@redhat.com>:
mm/userfaultfd: disable userfaultfd-wp on x86_32
Bartosz Golaszewski <bgolaszewski@baylibre.com>:
MAINTAINERS: add an entry for kfifo
Longpeng <longpeng2@huawei.com>:
mm/hugetlb: fix a addressing exception caused by huge_pte_offset
Michal Hocko <mhocko@suse.com>:
mm, gup: return EINTR when gup is interrupted by fatal signals
Christophe JAILLET <christophe.jaillet@wanadoo.fr>:
checkpatch: fix a typo in the regex for $allocFunctions
George Burgess IV <gbiv@google.com>:
tools/build: tweak unused value workaround
Muchun Song <songmuchun@bytedance.com>:
mm/ksm: fix NULL pointer dereference when KSM zero page is enabled
Hugh Dickins <hughd@google.com>:
mm/shmem: fix build without THP
Jann Horn <jannh@google.com>:
vmalloc: fix remap_vmalloc_range() bounds checks
Hugh Dickins <hughd@google.com>:
shmem: fix possible deadlocks on shmlock_user_lock
Yang Shi <yang.shi@linux.alibaba.com>:
mm: shmem: disable interrupt when acquiring info->lock in userfaultfd_copy path
Sudip Mukherjee <sudipm.mukherjee@gmail.com>:
coredump: fix null pointer dereference on coredump
Lucas Stach <l.stach@pengutronix.de>:
tools/vm: fix cross-compile build
MAINTAINERS | 7 +++++++
arch/sh/mm/init.c | 2 +-
arch/x86/Kconfig | 2 +-
fs/coredump.c | 2 ++
fs/proc/vmcore.c | 5 +++--
include/linux/vmalloc.h | 2 +-
mm/gup.c | 2 +-
mm/hugetlb.c | 14 ++++++++------
mm/ksm.c | 12 ++++++++++--
mm/shmem.c | 13 ++++++++-----
mm/slub.c | 12 ++++++++++--
mm/vmalloc.c | 16 +++++++++++++---
samples/vfio-mdev/mdpy.c | 2 +-
scripts/checkpatch.pl | 2 +-
tools/build/feature/test-sync-compare-and-swap.c | 2 +-
tools/vm/Makefile | 2 ++
16 files changed, 70 insertions(+), 27 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-04-12 7:41 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-04-12 7:41 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
A straggler. This patch caused a lot of build errors on a lot of
architectures for a long time, but Anshuman believes it's all fixed up
now.
1 patch, based on GIT b032227c62939b5481bcd45442b36dfa263f4a7c.
Anshuman Khandual <anshuman.khandual@arm.com>:
mm/debug: add tests validating architecture page table helpers
Documentation/features/debug/debug-vm-pgtable/arch-support.txt | 34
arch/arc/Kconfig | 1
arch/arm64/Kconfig | 1
arch/powerpc/Kconfig | 1
arch/s390/Kconfig | 1
arch/x86/Kconfig | 1
arch/x86/include/asm/pgtable_64.h | 6
include/linux/mmdebug.h | 5
init/main.c | 2
lib/Kconfig.debug | 26
mm/Makefile | 1
mm/debug_vm_pgtable.c | 392 ++++++++++
12 files changed, 471 insertions(+)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-04-10 21:30 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-04-10 21:30 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
Almost all of the rest of MM. Various other things.
35 patches, based on c0cc271173b2e1c2d8d0ceaef14e4dfa79eefc0d.
Subsystems affected by this patch series:
hfs
mm/memcg
mm/slab-generic
mm/slab
mm/pagealloc
mm/gup
ocfs2
mm/hugetlb
mm/pagemap
mm/memremap
kmod
misc
seqfile
Subsystem: hfs
Simon Gander <simon@tuxera.com>:
hfsplus: fix crash and filesystem corruption when deleting files
Subsystem: mm/memcg
Jakub Kicinski <kuba@kernel.org>:
mm, memcg: do not high throttle allocators based on wraparound
Subsystem: mm/slab-generic
Qiujun Huang <hqjagain@gmail.com>:
mm, slab_common: fix a typo in comment "eariler"->"earlier"
Subsystem: mm/slab
Mauro Carvalho Chehab <mchehab+huawei@kernel.org>:
docs: mm: slab.h: fix a broken cross-reference
Subsystem: mm/pagealloc
Randy Dunlap <rdunlap@infradead.org>:
mm/page_alloc.c: fix kernel-doc warning
Jason Yan <yanaijie@huawei.com>:
mm/page_alloc: make pcpu_drain_mutex and pcpu_drain static
Subsystem: mm/gup
Miles Chen <miles.chen@mediatek.com>:
mm/gup: fix null pointer dereference detected by coverity
Subsystem: ocfs2
Changwei Ge <chge@linux.alibaba.com>:
ocfs2: no need try to truncate file beyond i_size
Subsystem: mm/hugetlb
Aslan Bakirov <aslan@fb.com>:
mm: cma: NUMA node interface
Roman Gushchin <guro@fb.com>:
mm: hugetlb: optionally allocate gigantic hugepages using cma
Subsystem: mm/pagemap
Jaewon Kim <jaewon31.kim@samsung.com>:
mm/mmap.c: initialize align_offset explicitly for vm_unmapped_area
Arjun Roy <arjunroy@google.com>:
mm/memory.c: refactor insert_page to prepare for batched-lock insert
mm: bring sparc pte_index() semantics inline with other platforms
mm: define pte_index as macro for x86
mm/memory.c: add vm_insert_pages()
Anshuman Khandual <anshuman.khandual@arm.com>:
mm/vma: define a default value for VM_DATA_DEFAULT_FLAGS
mm/vma: introduce VM_ACCESS_FLAGS
mm/special: create generic fallbacks for pte_special() and pte_mkspecial()
Subsystem: mm/memremap
Logan Gunthorpe <logang@deltatee.com>:
Patch series "Allow setting caching mode in arch_add_memory() for P2PDMA", v4:
mm/memory_hotplug: drop the flags field from struct mhp_restrictions
mm/memory_hotplug: rename mhp_restrictions to mhp_params
x86/mm: thread pgprot_t through init_memory_mapping()
x86/mm: introduce __set_memory_prot()
powerpc/mm: thread pgprot_t through create_section_mapping()
mm/memory_hotplug: add pgprot_t to mhp_params
mm/memremap: set caching mode for PCI P2PDMA memory to WC
Subsystem: kmod
Eric Biggers <ebiggers@google.com>:
Patch series "module autoloading fixes and cleanups", v5:
kmod: make request_module() return an error when autoloading is disabled
fs/filesystems.c: downgrade user-reachable WARN_ONCE() to pr_warn_once()
docs: admin-guide: document the kernel.modprobe sysctl
selftests: kmod: fix handling test numbers above 9
selftests: kmod: test disabling module autoloading
Subsystem: misc
Pali Rohár <pali@kernel.org>:
change email address for Pali Rohár
kbuild test robot <lkp@intel.com>:
drivers/dma/tegra20-apb-dma.c: fix platform_get_irq.cocci warnings
Subsystem: seqfile
Vasily Averin <vvs@virtuozzo.com>:
Patch series "seq_file .next functions should increase position index":
fs/seq_file.c: seq_read(): add info message about buggy .next functions
kernel/gcov/fs.c: gcov_seq_next() should increase position index
ipc/util.c: sysvipc_find_ipc() should increase position index
Documentation/ABI/testing/sysfs-platform-dell-laptop | 8
Documentation/admin-guide/kernel-parameters.txt | 8
Documentation/admin-guide/sysctl/kernel.rst | 21 ++
MAINTAINERS | 16 -
arch/alpha/include/asm/page.h | 3
arch/alpha/include/asm/pgtable.h | 2
arch/arc/include/asm/page.h | 2
arch/arm/include/asm/page.h | 4
arch/arm/include/asm/pgtable-2level.h | 2
arch/arm/include/asm/pgtable.h | 15 -
arch/arm/mach-omap2/omap-secure.c | 2
arch/arm/mach-omap2/omap-secure.h | 2
arch/arm/mach-omap2/omap-smc.S | 2
arch/arm/mm/fault.c | 2
arch/arm/mm/mmu.c | 14 +
arch/arm64/include/asm/page.h | 4
arch/arm64/mm/fault.c | 2
arch/arm64/mm/init.c | 6
arch/arm64/mm/mmu.c | 7
arch/c6x/include/asm/page.h | 5
arch/csky/include/asm/page.h | 3
arch/csky/include/asm/pgtable.h | 3
arch/h8300/include/asm/page.h | 2
arch/hexagon/include/asm/page.h | 3
arch/hexagon/include/asm/pgtable.h | 2
arch/ia64/include/asm/page.h | 5
arch/ia64/include/asm/pgtable.h | 2
arch/ia64/mm/init.c | 7
arch/m68k/include/asm/mcf_pgtable.h | 10 -
arch/m68k/include/asm/motorola_pgtable.h | 2
arch/m68k/include/asm/page.h | 3
arch/m68k/include/asm/sun3_pgtable.h | 2
arch/microblaze/include/asm/page.h | 2
arch/microblaze/include/asm/pgtable.h | 4
arch/mips/include/asm/page.h | 5
arch/mips/include/asm/pgtable.h | 44 +++-
arch/nds32/include/asm/page.h | 3
arch/nds32/include/asm/pgtable.h | 9 -
arch/nds32/mm/fault.c | 2
arch/nios2/include/asm/page.h | 3
arch/nios2/include/asm/pgtable.h | 3
arch/openrisc/include/asm/page.h | 5
arch/openrisc/include/asm/pgtable.h | 2
arch/parisc/include/asm/page.h | 3
arch/parisc/include/asm/pgtable.h | 2
arch/powerpc/include/asm/book3s/64/hash.h | 3
arch/powerpc/include/asm/book3s/64/radix.h | 3
arch/powerpc/include/asm/page.h | 9 -
arch/powerpc/include/asm/page_64.h | 7
arch/powerpc/include/asm/sparsemem.h | 3
arch/powerpc/mm/book3s64/hash_utils.c | 5
arch/powerpc/mm/book3s64/pgtable.c | 7
arch/powerpc/mm/book3s64/pkeys.c | 2
arch/powerpc/mm/book3s64/radix_pgtable.c | 18 +-
arch/powerpc/mm/mem.c | 12 -
arch/riscv/include/asm/page.h | 3
arch/s390/include/asm/page.h | 3
arch/s390/mm/fault.c | 2
arch/s390/mm/init.c | 9 -
arch/sh/include/asm/page.h | 3
arch/sh/mm/init.c | 7
arch/sparc/include/asm/page_32.h | 3
arch/sparc/include/asm/page_64.h | 3
arch/sparc/include/asm/pgtable_32.h | 7
arch/sparc/include/asm/pgtable_64.h | 10 -
arch/um/include/asm/pgtable.h | 10 -
arch/unicore32/include/asm/page.h | 3
arch/unicore32/include/asm/pgtable.h | 3
arch/unicore32/mm/fault.c | 2
arch/x86/include/asm/page_types.h | 7
arch/x86/include/asm/pgtable.h | 6
arch/x86/include/asm/set_memory.h | 1
arch/x86/kernel/amd_gart_64.c | 3
arch/x86/kernel/setup.c | 4
arch/x86/mm/init.c | 9 -
arch/x86/mm/init_32.c | 19 +-
arch/x86/mm/init_64.c | 42 ++--
arch/x86/mm/mm_internal.h | 3
arch/x86/mm/pat/set_memory.c | 13 +
arch/x86/mm/pkeys.c | 2
arch/x86/platform/uv/bios_uv.c | 3
arch/x86/um/asm/vm-flags.h | 10 -
arch/xtensa/include/asm/page.h | 3
arch/xtensa/include/asm/pgtable.h | 3
drivers/char/hw_random/omap3-rom-rng.c | 4
drivers/dma/tegra20-apb-dma.c | 1
drivers/hwmon/dell-smm-hwmon.c | 4
drivers/platform/x86/dell-laptop.c | 4
drivers/platform/x86/dell-rbtn.c | 4
drivers/platform/x86/dell-rbtn.h | 2
drivers/platform/x86/dell-smbios-base.c | 4
drivers/platform/x86/dell-smbios-smm.c | 2
drivers/platform/x86/dell-smbios.h | 2
drivers/platform/x86/dell-smo8800.c | 2
drivers/platform/x86/dell-wmi.c | 4
drivers/power/supply/bq2415x_charger.c | 4
drivers/power/supply/bq27xxx_battery.c | 2
drivers/power/supply/isp1704_charger.c | 2
drivers/power/supply/rx51_battery.c | 4
drivers/staging/gasket/gasket_core.c | 2
fs/filesystems.c | 4
fs/hfsplus/attributes.c | 4
fs/ocfs2/alloc.c | 4
fs/seq_file.c | 7
fs/udf/ecma_167.h | 2
fs/udf/osta_udf.h | 2
include/linux/cma.h | 14 +
include/linux/hugetlb.h | 12 +
include/linux/memblock.h | 3
include/linux/memory_hotplug.h | 21 +-
include/linux/mm.h | 34 +++
include/linux/power/bq2415x_charger.h | 2
include/linux/slab.h | 2
ipc/util.c | 2
kernel/gcov/fs.c | 2
kernel/kmod.c | 4
mm/cma.c | 16 +
mm/gup.c | 3
mm/hugetlb.c | 109 ++++++++++++
mm/memblock.c | 2
mm/memcontrol.c | 3
mm/memory.c | 168 +++++++++++++++++--
mm/memory_hotplug.c | 13 -
mm/memremap.c | 17 +
mm/mmap.c | 4
mm/mprotect.c | 4
mm/page_alloc.c | 5
mm/slab_common.c | 2
tools/laptop/freefall/freefall.c | 2
tools/testing/selftests/kmod/kmod.sh | 43 ++++
130 files changed, 710 insertions(+), 370 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-04-07 3:02 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-04-07 3:02 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
- a lot more of MM, quite a bit more yet to come.
- various other subsystems
166 patches based on 7e63420847ae5f1036e4f7c42f0b3282e73efbc2.
Subsystems affected by this patch series:
mm/memcg
mm/pagemap
mm/vmalloc
mm/pagealloc
mm/migration
mm/thp
mm/ksm
mm/madvise
mm/virtio
mm/userfaultfd
mm/memory-hotplug
mm/shmem
mm/rmap
mm/zswap
mm/zsmalloc
mm/cleanups
procfs
misc
MAINTAINERS
bitops
lib
checkpatch
epoll
binfmt
kallsyms
reiserfs
kmod
gcov
kconfig
kcov
ubsan
fault-injection
ipc
Subsystem: mm/memcg
Chris Down <chris@chrisdown.name>:
mm, memcg: bypass high reclaim iteration for cgroup hierarchy root
Subsystem: mm/pagemap
Li Xinhai <lixinhai.lxh@gmail.com>:
Patch series "mm: Fix misuse of parent anon_vma in dup_mmap path":
mm: don't prepare anon_vma if vma has VM_WIPEONFORK
Revert "mm/rmap.c: reuse mergeable anon_vma as parent when fork"
mm: set vm_next and vm_prev to NULL in vm_area_dup()
Anshuman Khandual <anshuman.khandual@arm.com>:
Patch series "mm/vma: Use all available wrappers when possible", v2:
mm/vma: add missing VMA flag readable name for VM_SYNC
mm/vma: make vma_is_accessible() available for general use
mm/vma: replace all remaining open encodings with is_vm_hugetlb_page()
mm/vma: replace all remaining open encodings with vma_is_anonymous()
mm/vma: append unlikely() while testing VMA access permissions
Subsystem: mm/vmalloc
Qiujun Huang <hqjagain@gmail.com>:
mm/vmalloc: fix a typo in comment
Subsystem: mm/pagealloc
Michal Hocko <mhocko@suse.com>:
mm: make it clear that gfp reclaim modifiers are valid only for sleepable allocations
Subsystem: mm/migration
Wei Yang <richardw.yang@linux.intel.com>:
Patch series "cleanup on do_pages_move()", v5:
mm/migrate.c: no need to check for i > start in do_pages_move()
mm/migrate.c: wrap do_move_pages_to_node() and store_status()
mm/migrate.c: check pagelist in move_pages_and_store_status()
mm/migrate.c: unify "not queued for migration" handling in do_pages_move()
Yang Shi <yang.shi@linux.alibaba.com>:
mm/migrate.c: migrate PG_readahead flag
Subsystem: mm/thp
David Rientjes <rientjes@google.com>:
mm, shmem: add vmstat for hugepage fallback
mm, thp: track fallbacks due to failed memcg charges separately
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
include/linux/pagemap.h: optimise find_subpage for !THP
mm: remove CONFIG_TRANSPARENT_HUGE_PAGECACHE
Subsystem: mm/ksm
Li Chen <chenli@uniontech.com>:
mm/ksm.c: update get_user_pages() argument in comment
Subsystem: mm/madvise
Huang Ying <ying.huang@intel.com>:
mm: code cleanup for MADV_FREE
Subsystem: mm/virtio
Alexander Duyck <alexander.h.duyck@linux.intel.com>:
Patch series "mm / virtio: Provide support for free page reporting", v17:
mm: adjust shuffle code to allow for future coalescing
mm: use zone and order instead of free area in free_list manipulators
mm: add function __putback_isolated_page
mm: introduce Reported pages
virtio-balloon: pull page poisoning config out of free page hinting
virtio-balloon: add support for providing free page reports to host
mm/page_reporting: rotate reported pages to the tail of the list
mm/page_reporting: add budget limit on how many pages can be reported per pass
mm/page_reporting: add free page reporting documentation
David Hildenbrand <david@redhat.com>:
virtio-balloon: switch back to OOM handler for VIRTIO_BALLOON_F_DEFLATE_ON_OOM
Subsystem: mm/userfaultfd
Shaohua Li <shli@fb.com>:
Patch series "userfaultfd: write protection support", v6:
userfaultfd: wp: add helper for writeprotect check
Andrea Arcangeli <aarcange@redhat.com>:
userfaultfd: wp: hook userfault handler to write protection fault
userfaultfd: wp: add WP pagetable tracking to x86
userfaultfd: wp: userfaultfd_pte/huge_pmd_wp() helpers
userfaultfd: wp: add UFFDIO_COPY_MODE_WP
Peter Xu <peterx@redhat.com>:
mm: merge parameters for change_protection()
userfaultfd: wp: apply _PAGE_UFFD_WP bit
userfaultfd: wp: drop _PAGE_UFFD_WP properly when fork
userfaultfd: wp: add pmd_swp_*uffd_wp() helpers
userfaultfd: wp: support swap and page migration
khugepaged: skip collapse if uffd-wp detected
Shaohua Li <shli@fb.com>:
userfaultfd: wp: support write protection for userfault vma range
Andrea Arcangeli <aarcange@redhat.com>:
userfaultfd: wp: add the writeprotect API to userfaultfd ioctl
Shaohua Li <shli@fb.com>:
userfaultfd: wp: enabled write protection in userfaultfd API
Peter Xu <peterx@redhat.com>:
userfaultfd: wp: don't wake up when doing write protect
Martin Cracauer <cracauer@cons.org>:
userfaultfd: wp: UFFDIO_REGISTER_MODE_WP documentation update
Peter Xu <peterx@redhat.com>:
userfaultfd: wp: declare _UFFDIO_WRITEPROTECT conditionally
userfaultfd: selftests: refactor statistics
userfaultfd: selftests: add write-protect test
Subsystem: mm/memory-hotplug
David Hildenbrand <david@redhat.com>:
Patch series "mm: drop superfluous section checks when onlining/offlining":
drivers/base/memory.c: drop section_count
drivers/base/memory.c: drop pages_correctly_probed()
mm/page_ext.c: drop pfn_present() check when onlining
Baoquan He <bhe@redhat.com>:
mm/memory_hotplug.c: only respect mem= parameter during boot stage
David Hildenbrand <david@redhat.com>:
mm/memory_hotplug.c: simplify calculation of number of pages in __remove_pages()
mm/memory_hotplug.c: cleanup __add_pages()
Baoquan He <bhe@redhat.com>:
Patch series "mm/hotplug: Only use subsection map for VMEMMAP", v4:
mm/sparse.c: introduce new function fill_subsection_map()
mm/sparse.c: introduce a new function clear_subsection_map()
mm/sparse.c: only use subsection map in VMEMMAP case
mm/sparse.c: add note about only VMEMMAP supporting sub-section hotplug
mm/sparse.c: move subsection_map related functions together
David Hildenbrand <david@redhat.com>:
Patch series "mm/memory_hotplug: allow to specify a default online_type", v3:
drivers/base/memory: rename MMOP_ONLINE_KEEP to MMOP_ONLINE
drivers/base/memory: map MMOP_OFFLINE to 0
drivers/base/memory: store mapping between MMOP_* and string in an array
powernv/memtrace: always online added memory blocks
hv_balloon: don't check for memhp_auto_online manually
mm/memory_hotplug: unexport memhp_auto_online
mm/memory_hotplug: convert memhp_auto_online to store an online_type
mm/memory_hotplug: allow to specify a default online_type
chenqiwu <chenqiwu@xiaomi.com>:
mm/memory_hotplug.c: use __pfn_to_section() instead of open-coding
Subsystem: mm/shmem
Kees Cook <keescook@chromium.org>:
mm/shmem.c: distribute switch variables for initialization
Mateusz Nosek <mateusznosek0@gmail.com>:
mm/shmem.c: clean code by removing unnecessary assignment
Hugh Dickins <hughd@google.com>:
mm: huge tmpfs: try to split_huge_page() when punching hole
Subsystem: mm/rmap
Palmer Dabbelt <palmerdabbelt@google.com>:
mm: prevent a warning when casting void* -> enum
Subsystem: mm/zswap
"Maciej S. Szmigiero" <mail@maciej.szmigiero.name>:
mm/zswap: allow setting default status, compressor and allocator in Kconfig
Subsystem: mm/zsmalloc
Subsystem: mm/cleanups
Jules Irenge <jbi.octave@gmail.com>:
mm/compaction: add missing annotation for compact_lock_irqsave
mm/hugetlb: add missing annotation for gather_surplus_pages()
mm/mempolicy: add missing annotation for queue_pages_pmd()
mm/slub: add missing annotation for get_map()
mm/slub: add missing annotation for put_map()
mm/zsmalloc: add missing annotation for migrate_read_lock()
mm/zsmalloc: add missing annotation for migrate_read_unlock()
mm/zsmalloc: add missing annotation for pin_tag()
mm/zsmalloc: add missing annotation for unpin_tag()
chenqiwu <chenqiwu@xiaomi.com>:
mm: fix ambiguous comments for better code readability
Mateusz Nosek <mateusznosek0@gmail.com>:
mm/mm_init.c: clean code. Use BUILD_BUG_ON when comparing compile time constant
Joe Perches <joe@perches.com>:
mm: use fallthrough;
Steven Price <steven.price@arm.com>:
include/linux/swapops.h: correct guards for non_swap_entry()
Ira Weiny <ira.weiny@intel.com>:
include/linux/memremap.h: remove stale comments
Mateusz Nosek <mateusznosek0@gmail.com>:
mm/dmapool.c: micro-optimisation remove unnecessary branch
Waiman Long <longman@redhat.com>:
mm: remove dummy struct bootmem_data/bootmem_data_t
Subsystem: procfs
Jules Irenge <jbi.octave@gmail.com>:
fs/proc/inode.c: annotate close_pdeo() for sparse
Alexey Dobriyan <adobriyan@gmail.com>:
proc: faster open/read/close with "permanent" files
proc: speed up /proc/*/statm
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
proc: inline vma_stop into m_stop
proc: remove m_cache_vma
proc: use ppos instead of m->version
seq_file: remove m->version
proc: inline m_next_vma into m_next
Subsystem: misc
Michal Simek <michal.simek@xilinx.com>:
asm-generic: fix unistd_32.h generation format
Nathan Chancellor <natechancellor@gmail.com>:
kernel/extable.c: use address-of operator on section symbols
Masahiro Yamada <masahiroy@kernel.org>:
sparc,x86: vdso: remove meaningless undefining CONFIG_OPTIMIZE_INLINING
compiler: remove CONFIG_OPTIMIZE_INLINING entirely
Vegard Nossum <vegard.nossum@oracle.com>:
compiler.h: fix error in BUILD_BUG_ON() reporting
Subsystem: MAINTAINERS
Joe Perches <joe@perches.com>:
MAINTAINERS: list the section entries in the preferred order
Subsystem: bitops
Josh Poimboeuf <jpoimboe@redhat.com>:
bitops: always inline sign extension helpers
Subsystem: lib
Konstantin Khlebnikov <khlebnikov@yandex-team.ru>:
lib/test_lockup: test module to generate lockups
Colin Ian King <colin.king@canonical.com>:
lib/test_lockup.c: fix spelling mistake "iteraions" -> "iterations"
Konstantin Khlebnikov <khlebnikov@yandex-team.ru>:
lib/test_lockup.c: add parameters for locking generic vfs locks
"Gustavo A. R. Silva" <gustavo@embeddedor.com>:
lib/bch.c: replace zero-length array with flexible-array member
lib/ts_bm.c: replace zero-length array with flexible-array member
lib/ts_fsm.c: replace zero-length array with flexible-array member
lib/ts_kmp.c: replace zero-length array with flexible-array member
Geert Uytterhoeven <geert+renesas@glider.be>:
lib/scatterlist: fix sg_copy_buffer() kerneldoc
Kees Cook <keescook@chromium.org>:
lib: test_stackinit.c: XFAIL switch variable init tests
Alexander Potapenko <glider@google.com>:
lib/stackdepot.c: check depot_index before accessing the stack slab
lib/stackdepot.c: fix a condition in stack_depot_fetch()
lib/stackdepot.c: build with -fno-builtin
kasan: stackdepot: move filter_irq_stacks() to stackdepot.c
Qian Cai <cai@lca.pw>:
percpu_counter: fix a data race at vm_committed_as
Andy Shevchenko <andriy.shevchenko@linux.intel.com>:
lib/test_bitmap.c: make use of EXP2_IN_BITS
chenqiwu <chenqiwu@xiaomi.com>:
lib/rbtree: fix coding style of assignments
Dan Carpenter <dan.carpenter@oracle.com>:
lib/test_kmod.c: remove a NULL test
Rikard Falkeborn <rikard.falkeborn@gmail.com>:
linux/bits.h: add compile time sanity check of GENMASK inputs
Chris Wilson <chris@chris-wilson.co.uk>:
lib/list: prevent compiler reloads inside 'safe' list iteration
Nathan Chancellor <natechancellor@gmail.com>:
lib/dynamic_debug.c: use address-of operator on section symbols
Subsystem: checkpatch
Joe Perches <joe@perches.com>:
checkpatch: remove email address comment from email address comparisons
Lubomir Rintel <lkundrak@v3.sk>:
checkpatch: check SPDX tags in YAML files
John Hubbard <jhubbard@nvidia.com>:
checkpatch: support "base-commit:" format
Joe Perches <joe@perches.com>:
checkpatch: prefer fallthrough; over fallthrough comments
Antonio Borneo <borneo.antonio@gmail.com>:
checkpatch: fix minor typo and mixed space+tab in indentation
checkpatch: fix multiple const * types
checkpatch: add command-line option for TAB size
Joe Perches <joe@perches.com>:
checkpatch: improve Gerrit Change-Id: test
Lubomir Rintel <lkundrak@v3.sk>:
checkpatch: check proper licensing of Devicetree bindings
Joe Perches <joe@perches.com>:
checkpatch: avoid warning about uninitialized_var()
Subsystem: epoll
Roman Penyaev <rpenyaev@suse.de>:
kselftest: introduce new epoll test case
Jason Baron <jbaron@akamai.com>:
fs/epoll: make nesting accounting safe for -rt kernel
Subsystem: binfmt
Alexey Dobriyan <adobriyan@gmail.com>:
fs/binfmt_elf.c: delete "loc" variable
fs/binfmt_elf.c: allocate less for static executable
fs/binfmt_elf.c: don't free interpreter's ELF pheaders on common path
Subsystem: kallsyms
Will Deacon <will@kernel.org>:
Patch series "Unexport kallsyms_lookup_name() and kallsyms_on_each_symbol()":
samples/hw_breakpoint: drop HW_BREAKPOINT_R when reporting writes
samples/hw_breakpoint: drop use of kallsyms_lookup_name()
kallsyms: unexport kallsyms_lookup_name() and kallsyms_on_each_symbol()
Subsystem: reiserfs
Colin Ian King <colin.king@canonical.com>:
reiserfs: clean up several indentation issues
Subsystem: kmod
Qiujun Huang <hqjagain@gmail.com>:
kernel/kmod.c: fix a typo "assuems" -> "assumes"
Subsystem: gcov
"Gustavo A. R. Silva" <gustavo@embeddedor.com>:
gcov: gcc_4_7: replace zero-length array with flexible-array member
gcov: gcc_3_4: replace zero-length array with flexible-array member
kernel/gcov/fs.c: replace zero-length array with flexible-array member
Subsystem: kconfig
Krzysztof Kozlowski <krzk@kernel.org>:
init/Kconfig: clean up ANON_INODES and old IO schedulers options
Subsystem: kcov
Andrey Konovalov <andreyknvl@google.com>:
Patch series "kcov: collect coverage from usb soft interrupts", v4:
kcov: cleanup debug messages
kcov: fix potential use-after-free in kcov_remote_start
kcov: move t->kcov assignments into kcov_start/stop
kcov: move t->kcov_sequence assignment
kcov: use t->kcov_mode as enabled indicator
kcov: collect coverage from interrupts
usb: core: kcov: collect coverage from usb complete callback
Subsystem: ubsan
Kees Cook <keescook@chromium.org>:
Patch series "ubsan: Split out bounds checker", v5:
ubsan: add trap instrumentation option
ubsan: split "bounds" checker from other options
drivers/misc/lkdtm/bugs.c: add arithmetic overflow and array bounds checks
ubsan: check panic_on_warn
kasan: unset panic_on_warn before calling panic()
ubsan: include bug type in report header
Subsystem: fault-injection
Qiujun Huang <hqjagain@gmail.com>:
lib/Kconfig.debug: fix a typo "capabilitiy" -> "capability"
Subsystem: ipc
Somala Swaraj <somalaswaraj@gmail.com>:
ipc/mqueue.c: fix a brace coding style issue
Jason Yan <yanaijie@huawei.com>:
ipc/shm.c: make compat_ksys_shmctl() static
Documentation/admin-guide/kernel-parameters.txt | 13
Documentation/admin-guide/mm/transhuge.rst | 14
Documentation/admin-guide/mm/userfaultfd.rst | 51
Documentation/dev-tools/kcov.rst | 17
Documentation/vm/free_page_reporting.rst | 41
Documentation/vm/zswap.rst | 20
MAINTAINERS | 35
arch/alpha/include/asm/mmzone.h | 2
arch/alpha/kernel/syscalls/syscallhdr.sh | 2
arch/csky/mm/fault.c | 4
arch/ia64/kernel/syscalls/syscallhdr.sh | 2
arch/ia64/kernel/vmlinux.lds.S | 2
arch/m68k/mm/fault.c | 4
arch/microblaze/kernel/syscalls/syscallhdr.sh | 2
arch/mips/kernel/syscalls/syscallhdr.sh | 3
arch/mips/mm/fault.c | 4
arch/nds32/kernel/vmlinux.lds.S | 1
arch/parisc/kernel/syscalls/syscallhdr.sh | 2
arch/powerpc/kernel/syscalls/syscallhdr.sh | 3
arch/powerpc/kvm/e500_mmu_host.c | 2
arch/powerpc/mm/fault.c | 2
arch/powerpc/platforms/powernv/memtrace.c | 14
arch/sh/kernel/syscalls/syscallhdr.sh | 2
arch/sh/mm/fault.c | 2
arch/sparc/kernel/syscalls/syscallhdr.sh | 2
arch/sparc/vdso/vdso32/vclock_gettime.c | 4
arch/x86/Kconfig | 1
arch/x86/configs/i386_defconfig | 1
arch/x86/configs/x86_64_defconfig | 1
arch/x86/entry/vdso/vdso32/vclock_gettime.c | 4
arch/x86/include/asm/pgtable.h | 67 +
arch/x86/include/asm/pgtable_64.h | 8
arch/x86/include/asm/pgtable_types.h | 12
arch/x86/mm/fault.c | 2
arch/xtensa/kernel/syscalls/syscallhdr.sh | 2
drivers/base/memory.c | 138 --
drivers/hv/hv_balloon.c | 25
drivers/misc/lkdtm/bugs.c | 75 +
drivers/misc/lkdtm/core.c | 3
drivers/misc/lkdtm/lkdtm.h | 3
drivers/usb/core/hcd.c | 3
drivers/virtio/Kconfig | 1
drivers/virtio/virtio_balloon.c | 190 ++-
fs/binfmt_elf.c | 56
fs/eventpoll.c | 64 -
fs/proc/array.c | 39
fs/proc/cpuinfo.c | 1
fs/proc/generic.c | 31
fs/proc/inode.c | 188 ++-
fs/proc/internal.h | 6
fs/proc/kmsg.c | 1
fs/proc/stat.c | 1
fs/proc/task_mmu.c | 97 -
fs/reiserfs/do_balan.c | 2
fs/reiserfs/ioctl.c | 11
fs/reiserfs/namei.c | 10
fs/seq_file.c | 28
fs/userfaultfd.c | 116 +
include/asm-generic/pgtable.h | 1
include/asm-generic/pgtable_uffd.h | 66 +
include/asm-generic/tlb.h | 3
include/linux/bitops.h | 4
include/linux/bits.h | 22
include/linux/compiler.h | 2
include/linux/compiler_types.h | 11
include/linux/gfp.h | 2
include/linux/huge_mm.h | 2
include/linux/list.h | 50
include/linux/memory.h | 1
include/linux/memory_hotplug.h | 13
include/linux/memremap.h | 2
include/linux/mm.h | 25
include/linux/mm_inline.h | 15
include/linux/mm_types.h | 4
include/linux/mmzone.h | 47
include/linux/page-flags.h | 16
include/linux/page_reporting.h | 26
include/linux/pagemap.h | 4
include/linux/percpu_counter.h | 4
include/linux/proc_fs.h | 17
include/linux/sched.h | 3
include/linux/seq_file.h | 1
include/linux/shmem_fs.h | 10
include/linux/stackdepot.h | 2
include/linux/swapops.h | 5
include/linux/userfaultfd_k.h | 42
include/linux/vm_event_item.h | 5
include/trace/events/huge_memory.h | 1
include/trace/events/mmflags.h | 1
include/trace/events/vmscan.h | 2
include/uapi/linux/userfaultfd.h | 40
include/uapi/linux/virtio_balloon.h | 1
init/Kconfig | 8
ipc/mqueue.c | 5
ipc/shm.c | 2
ipc/util.c | 1
kernel/configs/tiny.config | 1
kernel/events/core.c | 3
kernel/extable.c | 3
kernel/fork.c | 10
kernel/gcov/fs.c | 2
kernel/gcov/gcc_3_4.c | 6
kernel/gcov/gcc_4_7.c | 2
kernel/kallsyms.c | 2
kernel/kcov.c | 282 +++-
kernel/kmod.c | 2
kernel/module.c | 1
kernel/sched/fair.c | 2
lib/Kconfig.debug | 35
lib/Kconfig.ubsan | 51
lib/Makefile | 8
lib/bch.c | 2
lib/dynamic_debug.c | 2
lib/rbtree.c | 4
lib/scatterlist.c | 2
lib/stackdepot.c | 39
lib/test_bitmap.c | 2
lib/test_kmod.c | 2
lib/test_lockup.c | 601 +++++++++-
lib/test_stackinit.c | 28
lib/ts_bm.c | 2
lib/ts_fsm.c | 2
lib/ts_kmp.c | 2
lib/ubsan.c | 47
mm/Kconfig | 135 ++
mm/Makefile | 1
mm/compaction.c | 3
mm/dmapool.c | 4
mm/filemap.c | 14
mm/gup.c | 9
mm/huge_memory.c | 36
mm/hugetlb.c | 1
mm/hugetlb_cgroup.c | 6
mm/internal.h | 2
mm/kasan/common.c | 23
mm/kasan/report.c | 10
mm/khugepaged.c | 39
mm/ksm.c | 5
mm/list_lru.c | 2
mm/memcontrol.c | 5
mm/memory-failure.c | 2
mm/memory.c | 42
mm/memory_hotplug.c | 53
mm/mempolicy.c | 11
mm/migrate.c | 122 +-
mm/mm_init.c | 2
mm/mmap.c | 10
mm/mprotect.c | 76 -
mm/page_alloc.c | 174 ++
mm/page_ext.c | 5
mm/page_isolation.c | 6
mm/page_reporting.c | 384 ++++++
mm/page_reporting.h | 54
mm/rmap.c | 23
mm/shmem.c | 168 +-
mm/shuffle.c | 12
mm/shuffle.h | 6
mm/slab_common.c | 1
mm/slub.c | 3
mm/sparse.c | 236 ++-
mm/swap.c | 20
mm/swapfile.c | 1
mm/userfaultfd.c | 98 +
mm/vmalloc.c | 2
mm/vmscan.c | 12
mm/vmstat.c | 3
mm/zsmalloc.c | 10
mm/zswap.c | 24
samples/hw_breakpoint/data_breakpoint.c | 11
scripts/Makefile.ubsan | 16
scripts/checkpatch.pl | 155 +-
tools/lib/rbtree.c | 4
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c | 67 +
tools/testing/selftests/vm/userfaultfd.c | 233 +++
174 files changed, 3990 insertions(+), 1399 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-04-02 4:01 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-04-02 4:01 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
A large amount of MM, plenty more to come.
155 patches, based on GIT 1a323ea5356edbb3073dc59d51b9e6b86908857d
Subsystems affected by this patch series:
tools
kthread
kbuild
scripts
ocfs2
vfs
mm/slub
mm/kmemleak
mm/pagecache
mm/gup
mm/swap
mm/memcg
mm/pagemap
mm/mremap
mm/sparsemem
mm/kasan
mm/pagealloc
mm/vmscan
mm/compaction
mm/mempolicy
mm/hugetlbfs
mm/hugetlb
Subsystem: tools
David Ahern <dsahern@kernel.org>:
tools/accounting/getdelays.c: fix netlink attribute length
Subsystem: kthread
Petr Mladek <pmladek@suse.com>:
kthread: mark timer used by delayed kthread works as IRQ safe
Subsystem: kbuild
Masahiro Yamada <masahiroy@kernel.org>:
asm-generic: make more kernel-space headers mandatory
Subsystem: scripts
Jonathan Neuschäfer <j.neuschaefer@gmx.net>:
scripts/spelling.txt: add syfs/sysfs pattern
Colin Ian King <colin.king@canonical.com>:
scripts/spelling.txt: add more spellings to spelling.txt
Subsystem: ocfs2
Alex Shi <alex.shi@linux.alibaba.com>:
ocfs2: remove FS_OCFS2_NM
ocfs2: remove unused macros
ocfs2: use OCFS2_SEC_BITS in macro
ocfs2: remove dlm_lock_is_remote
wangyan <wangyan122@huawei.com>:
ocfs2: there is no need to log twice in several functions
ocfs2: correct annotation from "l_next_rec" to "l_next_free_rec"
Alex Shi <alex.shi@linux.alibaba.com>:
ocfs2: remove useless err
Jules Irenge <jbi.octave@gmail.com>:
ocfs2: Add missing annotations for ocfs2_refcount_cache_lock() and ocfs2_refcount_cache_unlock()
"Gustavo A. R. Silva" <gustavo@embeddedor.com>:
ocfs2: replace zero-length array with flexible-array member
ocfs2: cluster: replace zero-length array with flexible-array member
ocfs2: dlm: replace zero-length array with flexible-array member
ocfs2: ocfs2_fs.h: replace zero-length array with flexible-array member
wangjian <wangjian161@huawei.com>:
ocfs2: roll back the reference count modification of the parent directory if an error occurs
Takashi Iwai <tiwai@suse.de>:
ocfs2: use scnprintf() for avoiding potential buffer overflow
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
ocfs2: use memalloc_nofs_save instead of memalloc_noio_save
Subsystem: vfs
Kees Cook <keescook@chromium.org>:
fs_parse: Remove pr_notice() about each validation
Subsystem: mm/slub
chenqiwu <chenqiwu@xiaomi.com>:
mm/slub.c: replace cpu_slab->partial with wrapped APIs
mm/slub.c: replace kmem_cache->cpu_partial with wrapped APIs
Kees Cook <keescook@chromium.org>:
slub: improve bit diffusion for freelist ptr obfuscation
slub: relocate freelist pointer to middle of object
Vlastimil Babka <vbabka@suse.cz>:
Revert "topology: add support for node_to_mem_node() to determine the fallback node"
Subsystem: mm/kmemleak
Nathan Chancellor <natechancellor@gmail.com>:
mm/kmemleak.c: use address-of operator on section symbols
Qian Cai <cai@lca.pw>:
mm/Makefile: disable KCSAN for kmemleak
Subsystem: mm/pagecache
Jan Kara <jack@suse.cz>:
mm/filemap.c: don't bother dropping mmap_sem for zero size readahead
Mauricio Faria de Oliveira <mfo@canonical.com>:
mm/page-writeback.c: write_cache_pages(): deduplicate identical checks
Xianting Tian <xianting_tian@126.com>:
mm/filemap.c: clear page error before actual read
Souptick Joarder <jrdr.linux@gmail.com>:
mm/filemap.c: remove unused argument from shrink_readahead_size_eio()
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
mm/filemap.c: use vm_fault error code directly
include/linux/pagemap.h: rename arguments to find_subpage
mm/page-writeback.c: use VM_BUG_ON_PAGE in clear_page_dirty_for_io
mm/filemap.c: unexport find_get_entry
mm/filemap.c: rewrite pagecache_get_page documentation
Subsystem: mm/gup
John Hubbard <jhubbard@nvidia.com>:
Patch series "mm/gup: track FOLL_PIN pages", v6:
mm/gup: split get_user_pages_remote() into two routines
mm/gup: pass a flags arg to __gup_device_* functions
mm: introduce page_ref_sub_return()
mm/gup: pass gup flags to two more routines
mm/gup: require FOLL_GET for get_user_pages_fast()
mm/gup: track FOLL_PIN pages
mm/gup: page->hpage_pinned_refcount: exact pin counts for huge pages
mm/gup: /proc/vmstat: pin_user_pages (FOLL_PIN) reporting
mm/gup_benchmark: support pin_user_pages() and related calls
selftests/vm: run_vmtests: invoke gup_benchmark with basic FOLL_PIN coverage
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
mm: improve dump_page() for compound pages
John Hubbard <jhubbard@nvidia.com>:
mm: dump_page(): additional diagnostics for huge pinned pages
Claudio Imbrenda <imbrenda@linux.ibm.com>:
mm/gup/writeback: add callbacks for inaccessible pages
Pingfan Liu <kernelfans@gmail.com>:
mm/gup: rename nr as nr_pinned in get_user_pages_fast()
mm/gup: fix omission of check on FOLL_LONGTERM in gup fast path
Subsystem: mm/swap
Chen Wandun <chenwandun@huawei.com>:
mm/swapfile.c: fix comments for swapcache_prepare
Wei Yang <richardw.yang@linux.intel.com>:
mm/swap.c: not necessary to export __pagevec_lru_add()
Qian Cai <cai@lca.pw>:
mm/swapfile: fix data races in try_to_unuse()
Wei Yang <richard.weiyang@linux.alibaba.com>:
mm/swap_slots.c: assign|reset cache slot by value directly
Yang Shi <yang.shi@linux.alibaba.com>:
mm: swap: make page_evictable() inline
mm: swap: use smp_mb__after_atomic() to order LRU bit set
Wei Yang <richard.weiyang@gmail.com>:
mm/swap_state.c: use the same way to count page in [add_to|delete_from]_swap_cache
Subsystem: mm/memcg
Yafang Shao <laoar.shao@gmail.com>:
mm, memcg: fix build error around the usage of kmem_caches
Kirill Tkhai <ktkhai@virtuozzo.com>:
mm/memcontrol.c: allocate shrinker_map on appropriate NUMA node
Roman Gushchin <guro@fb.com>:
mm: memcg/slab: use mem_cgroup_from_obj()
Patch series "mm: memcg: kmem API cleanup", v2:
mm: kmem: cleanup (__)memcg_kmem_charge_memcg() arguments
mm: kmem: cleanup memcg_kmem_uncharge_memcg() arguments
mm: kmem: rename memcg_kmem_(un)charge() into memcg_kmem_(un)charge_page()
mm: kmem: switch to nr_pages in (__)memcg_kmem_charge_memcg()
mm: memcg/slab: cache page number in memcg_(un)charge_slab()
mm: kmem: rename (__)memcg_kmem_(un)charge_memcg() to __memcg_kmem_(un)charge()
Johannes Weiner <hannes@cmpxchg.org>:
Patch series "mm: memcontrol: recursive memory.low protection", v3:
mm: memcontrol: fix memory.low proportional distribution
mm: memcontrol: clean up and document effective low/min calculations
mm: memcontrol: recursive memory.low protection
Shakeel Butt <shakeelb@google.com>:
memcg: css_tryget_online cleanups
Vincenzo Frascino <vincenzo.frascino@arm.com>:
mm/memcontrol.c: make mem_cgroup_id_get_many() __maybe_unused
Chris Down <chris@chrisdown.name>:
mm, memcg: prevent memory.high load/store tearing
mm, memcg: prevent memory.max load tearing
mm, memcg: prevent memory.low load/store tearing
mm, memcg: prevent memory.min load/store tearing
mm, memcg: prevent memory.swap.max load tearing
mm, memcg: prevent mem_cgroup_protected store tearing
Roman Gushchin <guro@fb.com>:
mm: memcg: make memory.oom.group tolerable to task migration
Subsystem: mm/pagemap
Thomas Hellstrom <thellstrom@vmware.com>:
mm/mapping_dirty_helpers: Update huge page-table entry callbacks
Anshuman Khandual <anshuman.khandual@arm.com>:
Patch series "mm/vma: some more minor changes", v2:
mm/vma: move VM_NO_KHUGEPAGED into generic header
mm/vma: make vma_is_foreign() available for general use
mm/vma: make is_vma_temporary_stack() available for general use
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
mm: add pagemap.h to the fine documentation
Peter Xu <peterx@redhat.com>:
Patch series "mm: Page fault enhancements", v6:
mm/gup: rename "nonblocking" to "locked" where proper
mm/gup: fix __get_user_pages() on fault retry of hugetlb
mm: introduce fault_signal_pending()
x86/mm: use helper fault_signal_pending()
arc/mm: use helper fault_signal_pending()
arm64/mm: use helper fault_signal_pending()
powerpc/mm: use helper fault_signal_pending()
sh/mm: use helper fault_signal_pending()
mm: return faster for non-fatal signals in user mode faults
userfaultfd: don't retake mmap_sem to emulate NOPAGE
mm: introduce FAULT_FLAG_DEFAULT
mm: introduce FAULT_FLAG_INTERRUPTIBLE
mm: allow VM_FAULT_RETRY for multiple times
mm/gup: allow VM_FAULT_RETRY for multiple times
mm/gup: allow to react to fatal signals
mm/userfaultfd: honor FAULT_FLAG_KILLABLE in fault path
WANG Wenhu <wenhu.wang@vivo.com>:
mm: clarify a confusing comment for remap_pfn_range()
Wang Wenhu <wenhu.wang@vivo.com>:
mm/memory.c: clarify a confusing comment for vm_iomap_memory
Jaewon Kim <jaewon31.kim@samsung.com>:
Patch series "mm: mmap: add mmap trace point", v3:
mmap: remove inline of vm_unmapped_area
mm: mmap: add trace point of vm_unmapped_area
Subsystem: mm/mremap
Brian Geffon <bgeffon@google.com>:
mm/mremap: add MREMAP_DONTUNMAP to mremap()
selftests: add MREMAP_DONTUNMAP selftest
Subsystem: mm/sparsemem
Wei Yang <richardw.yang@linux.intel.com>:
mm/sparsemem: get address to page struct instead of address to pfn
Pingfan Liu <kernelfans@gmail.com>:
mm/sparse: rename pfn_present() to pfn_in_present_section()
Baoquan He <bhe@redhat.com>:
mm/sparse.c: use kvmalloc/kvfree to alloc/free memmap for the classic sparse
mm/sparse.c: allocate memmap preferring the given node
Subsystem: mm/kasan
Walter Wu <walter-zh.wu@mediatek.com>:
Patch series "fix the missing underflow in memory operation function", v4:
kasan: detect negative size in memory operation function
kasan: add test for invalid size in memmove
Subsystem: mm/pagealloc
Joel Savitz <jsavitz@redhat.com>:
mm/page_alloc: increase default min_free_kbytes bound
Mateusz Nosek <mateusznosek0@gmail.com>:
mm, pagealloc: micro-optimisation: save two branches on hot page allocation path
chenqiwu <chenqiwu@xiaomi.com>:
mm/page_alloc.c: use free_area_empty() instead of open-coding
Mateusz Nosek <mateusznosek0@gmail.com>:
mm/page_alloc.c: micro-optimisation Remove unnecessary branch
chenqiwu <chenqiwu@xiaomi.com>:
mm/page_alloc: simplify page_is_buddy() for better code readability
Subsystem: mm/vmscan
Yang Shi <yang.shi@linux.alibaba.com>:
mm: vmpressure: don't need call kfree if kstrndup fails
mm: vmpressure: use mem_cgroup_is_root API
mm: vmscan: replace open codings to NUMA_NO_NODE
Wei Yang <richardw.yang@linux.intel.com>:
mm/vmscan.c: remove cpu online notification for now
Qian Cai <cai@lca.pw>:
mm/vmscan.c: fix data races using kswapd_classzone_idx
Mateusz Nosek <mateusznosek0@gmail.com>:
mm/vmscan.c: Clean code by removing unnecessary assignment
Kirill Tkhai <ktkhai@virtuozzo.com>:
mm/vmscan.c: make may_enter_fs bool in shrink_page_list()
Mateusz Nosek <mateusznosek0@gmail.com>:
mm/vmscan.c: do_try_to_free_pages(): clean code by removing unnecessary assignment
Michal Hocko <mhocko@suse.com>:
selftests: vm: drop dependencies on page flags from mlock2 tests
Subsystem: mm/compaction
Rik van Riel <riel@surriel.com>:
Patch series "fix THP migration for CMA allocations", v2:
mm,compaction,cma: add alloc_contig flag to compact_control
mm,thp,compaction,cma: allow THP migration for CMA allocations
Vlastimil Babka <vbabka@suse.cz>:
mm, compaction: fully assume capture is not NULL in compact_zone_order()
Sebastian Andrzej Siewior <bigeasy@linutronix.de>:
mm/compaction: really limit compact_unevictable_allowed to 0 and 1
mm/compaction: Disable compact_unevictable_allowed on RT
Mateusz Nosek <mateusznosek0@gmail.com>:
mm/compaction.c: clean code by removing unnecessary assignment
Subsystem: mm/mempolicy
Li Xinhai <lixinhai.lxh@gmail.com>:
mm/mempolicy: support MPOL_MF_STRICT for huge page mapping
mm/mempolicy: check hugepage migration is supported by arch in vma_migratable()
Yang Shi <yang.shi@linux.alibaba.com>:
mm: mempolicy: use VM_BUG_ON_VMA in queue_pages_test_walk()
Randy Dunlap <rdunlap@infradead.org>:
mm: mempolicy: require at least one nodeid for MPOL_PREFERRED
Colin Ian King <colin.king@canonical.com>:
mm/memblock.c: remove redundant assignment to variable max_addr
Subsystem: mm/hugetlbfs
Mike Kravetz <mike.kravetz@oracle.com>:
Patch series "hugetlbfs: use i_mmap_rwsem for more synchronization", v2:
hugetlbfs: use i_mmap_rwsem for more pmd sharing synchronization
hugetlbfs: Use i_mmap_rwsem to address page fault/truncate race
Subsystem: mm/hugetlb
Mina Almasry <almasrymina@google.com>:
hugetlb_cgroup: add hugetlb_cgroup reservation counter
hugetlb_cgroup: add interface for charge/uncharge hugetlb reservations
mm/hugetlb_cgroup: fix hugetlb_cgroup migration
hugetlb_cgroup: add reservation accounting for private mappings
hugetlb: disable region_add file_region coalescing
hugetlb_cgroup: add accounting for shared mappings
hugetlb_cgroup: support noreserve mappings
hugetlb: support file_region coalescing again
hugetlb_cgroup: add hugetlb_cgroup reservation tests
hugetlb_cgroup: add hugetlb_cgroup reservation docs
Mateusz Nosek <mateusznosek0@gmail.com>:
mm/hugetlb.c: clean code by removing unnecessary initialization
Vlastimil Babka <vbabka@suse.cz>:
mm/hugetlb: remove unnecessary memory fetch in PageHeadHuge()
Christophe Leroy <christophe.leroy@c-s.fr>:
selftests/vm: fix map_hugetlb length used for testing read and write
mm/hugetlb: fix build failure with HUGETLB_PAGE but not HUGEBTLBFS
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
include/linux/huge_mm.h: check PageTail in hpage_nr_pages even when !THP
Documentation/admin-guide/cgroup-v1/hugetlb.rst | 103 +-
Documentation/admin-guide/cgroup-v2.rst | 11
Documentation/admin-guide/sysctl/vm.rst | 3
Documentation/core-api/mm-api.rst | 3
Documentation/core-api/pin_user_pages.rst | 86 +
arch/alpha/include/asm/Kbuild | 11
arch/alpha/mm/fault.c | 6
arch/arc/include/asm/Kbuild | 21
arch/arc/mm/fault.c | 37
arch/arm/include/asm/Kbuild | 12
arch/arm/mm/fault.c | 7
arch/arm64/include/asm/Kbuild | 18
arch/arm64/mm/fault.c | 26
arch/c6x/include/asm/Kbuild | 37
arch/csky/include/asm/Kbuild | 36
arch/h8300/include/asm/Kbuild | 46
arch/hexagon/include/asm/Kbuild | 33
arch/hexagon/mm/vm_fault.c | 5
arch/ia64/include/asm/Kbuild | 7
arch/ia64/mm/fault.c | 5
arch/m68k/include/asm/Kbuild | 24
arch/m68k/mm/fault.c | 7
arch/microblaze/include/asm/Kbuild | 29
arch/microblaze/mm/fault.c | 5
arch/mips/include/asm/Kbuild | 13
arch/mips/mm/fault.c | 5
arch/nds32/include/asm/Kbuild | 37
arch/nds32/mm/fault.c | 5
arch/nios2/include/asm/Kbuild | 38
arch/nios2/mm/fault.c | 7
arch/openrisc/include/asm/Kbuild | 36
arch/openrisc/mm/fault.c | 5
arch/parisc/include/asm/Kbuild | 18
arch/parisc/mm/fault.c | 8
arch/powerpc/include/asm/Kbuild | 4
arch/powerpc/mm/book3s64/pkeys.c | 12
arch/powerpc/mm/fault.c | 20
arch/powerpc/platforms/pseries/hotplug-memory.c | 2
arch/riscv/include/asm/Kbuild | 28
arch/riscv/mm/fault.c | 9
arch/s390/include/asm/Kbuild | 15
arch/s390/mm/fault.c | 10
arch/sh/include/asm/Kbuild | 16
arch/sh/mm/fault.c | 13
arch/sparc/include/asm/Kbuild | 14
arch/sparc/mm/fault_32.c | 5
arch/sparc/mm/fault_64.c | 5
arch/um/kernel/trap.c | 3
arch/unicore32/include/asm/Kbuild | 34
arch/unicore32/mm/fault.c | 8
arch/x86/include/asm/Kbuild | 2
arch/x86/include/asm/mmu_context.h | 15
arch/x86/mm/fault.c | 32
arch/xtensa/include/asm/Kbuild | 26
arch/xtensa/mm/fault.c | 5
drivers/base/node.c | 2
drivers/gpu/drm/ttm/ttm_bo_vm.c | 12
fs/fs_parser.c | 2
fs/hugetlbfs/inode.c | 30
fs/ocfs2/alloc.c | 3
fs/ocfs2/cluster/heartbeat.c | 12
fs/ocfs2/cluster/netdebug.c | 4
fs/ocfs2/cluster/tcp.c | 27
fs/ocfs2/cluster/tcp.h | 2
fs/ocfs2/dir.c | 4
fs/ocfs2/dlm/dlmcommon.h | 8
fs/ocfs2/dlm/dlmdebug.c | 100 -
fs/ocfs2/dlm/dlmmaster.c | 2
fs/ocfs2/dlm/dlmthread.c | 3
fs/ocfs2/dlmglue.c | 2
fs/ocfs2/journal.c | 2
fs/ocfs2/namei.c | 15
fs/ocfs2/ocfs2_fs.h | 18
fs/ocfs2/refcounttree.c | 2
fs/ocfs2/reservations.c | 3
fs/ocfs2/stackglue.c | 2
fs/ocfs2/suballoc.c | 5
fs/ocfs2/super.c | 46
fs/pipe.c | 2
fs/userfaultfd.c | 64 -
include/asm-generic/Kbuild | 52 +
include/linux/cgroup-defs.h | 5
include/linux/fs.h | 5
include/linux/gfp.h | 6
include/linux/huge_mm.h | 10
include/linux/hugetlb.h | 76 +
include/linux/hugetlb_cgroup.h | 175 +++
include/linux/kasan.h | 2
include/linux/kthread.h | 3
include/linux/memcontrol.h | 66 -
include/linux/mempolicy.h | 29
include/linux/mm.h | 243 +++-
include/linux/mm_types.h | 7
include/linux/mmzone.h | 6
include/linux/page_ref.h | 9
include/linux/pagemap.h | 29
include/linux/sched/signal.h | 18
include/linux/swap.h | 1
include/linux/topology.h | 17
include/trace/events/mmap.h | 48
include/uapi/linux/mman.h | 5
kernel/cgroup/cgroup.c | 17
kernel/fork.c | 9
kernel/sysctl.c | 31
lib/test_kasan.c | 19
mm/Makefile | 1
mm/compaction.c | 31
mm/debug.c | 54 -
mm/filemap.c | 77 -
mm/gup.c | 682 ++++++++++---
mm/gup_benchmark.c | 71 +
mm/huge_memory.c | 29
mm/hugetlb.c | 866 ++++++++++++-----
mm/hugetlb_cgroup.c | 347 +++++-
mm/internal.h | 32
mm/kasan/common.c | 26
mm/kasan/generic.c | 9
mm/kasan/generic_report.c | 11
mm/kasan/kasan.h | 2
mm/kasan/report.c | 5
mm/kasan/tags.c | 9
mm/kasan/tags_report.c | 11
mm/khugepaged.c | 4
mm/kmemleak.c | 2
mm/list_lru.c | 12
mm/mapping_dirty_helpers.c | 42
mm/memblock.c | 2
mm/memcontrol.c | 378 ++++---
mm/memory-failure.c | 29
mm/memory.c | 4
mm/mempolicy.c | 73 +
mm/migrate.c | 25
mm/mmap.c | 32
mm/mremap.c | 92 +
mm/page-writeback.c | 19
mm/page_alloc.c | 82 -
mm/page_counter.c | 29
mm/page_ext.c | 2
mm/rmap.c | 39
mm/shuffle.c | 2
mm/slab.h | 32
mm/slab_common.c | 2
mm/slub.c | 27
mm/sparse.c | 33
mm/swap.c | 5
mm/swap_slots.c | 12
mm/swap_state.c | 2
mm/swapfile.c | 10
mm/userfaultfd.c | 11
mm/vmpressure.c | 8
mm/vmscan.c | 111 --
mm/vmstat.c | 2
scripts/spelling.txt | 21
tools/accounting/getdelays.c | 2
tools/testing/selftests/vm/.gitignore | 1
tools/testing/selftests/vm/Makefile | 2
tools/testing/selftests/vm/charge_reserved_hugetlb.sh | 575 +++++++++++
tools/testing/selftests/vm/gup_benchmark.c | 15
tools/testing/selftests/vm/hugetlb_reparenting_test.sh | 244 ++++
tools/testing/selftests/vm/map_hugetlb.c | 14
tools/testing/selftests/vm/mlock2-tests.c | 233 ----
tools/testing/selftests/vm/mremap_dontunmap.c | 313 ++++++
tools/testing/selftests/vm/run_vmtests | 37
tools/testing/selftests/vm/write_hugetlb_memory.sh | 23
tools/testing/selftests/vm/write_to_hugetlbfs.c | 242 ++++
165 files changed, 5020 insertions(+), 2376 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-03-29 2:14 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-03-29 2:14 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
5 fixes, based on 83fd69c93340177dcd66fd26ce6441fb581c1dbf:
Naohiro Aota <naohiro.aota@wdc.com>:
mm/swapfile.c: move inode_lock out of claim_swapfile
David Hildenbrand <david@redhat.com>:
drivers/base/memory.c: indicate all memory blocks as removable
Mina Almasry <almasrymina@google.com>:
hugetlb_cgroup: fix illegal access to memory
Roman Gushchin <guro@fb.com>:
mm: fork: fix kernel_stack memcg stats for various stack implementations
"Aneesh Kumar K.V" <aneesh.kumar@linux.ibm.com>:
mm/sparse: fix kernel crash with pfn_section_valid check
drivers/base/memory.c | 23 +++--------------------
include/linux/memcontrol.h | 12 ++++++++++++
kernel/fork.c | 4 ++--
mm/hugetlb_cgroup.c | 3 +--
mm/memcontrol.c | 38 ++++++++++++++++++++++++++++++++++++++
mm/sparse.c | 6 ++++++
mm/swapfile.c | 41 ++++++++++++++++++++---------------------
7 files changed, 82 insertions(+), 45 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-03-22 1:19 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-03-22 1:19 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
10 fixes, based on c63c50fc2ec9afc4de21ef9ead2eac64b178cce1:
Chunguang Xu <brookxu@tencent.com>:
memcg: fix NULL pointer dereference in __mem_cgroup_usage_unregister_event
Baoquan He <bhe@redhat.com>:
mm/hotplug: fix hot remove failure in SPARSEMEM|!VMEMMAP case
Qian Cai <cai@lca.pw>:
page-flags: fix a crash at SetPageError(THP_SWAP)
Chris Down <chris@chrisdown.name>:
mm, memcg: fix corruption on 64-bit divisor in memory.high throttling
mm, memcg: throttle allocators based on ancestral memory.high
Michal Hocko <mhocko@suse.com>:
mm: do not allow MADV_PAGEOUT for CoW pages
Roman Penyaev <rpenyaev@suse.de>:
epoll: fix possible lost wakeup on epoll_ctl() path
Qian Cai <cai@lca.pw>:
mm/mmu_notifier: silence PROVE_RCU_LIST warnings
Vlastimil Babka <vbabka@suse.cz>:
mm, slub: prevent kmalloc_node crashes and memory leaks
Joerg Roedel <jroedel@suse.de>:
x86/mm: split vmalloc_sync_all()
arch/x86/mm/fault.c | 26 ++++++++++-
drivers/acpi/apei/ghes.c | 2
fs/eventpoll.c | 8 +--
include/linux/page-flags.h | 2
include/linux/vmalloc.h | 5 +-
kernel/notifier.c | 2
mm/madvise.c | 12 +++--
mm/memcontrol.c | 105 ++++++++++++++++++++++++++++-----------------
mm/mmu_notifier.c | 27 +++++++----
mm/nommu.c | 10 +++-
mm/slub.c | 26 +++++++----
mm/sparse.c | 8 ++-
mm/vmalloc.c | 11 +++-
13 files changed, 165 insertions(+), 79 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-03-06 6:27 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-03-06 6:27 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
7 fixes, based on 9f65ed5fe41ce08ed1cb1f6a950f9ec694c142ad:
Mel Gorman <mgorman@techsingularity.net>:
mm, numa: fix bad pmd by atomically check for pmd_trans_huge when marking page tables prot_numa
Huang Ying <ying.huang@intel.com>:
mm: fix possible PMD dirty bit lost in set_pmd_migration_entry()
"Kirill A. Shutemov" <kirill@shutemov.name>:
mm: avoid data corruption on CoW fault into PFN-mapped VMA
OGAWA Hirofumi <hirofumi@mail.parknet.co.jp>:
fat: fix uninit-memory access for partial initialized inode
Sebastian Andrzej Siewior <bigeasy@linutronix.de>:
mm/z3fold.c: do not include rwlock.h directly
Vlastimil Babka <vbabka@suse.cz>:
mm, hotplug: fix page online with DEBUG_PAGEALLOC compiled but not enabled
Miroslav Benes <mbenes@suse.cz>:
arch/Kconfig: update HAVE_RELIABLE_STACKTRACE description
arch/Kconfig | 5 +++--
fs/fat/inode.c | 19 +++++++------------
include/linux/mm.h | 4 ++++
mm/huge_memory.c | 3 +--
mm/memory.c | 35 +++++++++++++++++++++++++++--------
mm/memory_hotplug.c | 8 +++++++-
mm/mprotect.c | 38 ++++++++++++++++++++++++++++++++++++--
mm/z3fold.c | 1 -
8 files changed, 85 insertions(+), 28 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2020-02-21 18:32 ` incoming Konstantin Ryabitsev
@ 2020-02-27 9:59 ` Vlastimil Babka
0 siblings, 0 replies; 409+ messages in thread
From: Vlastimil Babka @ 2020-02-27 9:59 UTC (permalink / raw)
To: Konstantin Ryabitsev, Linus Torvalds; +Cc: Andrew Morton, Linux-MM, mm-commits
On 2/21/20 7:32 PM, Konstantin Ryabitsev wrote:
> On Fri, Feb 21, 2020 at 10:21:19AM -0800, Linus Torvalds wrote:
>> On Thu, Feb 20, 2020 at 8:00 PM Andrew Morton <akpm@linux-foundation.org> wrote:
>> >
>> > - A few y2038 fixes which missed the merge window whiole dependencies
>> > in NFS were being sorted out.
>> >
>> > - A bunch of fixes. Some minor, some not.
>>
>> Hmm. Konstantin's nice lore script _used_ to pick up your patches, but
>> now they don't.
>>
>> I'm not sure what changed. It worked with your big series of 118 patches.
>>
>> It doesn't work with this smaller series of fixes.
>>
>> I think the difference is that you've done something bad to your patch
>> sending. That big series was properly threaded with each of the
>> patches being a reply to the 'incoming' message.
>>
>> This series is not.
>
> This is correct -- each patch is posted without an in-reply-to, so
> public-inbox doesn't group them into a thread.
>
> E.g.:
> https://lore.kernel.org/linux-mm/20200221040350.84HaG%25akpm@linux-foundation.org/
>
>>
>> Please, Andrew, can you make your email flow more consistent so that I
>> can actually use the nice new tool to download a patch series?
>
> Andrew, I'll be happy to provide you with a helper tool if you can
> describe me your workflow. E.g. if you have a quilt directory of patches
> plus a series file, it could easily be a tiny wrapper like:
>
> send-patches --base-commit 1234abcd --cover cover.txt patchdir/series
Once/if there is such tool, could it perhaps instead of mass e-mailing create
git commits, push them to korg repo and send a pull request?
Thanks,
Vlastimil
> -K
>
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2020-02-21 18:21 ` incoming Linus Torvalds
2020-02-21 18:32 ` incoming Konstantin Ryabitsev
@ 2020-02-21 19:33 ` Linus Torvalds
1 sibling, 0 replies; 409+ messages in thread
From: Linus Torvalds @ 2020-02-21 19:33 UTC (permalink / raw)
To: Andrew Morton, Konstantin Ryabitsev; +Cc: Linux-MM, mm-commits
Side note: I've obviously picked it up the old-fashioned way, but I
had been looking forward to seeing if I could just automate this more.
Linus
On Fri, Feb 21, 2020 at 10:21 AM Linus Torvalds
<torvalds@linux-foundation.org> wrote:
>
> Please, Andrew, can you make your email flow more consistent so that I
> can actually use the nice new tool to download a patch series?
>
> Linus
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2020-02-21 18:21 ` incoming Linus Torvalds
@ 2020-02-21 18:32 ` Konstantin Ryabitsev
2020-02-27 9:59 ` incoming Vlastimil Babka
2020-02-21 19:33 ` incoming Linus Torvalds
1 sibling, 1 reply; 409+ messages in thread
From: Konstantin Ryabitsev @ 2020-02-21 18:32 UTC (permalink / raw)
To: Linus Torvalds; +Cc: Andrew Morton, Linux-MM, mm-commits
On Fri, Feb 21, 2020 at 10:21:19AM -0800, Linus Torvalds wrote:
> On Thu, Feb 20, 2020 at 8:00 PM Andrew Morton <akpm@linux-foundation.org> wrote:
> >
> > - A few y2038 fixes which missed the merge window whiole dependencies
> > in NFS were being sorted out.
> >
> > - A bunch of fixes. Some minor, some not.
>
> Hmm. Konstantin's nice lore script _used_ to pick up your patches, but
> now they don't.
>
> I'm not sure what changed. It worked with your big series of 118 patches.
>
> It doesn't work with this smaller series of fixes.
>
> I think the difference is that you've done something bad to your patch
> sending. That big series was properly threaded with each of the
> patches being a reply to the 'incoming' message.
>
> This series is not.
This is correct -- each patch is posted without an in-reply-to, so
public-inbox doesn't group them into a thread.
E.g.:
https://lore.kernel.org/linux-mm/20200221040350.84HaG%25akpm@linux-foundation.org/
>
> Please, Andrew, can you make your email flow more consistent so that I
> can actually use the nice new tool to download a patch series?
Andrew, I'll be happy to provide you with a helper tool if you can
describe me your workflow. E.g. if you have a quilt directory of patches
plus a series file, it could easily be a tiny wrapper like:
send-patches --base-commit 1234abcd --cover cover.txt patchdir/series
-K
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2020-02-21 4:00 incoming Andrew Morton
2020-02-21 4:03 ` incoming Andrew Morton
@ 2020-02-21 18:21 ` Linus Torvalds
2020-02-21 18:32 ` incoming Konstantin Ryabitsev
2020-02-21 19:33 ` incoming Linus Torvalds
1 sibling, 2 replies; 409+ messages in thread
From: Linus Torvalds @ 2020-02-21 18:21 UTC (permalink / raw)
To: Andrew Morton, Konstantin Ryabitsev; +Cc: Linux-MM, mm-commits
On Thu, Feb 20, 2020 at 8:00 PM Andrew Morton <akpm@linux-foundation.org> wrote:
>
> - A few y2038 fixes which missed the merge window whiole dependencies
> in NFS were being sorted out.
>
> - A bunch of fixes. Some minor, some not.
Hmm. Konstantin's nice lore script _used_ to pick up your patches, but
now they don't.
I'm not sure what changed. It worked with your big series of 118 patches.
It doesn't work with this smaller series of fixes.
I think the difference is that you've done something bad to your patch
sending. That big series was properly threaded with each of the
patches being a reply to the 'incoming' message.
This series is not.
Please, Andrew, can you make your email flow more consistent so that I
can actually use the nice new tool to download a patch series?
Linus
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2020-02-21 4:00 incoming Andrew Morton
@ 2020-02-21 4:03 ` Andrew Morton
2020-02-21 18:21 ` incoming Linus Torvalds
1 sibling, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-02-21 4:03 UTC (permalink / raw)
To: Linus Torvalds, linux-mm, mm-commits
On Thu, 20 Feb 2020 20:00:30 -0800 Andrew Morton <akpm@linux-foundation.org> wrote:
> - A few y2038 fixes which missed the merge window whiole dependencies
> in NFS were being sorted out.
>
> - A bunch of fixes. Some minor, some not.
15 patches, based on ca7e1fd1026c5af6a533b4b5447e1d2f153e28f2
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-02-21 4:00 Andrew Morton
2020-02-21 4:03 ` incoming Andrew Morton
2020-02-21 18:21 ` incoming Linus Torvalds
0 siblings, 2 replies; 409+ messages in thread
From: Andrew Morton @ 2020-02-21 4:00 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
- A few y2038 fixes which missed the merge window whiole dependencies
in NFS were being sorted out.
- A bunch of fixes. Some minor, some not.
Subsystems affected by this patch series:
Arnd Bergmann <arnd@arndb.de>:
y2038: remove ktime to/from timespec/timeval conversion
y2038: remove unused time32 interfaces
y2038: hide timeval/timespec/itimerval/itimerspec types
Ioanna Alifieraki <ioanna-maria.alifieraki@canonical.com>:
Revert "ipc,sem: remove uneeded sem_undo_list lock usage in exit_sem()"
Christian Borntraeger <borntraeger@de.ibm.com>:
include/uapi/linux/swab.h: fix userspace breakage, use __BITS_PER_LONG for swap
SeongJae Park <sjpark@amazon.de>:
selftests/vm: add missed tests in run_vmtests
Joe Perches <joe@perches.com>:
get_maintainer: remove uses of P: for maintainer name
Douglas Anderson <dianders@chromium.org>:
scripts/get_maintainer.pl: deprioritize old Fixes: addresses
Christoph Hellwig <hch@lst.de>:
mm/swapfile.c: fix a comment in sys_swapon()
Vasily Averin <vvs@virtuozzo.com>:
mm/memcontrol.c: lost css_put in memcg_expand_shrinker_maps()
Alexandru Ardelean <alexandru.ardelean@analog.com>:
lib/string.c: update match_string() doc-strings with correct behavior
Gavin Shan <gshan@redhat.com>:
mm/vmscan.c: don't round up scan size for online memory cgroup
Wei Yang <richardw.yang@linux.intel.com>:
mm/sparsemem: pfn_to_page is not valid yet on SPARSEMEM
Alexander Potapenko <glider@google.com>:
lib/stackdepot.c: fix global out-of-bounds in stack_slabs
Randy Dunlap <rdunlap@infradead.org>:
MAINTAINERS: use tabs for SAFESETID
MAINTAINERS | 8 -
include/linux/compat.h | 29 ------
include/linux/ktime.h | 37 -------
include/linux/time32.h | 154 ---------------------------------
include/linux/timekeeping32.h | 32 ------
include/linux/types.h | 5 -
include/uapi/asm-generic/posix_types.h | 2
include/uapi/linux/swab.h | 4
include/uapi/linux/time.h | 22 ++--
ipc/sem.c | 6 -
kernel/compat.c | 64 -------------
kernel/time/time.c | 43 ---------
lib/stackdepot.c | 8 +
lib/string.c | 16 +++
mm/memcontrol.c | 4
mm/sparse.c | 2
mm/swapfile.c | 2
mm/vmscan.c | 9 +
scripts/get_maintainer.pl | 32 ------
tools/testing/selftests/vm/run_vmtests | 33 +++++++
20 files changed, 93 insertions(+), 419 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2020-02-04 2:46 ` incoming Andrew Morton
@ 2020-02-04 3:11 ` Linus Torvalds
0 siblings, 0 replies; 409+ messages in thread
From: Linus Torvalds @ 2020-02-04 3:11 UTC (permalink / raw)
To: Andrew Morton; +Cc: mm-commits, Linux-MM
On Tue, Feb 4, 2020 at 2:46 AM Andrew Morton <akpm@linux-foundation.org> wrote:
>
> On Tue, 4 Feb 2020 02:27:48 +0000 Linus Torvalds <torvalds@linux-foundation.org> wrote:
>
> > What's the base? You've changed your scripts or something, and that
> > information is no longer in your cover letter..
>
> Crap, sorry, geriatric.
>
> d4e9056daedca3891414fe3c91de3449a5dad0f2
Ok, I've tentatively applied it with the MIME decoding fixes I found,
and I'll guess I'll let it build and sit for a while before merging it
into my tree.
I didn't find anything else odd in there. But...
Linus
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2020-02-04 2:27 ` incoming Linus Torvalds
@ 2020-02-04 2:46 ` Andrew Morton
2020-02-04 3:11 ` incoming Linus Torvalds
0 siblings, 1 reply; 409+ messages in thread
From: Andrew Morton @ 2020-02-04 2:46 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, Linux-MM
On Tue, 4 Feb 2020 02:27:48 +0000 Linus Torvalds <torvalds@linux-foundation.org> wrote:
> On Tue, Feb 4, 2020 at 1:33 AM Andrew Morton <akpm@linux-foundation.org> wrote:
> >
> > The rest of MM and the rest of everything else.
>
> What's the base? You've changed your scripts or something, and that
> information is no longer in your cover letter..
>
Crap, sorry, geriatric.
d4e9056daedca3891414fe3c91de3449a5dad0f2
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2020-02-04 1:33 incoming Andrew Morton
@ 2020-02-04 2:27 ` Linus Torvalds
2020-02-04 2:46 ` incoming Andrew Morton
0 siblings, 1 reply; 409+ messages in thread
From: Linus Torvalds @ 2020-02-04 2:27 UTC (permalink / raw)
To: Andrew Morton; +Cc: mm-commits, Linux-MM
On Tue, Feb 4, 2020 at 1:33 AM Andrew Morton <akpm@linux-foundation.org> wrote:
>
> The rest of MM and the rest of everything else.
What's the base? You've changed your scripts or something, and that
information is no longer in your cover letter..
Linus
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-02-04 1:33 Andrew Morton
2020-02-04 2:27 ` incoming Linus Torvalds
0 siblings, 1 reply; 409+ messages in thread
From: Andrew Morton @ 2020-02-04 1:33 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
The rest of MM and the rest of everything else.
Subsystems affected by this patch series:
hotfixes
mm/pagealloc
mm/memory-hotplug
ipc
misc
mm/cleanups
mm/pagemap
procfs
lib
cleanups
arm
Subsystem: hotfixes
Gang He <GHe@suse.com>:
ocfs2: fix oops when writing cloned file
David Hildenbrand <david@redhat.com>:
Patch series "mm: fix max_pfn not falling on section boundary", v2:
mm/page_alloc.c: fix uninitialized memmaps on a partially populated last section
fs/proc/page.c: allow inspection of last section and fix end detection
mm/page_alloc.c: initialize memmap of unavailable memory directly
Subsystem: mm/pagealloc
David Hildenbrand <david@redhat.com>:
mm/page_alloc: fix and rework pfn handling in memmap_init_zone()
mm: factor out next_present_section_nr()
Subsystem: mm/memory-hotplug
"Aneesh Kumar K.V" <aneesh.kumar@linux.ibm.com>:
Patch series "mm/memory_hotplug: Shrink zones before removing memory", v6:
mm/memmap_init: update variable name in memmap_init_zone
David Hildenbrand <david@redhat.com>:
mm/memory_hotplug: poison memmap in remove_pfn_range_from_zone()
mm/memory_hotplug: we always have a zone in find_(smallest|biggest)_section_pfn
mm/memory_hotplug: don't check for "all holes" in shrink_zone_span()
mm/memory_hotplug: drop local variables in shrink_zone_span()
mm/memory_hotplug: cleanup __remove_pages()
mm/memory_hotplug: drop valid_start/valid_end from test_pages_in_a_zone()
Subsystem: ipc
Manfred Spraul <manfred@colorfullife.com>:
smp_mb__{before,after}_atomic(): update Documentation
Davidlohr Bueso <dave@stgolabs.net>:
ipc/mqueue.c: remove duplicated code
Manfred Spraul <manfred@colorfullife.com>:
ipc/mqueue.c: update/document memory barriers
ipc/msg.c: update and document memory barriers
ipc/sem.c: document and update memory barriers
Lu Shuaibing <shuaibinglu@126.com>:
ipc/msg.c: consolidate all xxxctl_down() functions
drivers/block/null_blk_main.c: fix layout
Subsystem: misc
Andrew Morton <akpm@linux-foundation.org>:
drivers/block/null_blk_main.c: fix layout
drivers/block/null_blk_main.c: fix uninitialized var warnings
Randy Dunlap <rdunlap@infradead.org>:
pinctrl: fix pxa2xx.c build warnings
Subsystem: mm/cleanups
Florian Westphal <fw@strlen.de>:
mm: remove __krealloc
Subsystem: mm/pagemap
Steven Price <steven.price@arm.com>:
Patch series "Generic page walk and ptdump", v17:
mm: add generic p?d_leaf() macros
arc: mm: add p?d_leaf() definitions
arm: mm: add p?d_leaf() definitions
arm64: mm: add p?d_leaf() definitions
mips: mm: add p?d_leaf() definitions
powerpc: mm: add p?d_leaf() definitions
riscv: mm: add p?d_leaf() definitions
s390: mm: add p?d_leaf() definitions
sparc: mm: add p?d_leaf() definitions
x86: mm: add p?d_leaf() definitions
mm: pagewalk: add p4d_entry() and pgd_entry()
mm: pagewalk: allow walking without vma
mm: pagewalk: don't lock PTEs for walk_page_range_novma()
mm: pagewalk: fix termination condition in walk_pte_range()
mm: pagewalk: add 'depth' parameter to pte_hole
x86: mm: point to struct seq_file from struct pg_state
x86: mm+efi: convert ptdump_walk_pgd_level() to take a mm_struct
x86: mm: convert ptdump_walk_pgd_level_debugfs() to take an mm_struct
mm: add generic ptdump
x86: mm: convert dump_pagetables to use walk_page_range
arm64: mm: convert mm/dump.c to use walk_page_range()
arm64: mm: display non-present entries in ptdump
mm: ptdump: reduce level numbers by 1 in note_page()
x86: mm: avoid allocating struct mm_struct on the stack
"Aneesh Kumar K.V" <aneesh.kumar@linux.ibm.com>:
Patch series "Fixup page directory freeing", v4:
powerpc/mmu_gather: enable RCU_TABLE_FREE even for !SMP case
Peter Zijlstra <peterz@infradead.org>:
mm/mmu_gather: invalidate TLB correctly on batch allocation failure and flush
asm-generic/tlb: avoid potential double flush
asm-gemeric/tlb: remove stray function declarations
asm-generic/tlb: add missing CONFIG symbol
asm-generic/tlb: rename HAVE_RCU_TABLE_FREE
asm-generic/tlb: rename HAVE_MMU_GATHER_PAGE_SIZE
asm-generic/tlb: rename HAVE_MMU_GATHER_NO_GATHER
asm-generic/tlb: provide MMU_GATHER_TABLE_FREE
Subsystem: procfs
Alexey Dobriyan <adobriyan@gmail.com>:
proc: decouple proc from VFS with "struct proc_ops"
proc: convert everything to "struct proc_ops"
Subsystem: lib
Yury Norov <yury.norov@gmail.com>:
Patch series "lib: rework bitmap_parse", v5:
lib/string: add strnchrnul()
bitops: more BITS_TO_* macros
lib: add test for bitmap_parse()
lib: make bitmap_parse_user a wrapper on bitmap_parse
lib: rework bitmap_parse()
lib: new testcases for bitmap_parse{_user}
include/linux/cpumask.h: don't calculate length of the input string
Subsystem: cleanups
Masahiro Yamada <masahiroy@kernel.org>:
treewide: remove redundant IS_ERR() before error code check
Subsystem: arm
Chen-Yu Tsai <wens@csie.org>:
ARM: dma-api: fix max_pfn off-by-one error in __dma_supported()
Documentation/memory-barriers.txt | 14
arch/Kconfig | 17
arch/alpha/kernel/srm_env.c | 17
arch/arc/include/asm/pgtable.h | 1
arch/arm/Kconfig | 2
arch/arm/include/asm/pgtable-2level.h | 1
arch/arm/include/asm/pgtable-3level.h | 1
arch/arm/include/asm/tlb.h | 6
arch/arm/kernel/atags_proc.c | 8
arch/arm/mm/alignment.c | 14
arch/arm/mm/dma-mapping.c | 2
arch/arm64/Kconfig | 3
arch/arm64/Kconfig.debug | 19
arch/arm64/include/asm/pgtable.h | 2
arch/arm64/include/asm/ptdump.h | 8
arch/arm64/mm/Makefile | 4
arch/arm64/mm/dump.c | 152 ++----
arch/arm64/mm/mmu.c | 4
arch/arm64/mm/ptdump_debugfs.c | 2
arch/ia64/kernel/salinfo.c | 24 -
arch/m68k/kernel/bootinfo_proc.c | 8
arch/mips/include/asm/pgtable.h | 5
arch/mips/lasat/picvue_proc.c | 31 -
arch/powerpc/Kconfig | 7
arch/powerpc/include/asm/book3s/32/pgalloc.h | 8
arch/powerpc/include/asm/book3s/64/pgalloc.h | 2
arch/powerpc/include/asm/book3s/64/pgtable.h | 3
arch/powerpc/include/asm/nohash/pgalloc.h | 8
arch/powerpc/include/asm/tlb.h | 11
arch/powerpc/kernel/proc_powerpc.c | 10
arch/powerpc/kernel/rtas-proc.c | 70 +--
arch/powerpc/kernel/rtas_flash.c | 34 -
arch/powerpc/kernel/rtasd.c | 14
arch/powerpc/mm/book3s64/pgtable.c | 7
arch/powerpc/mm/numa.c | 12
arch/powerpc/platforms/pseries/lpar.c | 24 -
arch/powerpc/platforms/pseries/lparcfg.c | 14
arch/powerpc/platforms/pseries/reconfig.c | 8
arch/powerpc/platforms/pseries/scanlog.c | 15
arch/riscv/include/asm/pgtable-64.h | 7
arch/riscv/include/asm/pgtable.h | 7
arch/s390/Kconfig | 4
arch/s390/include/asm/pgtable.h | 2
arch/sh/mm/alignment.c | 17
arch/sparc/Kconfig | 3
arch/sparc/include/asm/pgtable_64.h | 2
arch/sparc/include/asm/tlb_64.h | 11
arch/sparc/kernel/led.c | 15
arch/um/drivers/mconsole_kern.c | 9
arch/um/kernel/exitcode.c | 15
arch/um/kernel/process.c | 15
arch/x86/Kconfig | 3
arch/x86/Kconfig.debug | 20
arch/x86/include/asm/pgtable.h | 10
arch/x86/include/asm/tlb.h | 4
arch/x86/kernel/cpu/mtrr/if.c | 21
arch/x86/mm/Makefile | 4
arch/x86/mm/debug_pagetables.c | 18
arch/x86/mm/dump_pagetables.c | 418 +++++-------------
arch/x86/platform/efi/efi_32.c | 2
arch/x86/platform/efi/efi_64.c | 4
arch/x86/platform/uv/tlb_uv.c | 14
arch/xtensa/platforms/iss/simdisk.c | 10
crypto/af_alg.c | 2
drivers/acpi/battery.c | 15
drivers/acpi/proc.c | 15
drivers/acpi/scan.c | 2
drivers/base/memory.c | 9
drivers/block/null_blk_main.c | 58 +-
drivers/char/hw_random/bcm2835-rng.c | 2
drivers/char/hw_random/omap-rng.c | 4
drivers/clk/clk.c | 2
drivers/dma/mv_xor_v2.c | 2
drivers/firmware/efi/arm-runtime.c | 2
drivers/gpio/gpiolib-devres.c | 2
drivers/gpio/gpiolib-of.c | 8
drivers/gpio/gpiolib.c | 2
drivers/hwmon/dell-smm-hwmon.c | 15
drivers/i2c/busses/i2c-mv64xxx.c | 5
drivers/i2c/busses/i2c-synquacer.c | 2
drivers/ide/ide-proc.c | 19
drivers/input/input.c | 28 -
drivers/isdn/capi/kcapi_proc.c | 6
drivers/macintosh/via-pmu.c | 17
drivers/md/md.c | 15
drivers/misc/sgi-gru/gruprocfs.c | 42 -
drivers/mtd/ubi/build.c | 2
drivers/net/wireless/cisco/airo.c | 126 ++---
drivers/net/wireless/intel/ipw2x00/libipw_module.c | 15
drivers/net/wireless/intersil/hostap/hostap_hw.c | 4
drivers/net/wireless/intersil/hostap/hostap_proc.c | 14
drivers/net/wireless/intersil/hostap/hostap_wlan.h | 2
drivers/net/wireless/ray_cs.c | 20
drivers/of/device.c | 2
drivers/parisc/led.c | 17
drivers/pci/controller/pci-tegra.c | 2
drivers/pci/proc.c | 25 -
drivers/phy/phy-core.c | 4
drivers/pinctrl/pxa/pinctrl-pxa2xx.c | 1
drivers/platform/x86/thinkpad_acpi.c | 15
drivers/platform/x86/toshiba_acpi.c | 60 +-
drivers/pnp/isapnp/proc.c | 9
drivers/pnp/pnpbios/proc.c | 17
drivers/s390/block/dasd_proc.c | 15
drivers/s390/cio/blacklist.c | 14
drivers/s390/cio/css.c | 11
drivers/scsi/esas2r/esas2r_main.c | 9
drivers/scsi/scsi_devinfo.c | 15
drivers/scsi/scsi_proc.c | 29 -
drivers/scsi/sg.c | 30 -
drivers/spi/spi-orion.c | 3
drivers/staging/rtl8192u/ieee80211/ieee80211_module.c | 14
drivers/tty/sysrq.c | 8
drivers/usb/gadget/function/rndis.c | 17
drivers/video/fbdev/imxfb.c | 2
drivers/video/fbdev/via/viafbdev.c | 105 ++--
drivers/zorro/proc.c | 9
fs/cifs/cifs_debug.c | 108 ++--
fs/cifs/dfs_cache.c | 13
fs/cifs/dfs_cache.h | 2
fs/ext4/super.c | 2
fs/f2fs/node.c | 2
fs/fscache/internal.h | 2
fs/fscache/object-list.c | 11
fs/fscache/proc.c | 2
fs/jbd2/journal.c | 13
fs/jfs/jfs_debug.c | 14
fs/lockd/procfs.c | 12
fs/nfsd/nfsctl.c | 13
fs/nfsd/stats.c | 12
fs/ocfs2/file.c | 14
fs/ocfs2/suballoc.c | 2
fs/proc/cpuinfo.c | 12
fs/proc/generic.c | 38 -
fs/proc/inode.c | 76 +--
fs/proc/internal.h | 5
fs/proc/kcore.c | 13
fs/proc/kmsg.c | 14
fs/proc/page.c | 54 +-
fs/proc/proc_net.c | 32 -
fs/proc/proc_sysctl.c | 2
fs/proc/root.c | 2
fs/proc/stat.c | 12
fs/proc/task_mmu.c | 4
fs/proc/vmcore.c | 10
fs/sysfs/group.c | 2
include/asm-generic/pgtable.h | 20
include/asm-generic/tlb.h | 138 +++--
include/linux/bitmap.h | 8
include/linux/bitops.h | 4
include/linux/cpumask.h | 4
include/linux/memory_hotplug.h | 4
include/linux/mm.h | 6
include/linux/mmzone.h | 10
include/linux/pagewalk.h | 49 +-
include/linux/proc_fs.h | 23
include/linux/ptdump.h | 24 -
include/linux/seq_file.h | 13
include/linux/slab.h | 1
include/linux/string.h | 1
include/linux/sunrpc/stats.h | 4
ipc/mqueue.c | 123 ++++-
ipc/msg.c | 62 +-
ipc/sem.c | 66 +-
ipc/util.c | 14
kernel/configs.c | 9
kernel/irq/proc.c | 42 -
kernel/kallsyms.c | 12
kernel/latencytop.c | 14
kernel/locking/lockdep_proc.c | 15
kernel/module.c | 12
kernel/profile.c | 24 -
kernel/sched/psi.c | 48 +-
lib/bitmap.c | 195 ++++----
lib/string.c | 17
lib/test_bitmap.c | 105 ++++
mm/Kconfig.debug | 21
mm/Makefile | 1
mm/gup.c | 2
mm/hmm.c | 66 +-
mm/memory_hotplug.c | 104 +---
mm/memremap.c | 2
mm/migrate.c | 5
mm/mincore.c | 1
mm/mmu_gather.c | 158 ++++--
mm/page_alloc.c | 75 +--
mm/pagewalk.c | 167 +++++--
mm/ptdump.c | 159 ++++++
mm/slab_common.c | 37 -
mm/sparse.c | 10
mm/swapfile.c | 14
net/atm/mpoa_proc.c | 17
net/atm/proc.c | 8
net/core/dev.c | 2
net/core/filter.c | 2
net/core/pktgen.c | 44 -
net/ipv4/ipconfig.c | 10
net/ipv4/netfilter/ipt_CLUSTERIP.c | 16
net/ipv4/route.c | 24 -
net/netfilter/xt_recent.c | 17
net/sunrpc/auth_gss/svcauth_gss.c | 10
net/sunrpc/cache.c | 45 -
net/sunrpc/stats.c | 21
net/xfrm/xfrm_policy.c | 2
samples/kfifo/bytestream-example.c | 11
samples/kfifo/inttype-example.c | 11
samples/kfifo/record-example.c | 11
scripts/coccinelle/free/devm_free.cocci | 4
sound/core/info.c | 34 -
sound/soc/codecs/ak4104.c | 3
sound/soc/codecs/cs4270.c | 3
sound/soc/codecs/tlv320aic32x4.c | 6
sound/soc/sunxi/sun4i-spdif.c | 2
tools/include/linux/bitops.h | 9
214 files changed, 2589 insertions(+), 2227 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-01-31 6:10 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-01-31 6:10 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
Most of -mm and quite a number of other subsystems.
MM is fairly quiet this time. Holidays, I assume.
119 patches, based on 39bed42de2e7d74686a2d5a45638d6a5d7e7d473:
Subsystems affected by this patch series:
hotfixes
scripts
ocfs2
mm/slub
mm/kmemleak
mm/debug
mm/pagecache
mm/gup
mm/swap
mm/memcg
mm/pagemap
mm/tracing
mm/kasan
mm/initialization
mm/pagealloc
mm/vmscan
mm/tools
mm/memblock
mm/oom-kill
mm/hugetlb
mm/migration
mm/mmap
mm/memory-hotplug
mm/zswap
mm/cleanups
mm/zram
misc
lib
binfmt
init
reiserfs
exec
dma-mapping
kcov
Subsystem: hotfixes
Andy Shevchenko <andriy.shevchenko@linux.intel.com>:
lib/test_bitmap: correct test data offsets for 32-bit
"Theodore Ts'o" <tytso@mit.edu>:
memcg: fix a crash in wb_workfn when a device disappears
Dan Carpenter <dan.carpenter@oracle.com>:
mm/mempolicy.c: fix out of bounds write in mpol_parse_str()
Pingfan Liu <kernelfans@gmail.com>:
mm/sparse.c: reset section's mem_map when fully deactivated
Wei Yang <richardw.yang@linux.intel.com>:
mm/migrate.c: also overwrite error when it is bigger than zero
Dan Williams <dan.j.williams@intel.com>:
mm/memory_hotplug: fix remove_memory() lockdep splat
Wei Yang <richardw.yang@linux.intel.com>:
mm: thp: don't need care deferred split queue in memcg charge move path
Yang Shi <yang.shi@linux.alibaba.com>:
mm: move_pages: report the number of non-attempted pages
Subsystem: scripts
Xiong <xndchn@gmail.com>:
scripts/spelling.txt: add more spellings to spelling.txt
Luca Ceresoli <luca@lucaceresoli.net>:
scripts/spelling.txt: add "issus" typo
Subsystem: ocfs2
Aditya Pakki <pakki001@umn.edu>:
fs: ocfs: remove unnecessary assertion in dlm_migrate_lockres
zhengbin <zhengbin13@huawei.com>:
ocfs2: remove unneeded semicolons
Masahiro Yamada <masahiroy@kernel.org>:
ocfs2: make local header paths relative to C files
Colin Ian King <colin.king@canonical.com>:
ocfs2/dlm: remove redundant assignment to ret
Andy Shevchenko <andriy.shevchenko@linux.intel.com>:
ocfs2/dlm: move BITS_TO_BYTES() to bitops.h for wider use
wangyan <wangyan122@huawei.com>:
ocfs2: fix a NULL pointer dereference when call ocfs2_update_inode_fsync_trans()
ocfs2: use ocfs2_update_inode_fsync_trans() to access t_tid in handle->h_transaction
Subsystem: mm/slub
Yu Zhao <yuzhao@google.com>:
mm/slub.c: avoid slub allocation while holding list_lock
Subsystem: mm/kmemleak
He Zhe <zhe.he@windriver.com>:
mm/kmemleak: turn kmemleak_lock and object->lock to raw_spinlock_t
Subsystem: mm/debug
Vlastimil Babka <vbabka@suse.cz>:
mm/debug.c: always print flags in dump_page()
Subsystem: mm/pagecache
Ira Weiny <ira.weiny@intel.com>:
mm/filemap.c: clean up filemap_write_and_wait()
Subsystem: mm/gup
Qiujun Huang <hqjagain@gmail.com>:
mm: fix gup_pud_range
Wei Yang <richardw.yang@linux.intel.com>:
mm/gup.c: use is_vm_hugetlb_page() to check whether to follow huge
John Hubbard <jhubbard@nvidia.com>:
Patch series "mm/gup: prereqs to track dma-pinned pages: FOLL_PIN", v12:
mm/gup: factor out duplicate code from four routines
mm/gup: move try_get_compound_head() to top, fix minor issues
Dan Williams <dan.j.williams@intel.com>:
mm: Cleanup __put_devmap_managed_page() vs ->page_free()
John Hubbard <jhubbard@nvidia.com>:
mm: devmap: refactor 1-based refcounting for ZONE_DEVICE pages
goldish_pipe: rename local pin_user_pages() routine
mm: fix get_user_pages_remote()'s handling of FOLL_LONGTERM
vfio: fix FOLL_LONGTERM use, simplify get_user_pages_remote() call
mm/gup: allow FOLL_FORCE for get_user_pages_fast()
IB/umem: use get_user_pages_fast() to pin DMA pages
media/v4l2-core: set pages dirty upon releasing DMA buffers
mm/gup: introduce pin_user_pages*() and FOLL_PIN
goldish_pipe: convert to pin_user_pages() and put_user_page()
IB/{core,hw,umem}: set FOLL_PIN via pin_user_pages*(), fix up ODP
mm/process_vm_access: set FOLL_PIN via pin_user_pages_remote()
drm/via: set FOLL_PIN via pin_user_pages_fast()
fs/io_uring: set FOLL_PIN via pin_user_pages()
net/xdp: set FOLL_PIN via pin_user_pages()
media/v4l2-core: pin_user_pages (FOLL_PIN) and put_user_page() conversion
vfio, mm: pin_user_pages (FOLL_PIN) and put_user_page() conversion
powerpc: book3s64: convert to pin_user_pages() and put_user_page()
mm/gup_benchmark: use proper FOLL_WRITE flags instead of hard-coding "1"
mm, tree-wide: rename put_user_page*() to unpin_user_page*()
Subsystem: mm/swap
Vasily Averin <vvs@virtuozzo.com>:
mm/swapfile.c: swap_next should increase position index
Subsystem: mm/memcg
Kaitao Cheng <pilgrimtao@gmail.com>:
mm/memcontrol.c: cleanup some useless code
Subsystem: mm/pagemap
Li Xinhai <lixinhai.lxh@gmail.com>:
mm/page_vma_mapped.c: explicitly compare pfn for normal, hugetlbfs and THP page
Subsystem: mm/tracing
Junyong Sun <sunjy516@gmail.com>:
mm, tracing: print symbol name for kmem_alloc_node call_site events
Subsystem: mm/kasan
"Gustavo A. R. Silva" <gustavo@embeddedor.com>:
lib/test_kasan.c: fix memory leak in kmalloc_oob_krealloc_more()
Subsystem: mm/initialization
Andy Shevchenko <andriy.shevchenko@linux.intel.com>:
mm/early_ioremap.c: use %pa to print resource_size_t variables
Subsystem: mm/pagealloc
"Kirill A. Shutemov" <kirill@shutemov.name>:
mm/page_alloc: skip non present sections on zone initialization
David Hildenbrand <david@redhat.com>:
mm: remove the memory isolate notifier
mm: remove "count" parameter from has_unmovable_pages()
Subsystem: mm/vmscan
Liu Song <liu.song11@zte.com.cn>:
mm/vmscan.c: remove unused return value of shrink_node
Alex Shi <alex.shi@linux.alibaba.com>:
mm/vmscan: remove prefetch_prev_lru_page
mm/vmscan: remove unused RECLAIM_OFF/RECLAIM_ZONE
Subsystem: mm/tools
Daniel Wagner <dwagner@suse.de>:
tools/vm/slabinfo: fix sanity checks enabling
Subsystem: mm/memblock
Anshuman Khandual <anshuman.khandual@arm.com>:
mm/memblock: define memblock_physmem_add()
memblock: Use __func__ in remaining memblock_dbg() call sites
Subsystem: mm/oom-kill
David Rientjes <rientjes@google.com>:
mm, oom: dump stack of victim when reaping failed
Subsystem: mm/hugetlb
Wei Yang <richardw.yang@linux.intel.com>:
mm/huge_memory.c: use head to check huge zero page
mm/huge_memory.c: use head to emphasize the purpose of page
mm/huge_memory.c: reduce critical section protected by split_queue_lock
Subsystem: mm/migration
Ralph Campbell <rcampbell@nvidia.com>:
mm/migrate: remove useless mask of start address
mm/migrate: clean up some minor coding style
mm/migrate: add stable check in migrate_vma_insert_page()
David Rientjes <rientjes@google.com>:
mm, thp: fix defrag setting if newline is not used
Subsystem: mm/mmap
Miaohe Lin <linmiaohe@huawei.com>:
mm/mmap.c: get rid of odd jump labels in find_mergeable_anon_vma()
Subsystem: mm/memory-hotplug
David Hildenbrand <david@redhat.com>:
Patch series "mm/memory_hotplug: pass in nid to online_pages()":
mm/memory_hotplug: pass in nid to online_pages()
Qian Cai <cai@lca.pw>:
mm/hotplug: silence a lockdep splat with printk()
mm/page_isolation: fix potential warning from user
Subsystem: mm/zswap
Vitaly Wool <vitaly.wool@konsulko.com>:
mm/zswap.c: add allocation hysteresis if pool limit is hit
Dan Carpenter <dan.carpenter@oracle.com>:
zswap: potential NULL dereference on error in init_zswap()
Subsystem: mm/cleanups
Yu Zhao <yuzhao@google.com>:
include/linux/mm.h: clean up obsolete check on space in page->flags
Wei Yang <richardw.yang@linux.intel.com>:
include/linux/mm.h: remove dead code totalram_pages_set()
Anshuman Khandual <anshuman.khandual@arm.com>:
include/linux/memory.h: drop fields 'hw' and 'phys_callback' from struct memory_block
Hao Lee <haolee.swjtu@gmail.com>:
mm: fix comments related to node reclaim
Subsystem: mm/zram
Taejoon Song <taejoon.song@lge.com>:
zram: try to avoid worst-case scenario on same element pages
Colin Ian King <colin.king@canonical.com>:
drivers/block/zram/zram_drv.c: fix error return codes not being returned in writeback_store
Subsystem: misc
Akinobu Mita <akinobu.mita@gmail.com>:
Patch series "add header file for kelvin to/from Celsius conversion:
include/linux/units.h: add helpers for kelvin to/from Celsius conversion
ACPI: thermal: switch to use <linux/units.h> helpers
platform/x86: asus-wmi: switch to use <linux/units.h> helpers
platform/x86: intel_menlow: switch to use <linux/units.h> helpers
thermal: int340x: switch to use <linux/units.h> helpers
thermal: intel_pch: switch to use <linux/units.h> helpers
nvme: hwmon: switch to use <linux/units.h> helpers
thermal: remove kelvin to/from Celsius conversion helpers from <linux/thermal.h>
iwlegacy: use <linux/units.h> helpers
iwlwifi: use <linux/units.h> helpers
thermal: armada: remove unused TO_MCELSIUS macro
iio: adc: qcom-vadc-common: use <linux/units.h> helpers
Subsystem: lib
Mikhail Zaslonko <zaslonko@linux.ibm.com>:
Patch series "S390 hardware support for kernel zlib", v3:
lib/zlib: add s390 hardware support for kernel zlib_deflate
s390/boot: rename HEAP_SIZE due to name collision
lib/zlib: add s390 hardware support for kernel zlib_inflate
s390/boot: add dfltcc= kernel command line parameter
lib/zlib: add zlib_deflate_dfltcc_enabled() function
btrfs: use larger zlib buffer for s390 hardware compression
Nathan Chancellor <natechancellor@gmail.com>:
lib/scatterlist.c: adjust indentation in __sg_alloc_table
Yury Norov <yury.norov@gmail.com>:
uapi: rename ext2_swab() to swab() and share globally in swab.h
lib/find_bit.c: join _find_next_bit{_le}
lib/find_bit.c: uninline helper _find_next_bit()
Subsystem: binfmt
Alexey Dobriyan <adobriyan@gmail.com>:
fs/binfmt_elf.c: smaller code generation around auxv vector fill
fs/binfmt_elf.c: fix ->start_code calculation
fs/binfmt_elf.c: don't copy ELF header around
fs/binfmt_elf.c: better codegen around current->mm
fs/binfmt_elf.c: make BAD_ADDR() unlikely
fs/binfmt_elf.c: coredump: allocate core ELF header on stack
fs/binfmt_elf.c: coredump: delete duplicated overflow check
fs/binfmt_elf.c: coredump: allow process with empty address space to coredump
Subsystem: init
Arvind Sankar <nivedita@alum.mit.edu>:
init/main.c: log arguments and environment passed to init
init/main.c: remove unnecessary repair_env_string in do_initcall_level
Patch series "init/main.c: minor cleanup/bugfix of envvar handling", v2:
init/main.c: fix quoted value handling in unknown_bootoption
Christophe Leroy <christophe.leroy@c-s.fr>:
init/main.c: fix misleading "This architecture does not have kernel memory protection" message
Subsystem: reiserfs
Yunfeng Ye <yeyunfeng@huawei.com>:
reiserfs: prevent NULL pointer dereference in reiserfs_insert_item()
Subsystem: exec
Alexey Dobriyan <adobriyan@gmail.com>:
execve: warn if process starts with executable stack
Subsystem: dma-mapping
Andy Shevchenko <andriy.shevchenko@linux.intel.com>:
include/linux/io-mapping.h-mapping: use PHYS_PFN() macro in io_mapping_map_atomic_wc()
Subsystem: kcov
Dmitry Vyukov <dvyukov@google.com>:
kcov: ignore fault-inject and stacktrace
Documentation/admin-guide/kernel-parameters.txt | 12
Documentation/core-api/index.rst | 1
Documentation/core-api/pin_user_pages.rst | 234 +++++
Documentation/vm/zswap.rst | 13
arch/powerpc/mm/book3s64/iommu_api.c | 14
arch/s390/boot/compressed/decompressor.c | 8
arch/s390/boot/ipl_parm.c | 14
arch/s390/include/asm/setup.h | 7
arch/s390/kernel/setup.c | 14
drivers/acpi/thermal.c | 34
drivers/base/memory.c | 25
drivers/block/zram/zram_drv.c | 10
drivers/gpu/drm/via/via_dmablit.c | 6
drivers/iio/adc/qcom-vadc-common.c | 6
drivers/iio/adc/qcom-vadc-common.h | 1
drivers/infiniband/core/umem.c | 21
drivers/infiniband/core/umem_odp.c | 13
drivers/infiniband/hw/hfi1/user_pages.c | 4
drivers/infiniband/hw/mthca/mthca_memfree.c | 8
drivers/infiniband/hw/qib/qib_user_pages.c | 4
drivers/infiniband/hw/qib/qib_user_sdma.c | 8
drivers/infiniband/hw/usnic/usnic_uiom.c | 4
drivers/infiniband/sw/siw/siw_mem.c | 4
drivers/media/v4l2-core/videobuf-dma-sg.c | 20
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init.h | 1
drivers/net/wireless/intel/iwlegacy/4965-mac.c | 3
drivers/net/wireless/intel/iwlegacy/4965.c | 17
drivers/net/wireless/intel/iwlegacy/common.h | 3
drivers/net/wireless/intel/iwlwifi/dvm/dev.h | 5
drivers/net/wireless/intel/iwlwifi/dvm/devices.c | 6
drivers/nvdimm/pmem.c | 6
drivers/nvme/host/hwmon.c | 13
drivers/platform/goldfish/goldfish_pipe.c | 39
drivers/platform/x86/asus-wmi.c | 7
drivers/platform/x86/intel_menlow.c | 9
drivers/thermal/armada_thermal.c | 2
drivers/thermal/intel/int340x_thermal/int340x_thermal_zone.c | 7
drivers/thermal/intel/intel_pch_thermal.c | 3
drivers/vfio/vfio_iommu_type1.c | 39
fs/binfmt_elf.c | 154 +--
fs/btrfs/compression.c | 2
fs/btrfs/zlib.c | 135 ++
fs/exec.c | 5
fs/fs-writeback.c | 2
fs/io_uring.c | 6
fs/ocfs2/cluster/quorum.c | 2
fs/ocfs2/dlm/Makefile | 2
fs/ocfs2/dlm/dlmast.c | 8
fs/ocfs2/dlm/dlmcommon.h | 4
fs/ocfs2/dlm/dlmconvert.c | 8
fs/ocfs2/dlm/dlmdebug.c | 8
fs/ocfs2/dlm/dlmdomain.c | 8
fs/ocfs2/dlm/dlmlock.c | 8
fs/ocfs2/dlm/dlmmaster.c | 10
fs/ocfs2/dlm/dlmrecovery.c | 10
fs/ocfs2/dlm/dlmthread.c | 8
fs/ocfs2/dlm/dlmunlock.c | 8
fs/ocfs2/dlmfs/Makefile | 2
fs/ocfs2/dlmfs/dlmfs.c | 4
fs/ocfs2/dlmfs/userdlm.c | 6
fs/ocfs2/dlmglue.c | 2
fs/ocfs2/journal.h | 8
fs/ocfs2/namei.c | 3
fs/reiserfs/stree.c | 3
include/linux/backing-dev.h | 10
include/linux/bitops.h | 1
include/linux/fs.h | 6
include/linux/io-mapping.h | 5
include/linux/memblock.h | 7
include/linux/memory.h | 29
include/linux/memory_hotplug.h | 3
include/linux/mm.h | 116 +-
include/linux/mmzone.h | 2
include/linux/page-isolation.h | 8
include/linux/swab.h | 1
include/linux/thermal.h | 11
include/linux/units.h | 84 +
include/linux/zlib.h | 6
include/trace/events/kmem.h | 4
include/trace/events/writeback.h | 37
include/uapi/linux/swab.h | 10
include/uapi/linux/sysctl.h | 2
init/main.c | 36
kernel/Makefile | 1
lib/Kconfig | 7
lib/Makefile | 2
lib/decompress_inflate.c | 13
lib/find_bit.c | 82 -
lib/scatterlist.c | 2
lib/test_bitmap.c | 9
lib/test_kasan.c | 1
lib/zlib_deflate/deflate.c | 85 +
lib/zlib_deflate/deflate_syms.c | 1
lib/zlib_deflate/deftree.c | 54 -
lib/zlib_deflate/defutil.h | 134 ++
lib/zlib_dfltcc/Makefile | 13
lib/zlib_dfltcc/dfltcc.c | 57 +
lib/zlib_dfltcc/dfltcc.h | 155 +++
lib/zlib_dfltcc/dfltcc_deflate.c | 280 ++++++
lib/zlib_dfltcc/dfltcc_inflate.c | 149 +++
lib/zlib_dfltcc/dfltcc_syms.c | 17
lib/zlib_dfltcc/dfltcc_util.h | 123 ++
lib/zlib_inflate/inflate.c | 32
lib/zlib_inflate/inflate.h | 8
lib/zlib_inflate/infutil.h | 18
mm/Makefile | 1
mm/backing-dev.c | 1
mm/debug.c | 18
mm/early_ioremap.c | 8
mm/filemap.c | 34
mm/gup.c | 503 ++++++-----
mm/gup_benchmark.c | 9
mm/huge_memory.c | 44
mm/kmemleak.c | 112 +-
mm/memblock.c | 22
mm/memcontrol.c | 25
mm/memory_hotplug.c | 24
mm/mempolicy.c | 6
mm/memremap.c | 95 --
mm/migrate.c | 77 +
mm/mmap.c | 30
mm/oom_kill.c | 2
mm/page_alloc.c | 83 +
mm/page_isolation.c | 69 -
mm/page_vma_mapped.c | 12
mm/process_vm_access.c | 32
mm/slub.c | 88 +
mm/sparse.c | 2
mm/swap.c | 27
mm/swapfile.c | 2
mm/vmscan.c | 24
mm/zswap.c | 88 +
net/xdp/xdp_umem.c | 4
scripts/spelling.txt | 14
tools/testing/selftests/vm/gup_benchmark.c | 6
tools/vm/slabinfo.c | 4
136 files changed, 2790 insertions(+), 1358 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-01-14 0:28 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-01-14 0:28 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
11 MM fixes, based on b3a987b0264d3ddbb24293ebff10eddfc472f653:
Vlastimil Babka <vbabka@suse.cz>:
mm, thp: tweak reclaim/compaction effort of local-only and all-node allocations
David Hildenbrand <david@redhat.com>:
mm/memory_hotplug: don't free usage map when removing a re-added early section
"Kirill A. Shutemov" <kirill@shutemov.name>:
Patch series "Fix two above-47bit hint address vs. THP bugs":
mm/huge_memory.c: thp: fix conflict of above-47bit hint address and PMD alignment
mm/shmem.c: thp, shmem: fix conflict of above-47bit hint address and PMD alignment
Roman Gushchin <guro@fb.com>:
mm: memcg/slab: fix percpu slab vmstats flushing
Vlastimil Babka <vbabka@suse.cz>:
mm, debug_pagealloc: don't rely on static keys too early
Wen Yang <wenyang@linux.alibaba.com>:
Patch series "use div64_ul() instead of div_u64() if the divisor is:
mm/page-writeback.c: avoid potential division by zero in wb_min_max_ratio()
mm/page-writeback.c: use div64_ul() for u64-by-unsigned-long divide
mm/page-writeback.c: improve arithmetic divisions
Adrian Huang <ahuang12@lenovo.com>:
mm: memcg/slab: call flush_memcg_workqueue() only if memcg workqueue is valid
Yang Shi <yang.shi@linux.alibaba.com>:
mm: khugepaged: add trace status description for SCAN_PAGE_HAS_PRIVATE
include/linux/mm.h | 18 +++++++++-
include/linux/mmzone.h | 5 +--
include/trace/events/huge_memory.h | 3 +
init/main.c | 1
mm/huge_memory.c | 38 ++++++++++++++---------
mm/memcontrol.c | 37 +++++-----------------
mm/mempolicy.c | 10 ++++--
mm/page-writeback.c | 10 +++---
mm/page_alloc.c | 61 ++++++++++---------------------------
mm/shmem.c | 7 ++--
mm/slab.c | 4 +-
mm/slab_common.c | 3 +
mm/slub.c | 2 -
mm/sparse.c | 9 ++++-
mm/vmalloc.c | 4 +-
15 files changed, 102 insertions(+), 110 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2020-01-04 20:55 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2020-01-04 20:55 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
17 fixes, base on 5613970af3f5f8372c596b138bd64f3918513515:
David Hildenbrand <david@redhat.com>:
mm/memory_hotplug: shrink zones when offlining memory
Chanho Min <chanho.min@lge.com>:
mm/zsmalloc.c: fix the migrated zspage statistics.
Andrey Konovalov <andreyknvl@google.com>:
kcov: fix struct layout for kcov_remote_arg
Shakeel Butt <shakeelb@google.com>:
memcg: account security cred as well to kmemcg
Yang Shi <yang.shi@linux.alibaba.com>:
mm: move_pages: return valid node id in status if the page is already on the target node
Eric Biggers <ebiggers@google.com>:
fs/direct-io.c: include fs/internal.h for missing prototype
fs/nsfs.c: include headers for missing declarations
fs/namespace.c: make to_mnt_ns() static
Nick Desaulniers <ndesaulniers@google.com>:
hexagon: parenthesize registers in asm predicates
hexagon: work around compiler crash
Randy Dunlap <rdunlap@infradead.org>:
fs/posix_acl.c: fix kernel-doc warnings
Ilya Dryomov <idryomov@gmail.com>:
mm/oom: fix pgtables units mismatch in Killed process message
Navid Emamdoost <navid.emamdoost@gmail.com>:
mm/gup: fix memory leak in __gup_benchmark_ioctl
Waiman Long <longman@redhat.com>:
mm/hugetlb: defer freeing of huge pages if in non-task context
Kai Li <li.kai4@h3c.com>:
ocfs2: call journal flush to mark journal as empty after journal recovery when mount
Gang He <GHe@suse.com>:
ocfs2: fix the crash due to call ocfs2_get_dlm_debug once less
Nick Desaulniers <ndesaulniers@google.com>:
hexagon: define ioremap_uc
Documentation/dev-tools/kcov.rst | 10 +++----
arch/arm64/mm/mmu.c | 4 --
arch/hexagon/include/asm/atomic.h | 8 ++---
arch/hexagon/include/asm/bitops.h | 8 ++---
arch/hexagon/include/asm/cmpxchg.h | 2 -
arch/hexagon/include/asm/futex.h | 6 ++--
arch/hexagon/include/asm/io.h | 1
arch/hexagon/include/asm/spinlock.h | 20 +++++++-------
arch/hexagon/kernel/stacktrace.c | 4 --
arch/hexagon/kernel/vm_entry.S | 2 -
arch/ia64/mm/init.c | 4 --
arch/powerpc/mm/mem.c | 3 --
arch/s390/mm/init.c | 4 --
arch/sh/mm/init.c | 4 --
arch/x86/mm/init_32.c | 4 --
arch/x86/mm/init_64.c | 4 --
fs/direct-io.c | 2 +
fs/namespace.c | 2 -
fs/nsfs.c | 3 ++
fs/ocfs2/dlmglue.c | 1
fs/ocfs2/journal.c | 8 +++++
fs/posix_acl.c | 7 +++-
include/linux/memory_hotplug.h | 7 +++-
include/uapi/linux/kcov.h | 10 +++----
kernel/cred.c | 6 ++--
mm/gup_benchmark.c | 8 ++++-
mm/hugetlb.c | 51 +++++++++++++++++++++++++++++++++++-
mm/memory_hotplug.c | 31 +++++++++++----------
mm/memremap.c | 2 -
mm/migrate.c | 23 ++++++++++++----
mm/oom_kill.c | 2 -
mm/zsmalloc.c | 5 +++
32 files changed, 166 insertions(+), 90 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2019-12-18 4:50 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2019-12-18 4:50 UTC (permalink / raw)
To: Linus Torvalds; +Cc: linux-mm, mm-commits
6 fixes based on 2187f215ebaac73ddbd814696d7c7fa34f0c3de0:
Andrey Ryabinin <aryabinin@virtuozzo.com>:
kasan: fix crashes on access to memory mapped by vm_map_ram()
Daniel Axtens <dja@axtens.net>:
mm/memory.c: add apply_to_existing_page_range() helper
kasan: use apply_to_existing_page_range() for releasing vmalloc shadow
kasan: don't assume percpu shadow allocations will succeed
Yang Shi <yang.shi@linux.alibaba.com>:
mm: vmscan: protect shrinker idr replace with CONFIG_MEMCG
Changbin Du <changbin.du@gmail.com>:
lib/Kconfig.debug: fix some messed up configurations
include/linux/kasan.h | 15 +++--
include/linux/mm.h | 3 +
lib/Kconfig.debug | 100 ++++++++++++++++++------------------
mm/kasan/common.c | 36 ++++++++-----
mm/memory.c | 136 ++++++++++++++++++++++++++++++++++----------------
mm/vmalloc.c | 133 ++++++++++++++++++++++++++++--------------------
mm/vmscan.c | 2
7 files changed, 260 insertions(+), 165 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2019-12-05 0:48 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2019-12-05 0:48 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
Most of the rest of MM and various other things. Some Kconfig rework
still awaits merges of dependent trees from linux-next.
86 patches, based on 63de37476ebd1e9bab6a9e17186dc5aa1da9ea99.
Subsystems affected by this patch series:
mm/hotfixes
mm/memcg
mm/vmstat
mm/thp
procfs
sysctl
misc
notifiers
core-kernel
bitops
lib
checkpatch
epoll
binfmt
init
rapidio
uaccess
kcov
ubsan
ipc
bitmap
mm/pagemap
Subsystem: mm/hotfixes
zhong jiang <zhongjiang@huawei.com>:
mm/kasan/common.c: fix compile error
Subsystem: mm/memcg
Roman Gushchin <guro@fb.com>:
mm: memcg/slab: wait for !root kmem_cache refcnt killing on root kmem_cache destruction
Subsystem: mm/vmstat
Konstantin Khlebnikov <khlebnikov@yandex-team.ru>:
mm/vmstat: add helpers to get vmstat item names for each enum type
mm/memcontrol: use vmstat names for printing statistics
Subsystem: mm/thp
Yu Zhao <yuzhao@google.com>:
mm/memory.c: replace is_zero_pfn with is_huge_zero_pmd for thp
Subsystem: procfs
Alexey Dobriyan <adobriyan@gmail.com>:
proc: change ->nlink under proc_subdir_lock
fs/proc/generic.c: delete useless "len" variable
fs/proc/internal.h: shuffle "struct pde_opener"
Miaohe Lin <linmiaohe@huawei.com>:
include/linux/proc_fs.h: fix confusing macro arg name
Krzysztof Kozlowski <krzk@kernel.org>:
fs/proc/Kconfig: fix indentation
Subsystem: sysctl
Alessio Balsini <balsini@android.com>:
include/linux/sysctl.h: inline braces for ctl_table and ctl_table_header
Subsystem: misc
Stephen Boyd <swboyd@chromium.org>:
.gitattributes: use 'dts' diff driver for dts files
Rikard Falkeborn <rikard.falkeborn@gmail.com>:
linux/build_bug.h: change type to int
Masahiro Yamada <yamada.masahiro@socionext.com>:
linux/scc.h: make uapi linux/scc.h self-contained
Krzysztof Kozlowski <krzk@kernel.org>:
arch/Kconfig: fix indentation
Joe Perches <joe@perches.com>:
scripts/get_maintainer.pl: add signatures from Fixes: <badcommit> lines in commit message
Andy Shevchenko <andriy.shevchenko@linux.intel.com>:
kernel.h: update comment about simple_strto<foo>() functions
auxdisplay: charlcd: deduplicate simple_strtoul()
Subsystem: notifiers
Xiaoming Ni <nixiaoming@huawei.com>:
kernel/notifier.c: intercept duplicate registrations to avoid infinite loops
kernel/notifier.c: remove notifier_chain_cond_register()
kernel/notifier.c: remove blocking_notifier_chain_cond_register()
Subsystem: core-kernel
Nathan Chancellor <natechancellor@gmail.com>:
kernel/profile.c: use cpumask_available to check for NULL cpumask
Joe Perches <joe@perches.com>:
kernel/sys.c: avoid copying possible padding bytes in copy_to_user
Subsystem: bitops
William Breathitt Gray <vilhelm.gray@gmail.com>:
bitops: introduce the for_each_set_clump8 macro
lib/test_bitmap.c: add for_each_set_clump8 test cases
gpio: 104-dio-48e: utilize for_each_set_clump8 macro
gpio: 104-idi-48: utilize for_each_set_clump8 macro
gpio: gpio-mm: utilize for_each_set_clump8 macro
gpio: ws16c48: utilize for_each_set_clump8 macro
gpio: pci-idio-16: utilize for_each_set_clump8 macro
gpio: pcie-idio-24: utilize for_each_set_clump8 macro
gpio: uniphier: utilize for_each_set_clump8 macro
gpio: 74x164: utilize the for_each_set_clump8 macro
thermal: intel: intel_soc_dts_iosf: Utilize for_each_set_clump8 macro
gpio: pisosr: utilize the for_each_set_clump8 macro
gpio: max3191x: utilize the for_each_set_clump8 macro
gpio: pca953x: utilize the for_each_set_clump8 macro
Subsystem: lib
Wei Yang <richardw.yang@linux.intel.com>:
lib/rbtree: set successor's parent unconditionally
lib/rbtree: get successor's color directly
Laura Abbott <labbott@redhat.com>:
lib/test_meminit.c: add bulk alloc/free tests
Trent Piepho <tpiepho@gmail.com>:
lib/math/rational.c: fix possible incorrect result from rational fractions helper
Huang Shijie <sjhuang@iluvatar.ai>:
lib/genalloc.c: export symbol addr_in_gen_pool
lib/genalloc.c: rename addr_in_gen_pool to gen_pool_has_addr
Subsystem: checkpatch
Joe Perches <joe@perches.com>:
checkpatch: improve ignoring CamelCase SI style variants like mA
checkpatch: reduce is_maintained_obsolete lookup runtime
Subsystem: epoll
Jason Baron <jbaron@akamai.com>:
epoll: simplify ep_poll_safewake() for CONFIG_DEBUG_LOCK_ALLOC
Heiher <r@hev.cc>:
fs/epoll: remove unnecessary wakeups of nested epoll
selftests: add epoll selftests
Subsystem: binfmt
Alexey Dobriyan <adobriyan@gmail.com>:
fs/binfmt_elf.c: delete unused "interp_map_addr" argument
fs/binfmt_elf.c: extract elf_read() function
Subsystem: init
Krzysztof Kozlowski <krzk@kernel.org>:
init/Kconfig: fix indentation
Subsystem: rapidio
"Ben Dooks (Codethink)" <ben.dooks@codethink.co.uk>:
drivers/rapidio/rio-driver.c: fix missing include of <linux/rio_drv.h>
drivers/rapidio/rio-access.c: fix missing include of <linux/rio_drv.h>
Subsystem: uaccess
Daniel Vetter <daniel.vetter@ffwll.ch>:
drm: limit to INT_MAX in create_blob ioctl
Kees Cook <keescook@chromium.org>:
uaccess: disallow > INT_MAX copy sizes
Subsystem: kcov
Andrey Konovalov <andreyknvl@google.com>:
Patch series " kcov: collect coverage from usb and vhost", v3:
kcov: remote coverage support
usb, kcov: collect coverage from hub_event
vhost, kcov: collect coverage from vhost_worker
Subsystem: ubsan
Julien Grall <julien.grall@arm.com>:
lib/ubsan: don't serialize UBSAN report
Subsystem: ipc
Masahiro Yamada <yamada.masahiro@socionext.com>:
arch: ipcbuf.h: make uapi asm/ipcbuf.h self-contained
arch: msgbuf.h: make uapi asm/msgbuf.h self-contained
arch: sembuf.h: make uapi asm/sembuf.h self-contained
Subsystem: bitmap
Andy Shevchenko <andriy.shevchenko@linux.intel.com>:
Patch series "gpio: pca953x: Convert to bitmap (extended) API", v2:
lib/test_bitmap: force argument of bitmap_parselist_user() to proper address space
lib/test_bitmap: undefine macros after use
lib/test_bitmap: name EXP_BYTES properly
lib/test_bitmap: rename exp to exp1 to avoid ambiguous name
lib/test_bitmap: move exp1 and exp2 upper for others to use
lib/test_bitmap: fix comment about this file
lib/bitmap: introduce bitmap_replace() helper
gpio: pca953x: remove redundant variable and check in IRQ handler
gpio: pca953x: use input from regs structure in pca953x_irq_pending()
gpio: pca953x: convert to use bitmap API
gpio: pca953x: tighten up indentation
Subsystem: mm/pagemap
Mike Rapoport <rppt@linux.ibm.com>:
Patch series "mm: remove __ARCH_HAS_4LEVEL_HACK", v13:
alpha: use pgtable-nopud instead of 4level-fixup
arm: nommu: use pgtable-nopud instead of 4level-fixup
c6x: use pgtable-nopud instead of 4level-fixup
m68k: nommu: use pgtable-nopud instead of 4level-fixup
m68k: mm: use pgtable-nopXd instead of 4level-fixup
microblaze: use pgtable-nopmd instead of 4level-fixup
nds32: use pgtable-nopmd instead of 4level-fixup
parisc: use pgtable-nopXd instead of 4level-fixup
Helge Deller <deller@gmx.de>:
parisc/hugetlb: use pgtable-nopXd instead of 4level-fixup
Mike Rapoport <rppt@linux.ibm.com>:
sparc32: use pgtable-nopud instead of 4level-fixup
um: remove unused pxx_offset_proc() and addr_pte() functions
um: add support for folded p4d page tables
mm: remove __ARCH_HAS_4LEVEL_HACK and include/asm-generic/4level-fixup.h
.gitattributes | 2
Documentation/core-api/genalloc.rst | 2
Documentation/dev-tools/kcov.rst | 129
arch/Kconfig | 22
arch/alpha/include/asm/mmzone.h | 1
arch/alpha/include/asm/pgalloc.h | 4
arch/alpha/include/asm/pgtable.h | 24
arch/alpha/mm/init.c | 12
arch/arm/include/asm/pgtable.h | 2
arch/arm/mm/dma-mapping.c | 2
arch/c6x/include/asm/pgtable.h | 2
arch/m68k/include/asm/mcf_pgalloc.h | 7
arch/m68k/include/asm/mcf_pgtable.h | 28
arch/m68k/include/asm/mmu_context.h | 12
arch/m68k/include/asm/motorola_pgalloc.h | 4
arch/m68k/include/asm/motorola_pgtable.h | 32
arch/m68k/include/asm/page.h | 9
arch/m68k/include/asm/pgtable_mm.h | 11
arch/m68k/include/asm/pgtable_no.h | 2
arch/m68k/include/asm/sun3_pgalloc.h | 5
arch/m68k/include/asm/sun3_pgtable.h | 18
arch/m68k/kernel/sys_m68k.c | 10
arch/m68k/mm/init.c | 6
arch/m68k/mm/kmap.c | 39
arch/m68k/mm/mcfmmu.c | 16
arch/m68k/mm/motorola.c | 17
arch/m68k/sun3x/dvma.c | 7
arch/microblaze/include/asm/page.h | 3
arch/microblaze/include/asm/pgalloc.h | 16
arch/microblaze/include/asm/pgtable.h | 32
arch/microblaze/kernel/signal.c | 10
arch/microblaze/mm/init.c | 7
arch/microblaze/mm/pgtable.c | 13
arch/mips/include/uapi/asm/msgbuf.h | 1
arch/mips/include/uapi/asm/sembuf.h | 2
arch/nds32/include/asm/page.h | 3
arch/nds32/include/asm/pgalloc.h | 3
arch/nds32/include/asm/pgtable.h | 12
arch/nds32/include/asm/tlb.h | 1
arch/nds32/kernel/pm.c | 4
arch/nds32/mm/fault.c | 16
arch/nds32/mm/init.c | 11
arch/nds32/mm/mm-nds32.c | 6
arch/nds32/mm/proc.c | 26
arch/parisc/include/asm/page.h | 30
arch/parisc/include/asm/pgalloc.h | 41
arch/parisc/include/asm/pgtable.h | 52
arch/parisc/include/asm/tlb.h | 2
arch/parisc/include/uapi/asm/msgbuf.h | 1
arch/parisc/include/uapi/asm/sembuf.h | 1
arch/parisc/kernel/cache.c | 13
arch/parisc/kernel/pci-dma.c | 9
arch/parisc/mm/fixmap.c | 10
arch/parisc/mm/hugetlbpage.c | 18
arch/powerpc/include/uapi/asm/msgbuf.h | 2
arch/powerpc/include/uapi/asm/sembuf.h | 2
arch/s390/include/uapi/asm/ipcbuf.h | 2
arch/sparc/include/asm/pgalloc_32.h | 6
arch/sparc/include/asm/pgtable_32.h | 28
arch/sparc/include/uapi/asm/ipcbuf.h | 2
arch/sparc/include/uapi/asm/msgbuf.h | 2
arch/sparc/include/uapi/asm/sembuf.h | 2
arch/sparc/mm/fault_32.c | 11
arch/sparc/mm/highmem.c | 6
arch/sparc/mm/io-unit.c | 6
arch/sparc/mm/iommu.c | 6
arch/sparc/mm/srmmu.c | 51
arch/um/include/asm/pgtable-2level.h | 1
arch/um/include/asm/pgtable-3level.h | 1
arch/um/include/asm/pgtable.h | 3
arch/um/kernel/mem.c | 8
arch/um/kernel/skas/mmu.c | 12
arch/um/kernel/skas/uaccess.c | 7
arch/um/kernel/tlb.c | 85
arch/um/kernel/trap.c | 4
arch/x86/include/uapi/asm/msgbuf.h | 3
arch/x86/include/uapi/asm/sembuf.h | 2
arch/xtensa/include/uapi/asm/ipcbuf.h | 2
arch/xtensa/include/uapi/asm/msgbuf.h | 2
arch/xtensa/include/uapi/asm/sembuf.h | 1
drivers/auxdisplay/charlcd.c | 34
drivers/base/node.c | 9
drivers/gpio/gpio-104-dio-48e.c | 75
drivers/gpio/gpio-104-idi-48.c | 36
drivers/gpio/gpio-74x164.c | 19
drivers/gpio/gpio-gpio-mm.c | 75
drivers/gpio/gpio-max3191x.c | 19
drivers/gpio/gpio-pca953x.c | 209
drivers/gpio/gpio-pci-idio-16.c | 75
drivers/gpio/gpio-pcie-idio-24.c | 111
drivers/gpio/gpio-pisosr.c | 12
drivers/gpio/gpio-uniphier.c | 13
drivers/gpio/gpio-ws16c48.c | 73
drivers/gpu/drm/drm_property.c | 2
drivers/misc/sram-exec.c | 2
drivers/rapidio/rio-access.c | 2
drivers/rapidio/rio-driver.c | 1
drivers/thermal/intel/intel_soc_dts_iosf.c | 31
drivers/thermal/intel/intel_soc_dts_iosf.h | 2
drivers/usb/core/hub.c | 5
drivers/vhost/vhost.c | 6
drivers/vhost/vhost.h | 1
fs/binfmt_elf.c | 56
fs/eventpoll.c | 52
fs/proc/Kconfig | 8
fs/proc/generic.c | 37
fs/proc/internal.h | 2
include/asm-generic/4level-fixup.h | 39
include/asm-generic/bitops/find.h | 17
include/linux/bitmap.h | 51
include/linux/bitops.h | 12
include/linux/build_bug.h | 4
include/linux/genalloc.h | 2
include/linux/kcov.h | 23
include/linux/kernel.h | 19
include/linux/mm.h | 10
include/linux/notifier.h | 4
include/linux/proc_fs.h | 4
include/linux/rbtree_augmented.h | 6
include/linux/sched.h | 8
include/linux/sysctl.h | 6
include/linux/thread_info.h | 2
include/linux/vmstat.h | 54
include/uapi/asm-generic/ipcbuf.h | 2
include/uapi/asm-generic/msgbuf.h | 2
include/uapi/asm-generic/sembuf.h | 1
include/uapi/linux/kcov.h | 28
include/uapi/linux/scc.h | 1
init/Kconfig | 78
kernel/dma/remap.c | 2
kernel/kcov.c | 547 +
kernel/notifier.c | 45
kernel/profile.c | 6
kernel/sys.c | 4
lib/bitmap.c | 12
lib/find_bit.c | 14
lib/genalloc.c | 7
lib/math/rational.c | 63
lib/test_bitmap.c | 206
lib/test_meminit.c | 20
lib/ubsan.c | 64
mm/kasan/common.c | 1
mm/memcontrol.c | 52
mm/memory.c | 10
mm/slab_common.c | 12
mm/vmstat.c | 60
net/sunrpc/rpc_pipe.c | 2
scripts/checkpatch.pl | 13
scripts/get_maintainer.pl | 38
tools/testing/selftests/Makefile | 1
tools/testing/selftests/filesystems/epoll/.gitignore | 1
tools/testing/selftests/filesystems/epoll/Makefile | 7
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c | 3074 ++++++++++
usr/include/Makefile | 4
154 files changed, 5270 insertions(+), 1360 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2019-12-01 21:07 ` incoming Linus Torvalds
@ 2019-12-02 8:21 ` Steven Price
0 siblings, 0 replies; 409+ messages in thread
From: Steven Price @ 2019-12-02 8:21 UTC (permalink / raw)
To: Linus Torvalds; +Cc: Andrew Morton, mm-commits, Linux-MM
On Sun, Dec 01, 2019 at 09:07:47PM +0000, Linus Torvalds wrote:
> On Sat, Nov 30, 2019 at 5:47 PM Andrew Morton <akpm@linux-foundation.org> wrote:
> >
> > Steven Price <steven.price@arm.com>:
> > Patch series "Generic page walk and ptdump", v15:
> > mm: add generic p?d_leaf() macros
> > arc: mm: add p?d_leaf() definitions
> > arm: mm: add p?d_leaf() definitions
> > arm64: mm: add p?d_leaf() definitions
> > mips: mm: add p?d_leaf() definitions
> > powerpc: mm: add p?d_leaf() definitions
> > riscv: mm: add p?d_leaf() definitions
> > s390: mm: add p?d_leaf() definitions
> > sparc: mm: add p?d_leaf() definitions
> > x86: mm: add p?d_leaf() definitions
> > mm: pagewalk: add p4d_entry() and pgd_entry()
> > mm: pagewalk: allow walking without vma
> > mm: pagewalk: add test_p?d callbacks
> > mm: pagewalk: add 'depth' parameter to pte_hole
> > x86: mm: point to struct seq_file from struct pg_state
> > x86: mm+efi: convert ptdump_walk_pgd_level() to take a mm_struct
> > x86: mm: convert ptdump_walk_pgd_level_debugfs() to take an mm_struct
> > x86: mm: convert ptdump_walk_pgd_level_core() to take an mm_struct
> > mm: add generic ptdump
> > x86: mm: convert dump_pagetables to use walk_page_range
> > arm64: mm: convert mm/dump.c to use walk_page_range()
> > arm64: mm: display non-present entries in ptdump
> > mm: ptdump: reduce level numbers by 1 in note_page()
>
> I've dropped these, and since they clearly weren't ready I don't want
> to see them re-sent for 5.5.
Sorry about this, I'll try to track down the cause of this and hopefully
resubmit for 5.6.
Thanks,
Steve
> If somebody figures out the bug, trying again for 5.6 sounds fine.
>
> Linus
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2019-12-01 1:47 incoming Andrew Morton
2019-12-01 5:17 ` incoming James Bottomley
@ 2019-12-01 21:07 ` Linus Torvalds
2019-12-02 8:21 ` incoming Steven Price
1 sibling, 1 reply; 409+ messages in thread
From: Linus Torvalds @ 2019-12-01 21:07 UTC (permalink / raw)
To: Andrew Morton, Steven Price; +Cc: mm-commits, Linux-MM
On Sat, Nov 30, 2019 at 5:47 PM Andrew Morton <akpm@linux-foundation.org> wrote:
>
> Steven Price <steven.price@arm.com>:
> Patch series "Generic page walk and ptdump", v15:
> mm: add generic p?d_leaf() macros
> arc: mm: add p?d_leaf() definitions
> arm: mm: add p?d_leaf() definitions
> arm64: mm: add p?d_leaf() definitions
> mips: mm: add p?d_leaf() definitions
> powerpc: mm: add p?d_leaf() definitions
> riscv: mm: add p?d_leaf() definitions
> s390: mm: add p?d_leaf() definitions
> sparc: mm: add p?d_leaf() definitions
> x86: mm: add p?d_leaf() definitions
> mm: pagewalk: add p4d_entry() and pgd_entry()
> mm: pagewalk: allow walking without vma
> mm: pagewalk: add test_p?d callbacks
> mm: pagewalk: add 'depth' parameter to pte_hole
> x86: mm: point to struct seq_file from struct pg_state
> x86: mm+efi: convert ptdump_walk_pgd_level() to take a mm_struct
> x86: mm: convert ptdump_walk_pgd_level_debugfs() to take an mm_struct
> x86: mm: convert ptdump_walk_pgd_level_core() to take an mm_struct
> mm: add generic ptdump
> x86: mm: convert dump_pagetables to use walk_page_range
> arm64: mm: convert mm/dump.c to use walk_page_range()
> arm64: mm: display non-present entries in ptdump
> mm: ptdump: reduce level numbers by 1 in note_page()
I've dropped these, and since they clearly weren't ready I don't want
to see them re-sent for 5.5.
If somebody figures out the bug, trying again for 5.6 sounds fine.
Linus
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2019-12-01 1:47 incoming Andrew Morton
@ 2019-12-01 5:17 ` James Bottomley
2019-12-01 21:07 ` incoming Linus Torvalds
1 sibling, 0 replies; 409+ messages in thread
From: James Bottomley @ 2019-12-01 5:17 UTC (permalink / raw)
To: Andrew Morton, Linus Torvalds; +Cc: mm-commits, linux-mm
On Sat, 2019-11-30 at 17:47 -0800, Andrew Morton wrote:
> - a small number of updates to scripts/, ocfs2 and fs/buffer.c
>
> - most of MM. I still have quite a lot of material (mostly not MM)
> staged after linux-next due to -next dependencies. I'll send thos
> across next week as the preprequisites get merged up.
>
> 158 patches, based on 32ef9553635ab1236c33951a8bd9b5af1c3b1646.
Hey, Andrew, would it be at all possible for you to thread these
patches under something like this incoming message? The selfish reason
I'm asking is so I can mark the thread as read instead of having to do
it individually for 158 messages ... my thumb would thank you for this.
Regards,
James
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2019-12-01 1:47 Andrew Morton
2019-12-01 5:17 ` incoming James Bottomley
2019-12-01 21:07 ` incoming Linus Torvalds
0 siblings, 2 replies; 409+ messages in thread
From: Andrew Morton @ 2019-12-01 1:47 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
- a small number of updates to scripts/, ocfs2 and fs/buffer.c
- most of MM. I still have quite a lot of material (mostly not MM)
staged after linux-next due to -next dependencies. I'll send thos
across next week as the preprequisites get merged up.
158 patches, based on 32ef9553635ab1236c33951a8bd9b5af1c3b1646.
Subsystems affected by this patch series:
scripts
ocfs2
vfs
mm/slab
mm/slub
mm/pagecache
mm/gup
mm/swap
mm/memcg
mm/pagemap
mm/memfd
mm/memory-failure
mm/memory-hotplug
mm/sparsemem
mm/vmalloc
mm/kasan
mm/pagealloc
mm/vmscan
mm/proc
mm/z3fold
mm/mempolicy
mm/memblock
mm/hugetlbfs
mm/hugetlb
mm/migration
mm/thp
mm/cma
mm/autonuma
mm/page-poison
mm/mmap
mm/madvise
mm/userfaultfd
mm/shmem
mm/cleanups
mm/support
Subsystem: scripts
Colin Ian King <colin.king@canonical.com>:
scripts/spelling.txt: add more spellings to spelling.txt
Subsystem: ocfs2
Ding Xiang <dingxiang@cmss.chinamobile.com>:
ocfs2: fix passing zero to 'PTR_ERR' warning
Subsystem: vfs
Saurav Girepunje <saurav.girepunje@gmail.com>:
fs/buffer.c: fix use true/false for bool type
Ben Dooks <ben.dooks@codethink.co.uk>:
fs/buffer.c: include internal.h for missing declarations
Subsystem: mm/slab
Pengfei Li <lpf.vector@gmail.com>:
Patch series "mm, slab: Make kmalloc_info[] contain all types of names", v6:
mm, slab: make kmalloc_info[] contain all types of names
mm, slab: remove unused kmalloc_size()
mm, slab_common: use enum kmalloc_cache_type to iterate over kmalloc caches
Subsystem: mm/slub
Miles Chen <miles.chen@mediatek.com>:
mm: slub: print the offset of fault addresses
Yu Zhao <yuzhao@google.com>:
mm/slub.c: update comments
mm/slub.c: clean up validate_slab()
Subsystem: mm/pagecache
Konstantin Khlebnikov <khlebnikov@yandex-team.ru>:
mm/filemap.c: remove redundant cache invalidation after async direct-io write
fs/direct-io.c: keep dio_warn_stale_pagecache() when CONFIG_BLOCK=n
mm/filemap.c: warn if stale pagecache is left after direct write
Subsystem: mm/gup
zhong jiang <zhongjiang@huawei.com>:
mm/gup.c: allow CMA migration to propagate errors back to caller
Liu Xiang <liuxiang_1999@126.com>:
mm/gup.c: fix comments of __get_user_pages() and get_user_pages_remote()
Subsystem: mm/swap
Naohiro Aota <naohiro.aota@wdc.com>:
mm, swap: disallow swapon() on zoned block devices
Fengguang Wu <fengguang.wu@intel.com>:
mm/swap.c: trivial mark_page_accessed() cleanup
Subsystem: mm/memcg
Yafang Shao <laoar.shao@gmail.com>:
mm, memcg: clean up reclaim iter array
Johannes Weiner <hannes@cmpxchg.org>:
mm: memcontrol: remove dead code from memory_max_write()
mm: memcontrol: try harder to set a new memory.high
Hao Lee <haolee.swjtu@gmail.com>:
include/linux/memcontrol.h: fix comments based on per-node memcg
Shakeel Butt <shakeelb@google.com>:
mm: vmscan: memcontrol: remove mem_cgroup_select_victim_node()
Chris Down <chris@chrisdown.name>:
Documentation/admin-guide/cgroup-v2.rst: document why inactive_X + active_X may not equal X
Subsystem: mm/pagemap
Johannes Weiner <hannes@cmpxchg.org>:
mm: drop mmap_sem before calling balance_dirty_pages() in write fault
"Kirill A. Shutemov" <kirill.shutemov@linux.intel.com>:
shmem: pin the file in shmem_fault() if mmap_sem is dropped
"Joel Fernandes (Google)" <joel@joelfernandes.org>:
mm: emit tracepoint when RSS changes
rss_stat: add support to detect RSS updates of external mm
Wei Yang <richardw.yang@linux.intel.com>:
mm/mmap.c: remove a never-triggered warning in __vma_adjust()
Konstantin Khlebnikov <khlebnikov@yandex-team.ru>:
mm/swap.c: piggyback lru_add_drain_all() calls
Wei Yang <richardw.yang@linux.intel.com>:
mm/mmap.c: prev could be retrieved from vma->vm_prev
mm/mmap.c: __vma_unlink_prev() is not necessary now
mm/mmap.c: extract __vma_unlink_list() as counterpart for __vma_link_list()
mm/mmap.c: rb_parent is not necessary in __vma_link_list()
mm/rmap.c: don't reuse anon_vma if we just want a copy
mm/rmap.c: reuse mergeable anon_vma as parent when fork
Gaowei Pu <pugaowei@gmail.com>:
mm/mmap.c: use IS_ERR_VALUE to check return value of get_unmapped_area
Vineet Gupta <Vineet.Gupta1@synopsys.com>:
Patch series "elide extraneous generated code for folded p4d/pud/pmd", v3:
ARC: mm: remove __ARCH_USE_5LEVEL_HACK
asm-generic/tlb: stub out pud_free_tlb() if nopud ...
asm-generic/tlb: stub out p4d_free_tlb() if nop4d ...
asm-generic/tlb: stub out pmd_free_tlb() if nopmd
asm-generic/mm: stub out p{4,u}d_clear_bad() if __PAGETABLE_P{4,U}D_FOLDED
Miles Chen <miles.chen@mediatek.com>:
mm/rmap.c: fix outdated comment in page_get_anon_vma()
Yang Shi <yang.shi@linux.alibaba.com>:
mm/rmap.c: use VM_BUG_ON_PAGE() in __page_check_anon_rmap()
Thomas Hellstrom <thellstrom@vmware.com>:
mm: move the backup x_devmap() functions to asm-generic/pgtable.h
mm/memory.c: fix a huge pud insertion race during faulting
Steven Price <steven.price@arm.com>:
Patch series "Generic page walk and ptdump", v15:
mm: add generic p?d_leaf() macros
arc: mm: add p?d_leaf() definitions
arm: mm: add p?d_leaf() definitions
arm64: mm: add p?d_leaf() definitions
mips: mm: add p?d_leaf() definitions
powerpc: mm: add p?d_leaf() definitions
riscv: mm: add p?d_leaf() definitions
s390: mm: add p?d_leaf() definitions
sparc: mm: add p?d_leaf() definitions
x86: mm: add p?d_leaf() definitions
mm: pagewalk: add p4d_entry() and pgd_entry()
mm: pagewalk: allow walking without vma
mm: pagewalk: add test_p?d callbacks
mm: pagewalk: add 'depth' parameter to pte_hole
x86: mm: point to struct seq_file from struct pg_state
x86: mm+efi: convert ptdump_walk_pgd_level() to take a mm_struct
x86: mm: convert ptdump_walk_pgd_level_debugfs() to take an mm_struct
x86: mm: convert ptdump_walk_pgd_level_core() to take an mm_struct
mm: add generic ptdump
x86: mm: convert dump_pagetables to use walk_page_range
arm64: mm: convert mm/dump.c to use walk_page_range()
arm64: mm: display non-present entries in ptdump
mm: ptdump: reduce level numbers by 1 in note_page()
Subsystem: mm/memfd
Nicolas Geoffray <ngeoffray@google.com>:
mm, memfd: fix COW issue on MAP_PRIVATE and F_SEAL_FUTURE_WRITE mappings
"Joel Fernandes (Google)" <joel@joelfernandes.org>:
memfd: add test for COW on MAP_PRIVATE and F_SEAL_FUTURE_WRITE mappings
Subsystem: mm/memory-failure
Jane Chu <jane.chu@oracle.com>:
mm/memory-failure.c clean up around tk pre-allocation
Naoya Horiguchi <nao.horiguchi@gmail.com>:
mm, soft-offline: convert parameter to pfn
Yunfeng Ye <yeyunfeng@huawei.com>:
mm/memory-failure.c: use page_shift() in add_to_kill()
Subsystem: mm/memory-hotplug
Anshuman Khandual <anshuman.khandual@arm.com>:
mm/hotplug: reorder memblock_[free|remove]() calls in try_remove_memory()
Alastair D'Silva <alastair@d-silva.org>:
mm/memory_hotplug.c: add a bounds check to __add_pages()
David Hildenbrand <david@redhat.com>:
Patch series "mm/memory_hotplug: Export generic_online_page()":
mm/memory_hotplug: export generic_online_page()
hv_balloon: use generic_online_page()
mm/memory_hotplug: remove __online_page_free() and __online_page_increment_counters()
Patch series "mm: Memory offlining + page isolation cleanups", v2:
mm/page_alloc.c: don't set pages PageReserved() when offlining
mm/page_isolation.c: convert SKIP_HWPOISON to MEMORY_OFFLINE
"Ben Dooks (Codethink)" <ben.dooks@codethink.co.uk>:
include/linux/memory_hotplug.h: move definitions of {set,clear}_zone_contiguous
David Hildenbrand <david@redhat.com>:
drivers/base/memory.c: drop the mem_sysfs_mutex
mm/memory_hotplug.c: don't allow to online/offline memory blocks with holes
Subsystem: mm/sparsemem
Vincent Whitchurch <vincent.whitchurch@axis.com>:
mm/sparse: consistently do not zero memmap
Ilya Leoshkevich <iii@linux.ibm.com>:
mm/sparse.c: mark populate_section_memmap as __meminit
Michal Hocko <mhocko@suse.com>:
mm/sparse.c: do not waste pre allocated memmap space
Subsystem: mm/vmalloc
Liu Xiang <liuxiang_1999@126.com>:
mm/vmalloc.c: remove unnecessary highmem_mask from parameter of gfpflags_allow_blocking()
"Uladzislau Rezki (Sony)" <urezki@gmail.com>:
mm/vmalloc: remove preempt_disable/enable when doing preloading
mm/vmalloc: respect passed gfp_mask when doing preloading
mm/vmalloc: add more comments to the adjust_va_to_fit_type()
Anders Roxell <anders.roxell@linaro.org>:
selftests: vm: add fragment CONFIG_TEST_VMALLOC
"Uladzislau Rezki (Sony)" <urezki@gmail.com>:
mm/vmalloc: rework vmap_area_lock
Subsystem: mm/kasan
Daniel Axtens <dja@axtens.net>:
Patch series "kasan: support backing vmalloc space with real shadow:
kasan: support backing vmalloc space with real shadow memory
kasan: add test for vmalloc
fork: support VMAP_STACK with KASAN_VMALLOC
x86/kasan: support KASAN_VMALLOC
Subsystem: mm/pagealloc
Anshuman Khandual <anshuman.khandual@arm.com>:
mm/page_alloc: add alloc_contig_pages()
Mel Gorman <mgorman@techsingularity.net>:
mm, pcp: share common code between memory hotplug and percpu sysctl handler
mm, pcpu: make zone pcp updates and reset internal to the mm
Hao Lee <haolee.swjtu@gmail.com>:
include/linux/mmzone.h: fix comment for ISOLATE_UNMAPPED macro
lijiazi <jqqlijiazi@gmail.com>:
mm/page_alloc.c: print reserved_highatomic info
Subsystem: mm/vmscan
Andrey Ryabinin <aryabinin@virtuozzo.com>:
mm/vmscan: remove unused lru_pages argument
Yang Shi <yang.shi@linux.alibaba.com>:
mm/vmscan.c: remove unused scan_control parameter from pageout()
Johannes Weiner <hannes@cmpxchg.org>:
Patch series "mm: vmscan: cgroup-related cleanups":
mm: vmscan: simplify lruvec_lru_size()
mm: clean up and clarify lruvec lookup procedure
mm: vmscan: move inactive_list_is_low() swap check to the caller
mm: vmscan: naming fixes: global_reclaim() and sane_reclaim()
mm: vmscan: replace shrink_node() loop with a retry jump
mm: vmscan: turn shrink_node_memcg() into shrink_lruvec()
mm: vmscan: split shrink_node() into node part and memcgs part
mm: vmscan: harmonize writeback congestion tracking for nodes & memcgs
Patch series "mm: fix page aging across multiple cgroups":
mm: vmscan: move file exhaustion detection to the node level
mm: vmscan: detect file thrashing at the reclaim root
mm: vmscan: enforce inactive:active ratio at the reclaim root
Xianting Tian <xianting_tian@126.com>:
mm/vmscan.c: fix typo in comment
Subsystem: mm/proc
Johannes Weiner <hannes@cmpxchg.org>:
kernel: sysctl: make drop_caches write-only
Subsystem: mm/z3fold
Vitaly Wool <vitaly.wool@konsulko.com>:
mm/z3fold.c: add inter-page compaction
Subsystem: mm/mempolicy
Li Xinhai <lixinhai.lxh@gmail.com>:
Patch series "mm: Fix checking unmapped holes for mbind", v4:
mm/mempolicy.c: check range first in queue_pages_test_walk
mm/mempolicy.c: fix checking unmapped holes for mbind
Subsystem: mm/memblock
Cao jin <caoj.fnst@cn.fujitsu.com>:
mm/memblock.c: cleanup doc
mm/memblock: correct doc for function
Yunfeng Ye <yeyunfeng@huawei.com>:
mm: support memblock alloc on the exact node for sparse_buffer_init()
Subsystem: mm/hugetlbfs
Mike Kravetz <mike.kravetz@oracle.com>:
hugetlbfs: hugetlb_fault_mutex_hash() cleanup
mm/hugetlbfs: fix error handling when setting up mounts
Patch series "hugetlbfs: convert macros to static inline, fix sparse warning":
powerpc/mm: remove pmd_huge/pud_huge stubs and include hugetlb.h
hugetlbfs: convert macros to static inline, fix sparse warning
Piotr Sarna <p.sarna@tlen.pl>:
hugetlbfs: add O_TMPFILE support
Waiman Long <longman@redhat.com>:
hugetlbfs: take read_lock on i_mmap for PMD sharing
Subsystem: mm/hugetlb
Mina Almasry <almasrymina@google.com>:
hugetlb: region_chg provides only cache entry
hugetlb: remove duplicated code
Wei Yang <richardw.yang@linux.intel.com>:
hugetlb: remove unused hstate in hugetlb_fault_mutex_hash()
Zhigang Lu <tonnylu@tencent.com>:
mm/hugetlb: avoid looping to the same hugepage if !pages and !vmas
zhong jiang <zhongjiang@huawei.com>:
mm/huge_memory.c: split_huge_pages_fops should be defined with DEFINE_DEBUGFS_ATTRIBUTE
Subsystem: mm/migration
Yang Shi <yang.shi@linux.alibaba.com>:
mm/migrate.c: handle freed page at the first place
Subsystem: mm/thp
"Kirill A. Shutemov" <kirill@shutemov.name>:
mm, thp: do not queue fully unmapped pages for deferred split
Song Liu <songliubraving@fb.com>:
mm/thp: flush file for !is_shmem PageDirty() case in collapse_file()
Subsystem: mm/cma
Yunfeng Ye <yeyunfeng@huawei.com>:
mm/cma.c: switch to bitmap_zalloc() for cma bitmap allocation
zhong jiang <zhongjiang@huawei.com>:
mm/cma_debug.c: use DEFINE_DEBUGFS_ATTRIBUTE to define debugfs fops
Subsystem: mm/autonuma
Huang Ying <ying.huang@intel.com>:
autonuma: fix watermark checking in migrate_balanced_pgdat()
autonuma: reduce cache footprint when scanning page tables
Subsystem: mm/page-poison
zhong jiang <zhongjiang@huawei.com>:
mm/hwpoison-inject: use DEFINE_DEBUGFS_ATTRIBUTE to define debugfs fops
Subsystem: mm/mmap
Wei Yang <richardw.yang@linux.intel.com>:
mm/mmap.c: make vma_merge() comment more easy to understand
Subsystem: mm/madvise
Yunfeng Ye <yeyunfeng@huawei.com>:
mm/madvise.c: replace with page_size() in madvise_inject_error()
Wei Yang <richardw.yang@linux.intel.com>:
mm/madvise.c: use PAGE_ALIGN[ED] for range checking
Subsystem: mm/userfaultfd
Wei Yang <richardw.yang@linux.intel.com>:
userfaultfd: use vma_pagesize for all huge page size calculation
userfaultfd: remove unnecessary WARN_ON() in __mcopy_atomic_hugetlb()
userfaultfd: wrap the common dst_vma check into an inlined function
Andrea Arcangeli <aarcange@redhat.com>:
fs/userfaultfd.c: wp: clear VM_UFFD_MISSING or VM_UFFD_WP during userfaultfd_register()
Mike Rapoport <rppt@linux.ibm.com>:
userfaultfd: require CAP_SYS_PTRACE for UFFD_FEATURE_EVENT_FORK
Subsystem: mm/shmem
Colin Ian King <colin.king@canonical.com>:
mm/shmem.c: make array 'values' static const, makes object smaller
Yang Shi <yang.shi@linux.alibaba.com>:
mm: shmem: use proper gfp flags for shmem_writepage()
Chen Jun <chenjun102@huawei.com>:
mm/shmem.c: cast the type of unmap_start to u64
Subsystem: mm/cleanups
Hao Lee <haolee.swjtu@gmail.com>:
mm: fix struct member name in function comments
Wei Yang <richardw.yang@linux.intel.com>:
mm: fix typos in comments when calling __SetPageUptodate()
Souptick Joarder <jrdr.linux@gmail.com>:
mm/memory_hotplug.c: remove __online_page_set_limits()
Krzysztof Kozlowski <krzk@kernel.org>:
mm/Kconfig: fix indentation
Randy Dunlap <rdunlap@infradead.org>:
mm/Kconfig: fix trivial help text punctuation
Subsystem: mm/support
Minchan Kim <minchan@google.com>:
mm/page_io.c: annotate refault stalls from swap_readpage
Documentation/admin-guide/cgroup-v2.rst | 7
Documentation/dev-tools/kasan.rst | 63 +
arch/Kconfig | 9
arch/arc/include/asm/pgtable.h | 2
arch/arc/mm/fault.c | 10
arch/arc/mm/highmem.c | 4
arch/arm/include/asm/pgtable-2level.h | 1
arch/arm/include/asm/pgtable-3level.h | 1
arch/arm64/Kconfig | 1
arch/arm64/Kconfig.debug | 19
arch/arm64/include/asm/pgtable.h | 2
arch/arm64/include/asm/ptdump.h | 8
arch/arm64/mm/Makefile | 4
arch/arm64/mm/dump.c | 148 +---
arch/arm64/mm/mmu.c | 4
arch/arm64/mm/ptdump_debugfs.c | 2
arch/mips/include/asm/pgtable.h | 5
arch/powerpc/include/asm/book3s/64/pgtable-4k.h | 3
arch/powerpc/include/asm/book3s/64/pgtable-64k.h | 3
arch/powerpc/include/asm/book3s/64/pgtable.h | 30
arch/powerpc/mm/book3s64/radix_pgtable.c | 1
arch/riscv/include/asm/pgtable-64.h | 7
arch/riscv/include/asm/pgtable.h | 7
arch/s390/include/asm/pgtable.h | 2
arch/sparc/include/asm/pgtable_64.h | 2
arch/x86/Kconfig | 2
arch/x86/Kconfig.debug | 20
arch/x86/include/asm/pgtable.h | 10
arch/x86/mm/Makefile | 4
arch/x86/mm/debug_pagetables.c | 8
arch/x86/mm/dump_pagetables.c | 431 +++---------
arch/x86/mm/kasan_init_64.c | 61 +
arch/x86/platform/efi/efi_32.c | 2
arch/x86/platform/efi/efi_64.c | 4
drivers/base/memory.c | 40 -
drivers/firmware/efi/arm-runtime.c | 2
drivers/hv/hv_balloon.c | 4
drivers/xen/balloon.c | 1
fs/buffer.c | 6
fs/direct-io.c | 21
fs/hugetlbfs/inode.c | 67 +
fs/ocfs2/acl.c | 4
fs/proc/task_mmu.c | 4
fs/userfaultfd.c | 21
include/asm-generic/4level-fixup.h | 1
include/asm-generic/5level-fixup.h | 1
include/asm-generic/pgtable-nop4d.h | 2
include/asm-generic/pgtable-nopmd.h | 2
include/asm-generic/pgtable-nopud.h | 2
include/asm-generic/pgtable.h | 71 ++
include/asm-generic/tlb.h | 4
include/linux/fs.h | 6
include/linux/gfp.h | 2
include/linux/hugetlb.h | 142 +++-
include/linux/kasan.h | 31
include/linux/memblock.h | 3
include/linux/memcontrol.h | 51 -
include/linux/memory_hotplug.h | 11
include/linux/mm.h | 42 -
include/linux/mmzone.h | 34
include/linux/moduleloader.h | 2
include/linux/page-isolation.h | 4
include/linux/pagewalk.h | 42 -
include/linux/ptdump.h | 22
include/linux/slab.h | 20
include/linux/string.h | 2
include/linux/swap.h | 2
include/linux/vmalloc.h | 12
include/trace/events/kmem.h | 53 +
kernel/events/uprobes.c | 2
kernel/fork.c | 4
kernel/sysctl.c | 2
lib/Kconfig.kasan | 16
lib/test_kasan.c | 26
lib/vsprintf.c | 40 -
mm/Kconfig | 40 -
mm/Kconfig.debug | 21
mm/Makefile | 1
mm/cma.c | 6
mm/cma_debug.c | 10
mm/filemap.c | 56 -
mm/gup.c | 40 -
mm/hmm.c | 8
mm/huge_memory.c | 2
mm/hugetlb.c | 298 ++------
mm/hwpoison-inject.c | 4
mm/internal.h | 27
mm/kasan/common.c | 233 ++++++
mm/kasan/generic_report.c | 3
mm/kasan/kasan.h | 1
mm/khugepaged.c | 18
mm/madvise.c | 14
mm/memblock.c | 113 ++-
mm/memcontrol.c | 167 ----
mm/memory-failure.c | 61 -
mm/memory.c | 56 +
mm/memory_hotplug.c | 86 +-
mm/mempolicy.c | 59 +
mm/migrate.c | 21
mm/mincore.c | 1
mm/mmap.c | 75 --
mm/mprotect.c | 8
mm/mremap.c | 4
mm/nommu.c | 10
mm/page_alloc.c | 137 +++
mm/page_io.c | 15
mm/page_isolation.c | 12
mm/pagewalk.c | 126 ++-
mm/pgtable-generic.c | 9
mm/ptdump.c | 167 ++++
mm/rmap.c | 65 +
mm/shmem.c | 29
mm/slab.c | 7
mm/slab.h | 6
mm/slab_common.c | 101 +-
mm/slub.c | 36 -
mm/sparse.c | 22
mm/swap.c | 29
mm/swapfile.c | 7
mm/userfaultfd.c | 77 +-
mm/util.c | 22
mm/vmalloc.c | 196 +++--
mm/vmscan.c | 798 +++++++++++------------
mm/workingset.c | 75 +-
mm/z3fold.c | 375 ++++++++--
scripts/spelling.txt | 28
tools/testing/selftests/memfd/memfd_test.c | 36 +
tools/testing/selftests/vm/config | 1
128 files changed, 3409 insertions(+), 2121 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2019-11-22 1:53 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2019-11-22 1:53 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
4 fixes, based on 81429eb8d9ca40b0c65bb739d29fa856c5d5e958:
Vincent Whitchurch <vincent.whitchurch@axis.com>:
mm/sparse: consistently do not zero memmap
Joseph Qi <joseph.qi@linux.alibaba.com>:
Revert "fs: ocfs2: fix possible null-pointer dereferences in ocfs2_xa_prepare_entry()"
David Hildenbrand <david@redhat.com>:
mm/memory_hotplug: don't access uninitialized memmaps in shrink_zone_span()
Andrey Ryabinin <aryabinin@virtuozzo.com>:
mm/ksm.c: don't WARN if page is still mapped in remove_stable_node()
fs/ocfs2/xattr.c | 56 ++++++++++++++++++++++++++++++----------------------
mm/ksm.c | 14 ++++++-------
mm/memory_hotplug.c | 16 ++++++++++++--
mm/sparse.c | 2 -
4 files changed, 54 insertions(+), 34 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2019-11-16 1:34 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2019-11-16 1:34 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
11 fixes, based on 875fef493f21e54d20d71a581687990aaa50268c:
Yang Shi <yang.shi@linux.alibaba.com>:
mm: mempolicy: fix the wrong return value and potential pages leak of mbind
zhong jiang <zhongjiang@huawei.com>:
mm: fix trying to reclaim unevictable lru page when calling madvise_pageout
Lasse Collin <lasse.collin@tukaani.org>:
lib/xz: fix XZ_DYNALLOC to avoid useless memory reallocations
Roman Gushchin <guro@fb.com>:
mm: memcg: switch to css_tryget() in get_mem_cgroup_from_mm()
mm: hugetlb: switch to css_tryget() in hugetlb_cgroup_charge_cgroup()
Laura Abbott <labbott@redhat.com>:
mm: slub: really fix slab walking for init_on_free
Song Liu <songliubraving@fb.com>:
mm,thp: recheck each page before collapsing file THP
David Hildenbrand <david@redhat.com>:
mm/memory_hotplug: fix try_offline_node()
Vinayak Menon <vinmenon@codeaurora.org>:
mm/page_io.c: do not free shared swap slots
Ralph Campbell <rcampbell@nvidia.com>:
mm/debug.c: __dump_page() prints an extra line
mm/debug.c: PageAnon() is true for PageKsm() pages
drivers/base/memory.c | 36 ++++++++++++++++++++++++++++++++++++
include/linux/memory.h | 1 +
lib/xz/xz_dec_lzma2.c | 1 +
mm/debug.c | 33 ++++++++++++++++++---------------
mm/hugetlb_cgroup.c | 2 +-
mm/khugepaged.c | 28 ++++++++++++++++------------
mm/madvise.c | 16 ++++++++++++----
mm/memcontrol.c | 2 +-
mm/memory_hotplug.c | 47 +++++++++++++++++++++++++++++------------------
mm/mempolicy.c | 14 +++++++++-----
mm/page_io.c | 6 +++---
mm/slub.c | 39 +++++++++------------------------------
12 files changed, 136 insertions(+), 89 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2019-11-06 5:16 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2019-11-06 5:16 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
17 fixes, based on 26bc672134241a080a83b2ab9aa8abede8d30e1c:
Shakeel Butt <shakeelb@google.com>:
mm: memcontrol: fix NULL-ptr deref in percpu stats flush
John Hubbard <jhubbard@nvidia.com>:
mm/gup_benchmark: fix MAP_HUGETLB case
Mel Gorman <mgorman@techsingularity.net>:
mm, meminit: recalculate pcpu batch and high limits after init completes
Yang Shi <yang.shi@linux.alibaba.com>:
mm: thp: handle page cache THP correctly in PageTransCompoundMap
Shuning Zhang <sunny.s.zhang@oracle.com>:
ocfs2: protect extent tree in ocfs2_prepare_inode_for_write()
Jason Gunthorpe <jgg@mellanox.com>:
mm/mmu_notifiers: use the right return code for WARN_ON
Michal Hocko <mhocko@suse.com>:
mm, vmstat: hide /proc/pagetypeinfo from normal users
mm, vmstat: reduce zone->lock holding time by /proc/pagetypeinfo
Ville Syrjälä <ville.syrjala@linux.intel.com>:
mm/khugepaged: fix might_sleep() warn with CONFIG_HIGHPTE=y
Johannes Weiner <hannes@cmpxchg.org>:
mm/page_alloc.c: ratelimit allocation failure warnings more aggressively
Vitaly Wool <vitaly.wool@konsulko.com>:
zswap: add Vitaly to the maintainers list
Kevin Hao <haokexin@gmail.com>:
dump_stack: avoid the livelock of the dump_lock
Song Liu <songliubraving@fb.com>:
MAINTAINERS: update information for "MEMORY MANAGEMENT"
Roman Gushchin <guro@fb.com>:
mm: slab: make page_cgroup_ino() to recognize non-compound slab pages properly
Ilya Leoshkevich <iii@linux.ibm.com>:
scripts/gdb: fix debugging modules compiled with hot/cold partitioning
David Hildenbrand <david@redhat.com>:
mm/memory_hotplug: fix updating the node span
Johannes Weiner <hannes@cmpxchg.org>:
mm: memcontrol: fix network errors from failing __GFP_ATOMIC charges
MAINTAINERS | 5 +
fs/ocfs2/file.c | 125 ++++++++++++++++++++++-------
include/linux/mm.h | 5 -
include/linux/mm_types.h | 5 +
include/linux/page-flags.h | 20 ++++
lib/dump_stack.c | 7 +
mm/khugepaged.c | 7 -
mm/memcontrol.c | 23 +++--
mm/memory_hotplug.c | 8 +
mm/mmu_notifier.c | 2
mm/page_alloc.c | 17 ++-
mm/slab.h | 4
mm/vmstat.c | 25 ++++-
scripts/gdb/linux/symbols.py | 3
tools/testing/selftests/vm/gup_benchmark.c | 2
15 files changed, 197 insertions(+), 61 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2019-10-19 3:19 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2019-10-19 3:19 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
Rather a lot of fixes, almost all affecting mm/.
26 patches, based on b9959c7a347d6adbb558fba7e36e9fef3cba3b07:
David Hildenbrand <david@redhat.com>:
drivers/base/memory.c: don't access uninitialized memmaps in soft_offline_page_store()
fs/proc/page.c: don't access uninitialized memmaps in fs/proc/page.c
mm/memory-failure.c: don't access uninitialized memmaps in memory_failure()
Joel Colledge <joel.colledge@linbit.com>:
scripts/gdb: fix lx-dmesg when CONFIG_PRINTK_CALLER is set
Qian Cai <cai@lca.pw>:
mm/page_owner: don't access uninitialized memmaps when reading /proc/pagetypeinfo
David Hildenbrand <david@redhat.com>:
mm/memory_hotplug: don't access uninitialized memmaps in shrink_pgdat_span()
"Aneesh Kumar K.V" <aneesh.kumar@linux.ibm.com>:
Patch series "mm/memory_hotplug: Shrink zones before removing memory", v6:
mm/memunmap: don't access uninitialized memmap in memunmap_pages()
Roman Gushchin <guro@fb.com>:
mm: memcg/slab: fix panic in __free_slab() caused by premature memcg pointer release
Chengguang Xu <cgxu519@mykernel.net>:
ocfs2: fix error handling in ocfs2_setattr()
John Hubbard <jhubbard@nvidia.com>:
mm/gup_benchmark: add a missing "w" to getopt string
mm/gup: fix a misnamed "write" argument, and a related bug
Honglei Wang <honglei.wang@oracle.com>:
mm: memcg: get number of pages on the LRU list in memcgroup base on lru_zone_size
Mike Rapoport <rppt@linux.ibm.com>:
mm: memblock: do not enforce current limit for memblock_phys* family
David Hildenbrand <david@redhat.com>:
hugetlbfs: don't access uninitialized memmaps in pfn_range_valid_gigantic()
Yi Li <yilikernel@gmail.com>:
ocfs2: fix panic due to ocfs2_wq is null
Konstantin Khlebnikov <khlebnikov@yandex-team.ru>:
mm/memcontrol: update lruvec counters in mem_cgroup_move_account
Chenwandun <chenwandun@huawei.com>:
zram: fix race between backing_dev_show and backing_dev_store
Ben Dooks <ben.dooks@codethink.co.uk>:
mm: include <linux/huge_mm.h> for is_vma_temporary_stack
mm/filemap.c: include <linux/ramfs.h> for generic_file_vm_ops definition
"Ben Dooks (Codethink)" <ben.dooks@codethink.co.uk>:
mm/init-mm.c: include <linux/mman.h> for vm_committed_as_batch
"Kirill A. Shutemov" <kirill.shutemov@linux.intel.com>:
Patch series "Fixes for THP in page cache", v2:
proc/meminfo: fix output alignment
mm/thp: fix node page state in split_huge_page_to_list()
William Kucharski <william.kucharski@oracle.com>:
mm/vmscan.c: support removing arbitrary sized pages from mapping
"Kirill A. Shutemov" <kirill.shutemov@linux.intel.com>:
mm/thp: allow dropping THP from page cache
Song Liu <songliubraving@fb.com>:
kernel/events/uprobes.c: only do FOLL_SPLIT_PMD for uprobe register
Ilya Leoshkevich <iii@linux.ibm.com>:
scripts/gdb: fix debugging modules on s390
drivers/base/memory.c | 3 +
drivers/block/zram/zram_drv.c | 5 +
fs/ocfs2/file.c | 2
fs/ocfs2/journal.c | 3 -
fs/ocfs2/localalloc.c | 3 -
fs/proc/meminfo.c | 4 -
fs/proc/page.c | 28 ++++++----
kernel/events/uprobes.c | 13 ++++-
mm/filemap.c | 1
mm/gup.c | 14 +++--
mm/huge_memory.c | 9 ++-
mm/hugetlb.c | 5 -
mm/init-mm.c | 1
mm/memblock.c | 6 +-
mm/memcontrol.c | 18 ++++---
mm/memory-failure.c | 14 +++--
mm/memory_hotplug.c | 74 ++++++-----------------------
mm/memremap.c | 11 ++--
mm/page_owner.c | 5 +
mm/rmap.c | 1
mm/slab_common.c | 9 +--
mm/truncate.c | 12 ++++
mm/vmscan.c | 14 ++---
scripts/gdb/linux/dmesg.py | 16 ++++--
scripts/gdb/linux/symbols.py | 8 ++-
scripts/gdb/linux/utils.py | 25 +++++----
tools/testing/selftests/vm/gup_benchmark.c | 2
27 files changed, 166 insertions(+), 140 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2019-10-14 21:11 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2019-10-14 21:11 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
The usual shower of hotfixes and some followups to the recently merged
page_owner enhancements.
16 patches, based on 2abd839aa7e615f2bbc50c8ba7deb9e40d186768.
Subsystems affected by this patch series:
Vlastimil Babka <vbabka@suse.cz>:
Patch series "followups to debug_pagealloc improvements through page_owner", v3:
mm, page_owner: fix off-by-one error in __set_page_owner_handle()
mm, page_owner: decouple freeing stack trace from debug_pagealloc
mm, page_owner: rename flag indicating that page is allocated
Qian Cai <cai@lca.pw>:
mm/slub: fix a deadlock in show_slab_objects()
Eric Biggers <ebiggers@google.com>:
lib/generic-radix-tree.c: add kmemleak annotations
Alexander Potapenko <glider@google.com>:
mm/slub.c: init_on_free=1 should wipe freelist ptr for bulk allocations
lib/test_meminit: add a kmem_cache_alloc_bulk() test
David Rientjes <rientjes@google.com>:
mm, hugetlb: allow hugepage allocations to reclaim as needed
Vlastimil Babka <vbabka@suse.cz>:
mm, compaction: fix wrong pfn handling in __reset_isolation_pfn()
Randy Dunlap <rdunlap@infradead.org>:
fs/direct-io.c: fix kernel-doc warning
fs/libfs.c: fix kernel-doc warning
fs/fs-writeback.c: fix kernel-doc warning
bitmap.h: fix kernel-doc warning and typo
xarray.h: fix kernel-doc warning
mm/slab.c: fix kernel-doc warning for __ksize()
Jane Chu <jane.chu@oracle.com>:
mm/memory-failure: poison read receives SIGKILL instead of SIGBUS if mmaped more than once
Documentation/dev-tools/kasan.rst | 3 ++
fs/direct-io.c | 3 --
fs/fs-writeback.c | 2 -
fs/libfs.c | 3 --
include/linux/bitmap.h | 3 +-
include/linux/page_ext.h | 10 ++++++
include/linux/xarray.h | 4 +-
lib/generic-radix-tree.c | 32 +++++++++++++++++-----
lib/test_meminit.c | 27 ++++++++++++++++++
mm/compaction.c | 7 ++--
mm/memory-failure.c | 22 ++++++++-------
mm/page_alloc.c | 6 ++--
mm/page_ext.c | 23 ++++++---------
mm/page_owner.c | 55 +++++++++++++-------------------------
mm/slab.c | 3 ++
mm/slub.c | 35 ++++++++++++++++++------
16 files changed, 152 insertions(+), 86 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2019-10-07 0:57 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2019-10-07 0:57 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
The usual shower of hotfixes.
Chris's memcg patches aren't actually fixes - they're mature but a few
niggling review issues were late to arrive.
The ocfs2 fixes are quite old - those took some time to get
reviewer attention.
18 patches, based on 4ea655343ce4180fe9b2c7ec8cb8ef9884a47901.
Subsystems affected by this patch series:
ocfs2
hotfixes
mm/memcg
mm/slab-generic
Subsystem: ocfs2
Jia Guo <guojia12@huawei.com>:
ocfs2: clear zero in unaligned direct IO
Jia-Ju Bai <baijiaju1990@gmail.com>:
fs: ocfs2: fix possible null-pointer dereferences in ocfs2_xa_prepare_entry()
fs: ocfs2: fix a possible null-pointer dereference in ocfs2_write_end_nolock()
fs: ocfs2: fix a possible null-pointer dereference in ocfs2_info_scan_inode_alloc()
Subsystem: hotfixes
Will Deacon <will@kernel.org>:
panic: ensure preemption is disabled during panic()
Anshuman Khandual <anshuman.khandual@arm.com>:
mm/memremap: drop unused SECTION_SIZE and SECTION_MASK
Tejun Heo <tj@kernel.org>:
writeback: fix use-after-free in finish_writeback_work()
Yi Wang <wang.yi59@zte.com.cn>:
mm: fix -Wmissing-prototypes warnings
Baoquan He <bhe@redhat.com>:
memcg: only record foreign writebacks with dirty pages when memcg is not disabled
Michal Hocko <mhocko@suse.com>:
kernel/sysctl.c: do not override max_threads provided by userspace
Vitaly Wool <vitalywool@gmail.com>:
mm/z3fold.c: claim page in the beginning of free
Qian Cai <cai@lca.pw>:
mm/page_alloc.c: fix a crash in free_pages_prepare()
Dan Carpenter <dan.carpenter@oracle.com>:
mm/vmpressure.c: fix a signedness bug in vmpressure_register_event()
Subsystem: mm/memcg
Chris Down <chris@chrisdown.name>:
mm, memcg: proportional memory.{low,min} reclaim
mm, memcg: make memory.emin the baseline for utilisation determination
mm, memcg: make scan aggression always exclude protection
Subsystem: mm/slab-generic
Vlastimil Babka <vbabka@suse.cz>:
Patch series "guarantee natural alignment for kmalloc()", v2:
mm, sl[ou]b: improve memory accounting
mm, sl[aou]b: guarantee natural alignment for kmalloc(power-of-two)
Documentation/admin-guide/cgroup-v2.rst | 20 +-
Documentation/core-api/memory-allocation.rst | 4
fs/fs-writeback.c | 9 -
fs/ocfs2/aops.c | 25 +++
fs/ocfs2/ioctl.c | 2
fs/ocfs2/xattr.c | 56 +++----
include/linux/memcontrol.h | 67 ++++++---
include/linux/slab.h | 4
kernel/fork.c | 4
kernel/panic.c | 1
mm/memcontrol.c | 5
mm/memremap.c | 2
mm/page_alloc.c | 8 -
mm/shuffle.c | 2
mm/slab_common.c | 19 ++
mm/slob.c | 62 ++++++--
mm/slub.c | 14 +
mm/sparse.c | 2
mm/vmpressure.c | 20 +-
mm/vmscan.c | 198 +++++++++++++++++----------
mm/z3fold.c | 10 +
21 files changed, 363 insertions(+), 171 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2019-09-25 23:45 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2019-09-25 23:45 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
- almost all of the rest of -mm
- various other subsystems
76 patches, based on 351c8a09b00b5c51c8f58b016fffe51f87e2d820:
Subsystems affected by this patch series:
memcg
misc
core-kernel
lib
checkpatch
reiserfs
fat
fork
cpumask
kexec
uaccess
kconfig
kgdb
bug
ipc
lzo
kasan
madvise
cleanups
pagemap
Subsystem: memcg
Michal Hocko <mhocko@suse.com>:
memcg, kmem: do not fail __GFP_NOFAIL charges
Subsystem: misc
Masahiro Yamada <yamada.masahiro@socionext.com>:
linux/coff.h: add include guard
Subsystem: core-kernel
Valdis Kletnieks <valdis.kletnieks@vt.edu>:
kernel/elfcore.c: include proper prototypes
Subsystem: lib
Michel Lespinasse <walken@google.com>:
rbtree: avoid generating code twice for the cached versions (tools copy)
Patch series "make RB_DECLARE_CALLBACKS more generic", v3:
augmented rbtree: add comments for RB_DECLARE_CALLBACKS macro
augmented rbtree: add new RB_DECLARE_CALLBACKS_MAX macro
augmented rbtree: rework the RB_DECLARE_CALLBACKS macro definition
Joe Perches <joe@perches.com>:
kernel-doc: core-api: include string.h into core-api
Qian Cai <cai@lca.pw>:
include/trace/events/writeback.h: fix -Wstringop-truncation warnings
Kees Cook <keescook@chromium.org>:
strscpy: reject buffer sizes larger than INT_MAX
Valdis Kletnieks <valdis.kletnieks@vt.edu>:
lib/generic-radix-tree.c: make 2 functions static inline
lib/extable.c: add missing prototypes
Stephen Boyd <swboyd@chromium.org>:
lib/hexdump: make print_hex_dump_bytes() a nop on !DEBUG builds
Subsystem: checkpatch
Joe Perches <joe@perches.com>:
checkpatch: don't interpret stack dumps as commit IDs
checkpatch: improve SPDX license checking
Matteo Croce <mcroce@redhat.com>:
checkpatch.pl: warn on invalid commit id
Brendan Jackman <brendan.jackman@bluwireless.co.uk>:
checkpatch: exclude sizeof sub-expressions from MACRO_ARG_REUSE
Joe Perches <joe@perches.com>:
checkpatch: prefer __section over __attribute__((section(...)))
checkpatch: allow consecutive close braces
Sean Christopherson <sean.j.christopherson@intel.com>:
checkpatch: remove obsolete period from "ambiguous SHA1" query
Joe Perches <joe@perches.com>:
checkpatch: make git output use LANGUAGE=en_US.utf8
Subsystem: reiserfs
Jia-Ju Bai <baijiaju1990@gmail.com>:
fs: reiserfs: remove unnecessary check of bh in remove_from_transaction()
zhengbin <zhengbin13@huawei.com>:
fs/reiserfs/journal.c: remove set but not used variables
fs/reiserfs/stree.c: remove set but not used variables
fs/reiserfs/lbalance.c: remove set but not used variables
fs/reiserfs/objectid.c: remove set but not used variables
fs/reiserfs/prints.c: remove set but not used variables
fs/reiserfs/fix_node.c: remove set but not used variables
fs/reiserfs/do_balan.c: remove set but not used variables
Jason Yan <yanaijie@huawei.com>:
fs/reiserfs/journal.c: remove set but not used variable
fs/reiserfs/do_balan.c: remove set but not used variable
Subsystem: fat
Markus Elfring <elfring@users.sourceforge.net>:
fat: delete an unnecessary check before brelse()
Subsystem: fork
Sai Praneeth Prakhya <sai.praneeth.prakhya@intel.com>:
fork: improve error message for corrupted page tables
Subsystem: cpumask
Alexey Dobriyan <adobriyan@gmail.com>:
cpumask: nicer for_each_cpumask_and() signature
Subsystem: kexec
Tetsuo Handa <penguin-kernel@I-love.SAKURA.ne.jp>:
kexec: bail out upon SIGKILL when allocating memory.
Vasily Gorbik <gor@linux.ibm.com>:
kexec: restore arch_kexec_kernel_image_probe declaration
Subsystem: uaccess
Kees Cook <keescook@chromium.org>:
uaccess: add missing __must_check attributes
Subsystem: kconfig
Masahiro Yamada <yamada.masahiro@socionext.com>:
compiler: enable CONFIG_OPTIMIZE_INLINING forcibly
Subsystem: kgdb
Douglas Anderson <dianders@chromium.org>:
kgdb: don't use a notifier to enter kgdb at panic; call directly
scripts/gdb: handle split debug
Subsystem: bug
Kees Cook <keescook@chromium.org>:
Patch series "Clean up WARN() "cut here" handling", v2:
bug: refactor away warn_slowpath_fmt_taint()
bug: rename __WARN_printf_taint() to __WARN_printf()
bug: consolidate warn_slowpath_fmt() usage
bug: lift "cut here" out of __warn()
bug: clean up helper macros to remove __WARN_TAINT()
bug: consolidate __WARN_FLAGS usage
bug: move WARN_ON() "cut here" into exception handler
Subsystem: ipc
Markus Elfring <elfring@users.sourceforge.net>:
ipc/mqueue.c: delete an unnecessary check before the macro call dev_kfree_skb()
ipc/mqueue: improve exception handling in do_mq_notify()
"Joel Fernandes (Google)" <joel@joelfernandes.org>:
ipc/sem.c: convert to use built-in RCU list checking
Subsystem: lzo
Dave Rodgman <dave.rodgman@arm.com>:
lib/lzo/lzo1x_compress.c: fix alignment bug in lzo-rle
Subsystem: kasan
Andrey Konovalov <andreyknvl@google.com>:
Patch series "arm64: untag user pointers passed to the kernel", v19:
lib: untag user pointers in strn*_user
mm: untag user pointers passed to memory syscalls
mm: untag user pointers in mm/gup.c
mm: untag user pointers in get_vaddr_frames
fs/namespace: untag user pointers in copy_mount_options
userfaultfd: untag user pointers
drm/amdgpu: untag user pointers
drm/radeon: untag user pointers in radeon_gem_userptr_ioctl
media/v4l2-core: untag user pointers in videobuf_dma_contig_user_get
tee/shm: untag user pointers in tee_shm_register
vfio/type1: untag user pointers in vaddr_get_pfn
Catalin Marinas <catalin.marinas@arm.com>:
mm: untag user pointers in mmap/munmap/mremap/brk
Subsystem: madvise
Minchan Kim <minchan@kernel.org>:
Patch series "Introduce MADV_COLD and MADV_PAGEOUT", v7:
mm: introduce MADV_COLD
mm: change PAGEREF_RECLAIM_CLEAN with PAGE_REFRECLAIM
mm: introduce MADV_PAGEOUT
mm: factor out common parts between MADV_COLD and MADV_PAGEOUT
Subsystem: cleanups
Mike Rapoport <rppt@linux.ibm.com>:
hexagon: drop empty and unused free_initrd_mem
Denis Efremov <efremov@linux.com>:
checkpatch: check for nested (un)?likely() calls
xen/events: remove unlikely() from WARN() condition
fs: remove unlikely() from WARN_ON() condition
wimax/i2400m: remove unlikely() from WARN*() condition
xfs: remove unlikely() from WARN_ON() condition
IB/hfi1: remove unlikely() from IS_ERR*() condition
ntfs: remove (un)?likely() from IS_ERR() conditions
Subsystem: pagemap
Mark Rutland <mark.rutland@arm.com>:
mm: treewide: clarify pgtable_page_{ctor,dtor}() naming
Documentation/core-api/kernel-api.rst | 3
Documentation/vm/split_page_table_lock.rst | 10
arch/alpha/include/uapi/asm/mman.h | 3
arch/arc/include/asm/pgalloc.h | 4
arch/arm/include/asm/tlb.h | 2
arch/arm/mm/mmu.c | 2
arch/arm64/include/asm/tlb.h | 2
arch/arm64/mm/mmu.c | 2
arch/csky/include/asm/pgalloc.h | 2
arch/hexagon/include/asm/pgalloc.h | 2
arch/hexagon/mm/init.c | 13
arch/m68k/include/asm/mcf_pgalloc.h | 6
arch/m68k/include/asm/motorola_pgalloc.h | 6
arch/m68k/include/asm/sun3_pgalloc.h | 2
arch/mips/include/asm/pgalloc.h | 2
arch/mips/include/uapi/asm/mman.h | 3
arch/nios2/include/asm/pgalloc.h | 2
arch/openrisc/include/asm/pgalloc.h | 6
arch/parisc/include/uapi/asm/mman.h | 3
arch/powerpc/mm/pgtable-frag.c | 6
arch/riscv/include/asm/pgalloc.h | 2
arch/s390/mm/pgalloc.c | 6
arch/sh/include/asm/pgalloc.h | 2
arch/sparc/include/asm/pgtable_64.h | 5
arch/sparc/mm/init_64.c | 4
arch/sparc/mm/srmmu.c | 4
arch/um/include/asm/pgalloc.h | 2
arch/unicore32/include/asm/tlb.h | 2
arch/x86/mm/pat_rbtree.c | 19
arch/x86/mm/pgtable.c | 2
arch/xtensa/include/asm/pgalloc.h | 4
arch/xtensa/include/uapi/asm/mman.h | 3
drivers/block/drbd/drbd_interval.c | 29 -
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c | 2
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c | 2
drivers/gpu/drm/radeon/radeon_gem.c | 2
drivers/infiniband/hw/hfi1/verbs.c | 2
drivers/media/v4l2-core/videobuf-dma-contig.c | 9
drivers/net/wimax/i2400m/tx.c | 3
drivers/tee/tee_shm.c | 1
drivers/vfio/vfio_iommu_type1.c | 2
drivers/xen/events/events_base.c | 2
fs/fat/dir.c | 4
fs/namespace.c | 2
fs/ntfs/mft.c | 12
fs/ntfs/namei.c | 2
fs/ntfs/runlist.c | 2
fs/ntfs/super.c | 2
fs/open.c | 2
fs/reiserfs/do_balan.c | 15
fs/reiserfs/fix_node.c | 6
fs/reiserfs/journal.c | 22
fs/reiserfs/lbalance.c | 3
fs/reiserfs/objectid.c | 3
fs/reiserfs/prints.c | 3
fs/reiserfs/stree.c | 4
fs/userfaultfd.c | 22
fs/xfs/xfs_buf.c | 4
include/asm-generic/bug.h | 71 +-
include/asm-generic/pgalloc.h | 8
include/linux/cpumask.h | 14
include/linux/interval_tree_generic.h | 22
include/linux/kexec.h | 2
include/linux/kgdb.h | 2
include/linux/mm.h | 4
include/linux/mm_types_task.h | 4
include/linux/printk.h | 22
include/linux/rbtree_augmented.h | 114 +++-
include/linux/string.h | 5
include/linux/swap.h | 2
include/linux/thread_info.h | 2
include/linux/uaccess.h | 21
include/trace/events/writeback.h | 38 -
include/uapi/asm-generic/mman-common.h | 3
include/uapi/linux/coff.h | 5
ipc/mqueue.c | 22
ipc/sem.c | 3
kernel/debug/debug_core.c | 31 -
kernel/elfcore.c | 1
kernel/fork.c | 16
kernel/kexec_core.c | 2
kernel/panic.c | 48 -
lib/Kconfig.debug | 4
lib/bug.c | 11
lib/extable.c | 1
lib/generic-radix-tree.c | 4
lib/hexdump.c | 21
lib/lzo/lzo1x_compress.c | 14
lib/rbtree_test.c | 37 -
lib/string.c | 12
lib/strncpy_from_user.c | 3
lib/strnlen_user.c | 3
mm/frame_vector.c | 2
mm/gup.c | 4
mm/internal.h | 2
mm/madvise.c | 562 ++++++++++++++++-------
mm/memcontrol.c | 10
mm/mempolicy.c | 3
mm/migrate.c | 2
mm/mincore.c | 2
mm/mlock.c | 4
mm/mmap.c | 34 -
mm/mprotect.c | 2
mm/mremap.c | 13
mm/msync.c | 2
mm/oom_kill.c | 2
mm/swap.c | 42 +
mm/vmalloc.c | 5
mm/vmscan.c | 62 ++
scripts/checkpatch.pl | 69 ++
scripts/gdb/linux/symbols.py | 4
tools/include/linux/rbtree.h | 71 +-
tools/include/linux/rbtree_augmented.h | 145 +++--
tools/lib/rbtree.c | 37 -
114 files changed, 1195 insertions(+), 754 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2019-09-24 15:34 ` incoming Linus Torvalds
@ 2019-09-25 6:36 ` Michal Hocko
0 siblings, 0 replies; 409+ messages in thread
From: Michal Hocko @ 2019-09-25 6:36 UTC (permalink / raw)
To: Linus Torvalds
Cc: Andrew Morton, David Rientjes, Vlastimil Babka, Andrea Arcangeli,
mm-commits, Linux-MM
On Tue 24-09-19 08:34:20, Linus Torvalds wrote:
> On Tue, Sep 24, 2019 at 12:48 AM Michal Hocko <mhocko@kernel.org> wrote:
> >
> > The patch proposed by David is really non trivial wrt. potential side
> > effects.
>
> The thing is, that's not an argument when we know that the current
> state is garbage and has a lot of these non-trivial side effects that
> are bad.
>
> So the patch by David _fixes_ a non-trivial bad side effect.
>
> You can't then say "there may be other non-trivial side effects that I
> don't even know about" as an argument for saying it's bad. David at
> least has numbers and an argument for his patch.
All I am saying is that I am not able to wrap my head around this patch
to provide a competent Ack. I also believe that the fix is targetting a
wrong layer of the problem as explained in my review feedback. Appart
from reclaim/compaction interaction mentioned by Vlastimil, it seems
that it is an overly eager fallback to a remote node in the fast path
that is causing a large part of the problem as well. Kcompactd is not
eager enough to keep high order allocations ready for the fast path.
This is not specific to THP we have many other high order allocations
which are going to follow the same pattern, likely not visible in any
counters but still having performance implications.
Let's discuss technical details in the respective email thread
--
Michal Hocko
SUSE Labs
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2019-09-24 7:48 ` incoming Michal Hocko
2019-09-24 15:34 ` incoming Linus Torvalds
@ 2019-09-24 19:55 ` Vlastimil Babka
1 sibling, 0 replies; 409+ messages in thread
From: Vlastimil Babka @ 2019-09-24 19:55 UTC (permalink / raw)
To: Michal Hocko, Andrew Morton
Cc: Linus Torvalds, David Rientjes, Andrea Arcangeli, mm-commits, Linux-MM
On 9/24/19 9:48 AM, Michal Hocko wrote:
> On Mon 23-09-19 21:31:53, Andrew Morton wrote:
>> On Mon, 23 Sep 2019 17:55:24 -0700 Linus Torvalds
>> <torvalds@linux-foundation.org> wrote:
>>
>>> On Mon, Sep 23, 2019 at 3:31 PM Andrew Morton
>>> <akpm@linux-foundation.org> wrote:
>>>>
>>>> - almost all of -mm, as below.
>>>
>>> I was hoping that we could at least test the THP locality thing?
>>> Is it in your queue at all, or am I supposed to just do it
>>> myself?
>>>
>>
>> Confused. I saw a privately emailed patch from David which nobody
>> seems to have tested yet. I parked that for consideration after
>> -rc1. Or are you referring to something else?
>>
>> This thing keeps stalling. It would be nice to push this along and
>> get something nailed down which we can at least get into 5.4-rc,
>> perhaps with a backport-this tag?
>
> The patch proposed by David is really non trivial wrt. potential
> side effects. I have provided my review feedback [1] and it didn't
> get any reaction. I really believe that we need to debug this
> properly. A reproducer would be useful for others to work on that.
>
> There is a more fundamental problem here and we need to address it
> rather than to duck tape it and whack a mole afterwards.
I believe we found a problem when investigating over-reclaim in this
thread [1] where it seems madvised THP allocation attempt can result in
4MB reclaimed, if there is a small zone such as ZONE_DMA on the node. As
it happens, the patch "[patch 090/134] mm, reclaim: make
should_continue_reclaim perform dryrun detection" in Andrew's pile
should change this 4MB to 32 pages reclaimed (as a side-effect), but
that has to be tested. I'm also working on a patch to not reclaim even
those few pages. Of course there might be more fundamental issues with
reclaim/compaction interaction, but this one seems to become hopefully
clear now.
[1]
https://lore.kernel.org/linux-mm/4b4ba042-3741-7b16-2292-198c569da2aa@profihost.ag/
> [1] http://lkml.kernel.org/r/20190909193020.GD2063@dhcp22.suse.cz
>
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2019-09-24 7:48 ` incoming Michal Hocko
@ 2019-09-24 15:34 ` Linus Torvalds
2019-09-25 6:36 ` incoming Michal Hocko
2019-09-24 19:55 ` incoming Vlastimil Babka
1 sibling, 1 reply; 409+ messages in thread
From: Linus Torvalds @ 2019-09-24 15:34 UTC (permalink / raw)
To: Michal Hocko
Cc: Andrew Morton, David Rientjes, Vlastimil Babka, Andrea Arcangeli,
mm-commits, Linux-MM
On Tue, Sep 24, 2019 at 12:48 AM Michal Hocko <mhocko@kernel.org> wrote:
>
> The patch proposed by David is really non trivial wrt. potential side
> effects.
The thing is, that's not an argument when we know that the current
state is garbage and has a lot of these non-trivial side effects that
are bad.
So the patch by David _fixes_ a non-trivial bad side effect.
You can't then say "there may be other non-trivial side effects that I
don't even know about" as an argument for saying it's bad. David at
least has numbers and an argument for his patch.
Linus
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2019-09-24 4:31 ` incoming Andrew Morton
@ 2019-09-24 7:48 ` Michal Hocko
2019-09-24 15:34 ` incoming Linus Torvalds
2019-09-24 19:55 ` incoming Vlastimil Babka
0 siblings, 2 replies; 409+ messages in thread
From: Michal Hocko @ 2019-09-24 7:48 UTC (permalink / raw)
To: Andrew Morton
Cc: Linus Torvalds, David Rientjes, Vlastimil Babka,
Andrea Arcangeli, mm-commits, Linux-MM
On Mon 23-09-19 21:31:53, Andrew Morton wrote:
> On Mon, 23 Sep 2019 17:55:24 -0700 Linus Torvalds <torvalds@linux-foundation.org> wrote:
>
> > On Mon, Sep 23, 2019 at 3:31 PM Andrew Morton <akpm@linux-foundation.org> wrote:
> > >
> > > - almost all of -mm, as below.
> >
> > I was hoping that we could at least test the THP locality thing? Is it
> > in your queue at all, or am I supposed to just do it myself?
> >
>
> Confused. I saw a privately emailed patch from David which nobody
> seems to have tested yet. I parked that for consideration after -rc1.
> Or are you referring to something else?
>
> This thing keeps stalling. It would be nice to push this along and get
> something nailed down which we can at least get into 5.4-rc, perhaps
> with a backport-this tag?
The patch proposed by David is really non trivial wrt. potential side
effects. I have provided my review feedback [1] and it didn't get
any reaction. I really believe that we need to debug this properly. A
reproducer would be useful for others to work on that.
There is a more fundamental problem here and we need to address it
rather than to duck tape it and whack a mole afterwards.
[1] http://lkml.kernel.org/r/20190909193020.GD2063@dhcp22.suse.cz
--
Michal Hocko
SUSE Labs
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2019-09-24 0:55 ` incoming Linus Torvalds
@ 2019-09-24 4:31 ` Andrew Morton
2019-09-24 7:48 ` incoming Michal Hocko
0 siblings, 1 reply; 409+ messages in thread
From: Andrew Morton @ 2019-09-24 4:31 UTC (permalink / raw)
To: Linus Torvalds
Cc: David Rientjes, Vlastimil Babka, Michal Hocko, Andrea Arcangeli,
mm-commits, Linux-MM
On Mon, 23 Sep 2019 17:55:24 -0700 Linus Torvalds <torvalds@linux-foundation.org> wrote:
> On Mon, Sep 23, 2019 at 3:31 PM Andrew Morton <akpm@linux-foundation.org> wrote:
> >
> > - almost all of -mm, as below.
>
> I was hoping that we could at least test the THP locality thing? Is it
> in your queue at all, or am I supposed to just do it myself?
>
Confused. I saw a privately emailed patch from David which nobody
seems to have tested yet. I parked that for consideration after -rc1.
Or are you referring to something else?
This thing keeps stalling. It would be nice to push this along and get
something nailed down which we can at least get into 5.4-rc, perhaps
with a backport-this tag?
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2019-09-23 22:31 incoming Andrew Morton
@ 2019-09-24 0:55 ` Linus Torvalds
2019-09-24 4:31 ` incoming Andrew Morton
0 siblings, 1 reply; 409+ messages in thread
From: Linus Torvalds @ 2019-09-24 0:55 UTC (permalink / raw)
To: Andrew Morton, David Rientjes, Vlastimil Babka, Michal Hocko,
Andrea Arcangeli
Cc: mm-commits, Linux-MM
On Mon, Sep 23, 2019 at 3:31 PM Andrew Morton <akpm@linux-foundation.org> wrote:
>
> - almost all of -mm, as below.
I was hoping that we could at least test the THP locality thing? Is it
in your queue at all, or am I supposed to just do it myself?
Linus
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2019-09-23 22:31 Andrew Morton
2019-09-24 0:55 ` incoming Linus Torvalds
0 siblings, 1 reply; 409+ messages in thread
From: Andrew Morton @ 2019-09-23 22:31 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
- a few hot fixes
- ocfs2 updates
- almost all of -mm, as below.
134 patches, based on 619e17cf75dd58905aa67ccd494a6ba5f19d6cc6:
Subsystems affected by this patch series:
hotfixes
ocfs2
slab-generic
slab
slub
kmemleak
kasan
cleanups
debug
pagecache
memcg
gup
pagemap
memory-hotplug
sparsemem
vmalloc
initialization
z3fold
compaction
mempolicy
oom-kill
hugetlb
migration
thp
mmap
madvise
shmem
zswap
zsmalloc
Subsystem: hotfixes
OGAWA Hirofumi <hirofumi@mail.parknet.co.jp>:
fat: work around race with userspace's read via blockdev while mounting
Vitaly Wool <vitalywool@gmail.com>:
Revert "mm/z3fold.c: fix race between migration and destruction"
Arnd Bergmann <arnd@arndb.de>:
mm: add dummy can_do_mlock() helper
Vitaly Wool <vitalywool@gmail.com>:
z3fold: fix retry mechanism in page reclaim
Greg Thelen <gthelen@google.com>:
kbuild: clean compressed initramfs image
Subsystem: ocfs2
Joseph Qi <joseph.qi@linux.alibaba.com>:
ocfs2: use jbd2_inode dirty range scoping
jbd2: remove jbd2_journal_inode_add_[write|wait]
Greg Kroah-Hartman <gregkh@linuxfoundation.org>:
ocfs2: further debugfs cleanups
Guozhonghua <guozhonghua@h3c.com>:
ocfs2: remove unused ocfs2_calc_tree_trunc_credits()
ocfs2: remove unused ocfs2_orphan_scan_exit() declaration
zhengbin <zhengbin13@huawei.com>:
fs/ocfs2/namei.c: remove set but not used variables
fs/ocfs2/file.c: remove set but not used variables
fs/ocfs2/dir.c: remove set but not used variables
Markus Elfring <elfring@users.sourceforge.net>:
ocfs2: delete unnecessary checks before brelse()
Changwei Ge <gechangwei@live.cn>:
ocfs2: wait for recovering done after direct unlock request
ocfs2: checkpoint appending truncate log transaction before flushing
Colin Ian King <colin.king@canonical.com>:
ocfs2: fix spelling mistake "ambigous" -> "ambiguous"
Subsystem: slab-generic
Waiman Long <longman@redhat.com>:
mm, slab: extend slab/shrink to shrink all memcg caches
Subsystem: slab
Waiman Long <longman@redhat.com>:
mm, slab: move memcg_cache_params structure to mm/slab.h
Subsystem: slub
Qian Cai <cai@lca.pw>:
mm/slub.c: fix -Wunused-function compiler warnings
Subsystem: kmemleak
Nicolas Boichat <drinkcat@chromium.org>:
kmemleak: increase DEBUG_KMEMLEAK_EARLY_LOG_SIZE default to 16K
Catalin Marinas <catalin.marinas@arm.com>:
Patch series "mm: kmemleak: Use a memory pool for kmemleak object:
mm: kmemleak: make the tool tolerant to struct scan_area allocation failures
mm: kmemleak: simple memory allocation pool for kmemleak objects
mm: kmemleak: use the memory pool for early allocations
Qian Cai <cai@lca.pw>:
mm/kmemleak.c: record the current memory pool size
mm/kmemleak: increase the max mem pool to 1M
Subsystem: kasan
Walter Wu <walter-zh.wu@mediatek.com>:
kasan: add memory corruption identification for software tag-based mode
Mark Rutland <mark.rutland@arm.com>:
lib/test_kasan.c: add roundtrip tests
Subsystem: cleanups
Christophe JAILLET <christophe.jaillet@wanadoo.fr>:
mm/page_poison.c: fix a typo in a comment
YueHaibing <yuehaibing@huawei.com>:
mm/rmap.c: remove set but not used variable 'cstart'
Matthew Wilcox (Oracle) <willy@infradead.org>:
Patch series "Make working with compound pages easier", v2:
mm: introduce page_size()
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
mm: introduce page_shift()
Matthew Wilcox (Oracle) <willy@infradead.org>:
mm: introduce compound_nr()
Yu Zhao <yuzhao@google.com>:
mm: replace list_move_tail() with add_page_to_lru_list_tail()
Subsystem: debug
Vlastimil Babka <vbabka@suse.cz>:
Patch series "debug_pagealloc improvements through page_owner", v2:
mm, page_owner: record page owner for each subpage
mm, page_owner: keep owner info when freeing the page
mm, page_owner, debug_pagealloc: save and dump freeing stack trace
Subsystem: pagecache
Konstantin Khlebnikov <khlebnikov@yandex-team.ru>:
mm/filemap.c: don't initiate writeback if mapping has no dirty pages
mm/filemap.c: rewrite mapping_needs_writeback in less fancy manner
"Matthew Wilcox (Oracle)" <willy@infradead.org>:
mm: page cache: store only head pages in i_pages
Subsystem: memcg
Chris Down <chris@chrisdown.name>:
mm, memcg: throttle allocators when failing reclaim over memory.high
Roman Gushchin <guro@fb.com>:
mm: memcontrol: switch to rcu protection in drain_all_stock()
Johannes Weiner <hannes@cmpxchg.org>:
mm: vmscan: do not share cgroup iteration between reclaimers
Subsystem: gup
[11~From: John Hubbard <jhubbard@nvidia.com>:
Patch series "mm/gup: add make_dirty arg to put_user_pages_dirty_lock()",:
mm/gup: add make_dirty arg to put_user_pages_dirty_lock()
John Hubbard <jhubbard@nvidia.com>:
drivers/gpu/drm/via: convert put_page() to put_user_page*()
net/xdp: convert put_page() to put_user_page*()
Subsystem: pagemap
Wei Yang <richardw.yang@linux.intel.com>:
mm: remove redundant assignment of entry
Minchan Kim <minchan@kernel.org>:
mm: release the spinlock on zap_pte_range
Nicholas Piggin <npiggin@gmail.com>:
Patch series "mm: remove quicklist page table caches":
mm: remove quicklist page table caches
Mike Rapoport <rppt@linux.ibm.com>:
ia64: switch to generic version of pte allocation
sh: switch to generic version of pte allocation
microblaze: switch to generic version of pte allocation
mm: consolidate pgtable_cache_init() and pgd_cache_init()
Kefeng Wang <wangkefeng.wang@huawei.com>:
mm: do not hash address in print_bad_pte()
Subsystem: memory-hotplug
David Hildenbrand <david@redhat.com>:
mm/memory_hotplug: remove move_pfn_range()
drivers/base/node.c: simplify unregister_memory_block_under_nodes()
drivers/base/memory.c: fixup documentation of removable/phys_index/block_size_bytes
driver/base/memory.c: validate memory block size early
drivers/base/memory.c: don't store end_section_nr in memory blocks
Wei Yang <richardw.yang@linux.intel.com>:
mm/memory_hotplug.c: prevent memory leak when reusing pgdat
David Hildenbrand <david@redhat.com>:
Patch series "mm/memory_hotplug: online_pages() cleanups", v2:
mm/memory_hotplug.c: use PFN_UP / PFN_DOWN in walk_system_ram_range()
mm/memory_hotplug: drop PageReserved() check in online_pages_range()
mm/memory_hotplug: simplify online_pages_range()
mm/memory_hotplug: make sure the pfn is aligned to the order when onlining
mm/memory_hotplug: online_pages cannot be 0 in online_pages()
Alastair D'Silva <alastair@d-silva.org>:
Patch series "Add bounds check for Hotplugged memory", v3:
mm/memory_hotplug.c: add a bounds check to check_hotplug_memory_range()
mm/memremap.c: add a bounds check in devm_memremap_pages()
Souptick Joarder <jrdr.linux@gmail.com>:
mm/memory_hotplug.c: s/is/if
Subsystem: sparsemem
Lecopzer Chen <lecopzer.chen@mediatek.com>:
mm/sparse.c: fix memory leak of sparsemap_buf in aligned memory
mm/sparse.c: fix ALIGN() without power of 2 in sparse_buffer_alloc()
Wei Yang <richardw.yang@linux.intel.com>:
mm/sparse.c: use __nr_to_section(section_nr) to get mem_section
Alastair D'Silva <alastair@d-silva.org>:
mm/sparse.c: don't manually decrement num_poisoned_pages
"Alastair D'Silva" <alastair@d-silva.org>:
mm/sparse.c: remove NULL check in clear_hwpoisoned_pages()
Subsystem: vmalloc
"Uladzislau Rezki (Sony)" <urezki@gmail.com>:
mm/vmalloc: do not keep unpurged areas in the busy tree
Pengfei Li <lpf.vector@gmail.com>:
mm/vmalloc: modify struct vmap_area to reduce its size
Austin Kim <austindh.kim@gmail.com>:
mm/vmalloc.c: move 'area->pages' after if statement
Subsystem: initialization
Mike Rapoport <rppt@linux.ibm.com>:
mm: use CPU_BITS_NONE to initialize init_mm.cpu_bitmask
Qian Cai <cai@lca.pw>:
mm: silence -Woverride-init/initializer-overrides
Subsystem: z3fold
Vitaly Wool <vitalywool@gmail.com>:
z3fold: fix memory leak in kmem cache
Subsystem: compaction
Yafang Shao <laoar.shao@gmail.com>:
mm/compaction.c: clear total_{migrate,free}_scanned before scanning a new zone
Pengfei Li <lpf.vector@gmail.com>:
mm/compaction.c: remove unnecessary zone parameter in isolate_migratepages()
Subsystem: mempolicy
Kefeng Wang <wangkefeng.wang@huawei.com>:
mm/mempolicy.c: remove unnecessary nodemask check in kernel_migrate_pages()
Subsystem: oom-kill
Joel Savitz <jsavitz@redhat.com>:
mm/oom_kill.c: add task UID to info message on an oom kill
Tetsuo Handa <penguin-kernel@i-love.sakura.ne.jp>:
memcg, oom: don't require __GFP_FS when invoking memcg OOM killer
Edward Chron <echron@arista.com>:
mm/oom: add oom_score_adj and pgtables to Killed process message
Yi Wang <wang.yi59@zte.com.cn>:
mm/oom_kill.c: fix oom_cpuset_eligible() comment
Michal Hocko <mhocko@suse.com>:
mm, oom: consider present pages for the node size
Qian Cai <cai@lca.pw>:
mm/memcontrol.c: fix a -Wunused-function warning
Michal Hocko <mhocko@suse.com>:
memcg, kmem: deprecate kmem.limit_in_bytes
Subsystem: hugetlb
Hillf Danton <hdanton@sina.com>:
Patch series "address hugetlb page allocation stalls", v2:
mm, reclaim: make should_continue_reclaim perform dryrun detection
Vlastimil Babka <vbabka@suse.cz>:
mm, reclaim: cleanup should_continue_reclaim()
mm, compaction: raise compaction priority after it withdrawns
Mike Kravetz <mike.kravetz@oracle.com>:
hugetlbfs: don't retry when pool page allocations start to fail
Subsystem: migration
Pingfan Liu <kernelfans@gmail.com>:
mm/migrate.c: clean up useless code in migrate_vma_collect_pmd()
Subsystem: thp
Kefeng Wang <wangkefeng.wang@huawei.com>:
thp: update split_huge_page_pmd() comment
Song Liu <songliubraving@fb.com>:
Patch series "Enable THP for text section of non-shmem files", v10;:
filemap: check compound_head(page)->mapping in filemap_fault()
filemap: check compound_head(page)->mapping in pagecache_get_page()
filemap: update offset check in filemap_fault()
mm,thp: stats for file backed THP
khugepaged: rename collapse_shmem() and khugepaged_scan_shmem()
mm,thp: add read-only THP support for (non-shmem) FS
mm,thp: avoid writes to file with THP in pagecache
Yang Shi <yang.shi@linux.alibaba.com>:
Patch series "Make deferred split shrinker memcg aware", v6:
mm: thp: extract split_queue_* into a struct
mm: move mem_cgroup_uncharge out of __page_cache_release()
mm: shrinker: make shrinker not depend on memcg kmem
mm: thp: make deferred split shrinker memcg aware
Song Liu <songliubraving@fb.com>:
Patch series "THP aware uprobe", v13:
mm: move memcmp_pages() and pages_identical()
uprobe: use original page when all uprobes are removed
mm, thp: introduce FOLL_SPLIT_PMD
uprobe: use FOLL_SPLIT_PMD instead of FOLL_SPLIT
khugepaged: enable collapse pmd for pte-mapped THP
uprobe: collapse THP pmd after removing all uprobes
Subsystem: mmap
Alexandre Ghiti <alex@ghiti.fr>:
Patch series "Provide generic top-down mmap layout functions", v6:
mm, fs: move randomize_stack_top from fs to mm
arm64: make use of is_compat_task instead of hardcoding this test
arm64: consider stack randomization for mmap base only when necessary
arm64, mm: move generic mmap layout functions to mm
arm64, mm: make randomization selected by generic topdown mmap layout
arm: properly account for stack randomization and stack guard gap
arm: use STACK_TOP when computing mmap base address
arm: use generic mmap top-down layout and brk randomization
mips: properly account for stack randomization and stack guard gap
mips: use STACK_TOP when computing mmap base address
mips: adjust brk randomization offset to fit generic version
mips: replace arch specific way to determine 32bit task with generic version
mips: use generic mmap top-down layout and brk randomization
riscv: make mmap allocation top-down by default
Wei Yang <richardw.yang@linux.intel.com>:
mm/mmap.c: refine find_vma_prev() with rb_last()
Ivan Khoronzhuk <ivan.khoronzhuk@linaro.org>:
mm: mmap: increase sockets maximum memory size pgoff for 32bits
Subsystem: madvise
Mike Rapoport <rppt@linux.ibm.com>:
mm/madvise: reduce code duplication in error handling paths
Subsystem: shmem
Miles Chen <miles.chen@mediatek.com>:
shmem: fix obsolete comment in shmem_getpage_gfp()
Subsystem: zswap
Hui Zhu <teawaterz@linux.alibaba.com>:
zpool: add malloc_support_movable to zpool_driver
zswap: use movable memory if zpool support allocate movable memory
Vitaly Wool <vitalywool@gmail.com>:
zswap: do not map same object twice
Subsystem: zsmalloc
Qian Cai <cai@lca.pw>:
mm/zsmalloc.c: fix a -Wunused-function warning
Documentation/ABI/testing/sysfs-kernel-slab | 13
Documentation/admin-guide/cgroup-v1/memory.rst | 4
Documentation/admin-guide/kernel-parameters.txt | 2
arch/Kconfig | 11
arch/alpha/include/asm/pgalloc.h | 2
arch/alpha/include/asm/pgtable.h | 5
arch/arc/include/asm/pgalloc.h | 1
arch/arc/include/asm/pgtable.h | 5
arch/arm/Kconfig | 1
arch/arm/include/asm/pgalloc.h | 2
arch/arm/include/asm/pgtable-nommu.h | 5
arch/arm/include/asm/pgtable.h | 2
arch/arm/include/asm/processor.h | 2
arch/arm/kernel/process.c | 5
arch/arm/mm/flush.c | 7
arch/arm/mm/mmap.c | 80 -----
arch/arm64/Kconfig | 2
arch/arm64/include/asm/pgalloc.h | 2
arch/arm64/include/asm/pgtable.h | 2
arch/arm64/include/asm/processor.h | 2
arch/arm64/kernel/process.c | 8
arch/arm64/mm/flush.c | 3
arch/arm64/mm/mmap.c | 84 -----
arch/arm64/mm/pgd.c | 2
arch/c6x/include/asm/pgtable.h | 5
arch/csky/include/asm/pgalloc.h | 2
arch/csky/include/asm/pgtable.h | 5
arch/h8300/include/asm/pgtable.h | 6
arch/hexagon/include/asm/pgalloc.h | 2
arch/hexagon/include/asm/pgtable.h | 3
arch/hexagon/mm/Makefile | 2
arch/hexagon/mm/pgalloc.c | 10
arch/ia64/Kconfig | 4
arch/ia64/include/asm/pgalloc.h | 64 ----
arch/ia64/include/asm/pgtable.h | 5
arch/ia64/mm/init.c | 2
arch/m68k/include/asm/pgtable_mm.h | 7
arch/m68k/include/asm/pgtable_no.h | 7
arch/microblaze/include/asm/pgalloc.h | 128 --------
arch/microblaze/include/asm/pgtable.h | 7
arch/microblaze/mm/pgtable.c | 4
arch/mips/Kconfig | 2
arch/mips/include/asm/pgalloc.h | 2
arch/mips/include/asm/pgtable.h | 5
arch/mips/include/asm/processor.h | 5
arch/mips/mm/mmap.c | 124 +-------
arch/nds32/include/asm/pgalloc.h | 2
arch/nds32/include/asm/pgtable.h | 2
arch/nios2/include/asm/pgalloc.h | 2
arch/nios2/include/asm/pgtable.h | 2
arch/openrisc/include/asm/pgalloc.h | 2
arch/openrisc/include/asm/pgtable.h | 5
arch/parisc/include/asm/pgalloc.h | 2
arch/parisc/include/asm/pgtable.h | 2
arch/powerpc/include/asm/pgalloc.h | 2
arch/powerpc/include/asm/pgtable.h | 1
arch/powerpc/mm/book3s64/hash_utils.c | 2
arch/powerpc/mm/book3s64/iommu_api.c | 7
arch/powerpc/mm/hugetlbpage.c | 2
arch/riscv/Kconfig | 12
arch/riscv/include/asm/pgalloc.h | 4
arch/riscv/include/asm/pgtable.h | 5
arch/s390/include/asm/pgtable.h | 6
arch/sh/include/asm/pgalloc.h | 56 ---
arch/sh/include/asm/pgtable.h | 5
arch/sh/mm/Kconfig | 3
arch/sh/mm/nommu.c | 4
arch/sparc/include/asm/pgalloc_32.h | 2
arch/sparc/include/asm/pgalloc_64.h | 2
arch/sparc/include/asm/pgtable_32.h | 5
arch/sparc/include/asm/pgtable_64.h | 1
arch/sparc/mm/init_32.c | 1
arch/um/include/asm/pgalloc.h | 2
arch/um/include/asm/pgtable.h | 2
arch/unicore32/include/asm/pgalloc.h | 2
arch/unicore32/include/asm/pgtable.h | 2
arch/x86/include/asm/pgtable_32.h | 2
arch/x86/include/asm/pgtable_64.h | 3
arch/x86/mm/pgtable.c | 6
arch/xtensa/include/asm/pgtable.h | 1
arch/xtensa/include/asm/tlbflush.h | 3
drivers/base/memory.c | 44 +-
drivers/base/node.c | 55 +--
drivers/crypto/chelsio/chtls/chtls_io.c | 5
drivers/gpu/drm/via/via_dmablit.c | 10
drivers/infiniband/core/umem.c | 5
drivers/infiniband/hw/hfi1/user_pages.c | 5
drivers/infiniband/hw/qib/qib_user_pages.c | 5
drivers/infiniband/hw/usnic/usnic_uiom.c | 5
drivers/infiniband/sw/siw/siw_mem.c | 10
drivers/staging/android/ion/ion_system_heap.c | 4
drivers/target/tcm_fc/tfc_io.c | 3
drivers/vfio/vfio_iommu_spapr_tce.c | 8
fs/binfmt_elf.c | 20 -
fs/fat/dir.c | 13
fs/fat/fatent.c | 3
fs/inode.c | 3
fs/io_uring.c | 2
fs/jbd2/journal.c | 2
fs/jbd2/transaction.c | 12
fs/ocfs2/alloc.c | 20 +
fs/ocfs2/aops.c | 13
fs/ocfs2/blockcheck.c | 26 -
fs/ocfs2/cluster/heartbeat.c | 109 +------
fs/ocfs2/dir.c | 3
fs/ocfs2/dlm/dlmcommon.h | 1
fs/ocfs2/dlm/dlmdebug.c | 55 ---
fs/ocfs2/dlm/dlmdebug.h | 16 -
fs/ocfs2/dlm/dlmdomain.c | 7
fs/ocfs2/dlm/dlmunlock.c | 23 +
fs/ocfs2/dlmglue.c | 29 -
fs/ocfs2/extent_map.c | 3
fs/ocfs2/file.c | 13
fs/ocfs2/inode.c | 2
fs/ocfs2/journal.h | 42 --
fs/ocfs2/namei.c | 2
fs/ocfs2/ocfs2.h | 3
fs/ocfs2/super.c | 10
fs/open.c | 8
fs/proc/meminfo.c | 8
fs/proc/task_mmu.c | 6
include/asm-generic/pgalloc.h | 5
include/asm-generic/pgtable.h | 7
include/linux/compaction.h | 22 +
include/linux/fs.h | 32 ++
include/linux/huge_mm.h | 9
include/linux/hugetlb.h | 2
include/linux/jbd2.h | 2
include/linux/khugepaged.h | 12
include/linux/memcontrol.h | 23 -
include/linux/memory.h | 7
include/linux/memory_hotplug.h | 1
include/linux/mm.h | 37 ++
include/linux/mm_types.h | 1
include/linux/mmzone.h | 14
include/linux/page_ext.h | 1
include/linux/pagemap.h | 10
include/linux/quicklist.h | 94 ------
include/linux/shrinker.h | 7
include/linux/slab.h | 62 ----
include/linux/vmalloc.h | 20 -
include/linux/zpool.h | 3
init/main.c | 6
kernel/events/uprobes.c | 81 ++++-
kernel/resource.c | 4
kernel/sched/idle.c | 1
kernel/sysctl.c | 6
lib/Kconfig.debug | 15
lib/Kconfig.kasan | 8
lib/iov_iter.c | 2
lib/show_mem.c | 5
lib/test_kasan.c | 41 ++
mm/Kconfig | 16 -
mm/Kconfig.debug | 4
mm/Makefile | 4
mm/compaction.c | 50 +--
mm/filemap.c | 168 ++++------
mm/gup.c | 125 +++-----
mm/huge_memory.c | 129 ++++++--
mm/hugetlb.c | 89 +++++
mm/hugetlb_cgroup.c | 2
mm/init-mm.c | 2
mm/kasan/common.c | 32 +-
mm/kasan/kasan.h | 14
mm/kasan/report.c | 44 ++
mm/kasan/tags_report.c | 24 +
mm/khugepaged.c | 372 ++++++++++++++++++++----
mm/kmemleak.c | 338 +++++----------------
mm/ksm.c | 18 -
mm/madvise.c | 52 +--
mm/memcontrol.c | 188 ++++++++++--
mm/memfd.c | 2
mm/memory.c | 21 +
mm/memory_hotplug.c | 120 ++++---
mm/mempolicy.c | 4
mm/memremap.c | 5
mm/migrate.c | 13
mm/mmap.c | 12
mm/mmu_gather.c | 2
mm/nommu.c | 2
mm/oom_kill.c | 30 +
mm/page_alloc.c | 27 +
mm/page_owner.c | 127 +++++---
mm/page_poison.c | 2
mm/page_vma_mapped.c | 3
mm/quicklist.c | 103 ------
mm/rmap.c | 25 -
mm/shmem.c | 12
mm/slab.h | 64 ++++
mm/slab_common.c | 37 ++
mm/slob.c | 2
mm/slub.c | 22 -
mm/sparse.c | 25 +
mm/swap.c | 16 -
mm/swap_state.c | 6
mm/util.c | 126 +++++++-
mm/vmalloc.c | 84 +++--
mm/vmscan.c | 163 ++++------
mm/vmstat.c | 2
mm/z3fold.c | 154 ++-------
mm/zpool.c | 16 +
mm/zsmalloc.c | 23 -
mm/zswap.c | 15
net/xdp/xdp_umem.c | 9
net/xdp/xsk.c | 2
usr/Makefile | 3
206 files changed, 2385 insertions(+), 2533 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2019-08-30 23:04 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2019-08-30 23:04 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
7 fixes, based on 846d2db3e00048da3f650e0cfb0b8d67669cec3e:
Roman Gushchin <guro@fb.com>:
mm: memcontrol: flush percpu slab vmstats on kmem offlining
Andrew Morton <akpm@linux-foundation.org>:
mm/zsmalloc.c: fix build when CONFIG_COMPACTION=n
Roman Gushchin <guro@fb.com>:
mm, memcg: partially revert "mm/memcontrol.c: keep local VM counters in sync with the hierarchical ones"
"Gustavo A. R. Silva" <gustavo@embeddedor.com>:
mm/z3fold.c: fix lock/unlock imbalance in z3fold_page_isolate
Dmitry Safonov <dima@arista.com>:
mailmap: add aliases for Dmitry Safonov
Michal Hocko <mhocko@suse.com>:
mm, memcg: do not set reclaim_state on soft limit reclaim
Shakeel Butt <shakeelb@google.com>:
mm: memcontrol: fix percpu vmstats and vmevents flush
.mailmap | 3 ++
include/linux/mmzone.h | 5 ++--
mm/memcontrol.c | 53 ++++++++++++++++++++++++++++++++-----------------
mm/vmscan.c | 5 ++--
mm/z3fold.c | 1
mm/zsmalloc.c | 2 +
6 files changed, 47 insertions(+), 22 deletions(-)
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2019-08-25 0:54 Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2019-08-25 0:54 UTC (permalink / raw)
To: Linus Torvalds; +Cc: mm-commits, linux-mm
11 fixes, based on 361469211f876e67d7ca3d3d29e6d1c3e313d0f1:
Henry Burns <henryburns@google.com>:
mm/z3fold.c: fix race between migration and destruction
David Rientjes <rientjes@google.com>:
mm, page_alloc: move_freepages should not examine struct page of reserved memory
Qian Cai <cai@lca.pw>:
parisc: fix compilation errrors
Roman Gushchin <guro@fb.com>:
mm: memcontrol: flush percpu vmstats before releasing memcg
mm: memcontrol: flush percpu vmevents before releasing memcg
Jason Xing <kerneljasonxing@linux.alibaba.com>:
psi: get poll_work to run when calling poll syscall next time
Oleg Nesterov <oleg@redhat.com>:
userfaultfd_release: always remove uffd flags and clear vm_userfaultfd_ctx
Vlastimil Babka <vbabka@suse.cz>:
mm, page_owner: handle THP splits correctly
Henry Burns <henryburns@google.com>:
mm/zsmalloc.c: migration can leave pages in ZS_EMPTY indefinitely
mm/zsmalloc.c: fix race condition in zs_destroy_pool
Andrey Ryabinin <aryabinin@virtuozzo.com>:
mm/kasan: fix false positive invalid-free reports with CONFIG_KASAN_SW_TAGS=y
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2019-07-17 16:13 ` incoming Linus Torvalds
2019-07-17 17:09 ` incoming Christian Brauner
@ 2019-07-17 18:13 ` Vlastimil Babka
1 sibling, 0 replies; 409+ messages in thread
From: Vlastimil Babka @ 2019-07-17 18:13 UTC (permalink / raw)
To: Linus Torvalds
Cc: Linux List Kernel Mailing, linux-mm, Jonathan Corbet, Thorsten Leemhuis
On 7/17/19 6:13 PM, Linus Torvalds wrote:
> On Wed, Jul 17, 2019 at 1:47 AM Vlastimil Babka <vbabka@suse.cz> wrote:
>>
>> So I've tried now to provide an example what I had in mind, below.
>
> I'll take it as a trial. I added one-line notes about coda and the
> PTRACE_GET_SYSCALL_INFO interface too.
Thanks.
> I do hope that eventually I'll just get pull requests,
Very much agree, that was also discussed at length in the LSF/MM mm
process session I've linked.
> and they'll
> have more of a "theme" than this all (*)
I'll check if the first patch bomb would be more amenable to that, as I
plan to fill in the mm part for 5.3 on LinuxChanges wiki, but for a
merge commit it's too late.
> Linus
>
> (*) Although in many ways, the theme for Andrew is "falls through the
> cracks otherwise" so I'm not really complaining. This has been working
> for years and years.
Nevermind the misc stuff that much, but I think mm itself is more
important and deserves what other subsystems have.
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2019-07-17 16:13 ` incoming Linus Torvalds
@ 2019-07-17 17:09 ` Christian Brauner
2019-07-17 18:13 ` incoming Vlastimil Babka
1 sibling, 0 replies; 409+ messages in thread
From: Christian Brauner @ 2019-07-17 17:09 UTC (permalink / raw)
To: Linus Torvalds
Cc: Vlastimil Babka, Linux List Kernel Mailing, linux-mm,
Jonathan Corbet, Thorsten Leemhuis
On Wed, Jul 17, 2019 at 09:13:26AM -0700, Linus Torvalds wrote:
> On Wed, Jul 17, 2019 at 1:47 AM Vlastimil Babka <vbabka@suse.cz> wrote:
> >
> > So I've tried now to provide an example what I had in mind, below.
>
> I'll take it as a trial. I added one-line notes about coda and the
> PTRACE_GET_SYSCALL_INFO interface too.
>
> I do hope that eventually I'll just get pull requests, and they'll
> have more of a "theme" than this all (*)
>
> Linus
>
> (*) Although in many ways, the theme for Andrew is "falls through the
> cracks otherwise" so I'm not really complaining. This has been working
I put all pid{fd}/clone{3} which is mostly related to pid.c, exit.c,
fork.c into my tree and try to give it a consistent theme for the prs I
sent. And that at least from my perspective that worked and was pretty
easy to coordinate with Andrew. That should hopefully make it a little
easier to theme the -mm tree overall going forward.
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2019-07-17 8:47 ` incoming Vlastimil Babka
2019-07-17 8:57 ` incoming Bhaskar Chowdhury
@ 2019-07-17 16:13 ` Linus Torvalds
2019-07-17 17:09 ` incoming Christian Brauner
2019-07-17 18:13 ` incoming Vlastimil Babka
1 sibling, 2 replies; 409+ messages in thread
From: Linus Torvalds @ 2019-07-17 16:13 UTC (permalink / raw)
To: Vlastimil Babka
Cc: Linux List Kernel Mailing, linux-mm, Jonathan Corbet, Thorsten Leemhuis
On Wed, Jul 17, 2019 at 1:47 AM Vlastimil Babka <vbabka@suse.cz> wrote:
>
> So I've tried now to provide an example what I had in mind, below.
I'll take it as a trial. I added one-line notes about coda and the
PTRACE_GET_SYSCALL_INFO interface too.
I do hope that eventually I'll just get pull requests, and they'll
have more of a "theme" than this all (*)
Linus
(*) Although in many ways, the theme for Andrew is "falls through the
cracks otherwise" so I'm not really complaining. This has been working
for years and years.
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2019-07-17 8:47 ` incoming Vlastimil Babka
@ 2019-07-17 8:57 ` Bhaskar Chowdhury
2019-07-17 16:13 ` incoming Linus Torvalds
1 sibling, 0 replies; 409+ messages in thread
From: Bhaskar Chowdhury @ 2019-07-17 8:57 UTC (permalink / raw)
To: Vlastimil Babka
Cc: linux-kernel, Linus Torvalds, linux-mm, Jonathan Corbet,
Thorsten Leemhuis
[-- Attachment #1: Type: text/plain, Size: 2496 bytes --]
Cool !!
On 10:47 Wed 17 Jul , Vlastimil Babka wrote:
>On 7/17/19 1:25 AM, Andrew Morton wrote:
>>
>> Most of the rest of MM and just about all of the rest of everything
>> else.
>
>Hi,
>
>as I've mentioned at LSF/MM [1], I think it would be nice if mm pull
>requests had summaries similar to other subsystems. I see they are now
>more structured (thanks!), but they are now probably hitting the limit
>of what scripting can do to produce a high-level summary for human
>readers (unless patch authors themselves provide a blurb that can be
>extracted later?).
>
>So I've tried now to provide an example what I had in mind, below. Maybe
>it's too concise - if there were "larger" features in this pull request,
>they would probably benefit from more details. I'm CCing the known (to
>me) consumers of these mails to judge :) Note I've only covered mm, and
>core stuff that I think will be interesting to wide audience (change in
>LIST_POISON2 value? I'm sure as hell glad to know about that one :)
>
>Feel free to include this in the merge commit, if you find it useful.
>
>Thanks,
>Vlastimil
>
>[1] https://lwn.net/Articles/787705/
>
>-----
>
>- z3fold fixes and enhancements by Henry Burns and Vitaly Wool
>- more accurate reclaimed slab caches calculations by Yafang Shao
>- fix MAP_UNINITIALIZED UAPI symbol to not depend on config, by
>Christoph Hellwig
>- !CONFIG_MMU fixes by Christoph Hellwig
>- new novmcoredd parameter to omit device dumps from vmcore, by Kairui Song
>- new test_meminit module for testing heap and pagealloc initialization,
>by Alexander Potapenko
>- ioremap improvements for huge mappings, by Anshuman Khandual
>- generalize kprobe page fault handling, by Anshuman Khandual
>- device-dax hotplug fixes and improvements, by Pavel Tatashin
>- enable synchronous DAX fault on powerpc, by Aneesh Kumar K.V
>- add pte_devmap() support for arm64, by Robin Murphy
>- unify locked_vm accounting with a helper, by Daniel Jordan
>- several misc fixes
>
>core/lib
>- new typeof_member() macro including some users, by Alexey Dobriyan
>- make BIT() and GENMASK() available in asm, by Masahiro Yamada
>- changed LIST_POISON2 on x86_64 to 0xdead000000000122 for better code
>generation, by Alexey Dobriyan
>- rbtree code size optimizations, by Michel Lespinasse
>- convert struct pid count to refcount_t, by Joel Fernandes
>
>get_maintainer.pl
>- add --no-moderated switch to skip moderated ML's, by Joe Perches
>
>
[-- Attachment #2: signature.asc --]
[-- Type: application/pgp-signature, Size: 488 bytes --]
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
[not found] <20190716162536.bb52b8f34a8ecf5331a86a42@linux-foundation.org>
@ 2019-07-17 8:47 ` Vlastimil Babka
2019-07-17 8:57 ` incoming Bhaskar Chowdhury
2019-07-17 16:13 ` incoming Linus Torvalds
0 siblings, 2 replies; 409+ messages in thread
From: Vlastimil Babka @ 2019-07-17 8:47 UTC (permalink / raw)
To: linux-kernel, Linus Torvalds
Cc: linux-mm, Jonathan Corbet, Thorsten Leemhuis, LKML
On 7/17/19 1:25 AM, Andrew Morton wrote:
>
> Most of the rest of MM and just about all of the rest of everything
> else.
Hi,
as I've mentioned at LSF/MM [1], I think it would be nice if mm pull
requests had summaries similar to other subsystems. I see they are now
more structured (thanks!), but they are now probably hitting the limit
of what scripting can do to produce a high-level summary for human
readers (unless patch authors themselves provide a blurb that can be
extracted later?).
So I've tried now to provide an example what I had in mind, below. Maybe
it's too concise - if there were "larger" features in this pull request,
they would probably benefit from more details. I'm CCing the known (to
me) consumers of these mails to judge :) Note I've only covered mm, and
core stuff that I think will be interesting to wide audience (change in
LIST_POISON2 value? I'm sure as hell glad to know about that one :)
Feel free to include this in the merge commit, if you find it useful.
Thanks,
Vlastimil
[1] https://lwn.net/Articles/787705/
-----
- z3fold fixes and enhancements by Henry Burns and Vitaly Wool
- more accurate reclaimed slab caches calculations by Yafang Shao
- fix MAP_UNINITIALIZED UAPI symbol to not depend on config, by
Christoph Hellwig
- !CONFIG_MMU fixes by Christoph Hellwig
- new novmcoredd parameter to omit device dumps from vmcore, by Kairui Song
- new test_meminit module for testing heap and pagealloc initialization,
by Alexander Potapenko
- ioremap improvements for huge mappings, by Anshuman Khandual
- generalize kprobe page fault handling, by Anshuman Khandual
- device-dax hotplug fixes and improvements, by Pavel Tatashin
- enable synchronous DAX fault on powerpc, by Aneesh Kumar K.V
- add pte_devmap() support for arm64, by Robin Murphy
- unify locked_vm accounting with a helper, by Daniel Jordan
- several misc fixes
core/lib
- new typeof_member() macro including some users, by Alexey Dobriyan
- make BIT() and GENMASK() available in asm, by Masahiro Yamada
- changed LIST_POISON2 on x86_64 to 0xdead000000000122 for better code
generation, by Alexey Dobriyan
- rbtree code size optimizations, by Michel Lespinasse
- convert struct pid count to refcount_t, by Joel Fernandes
get_maintainer.pl
- add --no-moderated switch to skip moderated ML's, by Joe Perches
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2007-05-04 19:24 ` incoming Greg KH
@ 2007-05-04 19:29 ` Roland McGrath
0 siblings, 0 replies; 409+ messages in thread
From: Roland McGrath @ 2007-05-04 19:29 UTC (permalink / raw)
To: Greg KH
Cc: Andrew Morton, Linus Torvalds, Hugh Dickins, Christoph Lameter,
David S. Miller, Andi Kleen, Luck, Tony, Rik van Riel,
Benjamin Herrenschmidt, linux-kernel, linux-mm, Stephen Smalley
> ABI changes are not a problem for -stable, so don't let that stop anyone
> :)
In fact this is the harmless sort (changes only the error code of a
failure case) that might actually go in if there were any important
reason. But the smiley stands.
Thanks,
Roland
--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org. For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2007-05-04 18:57 ` incoming Roland McGrath
@ 2007-05-04 19:24 ` Greg KH
2007-05-04 19:29 ` incoming Roland McGrath
0 siblings, 1 reply; 409+ messages in thread
From: Greg KH @ 2007-05-04 19:24 UTC (permalink / raw)
To: Roland McGrath
Cc: Andrew Morton, Linus Torvalds, Hugh Dickins, Christoph Lameter,
David S. Miller, Andi Kleen, Luck, Tony, Rik van Riel,
Benjamin Herrenschmidt, linux-kernel, linux-mm, Stephen Smalley
On Fri, May 04, 2007 at 11:57:21AM -0700, Roland McGrath wrote:
> > Ah. The patch affects security code, but it doesn't actually address any
> > insecurity. I didn't think it was needed for -stable?
>
> I would not recommend it for -stable.
> It is an ABI change for the case of a security refusal.
ABI changes are not a problem for -stable, so don't let that stop anyone
:)
thanks,
greg k-h
--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org. For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2007-05-04 16:14 ` incoming Andrew Morton
2007-05-04 17:02 ` incoming Greg KH
@ 2007-05-04 18:57 ` Roland McGrath
2007-05-04 19:24 ` incoming Greg KH
1 sibling, 1 reply; 409+ messages in thread
From: Roland McGrath @ 2007-05-04 18:57 UTC (permalink / raw)
To: Andrew Morton
Cc: Greg KH, Linus Torvalds, Hugh Dickins, Christoph Lameter,
David S. Miller, Andi Kleen, Luck, Tony, Rik van Riel,
Benjamin Herrenschmidt, linux-kernel, linux-mm, Stephen Smalley
> Ah. The patch affects security code, but it doesn't actually address any
> insecurity. I didn't think it was needed for -stable?
I would not recommend it for -stable.
It is an ABI change for the case of a security refusal.
Thanks,
Roland
--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org. For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2007-05-04 16:14 ` incoming Andrew Morton
@ 2007-05-04 17:02 ` Greg KH
2007-05-04 18:57 ` incoming Roland McGrath
1 sibling, 0 replies; 409+ messages in thread
From: Greg KH @ 2007-05-04 17:02 UTC (permalink / raw)
To: Andrew Morton
Cc: Linus Torvalds, Hugh Dickins, Christoph Lameter, David S. Miller,
Andi Kleen, Luck, Tony, Rik van Riel, Benjamin Herrenschmidt,
linux-kernel, linux-mm, Roland McGrath, Stephen Smalley
On Fri, May 04, 2007 at 09:14:34AM -0700, Andrew Morton wrote:
> On Fri, 4 May 2007 06:37:28 -0700 Greg KH <greg@kroah.com> wrote:
>
> > On Wed, May 02, 2007 at 03:02:52PM -0700, Andrew Morton wrote:
> > > - One little security patch
> >
> > Care to cc: linux-stable with it so we can do a new 2.6.21 release with
> > it if needed?
> >
>
> Ah. The patch affects security code, but it doesn't actually address any
> insecurity. I didn't think it was needed for -stable?
Ah, ok, I read "security" as fixing a insecure problem, my mistake :)
thanks,
greg k-h
--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org. For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2007-05-04 13:37 ` incoming Greg KH
@ 2007-05-04 16:14 ` Andrew Morton
2007-05-04 17:02 ` incoming Greg KH
2007-05-04 18:57 ` incoming Roland McGrath
0 siblings, 2 replies; 409+ messages in thread
From: Andrew Morton @ 2007-05-04 16:14 UTC (permalink / raw)
To: Greg KH
Cc: Linus Torvalds, Hugh Dickins, Christoph Lameter, David S. Miller,
Andi Kleen, Luck, Tony, Rik van Riel, Benjamin Herrenschmidt,
linux-kernel, linux-mm, Roland McGrath, Stephen Smalley
On Fri, 4 May 2007 06:37:28 -0700 Greg KH <greg@kroah.com> wrote:
> On Wed, May 02, 2007 at 03:02:52PM -0700, Andrew Morton wrote:
> > - One little security patch
>
> Care to cc: linux-stable with it so we can do a new 2.6.21 release with
> it if needed?
>
Ah. The patch affects security code, but it doesn't actually address any
insecurity. I didn't think it was needed for -stable?
From: Roland McGrath <roland@redhat.com>
wait* syscalls return -ECHILD even when an individual PID of a live child
was requested explicitly, when security_task_wait denies the operation.
This means that something like a broken SELinux policy can produce an
unexpected failure that looks just like a bug with wait or ptrace or
something.
This patch makes do_wait return -EACCES (or other appropriate error returned
from security_task_wait() instead of -ECHILD if some children were ruled out
solely because security_task_wait failed.
[jmorris@namei.org: switch error code to EACCES]
Signed-off-by: Roland McGrath <roland@redhat.com>
Cc: Stephen Smalley <sds@tycho.nsa.gov>
Cc: Chris Wright <chrisw@sous-sol.org>
Cc: James Morris <jmorris@namei.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
kernel/exit.c | 17 +++++++++++++++--
1 files changed, 15 insertions(+), 2 deletions(-)
diff -puN kernel/exit.c~return-eperm-not-echild-on-security_task_wait-failure kernel/exit.c
--- a/kernel/exit.c~return-eperm-not-echild-on-security_task_wait-failure
+++ a/kernel/exit.c
@@ -1033,6 +1033,8 @@ asmlinkage void sys_exit_group(int error
static int eligible_child(pid_t pid, int options, struct task_struct *p)
{
+ int err;
+
if (pid > 0) {
if (p->pid != pid)
return 0;
@@ -1066,8 +1068,9 @@ static int eligible_child(pid_t pid, int
if (delay_group_leader(p))
return 2;
- if (security_task_wait(p))
- return 0;
+ err = security_task_wait(p);
+ if (err)
+ return err;
return 1;
}
@@ -1449,6 +1452,7 @@ static long do_wait(pid_t pid, int optio
DECLARE_WAITQUEUE(wait, current);
struct task_struct *tsk;
int flag, retval;
+ int allowed, denied;
add_wait_queue(¤t->signal->wait_chldexit,&wait);
repeat:
@@ -1457,6 +1461,7 @@ repeat:
* match our criteria, even if we are not able to reap it yet.
*/
flag = 0;
+ allowed = denied = 0;
current->state = TASK_INTERRUPTIBLE;
read_lock(&tasklist_lock);
tsk = current;
@@ -1472,6 +1477,12 @@ repeat:
if (!ret)
continue;
+ if (unlikely(ret < 0)) {
+ denied = ret;
+ continue;
+ }
+ allowed = 1;
+
switch (p->state) {
case TASK_TRACED:
/*
@@ -1570,6 +1581,8 @@ check_continued:
goto repeat;
}
retval = -ECHILD;
+ if (unlikely(denied) && !allowed)
+ retval = denied;
end:
current->state = TASK_RUNNING;
remove_wait_queue(¤t->signal->wait_chldexit,&wait);
_
--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org. For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2007-05-02 22:02 incoming Andrew Morton
2007-05-02 22:31 ` incoming Benjamin Herrenschmidt
2007-05-03 7:55 ` incoming Russell King
@ 2007-05-04 13:37 ` Greg KH
2007-05-04 16:14 ` incoming Andrew Morton
2 siblings, 1 reply; 409+ messages in thread
From: Greg KH @ 2007-05-04 13:37 UTC (permalink / raw)
To: Andrew Morton
Cc: Linus Torvalds, Hugh Dickins, Christoph Lameter, David S. Miller,
Andi Kleen, Luck, Tony, Rik van Riel, Benjamin Herrenschmidt,
linux-kernel, linux-mm
On Wed, May 02, 2007 at 03:02:52PM -0700, Andrew Morton wrote:
> - One little security patch
Care to cc: linux-stable with it so we can do a new 2.6.21 release with
it if needed?
thanks,
greg k-h
--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org. For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2007-05-03 7:55 ` incoming Russell King
@ 2007-05-03 8:05 ` Andrew Morton
0 siblings, 0 replies; 409+ messages in thread
From: Andrew Morton @ 2007-05-03 8:05 UTC (permalink / raw)
To: Russell King
Cc: Linus Torvalds, Hugh Dickins, Christoph Lameter, David S. Miller,
Andi Kleen, Luck, Tony, Rik van Riel, Benjamin Herrenschmidt,
linux-kernel, linux-mm
On Thu, 3 May 2007 08:55:43 +0100 Russell King <rmk+lkml@arm.linux.org.uk> wrote:
> On Wed, May 02, 2007 at 03:02:52PM -0700, Andrew Morton wrote:
> > So this is what I have lined up for the first mm->2.6.22 batch. I won't be
> > sending it off for another 12-24 hours yet. To give people time for final
> > comment and to give me time to see if it actually works.
>
> I assume you're going to update this list with my comments I sent
> yesterday?
>
Serial drivers? Well you saw me drop a bunch of them. I now have:
serial-driver-pmc-msp71xx.patch
rm9000-serial-driver.patch
serial-define-fixed_port-flag-for-serial_core.patch
mpsc-serial-driver-tx-locking.patch
serial-serial_core-use-pr_debug.patch
I'll also be holding off on MADV_FREE - Nick has some performance things to
share and I'm assuming they're not as good as he'd like.
--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org. For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2007-05-02 22:02 incoming Andrew Morton
2007-05-02 22:31 ` incoming Benjamin Herrenschmidt
@ 2007-05-03 7:55 ` Russell King
2007-05-03 8:05 ` incoming Andrew Morton
2007-05-04 13:37 ` incoming Greg KH
2 siblings, 1 reply; 409+ messages in thread
From: Russell King @ 2007-05-03 7:55 UTC (permalink / raw)
To: Andrew Morton
Cc: Linus Torvalds, Hugh Dickins, Christoph Lameter, David S. Miller,
Andi Kleen, Luck, Tony, Rik van Riel, Benjamin Herrenschmidt,
linux-kernel, linux-mm
On Wed, May 02, 2007 at 03:02:52PM -0700, Andrew Morton wrote:
> So this is what I have lined up for the first mm->2.6.22 batch. I won't be
> sending it off for another 12-24 hours yet. To give people time for final
> comment and to give me time to see if it actually works.
I assume you're going to update this list with my comments I sent
yesterday?
--
Russell King
Linux kernel 2.6 ARM Linux - http://www.arm.linux.org.uk/
maintainer of:
--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org. For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>
^ permalink raw reply [flat|nested] 409+ messages in thread
* Re: incoming
2007-05-02 22:02 incoming Andrew Morton
@ 2007-05-02 22:31 ` Benjamin Herrenschmidt
2007-05-03 7:55 ` incoming Russell King
2007-05-04 13:37 ` incoming Greg KH
2 siblings, 0 replies; 409+ messages in thread
From: Benjamin Herrenschmidt @ 2007-05-02 22:31 UTC (permalink / raw)
To: Andrew Morton
Cc: Linus Torvalds, Hugh Dickins, Christoph Lameter, David S. Miller,
Andi Kleen, Luck, Tony, Rik van Riel, linux-kernel, linux-mm
On Wed, 2007-05-02 at 15:02 -0700, Andrew Morton wrote:
> So this is what I have lined up for the first mm->2.6.22 batch. I won't be
> sending it off for another 12-24 hours yet. To give people time for final
> comment and to give me time to see if it actually works.
Thanks.
I have some powerpc bits that depend on that stuff that will go through
Paulus after these show up in git and I've rebased.
Cheers,
Ben.
--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org. For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>
^ permalink raw reply [flat|nested] 409+ messages in thread
* incoming
@ 2007-05-02 22:02 Andrew Morton
2007-05-02 22:31 ` incoming Benjamin Herrenschmidt
` (2 more replies)
0 siblings, 3 replies; 409+ messages in thread
From: Andrew Morton @ 2007-05-02 22:02 UTC (permalink / raw)
To: Linus Torvalds
Cc: Hugh Dickins, Christoph Lameter, David S. Miller, Andi Kleen,
Luck, Tony, Rik van Riel, Benjamin Herrenschmidt, linux-kernel,
linux-mm
So this is what I have lined up for the first mm->2.6.22 batch. I won't be
sending it off for another 12-24 hours yet. To give people time for final
comment and to give me time to see if it actually works.
- A few serial bits.
- A few pcmcia bits.
- Some of the MM queue. Includes:
- An enhancement to /proc/pid/smaps to permit monitoring of a running
program's working set.
There's another patchset which builds on this quite a lot from Matt
Mackall, but it's not quite ready yet.
- The SLUB allocator. It's pretty green but I do want to push ahead
with this pretty aggressively with a view to replacing slab altogether.
If it ends up not working out then we should remove slub altogether
again, but I doubt if that will occur.
If SLUB isn't in good shape by 2.6.22 we should hide it in Kconfig
to prevent people from hitting known problems. It'll remain
EXPERIMENTAL.
- generic pagetable quicklist management. We have x86_64 and ia64
and sparc64 implementations, but I'll only include David's sparc64
implementation here. I'll send the x86_64 and ia64 implementations
through maintainers.
- Various random MM bits
- Benh's teach-get_unmapped_area-about-MAP_FIXED changes
- madvise(MADV_FREE)
This means I'm holding back Mel's page allocator work, and Andy's
lumpy-reclaim.
A shame in a way - I have high hopes for lumpy reclaim against the
moveable zone, but these things are not to be done lightly.
A few MM things have been held back awaiting subsystem tree merges
(probably x86 - I didn't check).
- One little security patch
- the blackfin architecture
- small h8300 update
- small alpha update
- swsusp updates
- m68k bits
- cris udpates
- Lots of UML updates
- v850, xtensa
slab-introduce-krealloc.patch
at91_cf-minor-fix.patch
add-new_id-to-pcmcia-drivers.patch
ide-cs-recognize-2gb-compactflash-from-transcend.patch
serial-driver-pmc-msp71xx.patch
rm9000-serial-driver.patch
serial-define-fixed_port-flag-for-serial_core.patch
serial-use-resource_size_t-for-serial-port-io-addresses.patch
mpsc-serial-driver-tx-locking.patch
8250_pci-fix-pci-must_checks.patch
serial-serial_core-use-pr_debug.patch
add-apply_to_page_range-which-applies-a-function-to-a-pte-range.patch
safer-nr_node_ids-and-nr_node_ids-determination-and-initial.patch
use-zvc-counters-to-establish-exact-size-of-dirtyable-pages.patch
proper-prototype-for-hugetlb_get_unmapped_area.patch
mm-remove-gcc-workaround.patch
slab-ensure-cache_alloc_refill-terminates.patch
mm-make-read_cache_page-synchronous.patch
fs-buffer-dont-pageuptodate-without-page-locked.patch
allow-oom_adj-of-saintly-processes.patch
introduce-config_has_dma.patch
mm-slabc-proper-prototypes.patch
add-pfn_valid_within-helper-for-sub-max_order-hole-detection.patch
mm-simplify-filemap_nopage.patch
add-unitialized_var-macro-for-suppressing-gcc-warnings.patch
i386-add-ptep_test_and_clear_dirtyyoung.patch
i386-use-pte_update_defer-in-ptep_test_and_clear_dirtyyoung.patch
smaps-extract-pmd-walker-from-smaps-code.patch
smaps-add-pages-referenced-count-to-smaps.patch
smaps-add-clear_refs-file-to-clear-reference.patch
readahead-improve-heuristic-detecting-sequential-reads.patch
readahead-code-cleanup.patch
slab-use-num_possible_cpus-in-enable_cpucache.patch
slab-dont-allocate-empty-shared-caches.patch
slab-numa-kmem_cache-diet.patch
do-not-disable-interrupts-when-reading-min_free_kbytes.patch
slab-mark-set_up_list3s-__init.patch
cpusets-allow-tif_memdie-threads-to-allocate-anywhere.patch
i386-use-page-allocator-to-allocate-thread_info-structure.patch
slub-core.patch
make-page-private-usable-in-compound-pages-v1.patch
optimize-compound_head-by-avoiding-a-shared-page.patch
add-virt_to_head_page-and-consolidate-code-in-slab-and-slub.patch
slub-fix-object-tracking.patch
slub-enable-tracking-of-full-slabs.patch
slub-validation-of-slabs-metadata-and-guard-zones.patch
slub-add-min_partial.patch
slub-add-ability-to-list-alloc--free-callers-per-slab.patch
slub-free-slabs-and-sort-partial-slab-lists-in-kmem_cache_shrink.patch
slub-remove-object-activities-out-of-checking-functions.patch
slub-user-documentation.patch
slub-add-slabinfo-tool.patch
quicklists-for-page-table-pages.patch
quicklist-support-for-sparc64.patch
slob-handle-slab_panic-flag.patch
include-kern_-constant-in-printk-calls-in-mm-slabc.patch
mm-madvise-avoid-exclusive-mmap_sem.patch
mm-remove-destroy_dirty_buffers-from-invalidate_bdev.patch
mm-optimize-kill_bdev.patch
mm-optimize-acorn-partition-truncate.patch
slab-allocators-remove-obsolete-slab_must_hwcache_align.patch
kmem_cache-simplify-slab-cache-creation.patch
slab-allocators-remove-multiple-alignment-specifications.patch
fault-injection-fix-failslab-with-config_numa.patch
mm-fix-handling-of-panic_on_oom-when-cpusets-are-in-use.patch
oom-fix-constraint-deadlock.patch
get_unmapped_area-handles-map_fixed-on-powerpc.patch
get_unmapped_area-handles-map_fixed-on-alpha.patch
get_unmapped_area-handles-map_fixed-on-arm.patch
get_unmapped_area-handles-map_fixed-on-frv.patch
get_unmapped_area-handles-map_fixed-on-i386.patch
get_unmapped_area-handles-map_fixed-on-ia64.patch
get_unmapped_area-handles-map_fixed-on-parisc.patch
get_unmapped_area-handles-map_fixed-on-sparc64.patch
get_unmapped_area-handles-map_fixed-on-x86_64.patch
get_unmapped_area-handles-map_fixed-in-hugetlbfs.patch
get_unmapped_area-handles-map_fixed-in-generic-code.patch
get_unmapped_area-doesnt-need-hugetlbfs-hacks-anymore.patch
slab-allocators-remove-slab_debug_initial-flag.patch
slab-allocators-remove-slab_ctor_atomic.patch
slab-allocators-remove-useless-__gfp_no_grow-flag.patch
lazy-freeing-of-memory-through-madv_free.patch
restore-madv_dontneed-to-its-original-linux-behaviour.patch
hugetlbfs-add-null-check-in-hugetlb_zero_setup.patch
slob-fix-page-order-calculation-on-not-4kb-page.patch
page-migration-only-migrate-pages-if-allocation-in-the-highest-zone-is-possible.patch
return-eperm-not-echild-on-security_task_wait-failure.patch
blackfin-arch.patch
driver_bfin_serial_core.patch
blackfin-on-chip-ethernet-mac-controller-driver.patch
blackfin-patch-add-blackfin-support-in-smc91x.patch
blackfin-on-chip-rtc-controller-driver.patch
blackfin-blackfin-on-chip-spi-controller-driver.patch
convert-h8-300-to-generic-timekeeping.patch
h8300-generic-irq.patch
h8300-add-zimage-support.patch
round_up-macro-cleanup-in-arch-alpha-kernel-osf_sysc.patch
alpha-fix-bootp-image-creation.patch
alpha-prctl-macros.patch
srmcons-fix-kmallocgfp_kernel-inside-spinlock.patch
arm26-remove-useless-config-option-generic_bust_spinlock.patch
fix-refrigerator-vs-thaw_process-race.patch
swsusp-use-inline-functions-for-changing-page-flags.patch
swsusp-do-not-use-page-flags.patch
mm-remove-unused-page-flags.patch
swsusp-fix-error-paths-in-snapshot_open.patch
swsusp-use-gfp_kernel-for-creating-basic-data-structures.patch
freezer-remove-pf_nofreeze-from-handle_initrd.patch
swsusp-use-rbtree-for-tracking-allocated-swap.patch
freezer-fix-racy-usage-of-try_to_freeze-in-kswapd.patch
remove-software_suspend.patch
power-management-change-sys-power-disk-display.patch
kconfig-mentioneds-hibernation-not-just-swsusp.patch
swsusp-fix-snapshot_release.patch
swsusp-free-more-memory.patch
remove-unused-header-file-arch-m68k-atari-atasoundh.patch
spin_lock_unlocked-cleanup-in-arch-m68k.patch
remove-unused-header-file-drivers-serial-crisv10h.patch
cris-check-for-memory-allocation.patch
cris-remove-code-related-to-pre-22-kernel.patch
uml-delete-unused-code.patch
uml-formatting-fixes.patch
uml-host_info-tidying.patch
uml-mark-tt-mode-code-for-future-removal.patch
uml-print-coredump-limits.patch
uml-handle-block-device-hotplug-errors.patch
uml-driver-formatting-fixes.patch
uml-driver-formatting-fixes-fix.patch
uml-network-interface-hotplug-error-handling.patch
array_size-check-for-type.patch
uml-move-sigio-testing-to-sigioc.patch
uml-create-archh.patch
uml-create-as-layouth.patch
uml-move-remaining-useful-contents-of-user_utilh.patch
uml-remove-user_utilh.patch
uml-add-missing-__init-declarations.patch
remove-unused-header-file-arch-um-kernel-tt-include-mode_kern-tth.patch
uml-improve-checking-and-diagnostics-of-ethernet-macs.patch
uml-eliminate-temporary-buffer-in-eth_configure.patch
uml-replace-one-element-array-with-zero-element-array.patch
uml-fix-umid-in-xterm-titles.patch
uml-speed-up-exec.patch
uml-no-locking-needed-in-tlsc.patch
uml-tidy-processc.patch
uml-remove-page_size.patch
uml-kernel_thread-shouldnt-panic.patch
uml-tidy-fault-code.patch
uml-kernel-segfaults-should-dump-proper-registers.patch
uml-comment-early-boot-locking.patch
uml-irq-locking-commentary.patch
uml-delete-host_frame_size.patch
uml-drivers-get-release-methods.patch
uml-dump-registers-on-ptrace-or-wait-failure.patch
uml-speed-up-page-table-walking.patch
uml-remove-unused-x86_64-code.patch
uml-start-fixing-os_read_file-and-os_write_file.patch
uml-tidy-libc-code.patch
uml-convert-libc-layer-to-call-read-and-write.patch
uml-batch-i-o-requests.patch
uml-send-pointers-instead-of-structures-to-i-o-thread.patch
uml-send-pointers-instead-of-structures-to-i-o-thread-fix.patch
uml-dump-core-on-panic.patch
uml-dont-try-to-handle-signals-on-initial-process-stack.patch
uml-change-remaining-callers-of-os_read_write_file.patch
uml-formatting-fixes-around-os_read_write_file-callers.patch
uml-remove-debugging-remnants.patch
uml-rename-os_read_write_file_k-back-to-os_read_write_file.patch
uml-aio-deadlock-avoidance.patch
uml-speed-page-fault-path.patch
uml-eliminate-a-piece-of-debugging-code.patch
uml-more-page-fault-path-trimming.patch
uml-only-flush-areas-covered-by-vma.patch
uml-out-of-tmpfs-space-error-clarification.patch
uml-virtualized-time-fix.patch
uml-fix-prototypes.patch
v850-generic-timekeeping-conversion.patch
xtensa-strlcpy-is-smart-enough.patch
--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org. For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>
^ permalink raw reply [flat|nested] 409+ messages in thread
end of thread, other threads:[~2022-04-27 19:41 UTC | newest]
Thread overview: 409+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2021-06-29 2:32 incoming Andrew Morton
2021-06-29 2:33 ` [patch 001/192] mm/gup: fix try_grab_compound_head() race with split_huge_page() Andrew Morton
2021-06-29 2:33 ` [patch 002/192] mm/page_alloc: fix memory map initialization for descending nodes Andrew Morton
2021-06-29 2:33 ` [patch 003/192] mm/page_alloc: correct return value of populated elements if bulk array is populated Andrew Morton
2021-06-29 2:33 ` [patch 004/192] kthread: switch to new kerneldoc syntax for named variable macro argument Andrew Morton
2021-06-29 2:33 ` [patch 005/192] kthread_worker: fix return value when kthread_mod_delayed_work() races with kthread_cancel_delayed_work_sync() Andrew Morton
2021-06-29 2:33 ` [patch 006/192] ia64: headers: drop duplicated words Andrew Morton
2021-06-29 2:33 ` [patch 007/192] ia64: mca_drv: fix incorrect array size calculation Andrew Morton
2021-06-29 2:33 ` [patch 008/192] streamline_config.pl: make spacing consistent Andrew Morton
2021-06-29 2:33 ` [patch 009/192] streamline_config.pl: add softtabstop=4 for vim users Andrew Morton
2021-06-29 2:33 ` [patch 010/192] scripts/spelling.txt: add more spellings to spelling.txt Andrew Morton
2021-06-29 2:33 ` [patch 011/192] ntfs: fix validity check for file name attribute Andrew Morton
2021-06-29 2:33 ` [patch 012/192] squashfs: add option to panic on errors Andrew Morton
2021-06-29 2:33 ` [patch 013/192] ocfs2: remove unnecessary INIT_LIST_HEAD() Andrew Morton
2021-06-29 2:34 ` [patch 014/192] ocfs2: fix snprintf() checking Andrew Morton
2021-06-29 2:34 ` [patch 015/192] ocfs2: remove redundant assignment to pointer queue Andrew Morton
2021-06-29 2:34 ` [patch 016/192] ocfs2: remove repeated uptodate check for buffer Andrew Morton
2021-06-29 2:34 ` [patch 017/192] ocfs2: replace simple_strtoull() with kstrtoull() Andrew Morton
2021-06-29 2:34 ` [patch 018/192] ocfs2: remove redundant initialization of variable ret Andrew Morton
2021-06-29 2:34 ` [patch 019/192] kernel: watchdog: modify the explanation related to watchdog thread Andrew Morton
2021-06-29 2:34 ` [patch 020/192] doc: " Andrew Morton
2021-06-29 2:34 ` [patch 021/192] doc: watchdog: modify the doc related to "watchdog/%u" Andrew Morton
2021-06-29 2:34 ` [patch 022/192] slab: use __func__ to trace function name Andrew Morton
2021-06-29 2:34 ` [patch 023/192] kunit: make test->lock irq safe Andrew Morton
2021-06-29 2:34 ` [patch 024/192] mm/slub, kunit: add a KUnit test for SLUB debugging functionality Andrew Morton
2021-06-29 2:34 ` [patch 025/192] slub: remove resiliency_test() function Andrew Morton
2021-06-29 2:34 ` [patch 026/192] mm, slub: change run-time assertion in kmalloc_index() to compile-time Andrew Morton
2021-06-29 2:34 ` [patch 027/192] slub: restore slub_debug=- behavior Andrew Morton
2021-06-29 2:34 ` [patch 028/192] slub: actually use 'message' in restore_bytes() Andrew Morton
2021-06-29 2:34 ` [patch 029/192] slub: indicate slab_fix() uses printf formats Andrew Morton
2021-06-29 2:34 ` [patch 030/192] slub: force on no_hash_pointers when slub_debug is enabled Andrew Morton
2021-06-29 2:34 ` [patch 031/192] mm: slub: move sysfs slab alloc/free interfaces to debugfs Andrew Morton
2021-06-29 2:34 ` [patch 032/192] mm/slub: add taint after the errors are printed Andrew Morton
2021-06-29 2:35 ` [patch 033/192] mm/kmemleak: fix possible wrong memory scanning period Andrew Morton
2021-06-29 2:35 ` [patch 034/192] dax: fix ENOMEM handling in grab_mapping_entry() Andrew Morton
2021-06-29 2:35 ` [patch 035/192] tools/vm/page_owner_sort.c: check malloc() return Andrew Morton
2021-06-29 2:35 ` [patch 036/192] mm/debug_vm_pgtable: ensure THP availability via has_transparent_hugepage() Andrew Morton
2021-06-29 2:35 ` [patch 037/192] mm: mmap_lock: use local locks instead of disabling preemption Andrew Morton
2021-06-29 2:35 ` [patch 038/192] mm/page_reporting: fix code style in __page_reporting_request() Andrew Morton
2021-06-29 2:35 ` [patch 039/192] mm/page_reporting: export reporting order as module parameter Andrew Morton
2021-06-29 2:35 ` [patch 040/192] mm/page_reporting: allow driver to specify reporting order Andrew Morton
2021-06-29 2:35 ` [patch 041/192] virtio_balloon: specify page reporting order if needed Andrew Morton
2021-06-29 2:35 ` [patch 042/192] mm: page-writeback: kill get_writeback_state() comments Andrew Morton
2021-06-29 2:35 ` [patch 043/192] mm/page-writeback: Fix performance when BDI's share of ratio is 0 Andrew Morton
2021-06-29 2:35 ` [patch 044/192] mm/page-writeback: update the comment of Dirty position control Andrew Morton
2021-06-29 2:35 ` [patch 045/192] mm/page-writeback: use __this_cpu_inc() in account_page_dirtied() Andrew Morton
2021-06-29 2:35 ` [patch 046/192] writeback, cgroup: do not switch inodes with I_WILL_FREE flag Andrew Morton
2021-06-29 2:35 ` [patch 047/192] writeback, cgroup: add smp_mb() to cgroup_writeback_umount() Andrew Morton
2021-06-29 2:35 ` [patch 048/192] writeback, cgroup: increment isw_nr_in_flight before grabbing an inode Andrew Morton
2021-06-29 2:35 ` [patch 049/192] writeback, cgroup: switch to rcu_work API in inode_switch_wbs() Andrew Morton
2021-06-29 2:35 ` [patch 050/192] writeback, cgroup: keep list of inodes attached to bdi_writeback Andrew Morton
2021-06-29 2:35 ` [patch 051/192] writeback, cgroup: split out the functional part of inode_switch_wbs_work_fn() Andrew Morton
2021-06-29 2:35 ` [patch 052/192] writeback, cgroup: support switching multiple inodes at once Andrew Morton
2021-06-29 2:36 ` [patch 053/192] writeback, cgroup: release dying cgwbs by switching attached inodes Andrew Morton
2021-06-29 2:36 ` [patch 054/192] fs: unexport __set_page_dirty Andrew Morton
2021-06-29 2:36 ` [patch 055/192] fs: move ramfs_aops to libfs Andrew Morton
2021-06-29 2:36 ` [patch 056/192] mm: require ->set_page_dirty to be explicitly wired up Andrew Morton
2021-06-29 2:36 ` [patch 057/192] mm/writeback: move __set_page_dirty() to core mm Andrew Morton
2021-06-29 2:36 ` [patch 058/192] mm/writeback: use __set_page_dirty in __set_page_dirty_nobuffers Andrew Morton
2021-06-29 2:36 ` [patch 059/192] iomap: use __set_page_dirty_nobuffers Andrew Morton
2021-06-29 2:36 ` [patch 060/192] fs: remove anon_set_page_dirty() Andrew Morton
2021-06-29 2:36 ` [patch 061/192] fs: remove noop_set_page_dirty() Andrew Morton
2021-06-29 2:36 ` [patch 062/192] mm: move page dirtying prototypes from mm.h Andrew Morton
2021-06-29 2:36 ` [patch 063/192] mm/gup_benchmark: support threading Andrew Morton
2021-06-29 2:36 ` [patch 064/192] mm: gup: allow FOLL_PIN to scale in SMP Andrew Morton
2021-06-29 2:36 ` [patch 065/192] mm: gup: pack has_pinned in MMF_HAS_PINNED Andrew Morton
2021-06-29 2:36 ` [patch 066/192] mm: pagewalk: fix walk for hugepage tables Andrew Morton
2021-06-29 2:36 ` [patch 067/192] mm/swapfile: use percpu_ref to serialize against concurrent swapoff Andrew Morton
2021-06-29 2:36 ` [patch 068/192] swap: fix do_swap_page() race with swapoff Andrew Morton
2021-06-29 2:36 ` [patch 069/192] mm/swap: remove confusing checking for non_swap_entry() in swap_ra_info() Andrew Morton
2021-06-29 2:36 ` [patch 070/192] mm/shmem: fix shmem_swapin() race with swapoff Andrew Morton
2021-06-29 2:37 ` [patch 071/192] mm/swapfile: move get_swap_page_of_type() under CONFIG_HIBERNATION Andrew Morton
2021-06-29 2:37 ` [patch 072/192] mm/swap: remove unused local variable nr_shadows Andrew Morton
2021-06-29 2:37 ` [patch 073/192] mm/swap_slots.c: delete meaningless forward declarations Andrew Morton
2021-06-29 2:37 ` [patch 074/192] mm, swap: remove unnecessary smp_rmb() in swap_type_to_swap_info() Andrew Morton
2021-06-29 2:37 ` [patch 075/192] mm: free idle swap cache page after COW Andrew Morton
2021-06-29 2:37 ` [patch 076/192] swap: check mapping_empty() for swap cache before being freed Andrew Morton
2021-06-29 2:37 ` [patch 077/192] mm/memcg: move mod_objcg_state() to memcontrol.c Andrew Morton
2021-06-29 2:37 ` [patch 078/192] mm/memcg: cache vmstat data in percpu memcg_stock_pcp Andrew Morton
2021-06-29 2:37 ` [patch 079/192] mm/memcg: improve refill_obj_stock() performance Andrew Morton
2021-06-29 2:37 ` [patch 080/192] mm/memcg: optimize user context object stock access Andrew Morton
2021-06-29 2:37 ` [patch 081/192] mm: memcg/slab: properly set up gfp flags for objcg pointer array Andrew Morton
2021-06-29 2:37 ` [patch 082/192] mm: memcg/slab: create a new set of kmalloc-cg-<n> caches Andrew Morton
2021-06-29 2:37 ` [patch 083/192] mm: memcg/slab: disable cache merging for KMALLOC_NORMAL caches Andrew Morton
2021-06-29 2:37 ` [patch 084/192] mm: memcontrol: fix root_mem_cgroup charging Andrew Morton
2021-06-29 2:37 ` [patch 085/192] mm: memcontrol: fix page charging in page replacement Andrew Morton
2021-06-29 2:37 ` [patch 086/192] mm: memcontrol: bail out early when !mm in get_mem_cgroup_from_mm Andrew Morton
2021-06-29 2:37 ` [patch 087/192] mm: memcontrol: remove the pgdata parameter of mem_cgroup_page_lruvec Andrew Morton
2021-06-29 2:37 ` [patch 088/192] mm: memcontrol: simplify lruvec_holds_page_lru_lock Andrew Morton
2021-06-29 2:37 ` [patch 089/192] mm: memcontrol: rename lruvec_holds_page_lru_lock to page_matches_lruvec Andrew Morton
2021-06-29 2:38 ` [patch 090/192] mm: memcontrol: simplify the logic of objcg pinning memcg Andrew Morton
2021-06-29 2:38 ` [patch 091/192] mm: memcontrol: move obj_cgroup_uncharge_pages() out of css_set_lock Andrew Morton
2021-06-29 2:38 ` [patch 092/192] mm: vmscan: remove noinline_for_stack Andrew Morton
2021-06-29 2:38 ` [patch 093/192] memcontrol: use flexible-array member Andrew Morton
2021-06-29 2:38 ` [patch 094/192] loop: use worker per cgroup instead of kworker Andrew Morton
2021-06-29 2:38 ` [patch 095/192] mm: charge active memcg when no mm is set Andrew Morton
2021-06-29 2:38 ` [patch 096/192] loop: charge i/o to mem and blk cg Andrew Morton
2021-06-29 2:38 ` [patch 097/192] mm: memcontrol: remove trailing semicolon in macros Andrew Morton
2021-06-29 2:38 ` [patch 098/192] perf: MAP_EXECUTABLE does not indicate VM_MAYEXEC Andrew Morton
2021-06-29 2:38 ` [patch 099/192] binfmt: remove in-tree usage of MAP_EXECUTABLE Andrew Morton
2021-06-29 2:38 ` [patch 100/192] mm: ignore MAP_EXECUTABLE in ksys_mmap_pgoff() Andrew Morton
2021-06-29 2:38 ` [patch 101/192] mm/mmap.c: logic of find_vma_intersection repeated in __do_munmap Andrew Morton
2021-06-29 2:38 ` [patch 102/192] mm/mmap: introduce unlock_range() for code cleanup Andrew Morton
2021-06-29 2:38 ` [patch 103/192] mm/mmap: use find_vma_intersection() in do_mmap() for overlap Andrew Morton
2021-06-29 2:38 ` [patch 104/192] mm/memory.c: fix comment of finish_mkwrite_fault() Andrew Morton
2021-06-29 2:38 ` [patch 105/192] mm: add vma_lookup(), update find_vma_intersection() comments Andrew Morton
2021-06-29 2:38 ` [patch 106/192] drm/i915/selftests: use vma_lookup() in __igt_mmap() Andrew Morton
2021-06-29 2:38 ` [patch 107/192] arch/arc/kernel/troubleshoot: use vma_lookup() instead of find_vma() Andrew Morton
2021-06-29 2:38 ` [patch 108/192] arch/arm64/kvm: use vma_lookup() instead of find_vma_intersection() Andrew Morton
2021-06-29 2:39 ` [patch 109/192] arch/powerpc/kvm/book3s_hv_uvmem: " Andrew Morton
2021-06-29 2:39 ` [patch 110/192] arch/powerpc/kvm/book3s: use vma_lookup() in kvmppc_hv_setup_htab_rma() Andrew Morton
2021-06-29 2:39 ` [patch 111/192] arch/mips/kernel/traps: use vma_lookup() instead of find_vma() Andrew Morton
2021-06-29 2:39 ` [patch 112/192] arch/m68k/kernel/sys_m68k: use vma_lookup() in sys_cacheflush() Andrew Morton
2021-06-29 2:39 ` [patch 113/192] x86/sgx: use vma_lookup() in sgx_encl_find() Andrew Morton
2021-06-29 2:39 ` [patch 114/192] virt/kvm: use vma_lookup() instead of find_vma_intersection() Andrew Morton
2021-06-29 2:39 ` [patch 115/192] vfio: " Andrew Morton
2021-06-29 2:39 ` [patch 116/192] net/ipv5/tcp: use vma_lookup() in tcp_zerocopy_receive() Andrew Morton
2021-06-29 2:39 ` [patch 117/192] drm/amdgpu: use vma_lookup() in amdgpu_ttm_tt_get_user_pages() Andrew Morton
2021-06-29 2:39 ` [patch 118/192] media: videobuf2: use vma_lookup() in get_vaddr_frames() Andrew Morton
2021-06-29 2:39 ` [patch 119/192] misc/sgi-gru/grufault: use vma_lookup() in gru_find_vma() Andrew Morton
2021-06-29 2:39 ` [patch 120/192] kernel/events/uprobes: use vma_lookup() in find_active_uprobe() Andrew Morton
2021-06-29 2:39 ` [patch 121/192] lib/test_hmm: use vma_lookup() in dmirror_migrate() Andrew Morton
2021-06-29 2:39 ` [patch 122/192] mm/ksm: use vma_lookup() in find_mergeable_vma() Andrew Morton
2021-06-29 2:39 ` [patch 123/192] mm/migrate: use vma_lookup() in do_pages_stat_array() Andrew Morton
2021-06-29 2:39 ` [patch 124/192] mm/mremap: use vma_lookup() in vma_to_resize() Andrew Morton
2021-06-29 2:39 ` [patch 125/192] mm/memory.c: use vma_lookup() in __access_remote_vm() Andrew Morton
2021-06-29 2:39 ` [patch 126/192] mm/mempolicy: " Andrew Morton
2021-06-29 2:39 ` [patch 127/192] mm: update legacy flush_tlb_* to use vma Andrew Morton
2021-06-29 2:39 ` [patch 128/192] mm: improve mprotect(R|W) efficiency on pages referenced once Andrew Morton
2021-06-29 17:50 ` Linus Torvalds
2021-06-30 0:12 ` Peter Xu
2021-06-30 1:39 ` Peter Xu
2021-06-30 2:25 ` Linus Torvalds
2021-06-30 16:42 ` Peter Xu
2021-06-30 18:03 ` Linus Torvalds
2021-07-01 1:27 ` Peter Xu
2021-07-01 18:29 ` Linus Torvalds
2021-07-06 1:24 ` Peter Xu
2021-06-29 2:40 ` [patch 129/192] h8300: remove unused variable Andrew Morton
2021-06-29 2:40 ` [patch 130/192] mm/dmapool: use DEVICE_ATTR_RO macro Andrew Morton
2021-06-29 2:40 ` [patch 131/192] mm, tracing: unify PFN format strings Andrew Morton
2021-06-29 2:40 ` [patch 132/192] mm/page_alloc: add an alloc_pages_bulk_array_node() helper Andrew Morton
2021-06-29 2:40 ` [patch 133/192] mm/vmalloc: switch to bulk allocator in __vmalloc_area_node() Andrew Morton
2021-06-29 2:40 ` [patch 134/192] mm/vmalloc: print a warning message first on failure Andrew Morton
2021-06-29 2:40 ` [patch 135/192] mm/vmalloc: remove quoted strings split across lines Andrew Morton
2021-06-29 2:40 ` [patch 136/192] mm/vmalloc: fallback to a single page allocator Andrew Morton
2021-06-29 2:40 ` [patch 137/192] mm: vmalloc: add cond_resched() in __vunmap() Andrew Morton
2021-06-29 2:40 ` [patch 138/192] printk: introduce dump_stack_lvl() Andrew Morton
2021-06-29 2:40 ` [patch 139/192] kasan: use dump_stack_lvl(KERN_ERR) to print stacks Andrew Morton
2021-06-29 2:40 ` [patch 140/192] kasan: test: improve failure message in KUNIT_EXPECT_KASAN_FAIL() Andrew Morton
2021-06-29 2:40 ` [patch 141/192] kasan: allow an architecture to disable inline instrumentation Andrew Morton
2021-06-29 2:40 ` [patch 142/192] kasan: allow architectures to provide an outline readiness check Andrew Morton
2021-06-29 2:40 ` [patch 143/192] mm: define default MAX_PTRS_PER_* in include/pgtable.h Andrew Morton
2021-06-29 2:40 ` [patch 144/192] kasan: use MAX_PTRS_PER_* for early shadow tables Andrew Morton
2021-06-29 2:40 ` [patch 145/192] kasan: rename CONFIG_KASAN_SW_TAGS_IDENTIFY to CONFIG_KASAN_TAGS_IDENTIFY Andrew Morton
2021-06-29 2:40 ` [patch 146/192] kasan: integrate the common part of two KASAN tag-based modes Andrew Morton
2021-06-29 2:40 ` [patch 147/192] kasan: add memory corruption identification support for hardware tag-based mode Andrew Morton
2021-06-29 2:41 ` [patch 148/192] mm: report which part of mem is being freed on initmem case Andrew Morton
2021-06-29 2:41 ` [patch 149/192] mm/mmzone.h: simplify is_highmem_idx() Andrew Morton
2021-06-29 2:41 ` [patch 150/192] mm: make __dump_page static Andrew Morton
2021-06-29 2:41 ` [patch 151/192] mm/page_alloc: bail out on fatal signal during reclaim/compaction retry attempt Andrew Morton
2021-06-29 2:41 ` [patch 152/192] mm/debug: factor PagePoisoned out of __dump_page Andrew Morton
2021-06-29 2:41 ` [patch 153/192] mm/page_owner: constify dump_page_owner Andrew Morton
2021-06-29 2:41 ` [patch 154/192] mm: make compound_head const-preserving Andrew Morton
2021-06-29 2:41 ` [patch 155/192] mm: constify get_pfnblock_flags_mask and get_pfnblock_migratetype Andrew Morton
2021-06-29 2:41 ` [patch 156/192] mm: constify page_count and page_ref_count Andrew Morton
2021-06-29 2:41 ` [patch 157/192] mm: optimise nth_page for contiguous memmap Andrew Morton
2021-06-29 2:41 ` [patch 158/192] mm/page_alloc: switch to pr_debug Andrew Morton
2021-06-29 2:41 ` [patch 159/192] kbuild: skip per-CPU BTF generation for pahole v1.18-v1.21 Andrew Morton
2021-06-29 2:41 ` [patch 160/192] mm/page_alloc: split per cpu page lists and zone stats Andrew Morton
2021-06-29 2:41 ` [patch 161/192] mm/page_alloc: convert per-cpu list protection to local_lock Andrew Morton
2021-06-29 2:41 ` [patch 162/192] mm/vmstat: convert NUMA statistics to basic NUMA counters Andrew Morton
2021-06-29 2:41 ` [patch 163/192] mm/vmstat: inline NUMA event counter updates Andrew Morton
2021-06-29 2:41 ` [patch 164/192] mm/page_alloc: batch the accounting updates in the bulk allocator Andrew Morton
2021-06-29 2:41 ` [patch 165/192] mm/page_alloc: reduce duration that IRQs are disabled for VM counters Andrew Morton
2021-06-29 2:41 ` [patch 166/192] mm/page_alloc: explicitly acquire the zone lock in __free_pages_ok Andrew Morton
2021-06-29 2:42 ` [patch 167/192] mm/page_alloc: avoid conflating IRQs disabled with zone->lock Andrew Morton
2021-06-29 2:42 ` [patch 168/192] mm/page_alloc: update PGFREE outside the zone lock in __free_pages_ok Andrew Morton
2021-06-29 2:42 ` [patch 169/192] mm: page_alloc: dump migrate-failed pages only at -EBUSY Andrew Morton
2021-06-29 2:42 ` [patch 170/192] mm/page_alloc: delete vm.percpu_pagelist_fraction Andrew Morton
2021-06-29 2:42 ` [patch 171/192] mm/page_alloc: disassociate the pcp->high from pcp->batch Andrew Morton
2021-06-29 2:42 ` [patch 172/192] mm/page_alloc: adjust pcp->high after CPU hotplug events Andrew Morton
2021-06-29 2:42 ` [patch 173/192] mm/page_alloc: scale the number of pages that are batch freed Andrew Morton
2021-06-29 2:42 ` [patch 174/192] mm/page_alloc: limit the number of pages on PCP lists when reclaim is active Andrew Morton
2021-06-29 2:42 ` [patch 175/192] mm/page_alloc: introduce vm.percpu_pagelist_high_fraction Andrew Morton
2021-06-29 2:42 ` [patch 176/192] mm: drop SECTION_SHIFT in code comments Andrew Morton
2021-06-29 2:42 ` [patch 177/192] mm/page_alloc: improve memmap_pages dbg msg Andrew Morton
2021-06-29 2:42 ` [patch 178/192] mm/page_alloc: fix counting of managed_pages Andrew Morton
2021-06-29 2:42 ` [patch 179/192] mm/page_alloc: move free_the_page Andrew Morton
2021-06-29 2:42 ` [patch 180/192] alpha: remove DISCONTIGMEM and NUMA Andrew Morton
2021-06-29 2:42 ` [patch 181/192] arc: update comment about HIGHMEM implementation Andrew Morton
2021-06-29 2:42 ` [patch 182/192] arc: remove support for DISCONTIGMEM Andrew Morton
2021-06-29 2:42 ` [patch 183/192] m68k: " Andrew Morton
2021-06-29 2:42 ` [patch 184/192] mm: remove CONFIG_DISCONTIGMEM Andrew Morton
2021-06-29 2:42 ` [patch 185/192] arch, mm: remove stale mentions of DISCONIGMEM Andrew Morton
2021-06-29 2:42 ` [patch 186/192] docs: remove description of DISCONTIGMEM Andrew Morton
2021-06-29 2:43 ` [patch 187/192] mm: replace CONFIG_NEED_MULTIPLE_NODES with CONFIG_NUMA Andrew Morton
2021-06-29 2:43 ` [patch 188/192] mm: replace CONFIG_FLAT_NODE_MEM_MAP with CONFIG_FLATMEM Andrew Morton
2021-06-29 2:43 ` [patch 189/192] mm/page_alloc: allow high-order pages to be stored on the per-cpu lists Andrew Morton
2021-06-29 2:43 ` [patch 190/192] mm/page_alloc: split pcp->high across all online CPUs for cpuless nodes Andrew Morton
2021-06-29 2:43 ` [patch 191/192] mm,hwpoison: send SIGBUS with error virutal address Andrew Morton
2021-06-29 2:43 ` [patch 192/192] mm,hwpoison: make get_hwpoison_page() call get_any_page() Andrew Morton
-- strict thread matches above, loose matches on Subject: below --
2022-04-27 19:41 incoming Andrew Morton
2022-04-21 23:35 incoming Andrew Morton
2022-04-15 2:12 incoming Andrew Morton
2022-04-08 20:08 incoming Andrew Morton
2022-04-01 18:27 incoming Andrew Morton
2022-04-01 18:20 incoming Andrew Morton
2022-04-01 18:27 ` incoming Andrew Morton
2022-03-25 1:07 incoming Andrew Morton
2022-03-23 23:04 incoming Andrew Morton
2022-03-22 21:38 incoming Andrew Morton
2022-03-16 23:14 incoming Andrew Morton
2022-03-05 4:28 incoming Andrew Morton
2022-02-26 3:10 incoming Andrew Morton
2022-02-12 0:27 incoming Andrew Morton
2022-02-12 2:02 ` incoming Linus Torvalds
2022-02-12 5:24 ` incoming Andrew Morton
2022-02-04 4:48 incoming Andrew Morton
2022-01-29 21:40 incoming Andrew Morton
2022-01-29 2:13 incoming Andrew Morton
2022-01-29 4:25 ` incoming Matthew Wilcox
2022-01-29 6:23 ` incoming Andrew Morton
2022-01-22 6:10 incoming Andrew Morton
2022-01-20 2:07 incoming Andrew Morton
2022-01-14 22:02 incoming Andrew Morton
2021-12-31 4:12 incoming Andrew Morton
2021-12-25 5:11 incoming Andrew Morton
2021-12-10 22:45 incoming Andrew Morton
2021-11-20 0:42 incoming Andrew Morton
2021-11-11 4:32 incoming Andrew Morton
2021-11-09 2:30 incoming Andrew Morton
2021-11-05 20:34 incoming Andrew Morton
2021-10-28 21:35 incoming Andrew Morton
2021-10-18 22:14 incoming Andrew Morton
2021-09-24 22:42 incoming Andrew Morton
2021-09-10 3:09 incoming Andrew Morton
2021-09-10 17:11 ` incoming Kees Cook
2021-09-10 20:13 ` incoming Kees Cook
2021-09-09 1:08 incoming Andrew Morton
2021-09-08 22:17 incoming Andrew Morton
2021-09-08 2:52 incoming Andrew Morton
2021-09-08 8:57 ` incoming Vlastimil Babka
2021-09-02 21:48 incoming Andrew Morton
2021-09-02 21:49 ` incoming Andrew Morton
2021-08-25 19:17 incoming Andrew Morton
2021-08-20 2:03 incoming Andrew Morton
2021-08-13 23:53 incoming Andrew Morton
2021-07-29 21:52 incoming Andrew Morton
2021-07-23 22:49 incoming Andrew Morton
2021-07-15 4:26 incoming Andrew Morton
2021-07-08 0:59 incoming Andrew Morton
2021-07-01 1:46 incoming Andrew Morton
2021-07-03 0:28 ` incoming Linus Torvalds
2021-07-03 1:06 ` incoming Linus Torvalds
2021-06-25 1:38 incoming Andrew Morton
2021-06-16 1:22 incoming Andrew Morton
2021-06-05 3:00 incoming Andrew Morton
2021-05-23 0:41 incoming Andrew Morton
2021-05-15 0:26 incoming Andrew Morton
2021-05-07 1:01 incoming Andrew Morton
2021-05-07 7:12 ` incoming Linus Torvalds
2021-05-05 1:32 incoming Andrew Morton
2021-05-05 1:47 ` incoming Linus Torvalds
2021-05-05 3:16 ` incoming Andrew Morton
2021-05-05 17:10 ` incoming Linus Torvalds
2021-05-05 17:44 ` incoming Andrew Morton
2021-05-06 3:19 ` incoming Anshuman Khandual
2021-04-30 5:52 incoming Andrew Morton
2021-04-23 21:28 incoming Andrew Morton
2021-04-16 22:45 incoming Andrew Morton
2021-04-09 20:26 incoming Andrew Morton
2021-03-25 4:36 incoming Andrew Morton
2021-03-13 5:06 incoming Andrew Morton
2021-02-26 1:14 incoming Andrew Morton
2021-02-26 17:55 ` incoming Linus Torvalds
2021-02-26 19:16 ` incoming Andrew Morton
2021-02-24 19:58 incoming Andrew Morton
2021-02-24 21:30 ` incoming Linus Torvalds
2021-02-24 21:37 ` incoming Linus Torvalds
2021-02-25 8:53 ` incoming Arnd Bergmann
2021-02-25 9:12 ` incoming Andrey Ryabinin
2021-02-25 11:07 ` incoming Walter Wu
2021-02-13 4:52 incoming Andrew Morton
2021-02-09 21:41 incoming Andrew Morton
2021-02-10 19:30 ` incoming Linus Torvalds
2021-02-05 2:31 incoming Andrew Morton
2021-01-24 5:00 incoming Andrew Morton
2021-01-12 23:48 incoming Andrew Morton
2021-01-15 23:32 ` incoming Linus Torvalds
2020-12-29 23:13 incoming Andrew Morton
2020-12-22 19:58 incoming Andrew Morton
2020-12-22 21:43 ` incoming Linus Torvalds
2020-12-18 22:00 incoming Andrew Morton
2020-12-16 4:41 incoming Andrew Morton
2020-12-15 20:32 incoming Andrew Morton
2020-12-15 21:00 ` incoming Linus Torvalds
2020-12-15 22:48 ` incoming Linus Torvalds
2020-12-15 22:49 ` incoming Linus Torvalds
2020-12-15 22:55 ` incoming Andrew Morton
2020-12-15 3:02 incoming Andrew Morton
2020-12-15 3:25 ` incoming Linus Torvalds
2020-12-15 3:30 ` incoming Linus Torvalds
2020-12-15 14:04 ` incoming Konstantin Ryabitsev
2020-12-11 21:35 incoming Andrew Morton
2020-12-06 6:14 incoming Andrew Morton
2020-11-22 6:16 incoming Andrew Morton
2020-11-14 6:51 incoming Andrew Morton
2020-11-02 1:06 incoming Andrew Morton
2020-10-17 23:13 incoming Andrew Morton
2020-10-16 2:40 incoming Andrew Morton
2020-10-16 3:03 ` incoming Andrew Morton
2020-10-13 23:46 incoming Andrew Morton
2020-10-11 6:15 incoming Andrew Morton
2020-10-03 5:20 incoming Andrew Morton
2020-09-26 4:17 incoming Andrew Morton
2020-09-19 4:19 incoming Andrew Morton
2020-09-04 23:34 incoming Andrew Morton
2020-08-21 0:41 incoming Andrew Morton
2020-08-15 0:29 incoming Andrew Morton
2020-08-12 1:29 incoming Andrew Morton
2020-08-07 6:16 incoming Andrew Morton
2020-07-24 4:14 incoming Andrew Morton
2020-07-03 22:14 incoming Andrew Morton
2020-06-26 3:28 incoming Andrew Morton
2020-06-26 6:51 ` incoming Linus Torvalds
2020-06-26 7:31 ` incoming Linus Torvalds
2020-06-26 17:39 ` incoming Konstantin Ryabitsev
2020-06-26 17:40 ` incoming Konstantin Ryabitsev
2020-06-12 0:30 incoming Andrew Morton
2020-06-11 1:40 incoming Andrew Morton
2020-06-09 4:29 incoming Andrew Morton
2020-06-09 16:58 ` incoming Linus Torvalds
2020-06-08 4:35 incoming Andrew Morton
2020-06-04 23:45 incoming Andrew Morton
2020-06-03 22:55 incoming Andrew Morton
2020-06-02 20:09 incoming Andrew Morton
2020-06-02 4:44 incoming Andrew Morton
2020-06-02 20:08 ` incoming Andrew Morton
2020-06-02 20:45 ` incoming Linus Torvalds
2020-06-02 21:38 ` incoming Andrew Morton
2020-06-02 22:18 ` incoming Linus Torvalds
2020-05-28 5:20 incoming Andrew Morton
2020-05-28 20:10 ` incoming Linus Torvalds
2020-05-29 20:31 ` incoming Andrew Morton
2020-05-29 20:38 ` incoming Linus Torvalds
2020-05-29 21:12 ` incoming Andrew Morton
2020-05-29 21:20 ` incoming Linus Torvalds
2020-05-23 5:22 incoming Andrew Morton
2020-05-14 0:50 incoming Andrew Morton
2020-05-08 1:35 incoming Andrew Morton
2020-04-21 1:13 incoming Andrew Morton
2020-04-12 7:41 incoming Andrew Morton
2020-04-10 21:30 incoming Andrew Morton
2020-04-07 3:02 incoming Andrew Morton
2020-04-02 4:01 incoming Andrew Morton
2020-03-29 2:14 incoming Andrew Morton
2020-03-22 1:19 incoming Andrew Morton
2020-03-06 6:27 incoming Andrew Morton
2020-02-21 4:00 incoming Andrew Morton
2020-02-21 4:03 ` incoming Andrew Morton
2020-02-21 18:21 ` incoming Linus Torvalds
2020-02-21 18:32 ` incoming Konstantin Ryabitsev
2020-02-27 9:59 ` incoming Vlastimil Babka
2020-02-21 19:33 ` incoming Linus Torvalds
2020-02-04 1:33 incoming Andrew Morton
2020-02-04 2:27 ` incoming Linus Torvalds
2020-02-04 2:46 ` incoming Andrew Morton
2020-02-04 3:11 ` incoming Linus Torvalds
2020-01-31 6:10 incoming Andrew Morton
2020-01-14 0:28 incoming Andrew Morton
2020-01-04 20:55 incoming Andrew Morton
2019-12-18 4:50 incoming Andrew Morton
2019-12-05 0:48 incoming Andrew Morton
2019-12-01 1:47 incoming Andrew Morton
2019-12-01 5:17 ` incoming James Bottomley
2019-12-01 21:07 ` incoming Linus Torvalds
2019-12-02 8:21 ` incoming Steven Price
2019-11-22 1:53 incoming Andrew Morton
2019-11-16 1:34 incoming Andrew Morton
2019-11-06 5:16 incoming Andrew Morton
2019-10-19 3:19 incoming Andrew Morton
2019-10-14 21:11 incoming Andrew Morton
2019-10-07 0:57 incoming Andrew Morton
2019-09-25 23:45 incoming Andrew Morton
2019-09-23 22:31 incoming Andrew Morton
2019-09-24 0:55 ` incoming Linus Torvalds
2019-09-24 4:31 ` incoming Andrew Morton
2019-09-24 7:48 ` incoming Michal Hocko
2019-09-24 15:34 ` incoming Linus Torvalds
2019-09-25 6:36 ` incoming Michal Hocko
2019-09-24 19:55 ` incoming Vlastimil Babka
2019-08-30 23:04 incoming Andrew Morton
2019-08-25 0:54 incoming Andrew Morton
[not found] <20190716162536.bb52b8f34a8ecf5331a86a42@linux-foundation.org>
2019-07-17 8:47 ` incoming Vlastimil Babka
2019-07-17 8:57 ` incoming Bhaskar Chowdhury
2019-07-17 16:13 ` incoming Linus Torvalds
2019-07-17 17:09 ` incoming Christian Brauner
2019-07-17 18:13 ` incoming Vlastimil Babka
2007-05-02 22:02 incoming Andrew Morton
2007-05-02 22:31 ` incoming Benjamin Herrenschmidt
2007-05-03 7:55 ` incoming Russell King
2007-05-03 8:05 ` incoming Andrew Morton
2007-05-04 13:37 ` incoming Greg KH
2007-05-04 16:14 ` incoming Andrew Morton
2007-05-04 17:02 ` incoming Greg KH
2007-05-04 18:57 ` incoming Roland McGrath
2007-05-04 19:24 ` incoming Greg KH
2007-05-04 19:29 ` incoming Roland McGrath
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).