* vmemmap alloc failure in hot_add_req()
@ 2021-06-11 19:48 Nathan Chancellor
2021-06-12 3:42 ` Nathan Chancellor
[not found] ` <20210612021115.2136-1-hdanton@sina.com>
0 siblings, 2 replies; 7+ messages in thread
From: Nathan Chancellor @ 2021-06-11 19:48 UTC (permalink / raw)
To: K. Y. Srinivasan, Haiyang Zhang, Stephen Hemminger, Wei Liu, Dexuan Cui
Cc: linux-hyperv, linux-kernel
Hi all,
I am occasionally seeing a kernel warning when running virtual machines
in Hyper-V, which usually happens a minute or so after boot. It does not
happen on every boot and it is reproducible on at least v5.10. I think
it might have something to do with constant reboots, which I do when
testing various kernels.
The stack trace is as follows:
[ 49.215291] kworker/0:1: vmemmap alloc failure: order:9, mode:0x4cc0(GFP_KERNEL|__GFP_RETRY_MAYFAIL), nodemask=(null),cpuset=/,mems_allowed=0
[ 49.215299] CPU: 0 PID: 18 Comm: kworker/0:1 Not tainted 5.13.0-rc5 #1
[ 49.215301] Hardware name: Microsoft Corporation Virtual Machine/Virtual Machine, BIOS Hyper-V UEFI Release v4.0 11/01/2019
[ 49.215302] Workqueue: events hot_add_req [hv_balloon]
[ 49.215307] Call Trace:
[ 49.215310] dump_stack+0x76/0x94
[ 49.215314] warn_alloc.cold+0x78/0xdc
[ 49.215316] ? __alloc_pages+0x200/0x230
[ 49.215319] vmemmap_alloc_block+0x86/0xdc
[ 49.215323] vmemmap_populate+0x10e/0x31c
[ 49.215324] __populate_section_memmap+0x38/0x4e
[ 49.215326] sparse_add_section+0x12c/0x1cf
[ 49.215329] __add_pages+0xa9/0x130
[ 49.215330] add_pages+0x12/0x60
[ 49.215333] add_memory_resource+0x180/0x300
[ 49.215335] __add_memory+0x3b/0x80
[ 49.215336] add_memory+0x2e/0x50
[ 49.215337] hot_add_req+0x3fc/0x5a0 [hv_balloon]
[ 49.215340] process_one_work+0x214/0x3e0
[ 49.215342] worker_thread+0x4d/0x3d0
[ 49.215344] ? process_one_work+0x3e0/0x3e0
[ 49.215345] kthread+0x133/0x150
[ 49.215347] ? kthread_associate_blkcg+0xc0/0xc0
[ 49.215348] ret_from_fork+0x22/0x30
[ 49.215351] Mem-Info:
[ 49.215352] active_anon:251 inactive_anon:140868 isolated_anon:0
active_file:47497 inactive_file:88505 isolated_file:0
unevictable:8 dirty:14 writeback:0
slab_reclaimable:12013 slab_unreclaimable:11403
mapped:131701 shmem:12671 pagetables:3140 bounce:0
free:41388 free_pcp:37 free_cma:0
[ 49.215355] Node 0 active_anon:1004kB inactive_anon:563472kB active_file:189988kB inactive_file:354020kB unevictable:32kB isolated(anon):0kB isolated(file):0kB mapped:526804kB dirty:56kB writeback:0kB shmem:50684kB shmem_thp: 0kB shmem_pmdmapped: 0kB anon_thp: 0kB writeback_tmp:0kB kernel_stack:5904kB pagetables:12560kB all_unreclaimable? no
[ 49.215358] Node 0 DMA free:6496kB min:480kB low:600kB high:720kB reserved_highatomic:0KB active_anon:0kB inactive_anon:3120kB active_file:2584kB inactive_file:2792kB unevictable:0kB writepending:0kB present:15996kB managed:15360kB mlocked:0kB bounce:0kB free_pcp:0kB local_pcp:0kB free_cma:0kB
[ 49.215361] lowmem_reserve[]: 0 1384 1384 1384 1384
[ 49.215364] Node 0 DMA32 free:159056kB min:44572kB low:55712kB high:66852kB reserved_highatomic:0KB active_anon:1004kB inactive_anon:560352kB active_file:187004kB inactive_file:350864kB unevictable:32kB writepending:56kB present:1555760kB managed:1432388kB mlocked:32kB bounce:0kB free_pcp:172kB local_pcp:0kB free_cma:0kB
[ 49.215367] lowmem_reserve[]: 0 0 0 0 0
[ 49.215369] Node 0 DMA: 17*4kB (UM) 13*8kB (M) 10*16kB (M) 3*32kB (ME) 3*64kB (UME) 4*128kB (UME) 1*256kB (E) 2*512kB (UE) 2*1024kB (ME) 1*2048kB (E) 0*4096kB = 6508kB
[ 49.215377] Node 0 DMA32: 8061*4kB (UME) 5892*8kB (UME) 2449*16kB (UME) 604*32kB (UME) 207*64kB (UME) 49*128kB (UM) 7*256kB (M) 1*512kB (M) 0*1024kB 0*2048kB 0*4096kB = 159716kB
[ 49.215388] 148696 total pagecache pages
[ 49.215388] 0 pages in swap cache
[ 49.215389] Swap cache stats: add 0, delete 0, find 0/0
[ 49.215390] Free swap = 0kB
[ 49.215390] Total swap = 0kB
[ 49.215391] 392939 pages RAM
[ 49.215391] 0 pages HighMem/MovableOnly
[ 49.215391] 31002 pages reserved
[ 49.215392] 0 pages cma reserved
[ 49.215393] 0 pages hwpoisoned
Is this a known issue and/or am I doing something wrong? I only noticed
this because there are times when I am compiling something intensive in
the VM such as LLVM and the VM runs out of memory even though I have
plenty of free memory on the host but I am not sure if this warning is
related to that issue.
I do not have much experience with Hyper-V so it is possible I do not
have something configured properly or there is some other issue going
on. Let me know if there is any further information I can provide or
help debug in any way.
Cheers,
Nathan
^ permalink raw reply [flat|nested] 7+ messages in thread
* Re: vmemmap alloc failure in hot_add_req()
2021-06-11 19:48 vmemmap alloc failure in hot_add_req() Nathan Chancellor
@ 2021-06-12 3:42 ` Nathan Chancellor
[not found] ` <20210612021115.2136-1-hdanton@sina.com>
1 sibling, 0 replies; 7+ messages in thread
From: Nathan Chancellor @ 2021-06-12 3:42 UTC (permalink / raw)
To: K. Y. Srinivasan, Haiyang Zhang, Stephen Hemminger, Wei Liu, Dexuan Cui
Cc: linux-hyperv, linux-kernel
[-- Attachment #1: Type: text/plain, Size: 4785 bytes --]
On Fri, Jun 11, 2021 at 12:48:26PM -0700, Nathan Chancellor wrote:
> Hi all,
>
> I am occasionally seeing a kernel warning when running virtual machines
> in Hyper-V, which usually happens a minute or so after boot. It does not
> happen on every boot and it is reproducible on at least v5.10. I think
> it might have something to do with constant reboots, which I do when
> testing various kernels.
>
> The stack trace is as follows:
>
> [ 49.215291] kworker/0:1: vmemmap alloc failure: order:9, mode:0x4cc0(GFP_KERNEL|__GFP_RETRY_MAYFAIL), nodemask=(null),cpuset=/,mems_allowed=0
> [ 49.215299] CPU: 0 PID: 18 Comm: kworker/0:1 Not tainted 5.13.0-rc5 #1
> [ 49.215301] Hardware name: Microsoft Corporation Virtual Machine/Virtual Machine, BIOS Hyper-V UEFI Release v4.0 11/01/2019
> [ 49.215302] Workqueue: events hot_add_req [hv_balloon]
> [ 49.215307] Call Trace:
> [ 49.215310] dump_stack+0x76/0x94
> [ 49.215314] warn_alloc.cold+0x78/0xdc
> [ 49.215316] ? __alloc_pages+0x200/0x230
> [ 49.215319] vmemmap_alloc_block+0x86/0xdc
> [ 49.215323] vmemmap_populate+0x10e/0x31c
> [ 49.215324] __populate_section_memmap+0x38/0x4e
> [ 49.215326] sparse_add_section+0x12c/0x1cf
> [ 49.215329] __add_pages+0xa9/0x130
> [ 49.215330] add_pages+0x12/0x60
> [ 49.215333] add_memory_resource+0x180/0x300
> [ 49.215335] __add_memory+0x3b/0x80
> [ 49.215336] add_memory+0x2e/0x50
> [ 49.215337] hot_add_req+0x3fc/0x5a0 [hv_balloon]
> [ 49.215340] process_one_work+0x214/0x3e0
> [ 49.215342] worker_thread+0x4d/0x3d0
> [ 49.215344] ? process_one_work+0x3e0/0x3e0
> [ 49.215345] kthread+0x133/0x150
> [ 49.215347] ? kthread_associate_blkcg+0xc0/0xc0
> [ 49.215348] ret_from_fork+0x22/0x30
> [ 49.215351] Mem-Info:
> [ 49.215352] active_anon:251 inactive_anon:140868 isolated_anon:0
> active_file:47497 inactive_file:88505 isolated_file:0
> unevictable:8 dirty:14 writeback:0
> slab_reclaimable:12013 slab_unreclaimable:11403
> mapped:131701 shmem:12671 pagetables:3140 bounce:0
> free:41388 free_pcp:37 free_cma:0
> [ 49.215355] Node 0 active_anon:1004kB inactive_anon:563472kB active_file:189988kB inactive_file:354020kB unevictable:32kB isolated(anon):0kB isolated(file):0kB mapped:526804kB dirty:56kB writeback:0kB shmem:50684kB shmem_thp: 0kB shmem_pmdmapped: 0kB anon_thp: 0kB writeback_tmp:0kB kernel_stack:5904kB pagetables:12560kB all_unreclaimable? no
> [ 49.215358] Node 0 DMA free:6496kB min:480kB low:600kB high:720kB reserved_highatomic:0KB active_anon:0kB inactive_anon:3120kB active_file:2584kB inactive_file:2792kB unevictable:0kB writepending:0kB present:15996kB managed:15360kB mlocked:0kB bounce:0kB free_pcp:0kB local_pcp:0kB free_cma:0kB
> [ 49.215361] lowmem_reserve[]: 0 1384 1384 1384 1384
> [ 49.215364] Node 0 DMA32 free:159056kB min:44572kB low:55712kB high:66852kB reserved_highatomic:0KB active_anon:1004kB inactive_anon:560352kB active_file:187004kB inactive_file:350864kB unevictable:32kB writepending:56kB present:1555760kB managed:1432388kB mlocked:32kB bounce:0kB free_pcp:172kB local_pcp:0kB free_cma:0kB
> [ 49.215367] lowmem_reserve[]: 0 0 0 0 0
> [ 49.215369] Node 0 DMA: 17*4kB (UM) 13*8kB (M) 10*16kB (M) 3*32kB (ME) 3*64kB (UME) 4*128kB (UME) 1*256kB (E) 2*512kB (UE) 2*1024kB (ME) 1*2048kB (E) 0*4096kB = 6508kB
> [ 49.215377] Node 0 DMA32: 8061*4kB (UME) 5892*8kB (UME) 2449*16kB (UME) 604*32kB (UME) 207*64kB (UME) 49*128kB (UM) 7*256kB (M) 1*512kB (M) 0*1024kB 0*2048kB 0*4096kB = 159716kB
> [ 49.215388] 148696 total pagecache pages
> [ 49.215388] 0 pages in swap cache
> [ 49.215389] Swap cache stats: add 0, delete 0, find 0/0
> [ 49.215390] Free swap = 0kB
> [ 49.215390] Total swap = 0kB
> [ 49.215391] 392939 pages RAM
> [ 49.215391] 0 pages HighMem/MovableOnly
> [ 49.215391] 31002 pages reserved
> [ 49.215392] 0 pages cma reserved
> [ 49.215393] 0 pages hwpoisoned
>
> Is this a known issue and/or am I doing something wrong? I only noticed
> this because there are times when I am compiling something intensive in
> the VM such as LLVM and the VM runs out of memory even though I have
> plenty of free memory on the host but I am not sure if this warning is
> related to that issue.
I had one of these events just happen, journalctl output attached. I
have no idea what the "unhandled message: type:" messages mean, I guess
that corresponds to the switch statement in balloon_onchannelcallback().
> I do not have much experience with Hyper-V so it is possible I do not
> have something configured properly or there is some other issue going
> on. Let me know if there is any further information I can provide or
> help debug in any way.
Cheers,
Nathan
[-- Attachment #2: journalctl.log --]
[-- Type: text/plain, Size: 75079 bytes --]
-- Journal begins at Thu 2021-06-10 06:17:06 MST, ends at Fri 2021-06-11 20:34:01 MST. --
Jun 11 20:15:39 hyperv kernel: Linux version 5.13.0-rc5-next-20210611-llvm (nathan@hyperv) (ClangBuiltLinux clang version 13.0.0 (https://github.com/llvm/llvm-project 06c3d52aa2fa3b6a9aa019906ad8577813efe32a), LLD 13.0.0 (https://github.com/llvm/llvm-project 06c3d52aa2fa3b6a9aa019906ad8577813efe32a)) #1 SMP PREEMPT Fri Jun 11 20:10:16 MST 2021
Jun 11 20:15:39 hyperv kernel: Command line: initrd=\initramfs-linux-next-llvm.img root=PARTUUID=2660dacb-8872-49d2-8192-394e57557dd7 rw intel_pstate=no_hwp video=hyperv_fb:1920x1080
Jun 11 20:15:39 hyperv kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers'
Jun 11 20:15:39 hyperv kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers'
Jun 11 20:15:39 hyperv kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers'
Jun 11 20:15:39 hyperv kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256
Jun 11 20:15:39 hyperv kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format.
Jun 11 20:15:39 hyperv kernel: signal: max sigframe size: 1776
Jun 11 20:15:39 hyperv kernel: BIOS-provided physical RAM map:
Jun 11 20:15:39 hyperv kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable
Jun 11 20:15:39 hyperv kernel: BIOS-e820: [mem 0x00000000000c0000-0x00000000000fffff] reserved
Jun 11 20:15:39 hyperv kernel: BIOS-e820: [mem 0x0000000000100000-0x000000005ed1cfff] usable
Jun 11 20:15:39 hyperv kernel: BIOS-e820: [mem 0x000000005ed1d000-0x000000005ed1dfff] ACPI data
Jun 11 20:15:39 hyperv kernel: BIOS-e820: [mem 0x000000005ed1e000-0x000000005eecbfff] usable
Jun 11 20:15:39 hyperv kernel: BIOS-e820: [mem 0x000000005eecc000-0x000000005eeeafff] ACPI data
Jun 11 20:15:39 hyperv kernel: BIOS-e820: [mem 0x000000005eeeb000-0x000000005ef1afff] reserved
Jun 11 20:15:39 hyperv kernel: BIOS-e820: [mem 0x000000005ef1b000-0x000000005ff9afff] usable
Jun 11 20:15:39 hyperv kernel: BIOS-e820: [mem 0x000000005ff9b000-0x000000005fff2fff] reserved
Jun 11 20:15:39 hyperv kernel: BIOS-e820: [mem 0x000000005fff3000-0x000000005fffafff] ACPI data
Jun 11 20:15:39 hyperv kernel: BIOS-e820: [mem 0x000000005fffb000-0x000000005fffefff] ACPI NVS
Jun 11 20:15:39 hyperv kernel: BIOS-e820: [mem 0x000000005ffff000-0x000000005fffffff] usable
Jun 11 20:15:39 hyperv kernel: intel_pstate: HWP disabled
Jun 11 20:15:39 hyperv kernel: NX (Execute Disable) protection: active
Jun 11 20:15:39 hyperv kernel: efi: EFI v2.70 by Microsoft
Jun 11 20:15:39 hyperv kernel: efi: ACPI=0x5fffa000 ACPI 2.0=0x5fffa014 SMBIOS=0x5ffd8000 SMBIOS 3.0=0x5ffd6000 MEMATTR=0x5e541018 RNG=0x5ffda818
Jun 11 20:15:39 hyperv kernel: efi: seeding entropy pool
Jun 11 20:15:39 hyperv kernel: SMBIOS 3.1.0 present.
Jun 11 20:15:39 hyperv kernel: DMI: Microsoft Corporation Virtual Machine/Virtual Machine, BIOS Hyper-V UEFI Release v4.0 11/01/2019
Jun 11 20:15:39 hyperv kernel: Hypervisor detected: Microsoft Hyper-V
Jun 11 20:15:39 hyperv kernel: Hyper-V: privilege flags low 0x2e7f, high 0x3b8030, hints 0xc2c, misc 0xbed7b2
Jun 11 20:15:39 hyperv kernel: Hyper-V Host Build:19041-10.0-0-0.985
Jun 11 20:15:39 hyperv kernel: Hyper-V: LAPIC Timer Frequency: 0xa2c2a
Jun 11 20:15:39 hyperv kernel: tsc: Marking TSC unstable due to running on Hyper-V
Jun 11 20:15:39 hyperv kernel: Hyper-V: Using hypercall for remote TLB flush
Jun 11 20:15:39 hyperv kernel: clocksource: hyperv_clocksource_tsc_page: mask: 0xffffffffffffffff max_cycles: 0x24e6a1710, max_idle_ns: 440795202120 ns
Jun 11 20:15:39 hyperv kernel: tsc: Detected 3800.016 MHz processor
Jun 11 20:15:39 hyperv kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved
Jun 11 20:15:39 hyperv kernel: e820: remove [mem 0x000a0000-0x000fffff] usable
Jun 11 20:15:39 hyperv kernel: last_pfn = 0x60000 max_arch_pfn = 0x400000000
Jun 11 20:15:39 hyperv kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT
Jun 11 20:15:39 hyperv kernel: Using GB pages for direct mapping
Jun 11 20:15:39 hyperv kernel: Secure boot disabled
Jun 11 20:15:39 hyperv kernel: RAMDISK: [mem 0x5e543000-0x5ecb3fff]
Jun 11 20:15:39 hyperv kernel: ACPI: Early table checksum verification disabled
Jun 11 20:15:39 hyperv kernel: ACPI: RSDP 0x000000005FFFA014 000024 (v02 VRTUAL)
Jun 11 20:15:39 hyperv kernel: ACPI: XSDT 0x000000005FFF90E8 00005C (v01 VRTUAL MICROSFT 00000001 MSFT 00000001)
Jun 11 20:15:39 hyperv kernel: ACPI: FACP 0x000000005FFF8000 000114 (v06 VRTUAL MICROSFT 00000001 MSFT 00000001)
Jun 11 20:15:39 hyperv kernel: ACPI: DSDT 0x000000005EECC000 01E184 (v02 MSFTVM DSDT01 00000001 MSFT 05000000)
Jun 11 20:15:39 hyperv kernel: ACPI: FACS 0x000000005FFFE000 000040
Jun 11 20:15:39 hyperv kernel: ACPI: OEM0 0x000000005FFF7000 000064 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001)
Jun 11 20:15:39 hyperv kernel: ACPI: WAET 0x000000005FFF6000 000028 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001)
Jun 11 20:15:39 hyperv kernel: ACPI: APIC 0x000000005FFF5000 0000A8 (v04 VRTUAL MICROSFT 00000001 MSFT 00000001)
Jun 11 20:15:39 hyperv kernel: ACPI: SRAT 0x000000005FFF4000 000230 (v02 VRTUAL MICROSFT 00000001 MSFT 00000001)
Jun 11 20:15:39 hyperv kernel: ACPI: BGRT 0x000000005FFF3000 000038 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001)
Jun 11 20:15:39 hyperv kernel: ACPI: FPDT 0x000000005ED1D000 000034 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001)
Jun 11 20:15:39 hyperv kernel: ACPI: Reserving FACP table memory at [mem 0x5fff8000-0x5fff8113]
Jun 11 20:15:39 hyperv kernel: ACPI: Reserving DSDT table memory at [mem 0x5eecc000-0x5eeea183]
Jun 11 20:15:39 hyperv kernel: ACPI: Reserving FACS table memory at [mem 0x5fffe000-0x5fffe03f]
Jun 11 20:15:39 hyperv kernel: ACPI: Reserving OEM0 table memory at [mem 0x5fff7000-0x5fff7063]
Jun 11 20:15:39 hyperv kernel: ACPI: Reserving WAET table memory at [mem 0x5fff6000-0x5fff6027]
Jun 11 20:15:39 hyperv kernel: ACPI: Reserving APIC table memory at [mem 0x5fff5000-0x5fff50a7]
Jun 11 20:15:39 hyperv kernel: ACPI: Reserving SRAT table memory at [mem 0x5fff4000-0x5fff422f]
Jun 11 20:15:39 hyperv kernel: ACPI: Reserving BGRT table memory at [mem 0x5fff3000-0x5fff3037]
Jun 11 20:15:39 hyperv kernel: ACPI: Reserving FPDT table memory at [mem 0x5ed1d000-0x5ed1d033]
Jun 11 20:15:39 hyperv kernel: SRAT: PXM 0 -> APIC 0x00 -> Node 0
Jun 11 20:15:39 hyperv kernel: SRAT: PXM 0 -> APIC 0x01 -> Node 0
Jun 11 20:15:39 hyperv kernel: SRAT: PXM 0 -> APIC 0x02 -> Node 0
Jun 11 20:15:39 hyperv kernel: SRAT: PXM 0 -> APIC 0x03 -> Node 0
Jun 11 20:15:39 hyperv kernel: SRAT: PXM 0 -> APIC 0x04 -> Node 0
Jun 11 20:15:39 hyperv kernel: SRAT: PXM 0 -> APIC 0x05 -> Node 0
Jun 11 20:15:39 hyperv kernel: SRAT: PXM 0 -> APIC 0x06 -> Node 0
Jun 11 20:15:39 hyperv kernel: SRAT: PXM 0 -> APIC 0x07 -> Node 0
Jun 11 20:15:39 hyperv kernel: SRAT: PXM 0 -> APIC 0x08 -> Node 0
Jun 11 20:15:39 hyperv kernel: SRAT: PXM 0 -> APIC 0x09 -> Node 0
Jun 11 20:15:39 hyperv kernel: SRAT: PXM 0 -> APIC 0x0a -> Node 0
Jun 11 20:15:39 hyperv kernel: SRAT: PXM 0 -> APIC 0x0b -> Node 0
Jun 11 20:15:39 hyperv kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x5fffffff] hotplug
Jun 11 20:15:39 hyperv kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x60000000-0xf7ffffff] hotplug
Jun 11 20:15:39 hyperv kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000-0xfdfffffff] hotplug
Jun 11 20:15:39 hyperv kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x1000000000-0xfcffffffff] hotplug
Jun 11 20:15:39 hyperv kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x10000000000-0x1ffffffffff] hotplug
Jun 11 20:15:39 hyperv kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x20000000000-0x3ffffffffff] hotplug
Jun 11 20:15:39 hyperv kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x40000000000-0x7ffffffffff] hotplug
Jun 11 20:15:39 hyperv kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x80000000000-0xfffffffffff] hotplug
Jun 11 20:15:39 hyperv kernel: NODE_DATA(0) allocated [mem 0x5ece2000-0x5ece5fff]
Jun 11 20:15:39 hyperv kernel: Zone ranges:
Jun 11 20:15:39 hyperv kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff]
Jun 11 20:15:39 hyperv kernel: DMA32 [mem 0x0000000001000000-0x000000005fffffff]
Jun 11 20:15:39 hyperv kernel: Normal empty
Jun 11 20:15:39 hyperv kernel: Device empty
Jun 11 20:15:39 hyperv kernel: Movable zone start for each node
Jun 11 20:15:39 hyperv kernel: Early memory node ranges
Jun 11 20:15:39 hyperv kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff]
Jun 11 20:15:39 hyperv kernel: node 0: [mem 0x0000000000100000-0x000000005ed1cfff]
Jun 11 20:15:39 hyperv kernel: node 0: [mem 0x000000005ed1e000-0x000000005eecbfff]
Jun 11 20:15:39 hyperv kernel: node 0: [mem 0x000000005ef1b000-0x000000005ff9afff]
Jun 11 20:15:39 hyperv kernel: node 0: [mem 0x000000005ffff000-0x000000005fffffff]
Jun 11 20:15:39 hyperv kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000005fffffff]
Jun 11 20:15:39 hyperv kernel: DMA zone: 28769 pages in unavailable ranges
Jun 11 20:15:39 hyperv kernel: DMA32 zone: 180 pages in unavailable ranges
Jun 11 20:15:39 hyperv kernel: ACPI: PM-Timer IO Port: 0x408
Jun 11 20:15:39 hyperv kernel: ACPI: LAPIC_NMI (acpi_id[0x01] dfl dfl lint[0x1])
Jun 11 20:15:39 hyperv kernel: IOAPIC[0]: apic_id 12, version 17, address 0xfec00000, GSI 0-23
Jun 11 20:15:39 hyperv kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level)
Jun 11 20:15:39 hyperv kernel: ACPI: Using ACPI (MADT) for SMP configuration information
Jun 11 20:15:39 hyperv kernel: e820: update [mem 0x5ed1e000-0x5ed24fff] usable ==> reserved
Jun 11 20:15:39 hyperv kernel: smpboot: Allowing 12 CPUs, 0 hotplug CPUs
Jun 11 20:15:39 hyperv kernel: PM: hibernation: Registered nosave memory: [mem 0x00000000-0x00000fff]
Jun 11 20:15:39 hyperv kernel: PM: hibernation: Registered nosave memory: [mem 0x000a0000-0x000bffff]
Jun 11 20:15:39 hyperv kernel: PM: hibernation: Registered nosave memory: [mem 0x000c0000-0x000fffff]
Jun 11 20:15:39 hyperv kernel: PM: hibernation: Registered nosave memory: [mem 0x5ed1d000-0x5ed1dfff]
Jun 11 20:15:39 hyperv kernel: PM: hibernation: Registered nosave memory: [mem 0x5ed1e000-0x5ed24fff]
Jun 11 20:15:39 hyperv kernel: PM: hibernation: Registered nosave memory: [mem 0x5eecc000-0x5eeeafff]
Jun 11 20:15:39 hyperv kernel: PM: hibernation: Registered nosave memory: [mem 0x5eeeb000-0x5ef1afff]
Jun 11 20:15:39 hyperv kernel: PM: hibernation: Registered nosave memory: [mem 0x5ff9b000-0x5fff2fff]
Jun 11 20:15:39 hyperv kernel: PM: hibernation: Registered nosave memory: [mem 0x5fff3000-0x5fffafff]
Jun 11 20:15:39 hyperv kernel: PM: hibernation: Registered nosave memory: [mem 0x5fffb000-0x5fffefff]
Jun 11 20:15:39 hyperv kernel: [mem 0x60000000-0xffffffff] available for PCI devices
Jun 11 20:15:39 hyperv kernel: Booting paravirtualized kernel on Hyper-V
Jun 11 20:15:39 hyperv kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 6370452778343963 ns
Jun 11 20:15:39 hyperv kernel: setup_percpu: NR_CPUS:320 nr_cpumask_bits:320 nr_cpu_ids:12 nr_node_ids:1
Jun 11 20:15:39 hyperv kernel: percpu: Embedded 56 pages/cpu s192512 r8192 d28672 u262144
Jun 11 20:15:39 hyperv kernel: pcpu-alloc: s192512 r8192 d28672 u262144 alloc=1*2097152
Jun 11 20:15:39 hyperv kernel: pcpu-alloc: [0] 00 01 02 03 04 05 06 07 [0] 08 09 10 11 -- -- -- --
Jun 11 20:15:39 hyperv kernel: Hyper-V: PV spinlocks enabled
Jun 11 20:15:39 hyperv kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear)
Jun 11 20:15:39 hyperv kernel: Built 1 zonelists, mobility grouping on. Total pages: 385314
Jun 11 20:15:39 hyperv kernel: Policy zone: DMA32
Jun 11 20:15:39 hyperv kernel: Kernel command line: initrd=\initramfs-linux-next-llvm.img root=PARTUUID=2660dacb-8872-49d2-8192-394e57557dd7 rw intel_pstate=no_hwp video=hyperv_fb:1920x1080
Jun 11 20:15:39 hyperv kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear)
Jun 11 20:15:39 hyperv kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear)
Jun 11 20:15:39 hyperv kernel: mem auto-init: stack:off, heap alloc:on, heap free:off
Jun 11 20:15:39 hyperv kernel: Memory: 1394704K/1571756K available (14344K kernel code, 1921K rwdata, 4584K rodata, 1864K init, 4252K bss, 176792K reserved, 0K cma-reserved)
Jun 11 20:15:39 hyperv kernel: random: get_random_u64 called from kmem_cache_open+0x28/0x6c0 with crng_init=0
Jun 11 20:15:39 hyperv kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=12, Nodes=1
Jun 11 20:15:39 hyperv kernel: ftrace: allocating 40969 entries in 161 pages
Jun 11 20:15:39 hyperv kernel: ftrace: allocated 161 pages with 3 groups
Jun 11 20:15:39 hyperv kernel: rcu: Preemptible hierarchical RCU implementation.
Jun 11 20:15:39 hyperv kernel: rcu: RCU dyntick-idle grace-period acceleration is enabled.
Jun 11 20:15:39 hyperv kernel: rcu: RCU restricting CPUs from NR_CPUS=320 to nr_cpu_ids=12.
Jun 11 20:15:39 hyperv kernel: rcu: RCU priority boosting: priority 1 delay 500 ms.
Jun 11 20:15:39 hyperv kernel: Trampoline variant of Tasks RCU enabled.
Jun 11 20:15:39 hyperv kernel: Rude variant of Tasks RCU enabled.
Jun 11 20:15:39 hyperv kernel: Tracing variant of Tasks RCU enabled.
Jun 11 20:15:39 hyperv kernel: rcu: RCU calculated value of scheduler-enlistment delay is 30 jiffies.
Jun 11 20:15:39 hyperv kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=12
Jun 11 20:15:39 hyperv kernel: Using NULL legacy PIC
Jun 11 20:15:39 hyperv kernel: NR_IRQS: 20736, nr_irqs: 520, preallocated irqs: 0
Jun 11 20:15:39 hyperv kernel: Console: colour dummy device 80x25
Jun 11 20:15:39 hyperv kernel: printk: console [tty0] enabled
Jun 11 20:15:39 hyperv kernel: ACPI: Core revision 20210604
Jun 11 20:15:39 hyperv kernel: Failed to register legacy timer interrupt
Jun 11 20:15:39 hyperv kernel: APIC: Switch to symmetric I/O mode setup
Jun 11 20:15:39 hyperv kernel: Switched APIC routing to physical flat.
Jun 11 20:15:39 hyperv kernel: Hyper-V: Using IPI hypercalls
Jun 11 20:15:39 hyperv kernel: Hyper-V: Using enlightened APIC (xapic mode)
Jun 11 20:15:39 hyperv kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 7603.70 BogoMIPS (lpj=12666720)
Jun 11 20:15:39 hyperv kernel: pid_max: default: 32768 minimum: 301
Jun 11 20:15:39 hyperv kernel: LSM: Security Framework initializing
Jun 11 20:15:39 hyperv kernel: Yama: becoming mindful.
Jun 11 20:15:39 hyperv kernel: LSM support for eBPF active
Jun 11 20:15:39 hyperv kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear)
Jun 11 20:15:39 hyperv kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear)
Jun 11 20:15:39 hyperv kernel: Last level iTLB entries: 4KB 1024, 2MB 1024, 4MB 512
Jun 11 20:15:39 hyperv kernel: Last level dTLB entries: 4KB 2048, 2MB 2048, 4MB 1024, 1GB 0
Jun 11 20:15:39 hyperv kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization
Jun 11 20:15:39 hyperv kernel: Spectre V2 : Mitigation: Full AMD retpoline
Jun 11 20:15:39 hyperv kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch
Jun 11 20:15:39 hyperv kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier
Jun 11 20:15:39 hyperv kernel: Spectre V2 : User space: Mitigation: STIBP via seccomp and prctl
Jun 11 20:15:39 hyperv kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl and seccomp
Jun 11 20:15:39 hyperv kernel: Freeing SMP alternatives memory: 40K
Jun 11 20:15:39 hyperv kernel: smpboot: CPU0: AMD Ryzen 9 3900X 12-Core Processor (family: 0x17, model: 0x71, stepping: 0x0)
Jun 11 20:15:39 hyperv kernel: Performance Events: PMU not available due to virtualization, using software events only.
Jun 11 20:15:39 hyperv kernel: rcu: Hierarchical SRCU implementation.
Jun 11 20:15:39 hyperv kernel: NMI watchdog: Perf NMI watchdog permanently disabled
Jun 11 20:15:39 hyperv kernel: smp: Bringing up secondary CPUs ...
Jun 11 20:15:39 hyperv kernel: x86: Booting SMP configuration:
Jun 11 20:15:39 hyperv kernel: .... node #0, CPUs: #1 #2 #3 #4 #5 #6
Jun 11 20:15:39 hyperv kernel: random: fast init done
Jun 11 20:15:39 hyperv kernel: #7 #8 #9 #10 #11
Jun 11 20:15:39 hyperv kernel: smp: Brought up 1 node, 12 CPUs
Jun 11 20:15:39 hyperv kernel: smpboot: Max logical packages: 1
Jun 11 20:15:39 hyperv kernel: smpboot: Total of 12 processors activated (91330.19 BogoMIPS)
Jun 11 20:15:39 hyperv kernel: devtmpfs: initialized
Jun 11 20:15:39 hyperv kernel: x86/mm: Memory block size: 128MB
Jun 11 20:15:39 hyperv kernel: ACPI: PM: Registering ACPI NVS region [mem 0x5fffb000-0x5fffefff] (16384 bytes)
Jun 11 20:15:39 hyperv kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 6370867519511994 ns
Jun 11 20:15:39 hyperv kernel: futex hash table entries: 4096 (order: 6, 262144 bytes, linear)
Jun 11 20:15:39 hyperv kernel: pinctrl core: initialized pinctrl subsystem
Jun 11 20:15:39 hyperv kernel: PM: RTC time: 03:15:38, date: 2021-06-12
Jun 11 20:15:39 hyperv kernel: NET: Registered protocol family 16
Jun 11 20:15:39 hyperv kernel: DMA: preallocated 256 KiB GFP_KERNEL pool for atomic allocations
Jun 11 20:15:39 hyperv kernel: DMA: preallocated 256 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations
Jun 11 20:15:39 hyperv kernel: DMA: preallocated 256 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations
Jun 11 20:15:39 hyperv kernel: audit: initializing netlink subsys (disabled)
Jun 11 20:15:39 hyperv kernel: audit: type=2000 audit(1623467737.916:1): state=initialized audit_enabled=0 res=1
Jun 11 20:15:39 hyperv kernel: thermal_sys: Registered thermal governor 'fair_share'
Jun 11 20:15:39 hyperv kernel: thermal_sys: Registered thermal governor 'bang_bang'
Jun 11 20:15:39 hyperv kernel: thermal_sys: Registered thermal governor 'step_wise'
Jun 11 20:15:39 hyperv kernel: thermal_sys: Registered thermal governor 'user_space'
Jun 11 20:15:39 hyperv kernel: thermal_sys: Registered thermal governor 'power_allocator'
Jun 11 20:15:39 hyperv kernel: cpuidle: using governor ladder
Jun 11 20:15:39 hyperv kernel: cpuidle: using governor menu
Jun 11 20:15:39 hyperv kernel: ACPI: bus type PCI registered
Jun 11 20:15:39 hyperv kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5
Jun 11 20:15:39 hyperv kernel: Kprobes globally optimized
Jun 11 20:15:39 hyperv kernel: HugeTLB registered 1.00 GiB page size, pre-allocated 0 pages
Jun 11 20:15:39 hyperv kernel: HugeTLB registered 2.00 MiB page size, pre-allocated 0 pages
Jun 11 20:15:39 hyperv kernel: ACPI: Added _OSI(Module Device)
Jun 11 20:15:39 hyperv kernel: ACPI: Added _OSI(Processor Device)
Jun 11 20:15:39 hyperv kernel: ACPI: Added _OSI(3.0 _SCP Extensions)
Jun 11 20:15:39 hyperv kernel: ACPI: Added _OSI(Processor Aggregator Device)
Jun 11 20:15:39 hyperv kernel: ACPI: Added _OSI(Linux-Dell-Video)
Jun 11 20:15:39 hyperv kernel: ACPI: Added _OSI(Linux-Lenovo-NV-HDMI-Audio)
Jun 11 20:15:39 hyperv kernel: ACPI: Added _OSI(Linux-HPI-Hybrid-Graphics)
Jun 11 20:15:39 hyperv kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded
Jun 11 20:15:39 hyperv kernel: ACPI: Interpreter enabled
Jun 11 20:15:39 hyperv kernel: ACPI: PM: (supports S0 S5)
Jun 11 20:15:39 hyperv kernel: ACPI: Using IOAPIC for interrupt routing
Jun 11 20:15:39 hyperv kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug
Jun 11 20:15:39 hyperv kernel: ACPI: Enabled 1 GPEs in block 00 to 0F
Jun 11 20:15:39 hyperv kernel: iommu: Default domain type: Translated
Jun 11 20:15:39 hyperv kernel: vgaarb: loaded
Jun 11 20:15:39 hyperv kernel: SCSI subsystem initialized
Jun 11 20:15:39 hyperv kernel: libata version 3.00 loaded.
Jun 11 20:15:39 hyperv kernel: ACPI: bus type USB registered
Jun 11 20:15:39 hyperv kernel: usbcore: registered new interface driver usbfs
Jun 11 20:15:39 hyperv kernel: usbcore: registered new interface driver hub
Jun 11 20:15:39 hyperv kernel: usbcore: registered new device driver usb
Jun 11 20:15:39 hyperv kernel: pps_core: LinuxPPS API ver. 1 registered
Jun 11 20:15:39 hyperv kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti <giometti@linux.it>
Jun 11 20:15:39 hyperv kernel: PTP clock support registered
Jun 11 20:15:39 hyperv kernel: EDAC MC: Ver: 3.0.0
Jun 11 20:15:39 hyperv kernel: Registered efivars operations
Jun 11 20:15:39 hyperv kernel: NetLabel: Initializing
Jun 11 20:15:39 hyperv kernel: NetLabel: domain hash size = 128
Jun 11 20:15:39 hyperv kernel: NetLabel: protocols = UNLABELED CIPSOv4 CALIPSO
Jun 11 20:15:39 hyperv kernel: NetLabel: unlabeled traffic allowed by default
Jun 11 20:15:39 hyperv kernel: PCI: Using ACPI for IRQ routing
Jun 11 20:15:39 hyperv kernel: PCI: System does not support PCI
Jun 11 20:15:39 hyperv kernel: clocksource: Switched to clocksource hyperv_clocksource_tsc_page
Jun 11 20:15:39 hyperv kernel: VFS: Disk quotas dquot_6.6.0
Jun 11 20:15:39 hyperv kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes)
Jun 11 20:15:39 hyperv kernel: pnp: PnP ACPI init
Jun 11 20:15:39 hyperv kernel: pnp: PnP ACPI: found 1 devices
Jun 11 20:15:39 hyperv kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns
Jun 11 20:15:39 hyperv kernel: NET: Registered protocol family 2
Jun 11 20:15:39 hyperv kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear)
Jun 11 20:15:39 hyperv kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear)
Jun 11 20:15:39 hyperv kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear)
Jun 11 20:15:39 hyperv kernel: TCP bind hash table entries: 16384 (order: 6, 262144 bytes, linear)
Jun 11 20:15:39 hyperv kernel: TCP: Hash tables configured (established 16384 bind 16384)
Jun 11 20:15:39 hyperv kernel: MPTCP token hash table entries: 2048 (order: 3, 49152 bytes, linear)
Jun 11 20:15:39 hyperv kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear)
Jun 11 20:15:39 hyperv kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear)
Jun 11 20:15:39 hyperv kernel: NET: Registered protocol family 1
Jun 11 20:15:39 hyperv kernel: NET: Registered protocol family 44
Jun 11 20:15:39 hyperv kernel: PCI: CLS 0 bytes, default 64
Jun 11 20:15:39 hyperv kernel: PCI-DMA: Using software bounce buffering for IO (SWIOTLB)
Jun 11 20:15:39 hyperv kernel: software IO TLB: mapped [mem 0x0000000057800000-0x000000005b800000] (64MB)
Jun 11 20:15:39 hyperv kernel: Unpacking initramfs...
Jun 11 20:15:39 hyperv kernel: Initialise system trusted keyrings
Jun 11 20:15:39 hyperv kernel: Key type blacklist registered
Jun 11 20:15:39 hyperv kernel: workingset: timestamp_bits=41 max_order=19 bucket_order=0
Jun 11 20:15:39 hyperv kernel: zbud: loaded
Jun 11 20:15:39 hyperv kernel: Key type asymmetric registered
Jun 11 20:15:39 hyperv kernel: Asymmetric key parser 'x509' registered
Jun 11 20:15:39 hyperv kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 243)
Jun 11 20:15:39 hyperv kernel: io scheduler mq-deadline registered
Jun 11 20:15:39 hyperv kernel: io scheduler kyber registered
Jun 11 20:15:39 hyperv kernel: io scheduler bfq registered
Jun 11 20:15:39 hyperv kernel: shpchp: Standard Hot Plug PCI Controller Driver version: 0.4
Jun 11 20:15:39 hyperv kernel: efifb: probing for efifb
Jun 11 20:15:39 hyperv kernel: efifb: framebuffer at 0xf8000000, using 3072k, total 3072k
Jun 11 20:15:39 hyperv kernel: efifb: mode is 1024x768x32, linelength=4096, pages=1
Jun 11 20:15:39 hyperv kernel: efifb: scrolling: redraw
Jun 11 20:15:39 hyperv kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0
Jun 11 20:15:39 hyperv kernel: fbcon: Deferring console take-over
Jun 11 20:15:39 hyperv kernel: fb0: EFI VGA frame buffer device
Jun 11 20:15:39 hyperv kernel: Serial: 8250/16550 driver, 32 ports, IRQ sharing enabled
Jun 11 20:15:39 hyperv kernel: Non-volatile memory driver v1.3
Jun 11 20:15:39 hyperv kernel: AMD-Vi: AMD IOMMUv2 driver by Joerg Roedel <jroedel@suse.de>
Jun 11 20:15:39 hyperv kernel: AMD-Vi: AMD IOMMUv2 functionality not available on this system
Jun 11 20:15:39 hyperv kernel: ehci_hcd: USB 2.0 'Enhanced' Host Controller (EHCI) Driver
Jun 11 20:15:39 hyperv kernel: ehci-pci: EHCI PCI platform driver
Jun 11 20:15:39 hyperv kernel: ohci_hcd: USB 1.1 'Open' Host Controller (OHCI) Driver
Jun 11 20:15:39 hyperv kernel: ohci-pci: OHCI PCI platform driver
Jun 11 20:15:39 hyperv kernel: uhci_hcd: USB Universal Host Controller Interface driver
Jun 11 20:15:39 hyperv kernel: usbcore: registered new interface driver usbserial_generic
Jun 11 20:15:39 hyperv kernel: usbserial: USB Serial support registered for generic
Jun 11 20:15:39 hyperv kernel: rtc_cmos 00:00: RTC can wake from S4
Jun 11 20:15:39 hyperv kernel: rtc_cmos 00:00: registered as rtc0
Jun 11 20:15:39 hyperv kernel: rtc_cmos 00:00: setting system clock to 2021-06-12T03:15:38 UTC (1623467738)
Jun 11 20:15:39 hyperv kernel: rtc_cmos 00:00: alarms up to one month, 114 bytes nvram
Jun 11 20:15:39 hyperv kernel: ledtrig-cpu: registered to indicate activity on CPUs
Jun 11 20:15:39 hyperv kernel: hid: raw HID events driver (C) Jiri Kosina
Jun 11 20:15:39 hyperv kernel: drop_monitor: Initializing network drop monitor service
Jun 11 20:15:39 hyperv kernel: Initializing XFRM netlink socket
Jun 11 20:15:39 hyperv kernel: NET: Registered protocol family 10
Jun 11 20:15:39 hyperv kernel: Freeing initrd memory: 7620K
Jun 11 20:15:39 hyperv kernel: Segment Routing with IPv6
Jun 11 20:15:39 hyperv kernel: RPL Segment Routing with IPv6
Jun 11 20:15:39 hyperv kernel: NET: Registered protocol family 17
Jun 11 20:15:39 hyperv kernel: IPI shorthand broadcast: enabled
Jun 11 20:15:39 hyperv kernel: registered taskstats version 1
Jun 11 20:15:39 hyperv kernel: Loading compiled-in X.509 certificates
Jun 11 20:15:39 hyperv kernel: Loaded X.509 cert 'Build time autogenerated kernel key: f5b021a8b3f2573f50186584c9d698e8723ec3cb'
Jun 11 20:15:39 hyperv kernel: zswap: loaded using pool lz4/z3fold
Jun 11 20:15:39 hyperv kernel: Key type ._fscrypt registered
Jun 11 20:15:39 hyperv kernel: Key type .fscrypt registered
Jun 11 20:15:39 hyperv kernel: Key type fscrypt-provisioning registered
Jun 11 20:15:39 hyperv kernel: PM: Magic number: 9:934:259
Jun 11 20:15:39 hyperv kernel: RAS: Correctable Errors collector initialized.
Jun 11 20:15:39 hyperv kernel: Unstable clock detected, switching default tracing clock to "global"
If you want to keep using the local clock, then add:
"trace_clock=local"
on the kernel command line
Jun 11 20:15:39 hyperv kernel: Freeing unused decrypted memory: 2036K
Jun 11 20:15:39 hyperv kernel: Freeing unused kernel image (initmem) memory: 1864K
Jun 11 20:15:39 hyperv kernel: Write protecting the kernel read-only data: 22528k
Jun 11 20:15:39 hyperv kernel: Freeing unused kernel image (text/rodata gap) memory: 2036K
Jun 11 20:15:39 hyperv kernel: Freeing unused kernel image (rodata/data gap) memory: 1560K
Jun 11 20:15:39 hyperv kernel: x86/mm: Checked W+X mappings: passed, no W+X pages found.
Jun 11 20:15:39 hyperv kernel: rodata_test: all tests were successful
Jun 11 20:15:39 hyperv kernel: Run /init as init process
Jun 11 20:15:39 hyperv kernel: with arguments:
Jun 11 20:15:39 hyperv kernel: /init
Jun 11 20:15:39 hyperv kernel: with environment:
Jun 11 20:15:39 hyperv kernel: HOME=/
Jun 11 20:15:39 hyperv kernel: TERM=linux
Jun 11 20:15:39 hyperv kernel: fbcon: Taking over console
Jun 11 20:15:39 hyperv kernel: Console: switching to colour frame buffer device 128x48
Jun 11 20:15:39 hyperv kernel: hv_vmbus: Vmbus version:5.2
Jun 11 20:15:39 hyperv kernel: hv_vmbus: registering driver hid_hyperv
Jun 11 20:15:39 hyperv kernel: input: Microsoft Vmbus HID-compliant Mouse as /devices/0006:045E:0621.0001/input/input0
Jun 11 20:15:39 hyperv kernel: hid-generic 0006:045E:0621.0001: input: VIRTUAL HID v0.01 Mouse [Microsoft Vmbus HID-compliant Mouse] on
Jun 11 20:15:39 hyperv kernel: hv_vmbus: registering driver hyperv_keyboard
Jun 11 20:15:39 hyperv kernel: hv_vmbus: registering driver hv_storvsc
Jun 11 20:15:39 hyperv kernel: scsi host0: storvsc_host_t
Jun 11 20:15:39 hyperv kernel: input: AT Translated Set 2 keyboard as /devices/LNXSYSTM:00/LNXSYBUS:00/ACPI0004:00/VMBUS:00/d34b2567-b9b6-42b9-8778-0a4ec0b955bf/serio0/input/input1
Jun 11 20:15:39 hyperv kernel: scsi 0:0:0:0: Direct-Access Msft Virtual Disk 1.0 PQ: 0 ANSI: 5
Jun 11 20:15:39 hyperv kernel: scsi 0:0:0:1: CD-ROM Msft Virtual DVD-ROM 1.0 PQ: 0 ANSI: 0
Jun 11 20:15:39 hyperv kernel: sd 0:0:0:0: [sda] 314572800 512-byte logical blocks: (161 GB/150 GiB)
Jun 11 20:15:39 hyperv kernel: sd 0:0:0:0: [sda] 4096-byte physical blocks
Jun 11 20:15:39 hyperv kernel: sd 0:0:0:0: [sda] Write Protect is off
Jun 11 20:15:39 hyperv kernel: sd 0:0:0:0: [sda] Mode Sense: 0f 00 00 00
Jun 11 20:15:39 hyperv kernel: sd 0:0:0:0: [sda] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA
Jun 11 20:15:39 hyperv kernel: sda: sda1 sda2
Jun 11 20:15:39 hyperv kernel: sd 0:0:0:0: [sda] Attached SCSI disk
Jun 11 20:15:39 hyperv kernel: sr 0:0:0:1: [sr0] scsi-1 drive
Jun 11 20:15:39 hyperv kernel: cdrom: Uniform CD-ROM driver Revision: 3.20
Jun 11 20:15:39 hyperv kernel: sr 0:0:0:1: Attached scsi CD-ROM sr0
Jun 11 20:15:39 hyperv kernel: SGI XFS with ACLs, security attributes, realtime, scrub, repair, quota, no debug enabled
Jun 11 20:15:39 hyperv kernel: XFS (sda2): Mounting V5 Filesystem
Jun 11 20:15:39 hyperv kernel: XFS (sda2): Ending clean mount
Jun 11 20:15:39 hyperv kernel: xfs filesystem being mounted at /new_root supports timestamps until 2038 (0x7fffffff)
Jun 11 20:15:39 hyperv systemd[1]: systemd 248.3-2-arch running in system mode. (+PAM +AUDIT -SELINUX -APPARMOR -IMA +SMACK +SECCOMP +GCRYPT +GNUTLS +OPENSSL +ACL +BLKID +CURL +ELFUTILS +FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY +P11KIT -QRENCODE +BZIP2 +LZ4 +XZ +ZLIB +ZSTD +XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified)
Jun 11 20:15:39 hyperv systemd[1]: Detected virtualization microsoft.
Jun 11 20:15:39 hyperv systemd[1]: Detected architecture x86-64.
Jun 11 20:15:39 hyperv systemd[1]: Hostname set to <hyperv>.
Jun 11 20:15:39 hyperv systemd-fstab-generator[250]: Mount point is not a valid path, ignoring.
Jun 11 20:15:39 hyperv systemd-fstab-generator[250]: Mount point is not a valid path, ignoring.
Jun 11 20:15:39 hyperv kernel: random: lvmconfig: uninitialized urandom read (4 bytes read)
Jun 11 20:15:39 hyperv systemd[1]: Queued start job for default target Graphical Interface.
Jun 11 20:15:39 hyperv systemd[1]: Created slice system-getty.slice.
Jun 11 20:15:39 hyperv systemd[1]: Created slice system-modprobe.slice.
Jun 11 20:15:39 hyperv systemd[1]: Created slice system-systemd\x2dfsck.slice.
Jun 11 20:15:39 hyperv systemd[1]: Created slice User and Session Slice.
Jun 11 20:15:39 hyperv systemd[1]: Started Dispatch Password Requests to Console Directory Watch.
Jun 11 20:15:39 hyperv systemd[1]: Started Forward Password Requests to Wall Directory Watch.
Jun 11 20:15:39 hyperv systemd[1]: Set up automount Arbitrary Executable File Formats File System Automount Point.
Jun 11 20:15:39 hyperv systemd[1]: Reached target Local Encrypted Volumes.
Jun 11 20:15:39 hyperv systemd[1]: Reached target Login Prompts.
Jun 11 20:15:39 hyperv systemd[1]: Reached target Paths.
Jun 11 20:15:39 hyperv systemd[1]: Reached target Remote File Systems.
Jun 11 20:15:39 hyperv systemd[1]: Reached target Slices.
Jun 11 20:15:39 hyperv systemd[1]: Reached target Swap.
Jun 11 20:15:39 hyperv systemd[1]: Reached target Local Verity Integrity Protected Volumes.
Jun 11 20:15:39 hyperv systemd[1]: Listening on Device-mapper event daemon FIFOs.
Jun 11 20:15:39 hyperv systemd[1]: Listening on LVM2 poll daemon socket.
Jun 11 20:15:39 hyperv systemd[1]: Listening on Process Core Dump Socket.
Jun 11 20:15:39 hyperv systemd[1]: Listening on Journal Audit Socket.
Jun 11 20:15:39 hyperv systemd[1]: Listening on Journal Socket (/dev/log).
Jun 11 20:15:39 hyperv systemd[1]: Listening on Journal Socket.
Jun 11 20:15:39 hyperv systemd[1]: Listening on Network Service Netlink Socket.
Jun 11 20:15:39 hyperv systemd[1]: Listening on udev Control Socket.
Jun 11 20:15:39 hyperv systemd[1]: Listening on udev Kernel Socket.
Jun 11 20:15:39 hyperv systemd[1]: Mounting Huge Pages File System...
Jun 11 20:15:39 hyperv systemd[1]: Mounting POSIX Message Queue File System...
Jun 11 20:15:39 hyperv systemd[1]: Mounting Kernel Debug File System...
Jun 11 20:15:39 hyperv systemd[1]: Mounting Kernel Trace File System...
Jun 11 20:15:39 hyperv systemd[1]: Starting Create list of static device nodes for the current kernel...
Jun 11 20:15:39 hyperv systemd[1]: Starting Monitoring of LVM2 mirrors, snapshots etc. using dmeventd or progress polling...
Jun 11 20:15:39 hyperv systemd[1]: Starting Load Kernel Module configfs...
Jun 11 20:15:39 hyperv systemd[1]: Starting Load Kernel Module drm...
Jun 11 20:15:39 hyperv kernel: random: lvm: uninitialized urandom read (4 bytes read)
Jun 11 20:15:39 hyperv systemd[1]: Starting Load Kernel Module fuse...
Jun 11 20:15:39 hyperv systemd[1]: Starting Set Up Additional Binary Formats...
Jun 11 20:15:39 hyperv systemd[1]: Condition check resulted in File System Check on Root Device being skipped.
Jun 11 20:15:39 hyperv systemd[1]: Starting Journal Service...
Jun 11 20:15:39 hyperv systemd[1]: Starting Load Kernel Modules...
Jun 11 20:15:39 hyperv systemd[1]: Starting Remount Root and Kernel File Systems...
Jun 11 20:15:39 hyperv systemd[1]: Condition check resulted in Repartition Root Disk being skipped.
Jun 11 20:15:39 hyperv systemd[1]: Starting Coldplug All udev Devices...
Jun 11 20:15:39 hyperv systemd[1]: Mounted Huge Pages File System.
Jun 11 20:15:39 hyperv systemd[1]: Mounted POSIX Message Queue File System.
Jun 11 20:15:39 hyperv systemd[1]: Mounted Kernel Debug File System.
Jun 11 20:15:39 hyperv systemd[1]: Mounted Kernel Trace File System.
Jun 11 20:15:39 hyperv systemd[1]: Finished Create list of static device nodes for the current kernel.
Jun 11 20:15:39 hyperv systemd[1]: modprobe@configfs.service: Deactivated successfully.
Jun 11 20:15:39 hyperv systemd[1]: Finished Load Kernel Module configfs.
Jun 11 20:15:39 hyperv kernel: audit: type=1130 audit(1623467739.176:2): pid=1 uid=0 auid=4294967295 ses=4294967295 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
Jun 11 20:15:39 hyperv kernel: audit: type=1131 audit(1623467739.176:3): pid=1 uid=0 auid=4294967295 ses=4294967295 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
Jun 11 20:15:39 hyperv systemd[1]: proc-sys-fs-binfmt_misc.automount: Got automount request for /proc/sys/fs/binfmt_misc, triggered by 267 (systemd-binfmt)
Jun 11 20:15:39 hyperv kernel: fuse: init (API version 7.33)
Jun 11 20:15:39 hyperv kernel: xfs filesystem being remounted at / supports timestamps until 2038 (0x7fffffff)
Jun 11 20:15:39 hyperv systemd[1]: Mounting Arbitrary Executable File Formats File System...
Jun 11 20:15:39 hyperv systemd[1]: Mounting Kernel Configuration File System...
Jun 11 20:15:39 hyperv kernel: Asymmetric key parser 'pkcs8' registered
Jun 11 20:15:39 hyperv systemd[1]: modprobe@fuse.service: Deactivated successfully.
Jun 11 20:15:39 hyperv systemd[1]: Finished Load Kernel Module fuse.
Jun 11 20:15:39 hyperv kernel: audit: type=1130 audit(1623467739.179:4): pid=1 uid=0 auid=4294967295 ses=4294967295 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
Jun 11 20:15:39 hyperv kernel: audit: type=1131 audit(1623467739.179:5): pid=1 uid=0 auid=4294967295 ses=4294967295 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
Jun 11 20:15:39 hyperv systemd[1]: Finished Load Kernel Modules.
Jun 11 20:15:39 hyperv kernel: audit: type=1130 audit(1623467739.179:6): pid=1 uid=0 auid=4294967295 ses=4294967295 msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
Jun 11 20:15:39 hyperv systemd[1]: Finished Remount Root and Kernel File Systems.
Jun 11 20:15:39 hyperv kernel: audit: type=1130 audit(1623467739.179:7): pid=1 uid=0 auid=4294967295 ses=4294967295 msg='unit=systemd-remount-fs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
Jun 11 20:15:39 hyperv systemd[1]: Mounted Arbitrary Executable File Formats File System.
Jun 11 20:15:39 hyperv systemd[1]: Mounted Kernel Configuration File System.
Jun 11 20:15:39 hyperv systemd[1]: Mounting FUSE Control File System...
Jun 11 20:15:39 hyperv systemd[1]: Condition check resulted in First Boot Wizard being skipped.
Jun 11 20:15:39 hyperv systemd[1]: Condition check resulted in Rebuild Hardware Database being skipped.
Jun 11 20:15:39 hyperv systemd[1]: Starting Load/Save Random Seed...
Jun 11 20:15:39 hyperv systemd[1]: Starting Apply Kernel Variables...
Jun 11 20:15:39 hyperv systemd[1]: Starting Create System Users...
Jun 11 20:15:39 hyperv systemd[1]: Started Journal Service.
Jun 11 20:15:39 hyperv kernel: audit: type=1130 audit(1623467739.186:8): pid=1 uid=0 auid=4294967295 ses=4294967295 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
Jun 11 20:15:39 hyperv kernel: audit: type=1130 audit(1623467739.186:9): pid=1 uid=0 auid=4294967295 ses=4294967295 msg='unit=systemd-binfmt comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
Jun 11 20:15:39 hyperv kernel: audit: type=1130 audit(1623467739.189:10): pid=1 uid=0 auid=4294967295 ses=4294967295 msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
Jun 11 20:15:39 hyperv kernel: input: PC Speaker as /devices/platform/pcspkr/input/input2
Jun 11 20:15:39 hyperv kernel: mousedev: PS/2 mouse device common for all mice
Jun 11 20:15:42 hyperv kernel: hv_vmbus: registering driver hyperv_fb
Jun 11 20:15:42 hyperv kernel: hv_vmbus: registering driver hv_balloon
Jun 11 20:15:42 hyperv kernel: hv_utils: Registering HyperV Utility Driver
Jun 11 20:15:42 hyperv kernel: hv_vmbus: registering driver hv_utils
Jun 11 20:15:42 hyperv kernel: hyperv_fb: Synthvid Version major 3, minor 5
Jun 11 20:15:42 hyperv kernel: hv_balloon: Using Dynamic Memory protocol version 2.0
Jun 11 20:15:42 hyperv kernel: hyperv_fb: Screen resolution: 1920x1080, Color depth: 32
Jun 11 20:15:42 hyperv kernel: hv_utils: Shutdown IC version 3.2
Jun 11 20:15:42 hyperv kernel: hv_utils: TimeSync IC version 4.0
Jun 11 20:15:42 hyperv kernel: hv_utils: VSS IC version 5.0
Jun 11 20:15:42 hyperv kernel: checking generic (f8000000 300000) vs hw (f8000000 300000)
Jun 11 20:15:42 hyperv kernel: fb0: switching to hyperv_fb from EFI VGA
Jun 11 20:15:42 hyperv kernel: Console: switching to colour dummy device 80x25
Jun 11 20:15:42 hyperv kernel: Console: switching to colour frame buffer device 240x67
Jun 11 20:15:42 hyperv kernel: hv_vmbus: registering driver hv_netvsc
Jun 11 20:15:42 hyperv kernel: cryptd: max_cpu_qlen set to 1000
Jun 11 20:15:42 hyperv kernel: AVX2 version of gcm_enc/dec engaged.
Jun 11 20:15:42 hyperv kernel: AES CTR mode by8 optimization enabled
Jun 11 20:15:42 hyperv kernel: Decoding supported only on Scalable MCA processors.
Jun 11 20:15:42 hyperv kernel: Decoding supported only on Scalable MCA processors.
Jun 11 20:15:42 hyperv kernel: Decoding supported only on Scalable MCA processors.
Jun 11 20:15:43 hyperv kernel: Decoding supported only on Scalable MCA processors.
Jun 11 20:15:43 hyperv kernel: Decoding supported only on Scalable MCA processors.
Jun 11 20:15:43 hyperv kernel: Decoding supported only on Scalable MCA processors.
Jun 11 20:15:43 hyperv kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready
Jun 11 20:15:43 hyperv kernel: random: dbus-daemon: uninitialized urandom read (12 bytes read)
Jun 11 20:15:43 hyperv kernel: Decoding supported only on Scalable MCA processors.
Jun 11 20:15:43 hyperv kernel: Decoding supported only on Scalable MCA processors.
Jun 11 20:15:43 hyperv kernel: Decoding supported only on Scalable MCA processors.
Jun 11 20:15:43 hyperv kernel: Decoding supported only on Scalable MCA processors.
Jun 11 20:15:43 hyperv kernel: Decoding supported only on Scalable MCA processors.
Jun 11 20:15:43 hyperv kernel: Decoding supported only on Scalable MCA processors.
Jun 11 20:15:44 hyperv kernel: random: crng init done
Jun 11 20:15:44 hyperv kernel: random: 1 urandom warning(s) missed due to ratelimiting
Jun 11 20:15:44 hyperv kernel: hv_utils: Heartbeat IC version 3.0
Jun 11 20:15:56 hyperv kernel: kauditd_printk_skb: 48 callbacks suppressed
Jun 11 20:15:56 hyperv kernel: audit: type=1100 audit(1623467756.692:55): pid=450 uid=0 auid=4294967295 ses=4294967295 msg='op=PAM:authentication grantors=pam_shells,pam_faillock,pam_permit,pam_faillock acct="nathan" exe="/usr/lib/sddm/sddm-helper" hostname=? addr=? terminal=? res=success'
Jun 11 20:15:56 hyperv kernel: audit: type=1101 audit(1623467756.692:56): pid=450 uid=0 auid=4294967295 ses=4294967295 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_permit,pam_time acct="nathan" exe="/usr/lib/sddm/sddm-helper" hostname=? addr=? terminal=? res=success'
Jun 11 20:15:56 hyperv kernel: audit: type=1103 audit(1623467756.692:57): pid=450 uid=0 auid=4294967295 ses=4294967295 msg='op=PAM:setcred grantors=pam_shells,pam_faillock,pam_permit,pam_faillock acct="nathan" exe="/usr/lib/sddm/sddm-helper" hostname=? addr=? terminal=? res=success'
Jun 11 20:15:56 hyperv kernel: audit: type=1006 audit(1623467756.692:58): pid=450 uid=0 old-auid=4294967295 auid=1000 tty=(none) old-ses=4294967295 ses=2 res=1
Jun 11 20:15:56 hyperv kernel: audit: type=1300 audit(1623467756.692:58): arch=c000003e syscall=1 success=yes exit=4 a0=8 a1=7ffdeb98b870 a2=4 a3=3e8 items=0 ppid=368 pid=450 auid=1000 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=2 comm="sddm-helper" exe="/usr/lib/sddm/sddm-helper" key=(null)
Jun 11 20:15:56 hyperv kernel: audit: type=1327 audit(1623467756.692:58): proctitle=2F7573722F6C69622F7364646D2F7364646D2D68656C706572002D2D736F636B6574002F746D702F7364646D2D6175746837633433633864352D613933662D346131372D613131302D666363313237363465323963002D2D69640031002D2D7374617274002F7573722F62696E2F7374617274706C61736D612D783131002D2D
Jun 11 20:15:56 hyperv kernel: audit: type=1130 audit(1623467756.702:59): pid=1 uid=0 auid=4294967295 ses=4294967295 msg='unit=user-runtime-dir@1000 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
Jun 11 20:15:56 hyperv kernel: audit: type=1101 audit(1623467756.705:60): pid=452 uid=0 auid=4294967295 ses=4294967295 msg='op=PAM:accounting grantors=pam_access,pam_unix,pam_permit,pam_time acct="nathan" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
Jun 11 20:15:56 hyperv kernel: audit: type=1103 audit(1623467756.705:61): pid=452 uid=0 auid=4294967295 ses=4294967295 msg='op=PAM:setcred grantors=? acct="nathan" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=failed'
Jun 11 20:15:56 hyperv kernel: audit: type=1006 audit(1623467756.705:62): pid=452 uid=0 old-auid=4294967295 auid=1000 tty=(none) old-ses=4294967295 ses=3 res=1
Jun 11 20:16:06 hyperv kernel: kauditd_printk_skb: 17 callbacks suppressed
Jun 11 20:16:06 hyperv kernel: audit: type=1131 audit(1623467766.898:76): pid=1 uid=0 auid=4294967295 ses=4294967295 msg='unit=user@974 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
Jun 11 20:16:06 hyperv kernel: audit: type=1131 audit(1623467766.925:77): pid=1 uid=0 auid=4294967295 ses=4294967295 msg='unit=user-runtime-dir@974 comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success'
Jun 11 20:16:12 hyperv kernel: audit: type=1100 audit(1623467772.648:78): pid=997 uid=1000 auid=1000 ses=2 msg='op=PAM:authentication grantors=pam_faillock,pam_permit,pam_faillock acct="nathan" exe="/usr/bin/doas" hostname=hyperv addr=? terminal=pts/1 res=success'
Jun 11 20:16:12 hyperv kernel: audit: type=1101 audit(1623467772.655:79): pid=997 uid=1000 auid=1000 ses=2 msg='op=PAM:accounting grantors=pam_unix,pam_permit,pam_time acct="nathan" exe="/usr/bin/doas" hostname=hyperv addr=? terminal=pts/1 res=success'
Jun 11 20:16:12 hyperv kernel: audit: type=1110 audit(1623467772.655:80): pid=997 uid=1000 auid=1000 ses=2 msg='op=PAM:setcred grantors=pam_faillock,pam_permit,pam_faillock acct="root" exe="/usr/bin/doas" hostname=hyperv addr=? terminal=pts/1 res=success'
Jun 11 20:16:12 hyperv kernel: audit: type=1105 audit(1623467772.655:81): pid=997 uid=1000 auid=1000 ses=2 msg='op=PAM:session_open grantors=pam_limits,pam_unix,pam_permit acct="root" exe="/usr/bin/doas" hostname=hyperv addr=? terminal=pts/1 res=success'
Jun 11 20:16:13 hyperv kernel: audit: type=1106 audit(1623467773.948:82): pid=997 uid=1000 auid=1000 ses=2 msg='op=PAM:session_close grantors=pam_limits,pam_unix,pam_permit acct="root" exe="/usr/bin/doas" hostname=hyperv addr=? terminal=pts/1 res=success'
Jun 11 20:16:13 hyperv kernel: audit: type=1104 audit(1623467773.948:83): pid=997 uid=1000 auid=1000 ses=2 msg='op=PAM:setcred grantors=pam_faillock,pam_permit,pam_faillock acct="root" exe="/usr/bin/doas" hostname=hyperv addr=? terminal=pts/1 res=success'
Jun 11 20:16:16 hyperv kernel: audit: type=1101 audit(1623467776.955:84): pid=1041 uid=1000 auid=1000 ses=2 msg='op=PAM:accounting grantors=pam_unix,pam_permit,pam_time acct="nathan" exe="/usr/bin/doas" hostname=hyperv addr=? terminal=pts/1 res=success'
Jun 11 20:16:16 hyperv kernel: audit: type=1110 audit(1623467776.955:85): pid=1041 uid=1000 auid=1000 ses=2 msg='op=PAM:setcred grantors=pam_faillock,pam_permit,pam_env,pam_faillock acct="root" exe="/usr/bin/doas" hostname=hyperv addr=? terminal=pts/1 res=success'
Jun 11 20:16:17 hyperv kernel: audit: type=1105 audit(1623467776.955:86): pid=1041 uid=1000 auid=1000 ses=2 msg='op=PAM:session_open grantors=pam_limits,pam_unix,pam_permit acct="root" exe="/usr/bin/doas" hostname=hyperv addr=? terminal=pts/1 res=success'
Jun 11 20:16:31 hyperv kernel: hv_balloon: Max. dynamic memory size: 1048576 MB
Jun 11 20:16:32 hyperv kernel: kworker/0:2: vmemmap alloc failure: order:9, mode:0x4cc0(GFP_KERNEL|__GFP_RETRY_MAYFAIL), nodemask=(null),cpuset=/,mems_allowed=0
Jun 11 20:16:32 hyperv kernel: CPU: 0 PID: 201 Comm: kworker/0:2 Not tainted 5.13.0-rc5-next-20210611-llvm #1
Jun 11 20:16:32 hyperv kernel: Hardware name: Microsoft Corporation Virtual Machine/Virtual Machine, BIOS Hyper-V UEFI Release v4.0 11/01/2019
Jun 11 20:16:32 hyperv kernel: Workqueue: events hot_add_req [hv_balloon]
Jun 11 20:16:32 hyperv kernel: Call Trace:
Jun 11 20:16:32 hyperv kernel: dump_stack_lvl+0x64/0x95
Jun 11 20:16:32 hyperv kernel: warn_alloc+0x10a/0x170
Jun 11 20:16:32 hyperv kernel: vmemmap_alloc_block+0xd4/0xe5
Jun 11 20:16:32 hyperv kernel: vmemmap_populate_hugepages+0x151/0x377
Jun 11 20:16:32 hyperv kernel: vmemmap_populate+0xa6/0xa8
Jun 11 20:16:32 hyperv kernel: __populate_section_memmap+0x29/0x41
Jun 11 20:16:32 hyperv kernel: section_activate+0x136/0x1a8
Jun 11 20:16:32 hyperv kernel: sparse_add_section+0x65/0x103
Jun 11 20:16:32 hyperv kernel: __add_pages+0xbf/0x120
Jun 11 20:16:32 hyperv kernel: arch_add_memory+0x49/0xa0
Jun 11 20:16:32 hyperv kernel: add_memory_resource+0x145/0x380
Jun 11 20:16:32 hyperv kernel: __add_memory+0x3c/0x60
Jun 11 20:16:32 hyperv kernel: add_memory+0x2b/0x40
Jun 11 20:16:32 hyperv kernel: handle_pg_range+0x1cb/0x2c0 [hv_balloon]
Jun 11 20:16:32 hyperv kernel: hot_add_req+0x21e/0x340 [hv_balloon]
Jun 11 20:16:32 hyperv kernel: process_one_work+0x187/0x330
Jun 11 20:16:32 hyperv kernel: worker_thread+0x26d/0x4a0
Jun 11 20:16:32 hyperv kernel: kthread+0x154/0x180
Jun 11 20:16:32 hyperv kernel: ? worker_clr_flags+0x50/0x50
Jun 11 20:16:32 hyperv kernel: ? kthread_blkcg+0x30/0x30
Jun 11 20:16:32 hyperv kernel: ret_from_fork+0x22/0x30
Jun 11 20:16:32 hyperv kernel: Mem-Info:
Jun 11 20:16:32 hyperv kernel: active_anon:258 inactive_anon:179985 isolated_anon:0
active_file:3169 inactive_file:92358 isolated_file:0
unevictable:8 dirty:17 writeback:0
slab_reclaimable:24513 slab_unreclaimable:16311
mapped:69043 shmem:12642 pagetables:4097 bounce:0
free:18819 free_pcp:145 free_cma:0
Jun 11 20:16:32 hyperv kernel: Node 0 active_anon:1032kB inactive_anon:719940kB active_file:12676kB inactive_file:369432kB unevictable:32kB isolated(anon):0kB isolated(file):0kB mapped:276172kB dirty:68kB writeback:0kB shmem:50568kB shmem_thp: 0kB shmem_pmdmapped: 0kB anon_thp: 0kB writeback_tmp:0kB kernel_stack:7200kB pagetables:16388kB all_unreclaimable? no
Jun 11 20:16:32 hyperv kernel: Node 0 DMA free:6092kB min:484kB low:604kB high:724kB reserved_highatomic:0KB active_anon:0kB inactive_anon:8676kB active_file:0kB inactive_file:240kB unevictable:0kB writepending:0kB present:15996kB managed:15360kB mlocked:0kB bounce:0kB free_pcp:0kB local_pcp:0kB free_cma:0kB
Jun 11 20:16:32 hyperv kernel: lowmem_reserve[]: 0 1375 1375 1375 1375
Jun 11 20:16:32 hyperv kernel: Node 0 DMA32 free:69184kB min:44568kB low:55708kB high:66848kB reserved_highatomic:0KB active_anon:1032kB inactive_anon:711264kB active_file:12676kB inactive_file:369368kB unevictable:32kB writepending:68kB present:1555760kB managed:1423464kB mlocked:32kB bounce:0kB free_pcp:580kB local_pcp:0kB free_cma:0kB
Jun 11 20:16:32 hyperv kernel: lowmem_reserve[]: 0 0 0 0 0
Jun 11 20:16:32 hyperv kernel: Node 0 DMA: 4*4kB (UM) 6*8kB (UME) 5*16kB (UE) 4*32kB (UM) 3*64kB (ME) 2*128kB (ME) 1*256kB (E) 2*512kB (UE) 2*1024kB (ME) 1*2048kB (E) 0*4096kB = 6096kB
Jun 11 20:16:32 hyperv kernel: Node 0 DMA32: 3676*4kB (UME) 1221*8kB (UME) 776*16kB (UME) 497*32kB (UME) 192*64kB (UME) 41*128kB (UME) 0*256kB 0*512kB 0*1024kB 0*2048kB 0*4096kB = 70328kB
Jun 11 20:16:32 hyperv kernel: Node 0 hugepages_total=0 hugepages_free=0 hugepages_surp=0 hugepages_size=1048576kB
Jun 11 20:16:32 hyperv kernel: Node 0 hugepages_total=0 hugepages_free=0 hugepages_surp=0 hugepages_size=2048kB
Jun 11 20:16:32 hyperv kernel: 108229 total pagecache pages
Jun 11 20:16:32 hyperv kernel: 0 pages in swap cache
Jun 11 20:16:32 hyperv kernel: Swap cache stats: add 0, delete 0, find 0/0
Jun 11 20:16:32 hyperv kernel: Free swap = 0kB
Jun 11 20:16:32 hyperv kernel: Total swap = 0kB
Jun 11 20:16:32 hyperv kernel: 392939 pages RAM
Jun 11 20:16:32 hyperv kernel: 0 pages HighMem/MovableOnly
Jun 11 20:16:32 hyperv kernel: 33233 pages reserved
Jun 11 20:16:32 hyperv kernel: 0 pages cma reserved
Jun 11 20:16:32 hyperv kernel: 0 pages hwpoisoned
Jun 11 20:16:32 hyperv kernel: Built 1 zonelists, mobility grouping on. Total pages: 375581
Jun 11 20:16:32 hyperv kernel: Policy zone: Normal
Jun 11 20:16:50 hyperv kernel: TCP: eth0: Driver has suspect GRO implementation, TCP performance may be compromised.
Jun 11 20:17:20 hyperv kernel: llvm-tblgen invoked oom-killer: gfp_mask=0x1100cca(GFP_HIGHUSER_MOVABLE), order=0, oom_score_adj=0
Jun 11 20:17:20 hyperv kernel: CPU: 4 PID: 4749 Comm: llvm-tblgen Not tainted 5.13.0-rc5-next-20210611-llvm #1
Jun 11 20:17:20 hyperv kernel: Hardware name: Microsoft Corporation Virtual Machine/Virtual Machine, BIOS Hyper-V UEFI Release v4.0 11/01/2019
Jun 11 20:17:20 hyperv kernel: Call Trace:
Jun 11 20:17:20 hyperv kernel: dump_stack_lvl+0x64/0x95
Jun 11 20:17:20 hyperv kernel: dump_header+0x50/0x240
Jun 11 20:17:20 hyperv kernel: oom_kill_process+0x126/0x230
Jun 11 20:17:20 hyperv kernel: out_of_memory+0x2f7/0x410
Jun 11 20:17:20 hyperv kernel: __alloc_pages_slowpath+0x6e2/0x930
Jun 11 20:17:20 hyperv kernel: __alloc_pages+0x25d/0x350
Jun 11 20:17:20 hyperv kernel: pagecache_get_page+0x2d5/0x530
Jun 11 20:17:20 hyperv kernel: filemap_fault+0x1d3/0x360
Jun 11 20:17:20 hyperv kernel: __xfs_filemap_fault+0x174/0x250 [xfs]
Jun 11 20:17:20 hyperv kernel: __do_fault+0x46/0xa0
Jun 11 20:17:20 hyperv kernel: do_fault+0x12b/0x440
Jun 11 20:17:20 hyperv kernel: __handle_mm_fault+0x6db/0x880
Jun 11 20:17:20 hyperv kernel: handle_mm_fault+0xe5/0x240
Jun 11 20:17:20 hyperv kernel: do_user_addr_fault+0x25d/0x4f0
Jun 11 20:17:20 hyperv kernel: ? asm_exc_page_fault+0x8/0x30
Jun 11 20:17:20 hyperv kernel: exc_page_fault+0x75/0x190
Jun 11 20:17:20 hyperv kernel: asm_exc_page_fault+0x1e/0x30
Jun 11 20:17:20 hyperv kernel: RIP: 0033:0x7fa6cd5a76c0
Jun 11 20:17:20 hyperv kernel: Code: Unable to access opcode bytes at RIP 0x7fa6cd5a7696.
Jun 11 20:17:20 hyperv kernel: RSP: 002b:00007ffe4cc00c38 EFLAGS: 00010206
Jun 11 20:17:20 hyperv kernel: RAX: 0000000000000400 RBX: 00007fa6cd3b257b RCX: 0000556a667ef13e
Jun 11 20:17:20 hyperv kernel: RDX: 0000556a6652f304 RSI: 0000000000000001 RDI: 0000000000000065
Jun 11 20:17:20 hyperv kernel: RBP: 0000000000000065 R08: 0000556a70c453e0 R09: 0000556a6881f010
Jun 11 20:17:20 hyperv kernel: R10: 00007fa6cd732a00 R11: 0000000000000000 R12: 0000000000000001
Jun 11 20:17:20 hyperv kernel: R13: 00007ffe4cc013c0 R14: 00007ffe4cc013c0 R15: 00007fa6cd3b257a
Jun 11 20:17:20 hyperv kernel: Mem-Info:
Jun 11 20:17:20 hyperv kernel: active_anon:5909 inactive_anon:766016 isolated_anon:0
active_file:29 inactive_file:5 isolated_file:0
unevictable:8 dirty:0 writeback:0
slab_reclaimable:25628 slab_unreclaimable:19193
mapped:12421 shmem:18267 pagetables:7166 bounce:0
free:20288 free_pcp:407 free_cma:0
Jun 11 20:17:20 hyperv kernel: Node 0 active_anon:23636kB inactive_anon:3064064kB active_file:116kB inactive_file:20kB unevictable:32kB isolated(anon):0kB isolated(file):0kB mapped:49684kB dirty:0kB writeback:0kB shmem:73068kB shmem_thp: 0kB shmem_pmdmapped: 0kB anon_thp: 0kB writeback_tmp:0kB kernel_stack:7568kB pagetables:28664kB all_unreclaimable? no
Jun 11 20:17:20 hyperv kernel: Node 0 DMA free:6344kB min:296kB low:368kB high:440kB reserved_highatomic:0KB active_anon:0kB inactive_anon:8676kB active_file:0kB inactive_file:0kB unevictable:0kB writepending:0kB present:15996kB managed:15360kB mlocked:0kB bounce:0kB free_pcp:0kB local_pcp:0kB free_cma:0kB
Jun 11 20:17:20 hyperv kernel: lowmem_reserve[]: 0 1390 3396 3396 3396
Jun 11 20:17:20 hyperv kernel: Node 0 DMA32 free:35512kB min:27540kB low:34424kB high:41308kB reserved_highatomic:0KB active_anon:3744kB inactive_anon:1122792kB active_file:0kB inactive_file:776kB unevictable:32kB writepending:0kB present:1555760kB managed:1423464kB mlocked:32kB bounce:0kB free_pcp:132kB local_pcp:52kB free_cma:0kB
Jun 11 20:17:20 hyperv kernel: lowmem_reserve[]: 0 0 2006 2006 2006
Jun 11 20:17:20 hyperv kernel: Node 0 Normal free:39296kB min:39744kB low:49680kB high:59616kB reserved_highatomic:0KB active_anon:19892kB inactive_anon:1931836kB active_file:0kB inactive_file:0kB unevictable:0kB writepending:0kB present:2097152kB managed:2054144kB mlocked:0kB bounce:0kB free_pcp:1504kB local_pcp:356kB free_cma:0kB
Jun 11 20:17:20 hyperv kernel: lowmem_reserve[]: 0 0 0 0 0
Jun 11 20:17:20 hyperv kernel: Node 0 DMA: 20*4kB (UM) 11*8kB (UM) 8*16kB (UM) 5*32kB (UME) 2*64kB (ME) 1*128kB (E) 2*256kB (ME) 2*512kB (UE) 2*1024kB (ME) 1*2048kB (E) 0*4096kB = 6344kB
Jun 11 20:17:20 hyperv kernel: Node 0 DMA32: 1415*4kB (UME) 1175*8kB (UME) 754*16kB (UME) 219*32kB (UME) 56*64kB (UM) 0*128kB 0*256kB 0*512kB 0*1024kB 0*2048kB 0*4096kB = 37716kB
Jun 11 20:17:20 hyperv kernel: Node 0 Normal: 1353*4kB (UME) 887*8kB (UME) 520*16kB (UME) 341*32kB (UME) 113*64kB (UM) 20*128kB (UM) 0*256kB 0*512kB 0*1024kB 0*2048kB 0*4096kB = 41532kB
Jun 11 20:17:20 hyperv kernel: Node 0 hugepages_total=0 hugepages_free=0 hugepages_surp=0 hugepages_size=1048576kB
Jun 11 20:17:20 hyperv kernel: Node 0 hugepages_total=0 hugepages_free=0 hugepages_surp=0 hugepages_size=2048kB
Jun 11 20:17:20 hyperv kernel: 18401 total pagecache pages
Jun 11 20:17:20 hyperv kernel: 0 pages in swap cache
Jun 11 20:17:20 hyperv kernel: Swap cache stats: add 0, delete 0, find 0/0
Jun 11 20:17:20 hyperv kernel: Free swap = 0kB
Jun 11 20:17:20 hyperv kernel: Total swap = 0kB
Jun 11 20:17:20 hyperv kernel: 917227 pages RAM
Jun 11 20:17:20 hyperv kernel: 0 pages HighMem/MovableOnly
Jun 11 20:17:20 hyperv kernel: 43985 pages reserved
Jun 11 20:17:20 hyperv kernel: 0 pages cma reserved
Jun 11 20:17:20 hyperv kernel: 0 pages hwpoisoned
Jun 11 20:17:20 hyperv kernel: Tasks state (memory values in pages):
Jun 11 20:17:20 hyperv kernel: [ pid ] uid tgid total_vm rss pgtables_bytes swapents oom_score_adj name
Jun 11 20:17:20 hyperv kernel: [ 268] 0 268 22862 267 208896 0 -250 systemd-journal
Jun 11 20:17:20 hyperv kernel: [ 287] 0 287 6817 573 77824 0 -1000 systemd-udevd
Jun 11 20:17:20 hyperv kernel: [ 294] 980 294 4920 236 69632 0 0 systemd-network
Jun 11 20:17:20 hyperv kernel: [ 342] 978 342 4809 460 73728 0 0 systemd-resolve
Jun 11 20:17:20 hyperv kernel: [ 362] 81 362 3364 482 65536 0 -900 dbus-daemon
Jun 11 20:17:20 hyperv kernel: [ 363] 0 363 3714 261 69632 0 0 systemd-logind
Jun 11 20:17:20 hyperv kernel: [ 368] 0 368 35319 409 135168 0 0 sddm
Jun 11 20:17:20 hyperv kernel: [ 369] 87 369 19229 210 65536 0 0 ntpd
Jun 11 20:17:20 hyperv kernel: [ 373] 0 373 585290 23858 753664 0 0 Xorg
Jun 11 20:17:20 hyperv kernel: [ 450] 0 450 16258 423 114688 0 0 sddm-helper
Jun 11 20:17:20 hyperv kernel: [ 452] 1000 452 4323 425 73728 0 0 systemd
Jun 11 20:17:20 hyperv kernel: [ 453] 1000 453 5358 678 77824 0 0 (sd-pam)
Jun 11 20:17:20 hyperv kernel: [ 464] 1000 464 70579 1387 253952 0 0 kwalletd5
Jun 11 20:17:20 hyperv kernel: [ 465] 1000 465 35785 554 135168 0 0 startplasma-x11
Jun 11 20:17:20 hyperv kernel: [ 505] 1000 505 38723 69 65536 0 0 gpg-agent
Jun 11 20:17:20 hyperv kernel: [ 513] 1000 513 3088 272 69632 0 0 dbus-daemon
Jun 11 20:17:20 hyperv kernel: [ 563] 1000 563 570 25 40960 0 0 start_kdeinit
Jun 11 20:17:20 hyperv kernel: [ 564] 1000 564 27238 949 192512 0 0 kdeinit5
Jun 11 20:17:20 hyperv kernel: [ 565] 1000 565 70072 1333 249856 0 0 klauncher
Jun 11 20:17:20 hyperv kernel: [ 578] 1000 578 375839 2964 454656 0 0 kded5
Jun 11 20:17:20 hyperv kernel: [ 582] 1000 582 714839 13806 917504 0 0 kwin_x11
Jun 11 20:17:20 hyperv kernel: [ 588] 1000 588 38766 127 65536 0 0 dconf-service
Jun 11 20:17:20 hyperv kernel: [ 645] 1000 645 71164 1480 266240 0 0 ksmserver
Jun 11 20:17:20 hyperv kernel: [ 646] 1000 646 70374 1582 253952 0 0 kglobalaccel5
Jun 11 20:17:20 hyperv kernel: [ 654] 1000 654 71089 1413 266240 0 0 kaccess
Jun 11 20:17:20 hyperv kernel: [ 656] 1000 656 107956 1433 282624 0 0 polkit-kde-auth
Jun 11 20:17:20 hyperv kernel: [ 658] 1000 658 138994 1270 278528 0 0 org_kde_powerde
Jun 11 20:17:20 hyperv kernel: [ 660] 1000 660 56537 633 155648 0 0 xembedsniproxy
Jun 11 20:17:20 hyperv kernel: [ 662] 1000 662 67256497 71335 5390336 0 0 baloo_file
Jun 11 20:17:20 hyperv kernel: [ 664] 1000 664 795205 34461 1159168 0 0 plasmashell
Jun 11 20:17:20 hyperv kernel: [ 672] 1000 672 135409 1096 229376 0 0 kactivitymanage
Jun 11 20:17:20 hyperv kernel: [ 676] 0 676 98835 498 126976 0 0 udisksd
Jun 11 20:17:20 hyperv kernel: [ 687] 102 687 696775 1296 245760 0 0 polkitd
Jun 11 20:17:20 hyperv kernel: [ 709] 1000 709 90530 1481 286720 0 0 DiscoverNotifie
Jun 11 20:17:20 hyperv kernel: [ 712] 1000 712 56776 639 167936 0 0 gmenudbusmenupr
Jun 11 20:17:20 hyperv kernel: [ 785] 0 785 60546 303 102400 0 0 upowerd
Jun 11 20:17:20 hyperv kernel: [ 814] 1000 814 54778 564 155648 0 0 kscreen_backend
Jun 11 20:17:20 hyperv kernel: [ 817] 1000 817 11595 172 77824 0 0 obexd
Jun 11 20:17:20 hyperv kernel: [ 832] 1000 832 45850 999 200704 0 0 desktop.so
Jun 11 20:17:20 hyperv kernel: [ 834] 1000 834 27794 984 196608 0 0 file.so
Jun 11 20:17:20 hyperv kernel: [ 835] 1000 835 26679 664 77824 0 0 pipewire
Jun 11 20:17:20 hyperv kernel: [ 836] 1000 836 22843 389 86016 0 0 pipewire-media-
Jun 11 20:17:20 hyperv kernel: [ 837] 1000 837 21717 229 73728 0 0 pipewire-pulse
Jun 11 20:17:20 hyperv kernel: [ 841] 133 841 38519 68 65536 0 0 rtkit-daemon
Jun 11 20:17:20 hyperv kernel: [ 854] 1000 854 27271 953 192512 0 0 file.so
Jun 11 20:17:20 hyperv kernel: [ 856] 1000 856 67251577 65423 6680576 0 0 baloo_file_extr
Jun 11 20:17:20 hyperv kernel: [ 862] 1000 862 705374 7931 892928 0 0 konsole
Jun 11 20:17:20 hyperv kernel: [ 899] 1000 899 42211 401 86016 0 0 fish
Jun 11 20:17:20 hyperv kernel: [ 1041] 1000 1041 2537 187 61440 0 0 sudo
Jun 11 20:17:20 hyperv kernel: [ 1042] 0 1042 720 28 45056 0 0 dmesg
Jun 11 20:17:20 hyperv kernel: [ 1049] 1000 1049 42186 379 86016 0 0 fish
Jun 11 20:17:20 hyperv kernel: [ 1151] 1000 1151 2590 733 61440 0 0 htop
Jun 11 20:17:20 hyperv kernel: [ 1152] 1000 1152 40137 380 77824 0 0 fish
Jun 11 20:17:20 hyperv kernel: [ 1256] 1000 1256 6795 2710 98304 0 0 python3
Jun 11 20:17:20 hyperv kernel: [ 1329] 1000 1329 27271 953 192512 0 0 file.so
Jun 11 20:17:20 hyperv kernel: [ 2761] 1000 2761 6686 5087 90112 0 0 ninja
Jun 11 20:17:20 hyperv kernel: [ 3564] 1000 3564 2105 135 53248 0 0 ccache
Jun 11 20:17:20 hyperv kernel: [ 3610] 1000 3610 100429 52196 786432 0 0 clang++
Jun 11 20:17:20 hyperv kernel: [ 4037] 1000 4037 2106 136 65536 0 0 ccache
Jun 11 20:17:20 hyperv kernel: [ 4100] 1000 4100 99819 51760 774144 0 0 clang++
Jun 11 20:17:20 hyperv kernel: [ 4229] 1000 4229 2116 163 53248 0 0 ccache
Jun 11 20:17:20 hyperv kernel: [ 4247] 1000 4247 86088 38126 663552 0 0 clang++
Jun 11 20:17:20 hyperv kernel: [ 4359] 1000 4359 2105 133 53248 0 0 ccache
Jun 11 20:17:20 hyperv kernel: [ 4393] 1000 4393 76078 28074 577536 0 0 clang++
Jun 11 20:17:20 hyperv kernel: [ 4442] 1000 4442 2149 172 53248 0 0 ccache
Jun 11 20:17:20 hyperv kernel: [ 4473] 1000 4473 2148 176 53248 0 0 ccache
Jun 11 20:17:20 hyperv kernel: [ 4483] 1000 4483 69999 22528 528384 0 0 clang++
Jun 11 20:17:20 hyperv kernel: [ 4521] 1000 4521 70849 23427 528384 0 0 clang++
Jun 11 20:17:20 hyperv kernel: [ 4738] 1000 4738 35626 32417 327680 0 0 llvm-tblgen
Jun 11 20:17:20 hyperv kernel: [ 4739] 1000 4739 37487 34256 331776 0 0 llvm-tblgen
Jun 11 20:17:20 hyperv kernel: [ 4741] 1000 4741 34727 31542 315392 0 0 llvm-tblgen
Jun 11 20:17:20 hyperv kernel: [ 4742] 1000 4742 41070 37633 368640 0 0 llvm-tblgen
Jun 11 20:17:20 hyperv kernel: [ 4746] 1000 4746 38537 35229 352256 0 0 llvm-tblgen
Jun 11 20:17:20 hyperv kernel: [ 4747] 1000 4747 69690 64648 598016 0 0 llvm-tblgen
Jun 11 20:17:20 hyperv kernel: [ 4749] 1000 4749 37487 34224 339968 0 0 llvm-tblgen
Jun 11 20:17:20 hyperv kernel: [ 4750] 1000 4750 36071 32873 331776 0 0 llvm-tblgen
Jun 11 20:17:20 hyperv kernel: oom-kill:constraint=CONSTRAINT_NONE,nodemask=(null),cpuset=/,mems_allowed=0,global_oom,task_memcg=/user.slice/user-1000.slice/session-2.scope,task=baloo_file,pid=662,uid=1000
Jun 11 20:17:20 hyperv kernel: Out of memory: Killed process 662 (baloo_file) total-vm:269025988kB, anon-rss:285340kB, file-rss:0kB, shmem-rss:0kB, UID:1000 pgtables:5264kB oom_score_adj:0
Jun 11 20:17:20 hyperv kernel: oom_reaper: reaped process 662 (baloo_file), now anon-rss:0kB, file-rss:4kB, shmem-rss:0kB
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 61534
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 7412
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 35438
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 29934
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 11207
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 61534
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 11207
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 11207
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 7412
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 7412
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 61534
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 35438
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 7412
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 29934
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 29934
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 7412
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 11207
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 61534
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 11207
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 11207
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 35438
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 29934
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 61534
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 7412
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 29934
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 29934
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 35438
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 7412
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 7412
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 11207
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 11207
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 61534
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 61534
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 35438
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 11207
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 11207
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 61534
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 61534
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 61534
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 61534
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 29934
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 35438
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 7412
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 35438
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 61534
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 29934
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 7412
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 7412
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 7412
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 11207
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 7412
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 29934
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 61534
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 11207
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 7412
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 7412
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 7412
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 61534
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 35438
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 7412
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 61534
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 35438
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 7412
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 29934
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 35438
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 29934
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 61534
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 61534
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 35438
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 35438
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 7412
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 35438
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 29934
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 35438
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 11207
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 29934
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 61534
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 61534
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 35438
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 11207
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 7412
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 29934
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 35438
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 29934
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 35438
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 29934
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 7412
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 61534
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 35438
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 7412
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 11207
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 11207
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 61534
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 61534
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 11207
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 29934
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 7412
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 61534
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 61534
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 11207
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 61534
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 29934
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 11207
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 35438
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 11207
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 35438
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 61534
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 35438
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 11207
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 11207
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 61534
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 11207
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 29934
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 29934
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 61534
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 7412
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 7412
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 61534
Jun 11 20:24:50 hyperv kernel: hv_balloon: Unhandled message: type: 61534
^ permalink raw reply [flat|nested] 7+ messages in thread
* Re: vmemmap alloc failure in hot_add_req()
[not found] ` <20210612021115.2136-1-hdanton@sina.com>
@ 2021-06-14 7:38 ` David Hildenbrand
2021-06-17 1:08 ` Nathan Chancellor
0 siblings, 1 reply; 7+ messages in thread
From: David Hildenbrand @ 2021-06-14 7:38 UTC (permalink / raw)
To: Hillf Danton, Nathan Chancellor
Cc: K. Y. Srinivasan, Haiyang Zhang, Stephen Hemminger, Wei Liu,
Dexuan Cui, linux-hyperv, linux-mm, linux-kernel, Oscar Salvador
On 12.06.21 04:11, Hillf Danton wrote:
> On Fri, 11 Jun 2021 12:48:26 -0700 Nathan Chancellor wrote:
>> Hi all,
>>
>> I am occasionally seeing a kernel warning when running virtual machines
>> in Hyper-V, which usually happens a minute or so after boot. It does not
>> happen on every boot and it is reproducible on at least v5.10. I think
>> it might have something to do with constant reboots, which I do when
>> testing various kernels.
>>
>> The stack trace is as follows:
>>
>> [ 49.215291] kworker/0:1: vmemmap alloc failure: order:9, mode:0x4cc0(GFP_KERNEL|__GFP_RETRY_MAYFAIL), nodemask=(null),cpuset=/,mems_allowed=0
>> [ 49.215299] CPU: 0 PID: 18 Comm: kworker/0:1 Not tainted 5.13.0-rc5 #1
>> [ 49.215301] Hardware name: Microsoft Corporation Virtual Machine/Virtual Machine, BIOS Hyper-V UEFI Release v4.0 11/01/2019
>> [ 49.215302] Workqueue: events hot_add_req [hv_balloon]
>
> Apart from order:9 (mm Cced), events_unbound is the right workqueue instead
> because the report shows the risk that hot_add_req could block other pending
> events longer than thought. Any special reason for the events wq?
>
>> [ 49.215307] Call Trace:
>> [ 49.215310] dump_stack+0x76/0x94
>> [ 49.215314] warn_alloc.cold+0x78/0xdc
>> [ 49.215316] ? __alloc_pages+0x200/0x230
>> [ 49.215319] vmemmap_alloc_block+0x86/0xdc
>> [ 49.215323] vmemmap_populate+0x10e/0x31c
>> [ 49.215324] __populate_section_memmap+0x38/0x4e
>> [ 49.215326] sparse_add_section+0x12c/0x1cf
>> [ 49.215329] __add_pages+0xa9/0x130
>> [ 49.215330] add_pages+0x12/0x60
>> [ 49.215333] add_memory_resource+0x180/0x300
>> [ 49.215335] __add_memory+0x3b/0x80
>> [ 49.215336] add_memory+0x2e/0x50
>> [ 49.215337] hot_add_req+0x3fc/0x5a0 [hv_balloon]
>> [ 49.215340] process_one_work+0x214/0x3e0
>> [ 49.215342] worker_thread+0x4d/0x3d0
>> [ 49.215344] ? process_one_work+0x3e0/0x3e0
>> [ 49.215345] kthread+0x133/0x150
>> [ 49.215347] ? kthread_associate_blkcg+0xc0/0xc0
>> [ 49.215348] ret_from_fork+0x22/0x30
>> [ 49.215351] Mem-Info:
>> [ 49.215352] active_anon:251 inactive_anon:140868 isolated_anon:0
>> active_file:47497 inactive_file:88505 isolated_file:0
>> unevictable:8 dirty:14 writeback:0
>> slab_reclaimable:12013 slab_unreclaimable:11403
>> mapped:131701 shmem:12671 pagetables:3140 bounce:0
>> free:41388 free_pcp:37 free_cma:0
>> [ 49.215355] Node 0 active_anon:1004kB inactive_anon:563472kB active_file:189988kB inactive_file:354020kB unevictable:32kB isolated(anon):0kB isolated(file):0kB mapped:526804kB dirty:56kB writeback:0kB shmem:50684kB shmem_thp: 0kB shmem_pmdmapped: 0kB anon_thp: 0kB writeback_tmp:0kB kernel_stack:5904kB pagetables:12560kB all_unreclaimable? no
>> [ 49.215358] Node 0 DMA free:6496kB min:480kB low:600kB high:720kB reserved_highatomic:0KB active_anon:0kB inactive_anon:3120kB active_file:2584kB inactive_file:2792kB unevictable:0kB writepending:0kB present:15996kB managed:15360kB mlocked:0kB bounce:0kB free_pcp:0kB local_pcp:0kB free_cma:0kB
>> [ 49.215361] lowmem_reserve[]: 0 1384 1384 1384 1384
>> [ 49.215364] Node 0 DMA32 free:159056kB min:44572kB low:55712kB high:66852kB reserved_highatomic:0KB active_anon:1004kB inactive_anon:560352kB active_file:187004kB inactive_file:350864kB unevictable:32kB writepending:56kB present:1555760kB managed:1432388kB mlocked:32kB bounce:0kB free_pcp:172kB local_pcp:0kB free_cma:0kB
>> [ 49.215367] lowmem_reserve[]: 0 0 0 0 0
>> [ 49.215369] Node 0 DMA: 17*4kB (UM) 13*8kB (M) 10*16kB (M) 3*32kB (ME) 3*64kB (UME) 4*128kB (UME) 1*256kB (E) 2*512kB (UE) 2*1024kB (ME) 1*2048kB (E) 0*4096kB = 6508kB
>> [ 49.215377] Node 0 DMA32: 8061*4kB (UME) 5892*8kB (UME) 2449*16kB (UME) 604*32kB (UME) 207*64kB (UME) 49*128kB (UM) 7*256kB (M) 1*512kB (M) 0*1024kB 0*2048kB 0*4096kB = 159716kB
>> [ 49.215388] 148696 total pagecache pages
>> [ 49.215388] 0 pages in swap cache
>> [ 49.215389] Swap cache stats: add 0, delete 0, find 0/0
>> [ 49.215390] Free swap = 0kB
>> [ 49.215390] Total swap = 0kB
>> [ 49.215391] 392939 pages RAM
>> [ 49.215391] 0 pages HighMem/MovableOnly
>> [ 49.215391] 31002 pages reserved
>> [ 49.215392] 0 pages cma reserved
>> [ 49.215393] 0 pages hwpoisoned
>>
>> Is this a known issue and/or am I doing something wrong? I only noticed
>> this because there are times when I am compiling something intensive in
>> the VM such as LLVM and the VM runs out of memory even though I have
>> plenty of free memory on the host but I am not sure if this warning is
>> related to that issue.
Hi,
Is hotplugged memory getting onlined automatically (either from user
space via a udev script or via the kernel, for example, with
CONFIG_MEMORY_HOTPLUG_DEFAULT_ONLINE)?
If it's not getting onlined, you easily sport after hotplug e.g., via
"lsmem" that there are quite some offline memory blocks.
Note that x86_64 code will fallback from populating huge pages to
populating base pages for the vmemmap; this can happen easily when under
memory pressure.
If adding memory would fail completely, you'd see another "hot_add
memory failed error is ..." error message from hyper-v in the kernel
log. If that doesn't show up, it's simply suboptimal, but hotplugging
memory still succeeded.
Note: we could support "memmap_on_memory" in some cases (e.g., no memory
holes in hotadded range) when hotplugging memory blocks via hyper-v,
which would result in this warning less trigger less frequently.
--
Thanks,
David / dhildenb
^ permalink raw reply [flat|nested] 7+ messages in thread
* Re: vmemmap alloc failure in hot_add_req()
2021-06-14 7:38 ` David Hildenbrand
@ 2021-06-17 1:08 ` Nathan Chancellor
2021-06-17 8:42 ` David Hildenbrand
0 siblings, 1 reply; 7+ messages in thread
From: Nathan Chancellor @ 2021-06-17 1:08 UTC (permalink / raw)
To: David Hildenbrand
Cc: K. Y. Srinivasan, Haiyang Zhang, Stephen Hemminger, Wei Liu,
Dexuan Cui, linux-hyperv, linux-mm, linux-kernel, Oscar Salvador,
Hillf Danton
Hi David,
On 6/14/2021 12:38 AM, David Hildenbrand wrote:
> On 12.06.21 04:11, Hillf Danton wrote:
>> On Fri, 11 Jun 2021 12:48:26 -0700 Nathan Chancellor wrote:
>>> Hi all,
>>>
>>> I am occasionally seeing a kernel warning when running virtual machines
>>> in Hyper-V, which usually happens a minute or so after boot. It does not
>>> happen on every boot and it is reproducible on at least v5.10. I think
>>> it might have something to do with constant reboots, which I do when
>>> testing various kernels.
>>>
>>> The stack trace is as follows:
>>>
>>> [ 49.215291] kworker/0:1: vmemmap alloc failure: order:9,
>>> mode:0x4cc0(GFP_KERNEL|__GFP_RETRY_MAYFAIL),
>>> nodemask=(null),cpuset=/,mems_allowed=0
>>> [ 49.215299] CPU: 0 PID: 18 Comm: kworker/0:1 Not tainted
>>> 5.13.0-rc5 #1
>>> [ 49.215301] Hardware name: Microsoft Corporation Virtual
>>> Machine/Virtual Machine, BIOS Hyper-V UEFI Release v4.0 11/01/2019
>>> [ 49.215302] Workqueue: events hot_add_req [hv_balloon]
>>
>> Apart from order:9 (mm Cced), events_unbound is the right workqueue
>> instead
>> because the report shows the risk that hot_add_req could block other
>> pending
>> events longer than thought. Any special reason for the events wq?
>>
>>> [ 49.215307] Call Trace:
>>> [ 49.215310] dump_stack+0x76/0x94
>>> [ 49.215314] warn_alloc.cold+0x78/0xdc
>>> [ 49.215316] ? __alloc_pages+0x200/0x230
>>> [ 49.215319] vmemmap_alloc_block+0x86/0xdc
>>> [ 49.215323] vmemmap_populate+0x10e/0x31c
>>> [ 49.215324] __populate_section_memmap+0x38/0x4e
>>> [ 49.215326] sparse_add_section+0x12c/0x1cf
>>> [ 49.215329] __add_pages+0xa9/0x130
>>> [ 49.215330] add_pages+0x12/0x60
>>> [ 49.215333] add_memory_resource+0x180/0x300
>>> [ 49.215335] __add_memory+0x3b/0x80
>>> [ 49.215336] add_memory+0x2e/0x50
>>> [ 49.215337] hot_add_req+0x3fc/0x5a0 [hv_balloon]
>>> [ 49.215340] process_one_work+0x214/0x3e0
>>> [ 49.215342] worker_thread+0x4d/0x3d0
>>> [ 49.215344] ? process_one_work+0x3e0/0x3e0
>>> [ 49.215345] kthread+0x133/0x150
>>> [ 49.215347] ? kthread_associate_blkcg+0xc0/0xc0
>>> [ 49.215348] ret_from_fork+0x22/0x30
>>> [ 49.215351] Mem-Info:
>>> [ 49.215352] active_anon:251 inactive_anon:140868 isolated_anon:0
>>> active_file:47497 inactive_file:88505 isolated_file:0
>>> unevictable:8 dirty:14 writeback:0
>>> slab_reclaimable:12013 slab_unreclaimable:11403
>>> mapped:131701 shmem:12671 pagetables:3140 bounce:0
>>> free:41388 free_pcp:37 free_cma:0
>>> [ 49.215355] Node 0 active_anon:1004kB inactive_anon:563472kB
>>> active_file:189988kB inactive_file:354020kB unevictable:32kB
>>> isolated(anon):0kB isolated(file):0kB mapped:526804kB dirty:56kB
>>> writeback:0kB shmem:50684kB shmem_thp: 0kB shmem_pmdmapped: 0kB
>>> anon_thp: 0kB writeback_tmp:0kB kernel_stack:5904kB
>>> pagetables:12560kB all_unreclaimable? no
>>> [ 49.215358] Node 0 DMA free:6496kB min:480kB low:600kB high:720kB
>>> reserved_highatomic:0KB active_anon:0kB inactive_anon:3120kB
>>> active_file:2584kB inactive_file:2792kB unevictable:0kB
>>> writepending:0kB present:15996kB managed:15360kB mlocked:0kB
>>> bounce:0kB free_pcp:0kB local_pcp:0kB free_cma:0kB
>>> [ 49.215361] lowmem_reserve[]: 0 1384 1384 1384 1384
>>> [ 49.215364] Node 0 DMA32 free:159056kB min:44572kB low:55712kB
>>> high:66852kB reserved_highatomic:0KB active_anon:1004kB
>>> inactive_anon:560352kB active_file:187004kB inactive_file:350864kB
>>> unevictable:32kB writepending:56kB present:1555760kB
>>> managed:1432388kB mlocked:32kB bounce:0kB free_pcp:172kB
>>> local_pcp:0kB free_cma:0kB
>>> [ 49.215367] lowmem_reserve[]: 0 0 0 0 0
>>> [ 49.215369] Node 0 DMA: 17*4kB (UM) 13*8kB (M) 10*16kB (M) 3*32kB
>>> (ME) 3*64kB (UME) 4*128kB (UME) 1*256kB (E) 2*512kB (UE) 2*1024kB
>>> (ME) 1*2048kB (E) 0*4096kB = 6508kB
>>> [ 49.215377] Node 0 DMA32: 8061*4kB (UME) 5892*8kB (UME) 2449*16kB
>>> (UME) 604*32kB (UME) 207*64kB (UME) 49*128kB (UM) 7*256kB (M) 1*512kB
>>> (M) 0*1024kB 0*2048kB 0*4096kB = 159716kB
>>> [ 49.215388] 148696 total pagecache pages
>>> [ 49.215388] 0 pages in swap cache
>>> [ 49.215389] Swap cache stats: add 0, delete 0, find 0/0
>>> [ 49.215390] Free swap = 0kB
>>> [ 49.215390] Total swap = 0kB
>>> [ 49.215391] 392939 pages RAM
>>> [ 49.215391] 0 pages HighMem/MovableOnly
>>> [ 49.215391] 31002 pages reserved
>>> [ 49.215392] 0 pages cma reserved
>>> [ 49.215393] 0 pages hwpoisoned
>>>
>>> Is this a known issue and/or am I doing something wrong? I only noticed
>>> this because there are times when I am compiling something intensive in
>>> the VM such as LLVM and the VM runs out of memory even though I have
>>> plenty of free memory on the host but I am not sure if this warning is
>>> related to that issue.
>
> Hi,
>
> Is hotplugged memory getting onlined automatically (either from user
> space via a udev script or via the kernel, for example, with
> CONFIG_MEMORY_HOTPLUG_DEFAULT_ONLINE)?
It does look like this kernel configuration has
CONFIG_MEMORY_HOTPLUG_DEFAULT_ONLINE=y.
> If it's not getting onlined, you easily sport after hotplug e.g., via
> "lsmem" that there are quite some offline memory blocks.
>
> Note that x86_64 code will fallback from populating huge pages to
> populating base pages for the vmemmap; this can happen easily when under
> memory pressure.
Not sure if it is relevant or not but this warning can show up within a
minute of startup without me doing anything in particular.
> If adding memory would fail completely, you'd see another "hot_add
> memory failed error is ..." error message from hyper-v in the kernel
> log. If that doesn't show up, it's simply suboptimal, but hotplugging
> memory still succeeded.
I did notice that from the code in hv_balloon.c but I do not think I
have ever seen that message in my logs.
> Note: we could support "memmap_on_memory" in some cases (e.g., no memory
> holes in hotadded range) when hotplugging memory blocks via hyper-v,
> which would result in this warning less trigger less frequently.
Cheers,
Nathan
^ permalink raw reply [flat|nested] 7+ messages in thread
* Re: vmemmap alloc failure in hot_add_req()
2021-06-17 1:08 ` Nathan Chancellor
@ 2021-06-17 8:42 ` David Hildenbrand
2021-06-18 0:16 ` Michael Kelley
0 siblings, 1 reply; 7+ messages in thread
From: David Hildenbrand @ 2021-06-17 8:42 UTC (permalink / raw)
To: Nathan Chancellor
Cc: K. Y. Srinivasan, Haiyang Zhang, Stephen Hemminger, Wei Liu,
Dexuan Cui, linux-hyperv, linux-mm, linux-kernel, Oscar Salvador,
Hillf Danton
> It does look like this kernel configuration has
> CONFIG_MEMORY_HOTPLUG_DEFAULT_ONLINE=y.
Okay, so then it's most likely really more of an issue with fragmented
physical memory -- which is suboptimal but not a show blocker in your setup.
(there are still cases where memory onlining can fail, especially with
kasan running, but these are rather corner cases)
>
>> If it's not getting onlined, you easily sport after hotplug e.g., via
>> "lsmem" that there are quite some offline memory blocks.
>>
>> Note that x86_64 code will fallback from populating huge pages to
>> populating base pages for the vmemmap; this can happen easily when under
>> memory pressure.
>
> Not sure if it is relevant or not but this warning can show up within a
> minute of startup without me doing anything in particular.
I remember that Hyper-V will start with a certain (configured) boot VM
memory size and once the guest is up and running, use memory stats of
the guest to decide whether to add (hotplug) or remove (balloon inflate)
memory from the VM.
So this could just be Hyper-V trying to apply its heuristics.
>
>> If adding memory would fail completely, you'd see another "hot_add
>> memory failed error is ..." error message from hyper-v in the kernel
>> log. If that doesn't show up, it's simply suboptimal, but hotplugging
>> memory still succeeded.
>
> I did notice that from the code in hv_balloon.c but I do not think I
> have ever seen that message in my logs.
Okay, so at least hotplugging memory is working.
--
Thanks,
David / dhildenb
^ permalink raw reply [flat|nested] 7+ messages in thread
* RE: vmemmap alloc failure in hot_add_req()
2021-06-17 8:42 ` David Hildenbrand
@ 2021-06-18 0:16 ` Michael Kelley
2021-06-18 21:49 ` Nathan Chancellor
0 siblings, 1 reply; 7+ messages in thread
From: Michael Kelley @ 2021-06-18 0:16 UTC (permalink / raw)
To: David Hildenbrand, Nathan Chancellor
Cc: KY Srinivasan, Haiyang Zhang, Stephen Hemminger, Wei Liu,
Dexuan Cui, linux-hyperv, linux-mm, linux-kernel, Oscar Salvador,
Hillf Danton
From: David Hildenbrand <david@redhat.com> Sent: Thursday, June 17, 2021 1:43 AM
>
> > It does look like this kernel configuration has
> > CONFIG_MEMORY_HOTPLUG_DEFAULT_ONLINE=y.
>
> Okay, so then it's most likely really more of an issue with fragmented
> physical memory -- which is suboptimal but not a show blocker in your setup.
>
> (there are still cases where memory onlining can fail, especially with
> kasan running, but these are rather corner cases)
>
> >
> >> If it's not getting onlined, you easily sport after hotplug e.g., via
> >> "lsmem" that there are quite some offline memory blocks.
> >>
> >> Note that x86_64 code will fallback from populating huge pages to
> >> populating base pages for the vmemmap; this can happen easily when under
> >> memory pressure.
> >
> > Not sure if it is relevant or not but this warning can show up within a
> > minute of startup without me doing anything in particular.
>
> I remember that Hyper-V will start with a certain (configured) boot VM
> memory size and once the guest is up and running, use memory stats of
> the guest to decide whether to add (hotplug) or remove (balloon inflate)
> memory from the VM.
>
> So this could just be Hyper-V trying to apply its heuristics.
Nathan --
Could you clarify if your VM is running in the context of the Windows
Subsystem for Linux (WSL) v2 feature in Windows 10? Or are you
running a "traditional" VM created using the Hyper-V Manager UI
or Powershell?
If the latter, how do you have the memory configuration set up? In
the UI, first you can specify the RAM allocated to the VM. Then
separately, you can enable the "Dynamic Memory" feature, in which
case you also specify a "Minimum RAM" and "Maximum RAM". It
looks like you must have the "Dynamic Memory" feature enabled
since the original stack trace includes the hot_add_req() function
from the hv_balloon driver.
The Dynamic Memory feature is generally used only when you
need to allow Hyper-V to manage the allocation of physical memory
across multiple VMs. Dynamic Memory is essentially Hyper-V's way of
allowing memory overcommit. If you don't need that capability,
turning off Dynamic Memory and just specifying the amount of
memory you want to assign to the VM is the best course of action.
With Dynamic Memory enabled, you may have encountered a
situation where the memory needs of the VM grew very quickly,
and Hyper-V balloon driver got into a situation where it needed
to allocate memory in order to add memory, and it couldn't. If
you want to continue to use the Dynamic Memory feature, then
you probably need to increase the initial amount of RAM assigned
to the VM (the "RAM" setting in the Hyper-V Manager UI).
Michael
>
> >
> >> If adding memory would fail completely, you'd see another "hot_add
> >> memory failed error is ..." error message from hyper-v in the kernel
> >> log. If that doesn't show up, it's simply suboptimal, but hotplugging
> >> memory still succeeded.
> >
> > I did notice that from the code in hv_balloon.c but I do not think I
> > have ever seen that message in my logs.
>
> Okay, so at least hotplugging memory is working.
>
> --
> Thanks,
>
> David / dhildenb
^ permalink raw reply [flat|nested] 7+ messages in thread
* Re: vmemmap alloc failure in hot_add_req()
2021-06-18 0:16 ` Michael Kelley
@ 2021-06-18 21:49 ` Nathan Chancellor
0 siblings, 0 replies; 7+ messages in thread
From: Nathan Chancellor @ 2021-06-18 21:49 UTC (permalink / raw)
To: Michael Kelley, David Hildenbrand
Cc: KY Srinivasan, Haiyang Zhang, Stephen Hemminger, Wei Liu,
Dexuan Cui, linux-hyperv, linux-mm, linux-kernel, Oscar Salvador,
Hillf Danton
Hi Michael,
On 6/17/2021 5:16 PM, Michael Kelley wrote:
> From: David Hildenbrand <david@redhat.com> Sent: Thursday, June 17, 2021 1:43 AM
>>
>>> It does look like this kernel configuration has
>>> CONFIG_MEMORY_HOTPLUG_DEFAULT_ONLINE=y.
>>
>> Okay, so then it's most likely really more of an issue with fragmented
>> physical memory -- which is suboptimal but not a show blocker in your setup.
>>
>> (there are still cases where memory onlining can fail, especially with
>> kasan running, but these are rather corner cases)
>>
>>>
>>>> If it's not getting onlined, you easily sport after hotplug e.g., via
>>>> "lsmem" that there are quite some offline memory blocks.
>>>>
>>>> Note that x86_64 code will fallback from populating huge pages to
>>>> populating base pages for the vmemmap; this can happen easily when under
>>>> memory pressure.
>>>
>>> Not sure if it is relevant or not but this warning can show up within a
>>> minute of startup without me doing anything in particular.
>>
>> I remember that Hyper-V will start with a certain (configured) boot VM
>> memory size and once the guest is up and running, use memory stats of
>> the guest to decide whether to add (hotplug) or remove (balloon inflate)
>> memory from the VM.
>>
>> So this could just be Hyper-V trying to apply its heuristics.
>
> Nathan --
>
> Could you clarify if your VM is running in the context of the Windows
> Subsystem for Linux (WSL) v2 feature in Windows 10? Or are you
> running a "traditional" VM created using the Hyper-V Manager UI
> or Powershell?
This is a traditional VM created using the Hyper-V Manager.
> If the latter, how do you have the memory configuration set up? In
> the UI, first you can specify the RAM allocated to the VM. Then
> separately, you can enable the "Dynamic Memory" feature, in which
> case you also specify a "Minimum RAM" and "Maximum RAM". It
> looks like you must have the "Dynamic Memory" feature enabled
> since the original stack trace includes the hot_add_req() function
> from the hv_balloon driver.
That is correct. I believe Dynamic Memory is the default setting so I
just left that as it was. The startup memory for this virtual machine is
2GB as it is a lightweight Arch Linux Xfce4 configuration and aside from
occasionally compiling software, it will just be sitting there because
it is mainly there for testing kernels.
> The Dynamic Memory feature is generally used only when you
> need to allow Hyper-V to manage the allocation of physical memory
> across multiple VMs. Dynamic Memory is essentially Hyper-V's way of
> allowing memory overcommit. If you don't need that capability,
> turning off Dynamic Memory and just specifying the amount of
> memory you want to assign to the VM is the best course of action.
Ack. My workstation was occasionally memory constrained so I figured
relying on the Dynamic Memory feature would make sense. I upgraded the
amount of RAM that I had today so I will probably just end up disabling
the Dynamic Memory feature and allocating the amount of memory up front.
> With Dynamic Memory enabled, you may have encountered a
> situation where the memory needs of the VM grew very quickly,
> and Hyper-V balloon driver got into a situation where it needed
> to allocate memory in order to add memory, and it couldn't. If
> you want to continue to use the Dynamic Memory feature, then
> you probably need to increase the initial amount of RAM assigned
> to the VM (the "RAM" setting in the Hyper-V Manager UI).
I will keep that in mind and see if I can find a good number.
Thanks for the reply!
Cheers,
Nathan
^ permalink raw reply [flat|nested] 7+ messages in thread
end of thread, other threads:[~2021-06-18 21:49 UTC | newest]
Thread overview: 7+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2021-06-11 19:48 vmemmap alloc failure in hot_add_req() Nathan Chancellor
2021-06-12 3:42 ` Nathan Chancellor
[not found] ` <20210612021115.2136-1-hdanton@sina.com>
2021-06-14 7:38 ` David Hildenbrand
2021-06-17 1:08 ` Nathan Chancellor
2021-06-17 8:42 ` David Hildenbrand
2021-06-18 0:16 ` Michael Kelley
2021-06-18 21:49 ` Nathan Chancellor
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).