From: Alex Shi <alex.shi@linux.alibaba.com>
To: Daniel Jordan <daniel.m.jordan@oracle.com>
Cc: akpm@linux-foundation.org, mgorman@techsingularity.net,
tj@kernel.org, hughd@google.com, khlebnikov@yandex-team.ru,
willy@infradead.org, hannes@cmpxchg.org, lkp@intel.com,
linux-mm@kvack.org, linux-kernel@vger.kernel.org,
cgroups@vger.kernel.org, shakeelb@google.com,
iamjoonsoo.kim@lge.com, richard.weiyang@gmail.com,
kirill@shutemov.name, alexander.duyck@gmail.com,
rong.a.chen@intel.com, mhocko@suse.com, vdavydov.dev@gmail.com,
shy828301@gmail.com, aaron.lwe@gmail.com
Subject: Re: [PATCH v19 00/20] per memcg lru_lock
Date: Fri, 25 Sep 2020 08:01:08 +0800 [thread overview]
Message-ID: <c633b107-845a-1c70-b8bb-5d4163bfa274@linux.alibaba.com> (raw)
In-Reply-To: <20200924180621.2r62kv4lumnry4zm@ca-dmjordan1.us.oracle.com>
在 2020/9/25 上午2:06, Daniel Jordan 写道:
> On Thu, Sep 24, 2020 at 11:28:15AM +0800, Alex Shi wrote:
>> The new version rebased on v5.9-rc6 with line by line review by Hugh Dickins.
>
> These thpscale numbers are from a proto-v19, which now seems to be gone from
Hi Daniel,
Thanks a lot for testing!
Is the head of lruv19 branch 9ee69337d51d on my github? We may just hit a narrow
window, which I just removed the old lruv19, and updated a new one. But even the
old version has same performance behavior too. :)
> Alex Shi's github. Once again, this is a regression test where memcg is
> enabled but not used. There's a reasonable amount of isolations from
> compaction as expected, as well as reclaim to a lesser extent.
>
> I confined the runs to one node of a two-socket broadwell server, the same
> machine as my other results. thpscale's total_size argument is about 80% of
> the node's memory, its madvise option was enabled, and the system's thp
> settings are enabled=always and defrag=madvise.
>
> Both base and lru kernels show huge variance run to run, which is surprising
> because I was expecting a microbenchmark like this to be more stable. There
> seems to be an overall decrease in mean fault latency with the lru changes, but
> it's in the noise, so it's hard to say how much of an improvement it really is.
Yes, the lru lock replace has no clear performance changes, unless the
migration/compaction compete on different lru locks which require to run in
different memcgs, and then tphscale should get some benefit.
>
> I only ran up to four threads so these would be ready before I left. I'll be
> away for a few days.
>
>
> thpscale Fault Latencies
> thp thp thp thp
> base1 base2 lru1 lru2
> Min fault-base-1 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%)
> Min fault-base-3 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%)
> Min fault-base-4 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%)
> Min fault-huge-1 214.00 ( 0.00%) 227.00 ( -6.07%) 235.00 ( -9.81%) 240.00 ( -12.15%)
> Min fault-huge-3 321.00 ( 0.00%) 366.00 ( -14.02%) 323.00 ( -0.62%) 407.00 ( -26.79%)
> Min fault-huge-4 441.00 ( 0.00%) 401.00 ( 9.07%) 525.00 ( -19.05%) 434.00 ( 1.59%)
> Min fault-both-1 214.00 ( 0.00%) 227.00 ( -6.07%) 235.00 ( -9.81%) 240.00 ( -12.15%)
> Min fault-both-3 321.00 ( 0.00%) 366.00 ( -14.02%) 323.00 ( -0.62%) 407.00 ( -26.79%)
> Min fault-both-4 441.00 ( 0.00%) 401.00 ( 9.07%) 525.00 ( -19.05%) 434.00 ( 1.59%)
> Amean fault-base-1 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%)
> Amean fault-base-3 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%)
> Amean fault-base-4 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%)
> Amean fault-huge-1 1549.55 ( 0.00%) 1667.13 * -7.59%* 1361.50 * 12.14%* 1458.10 * 5.90%*
> Amean fault-huge-3 2582.29 ( 0.00%) 2556.45 ( 1.00%) 2756.65 * -6.75%* 2157.29 * 16.46%*
> Amean fault-huge-4 2352.21 ( 0.00%) 2709.37 * -15.18%* 2323.89 ( 1.20%) 1888.94 * 19.69%*
> Amean fault-both-1 1549.55 ( 0.00%) 1667.13 * -7.59%* 1361.50 * 12.14%* 1458.10 * 5.90%*
> Amean fault-both-3 2582.29 ( 0.00%) 2556.45 ( 1.00%) 2756.65 * -6.75%* 2157.29 * 16.46%*
> Amean fault-both-4 2352.21 ( 0.00%) 2709.37 * -15.18%* 2323.89 ( 1.20%) 1888.94 * 19.69%*
> Stddev fault-base-1 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%)
> Stddev fault-base-3 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%)
> Stddev fault-base-4 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%)
> Stddev fault-huge-1 2011.48 ( 0.00%) 3765.64 ( -87.21%) 3061.45 ( -52.20%) 1758.56 ( 12.57%)
> Stddev fault-huge-3 11153.49 ( 0.00%) 8339.83 ( 25.23%) 7017.28 ( 37.08%) 3976.86 ( 64.34%)
> Stddev fault-huge-4 8817.67 ( 0.00%) 16241.48 ( -84.19%) 11595.28 ( -31.50%) 6631.47 ( 24.79%)
> Stddev fault-both-1 2011.48 ( 0.00%) 3765.64 ( -87.21%) 3061.45 ( -52.20%) 1758.56 ( 12.57%)
> Stddev fault-both-3 11153.49 ( 0.00%) 8339.83 ( 25.23%) 7017.28 ( 37.08%) 3976.86 ( 64.34%)
> Stddev fault-both-4 8817.67 ( 0.00%) 16241.48 ( -84.19%) 11595.28 ( -31.50%) 6631.47 ( 24.79%)
> CoeffVar fault-base-1 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%)
> CoeffVar fault-base-3 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%)
> CoeffVar fault-base-4 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%)
> CoeffVar fault-huge-1 129.81 ( 0.00%) 225.88 ( -74.00%) 224.86 ( -73.22%) 120.61 ( 7.09%)
> CoeffVar fault-huge-3 431.92 ( 0.00%) 326.23 ( 24.47%) 254.56 ( 41.06%) 184.35 ( 57.32%)
> CoeffVar fault-huge-4 374.87 ( 0.00%) 599.46 ( -59.91%) 498.96 ( -33.10%) 351.07 ( 6.35%)
> CoeffVar fault-both-1 129.81 ( 0.00%) 225.88 ( -74.00%) 224.86 ( -73.22%) 120.61 ( 7.09%)
> CoeffVar fault-both-3 431.92 ( 0.00%) 326.23 ( 24.47%) 254.56 ( 41.06%) 184.35 ( 57.32%)
> CoeffVar fault-both-4 374.87 ( 0.00%) 599.46 ( -59.91%) 498.96 ( -33.10%) 351.07 ( 6.35%)
> Max fault-base-1 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%)
> Max fault-base-3 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%)
> Max fault-base-4 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%)
> Max fault-huge-1 90549.00 ( 0.00%) 166062.00 ( -83.39%) 134242.00 ( -48.25%) 82424.00 ( 8.97%)
> Max fault-huge-3 1229237.00 ( 0.00%) 392796.00 ( 68.05%) 508290.00 ( 58.65%) 433992.00 ( 64.69%)
> Max fault-huge-4 418414.00 ( 0.00%) 1147308.00 (-174.20%) 792153.00 ( -89.32%) 433393.00 ( -3.58%)
> Max fault-both-1 90549.00 ( 0.00%) 166062.00 ( -83.39%) 134242.00 ( -48.25%) 82424.00 ( 8.97%)
> Max fault-both-3 1229237.00 ( 0.00%) 392796.00 ( 68.05%) 508290.00 ( 58.65%) 433992.00 ( 64.69%)
> Max fault-both-4 418414.00 ( 0.00%) 1147308.00 (-174.20%) 792153.00 ( -89.32%) 433393.00 ( -3.58%)
> BAmean-50 fault-base-1 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%)
> BAmean-50 fault-base-3 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%)
> BAmean-50 fault-base-4 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%)
> BAmean-50 fault-huge-1 1329.84 ( 0.00%) 1086.23 ( 18.32%) 765.68 ( 42.42%) 1238.41 ( 6.88%)
> BAmean-50 fault-huge-3 1410.46 ( 0.00%) 1406.43 ( 0.29%) 1749.95 ( -24.07%) 1362.23 ( 3.42%)
> BAmean-50 fault-huge-4 1453.01 ( 0.00%) 1373.01 ( 5.51%) 1268.61 ( 12.69%) 1157.08 ( 20.37%)
> BAmean-50 fault-both-1 1329.84 ( 0.00%) 1086.23 ( 18.32%) 765.68 ( 42.42%) 1238.41 ( 6.88%)
> BAmean-50 fault-both-3 1410.46 ( 0.00%) 1406.43 ( 0.29%) 1749.95 ( -24.07%) 1362.23 ( 3.42%)
> BAmean-50 fault-both-4 1453.01 ( 0.00%) 1373.01 ( 5.51%) 1268.61 ( 12.69%) 1157.08 ( 20.37%)
> BAmean-95 fault-base-1 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%)
> BAmean-95 fault-base-3 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%)
> BAmean-95 fault-base-4 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%)
> BAmean-95 fault-huge-1 1428.93 ( 0.00%) 1336.45 ( 6.47%) 1125.72 ( 21.22%) 1324.67 ( 7.30%)
> BAmean-95 fault-huge-3 1881.39 ( 0.00%) 1868.23 ( 0.70%) 2257.90 ( -20.01%) 1794.46 ( 4.62%)
> BAmean-95 fault-huge-4 1853.49 ( 0.00%) 2038.27 ( -9.97%) 1758.37 ( 5.13%) 1522.69 ( 17.85%)
> BAmean-95 fault-both-1 1428.93 ( 0.00%) 1336.45 ( 6.47%) 1125.72 ( 21.22%) 1324.67 ( 7.30%)
> BAmean-95 fault-both-3 1881.39 ( 0.00%) 1868.23 ( 0.70%) 2257.90 ( -20.01%) 1794.46 ( 4.62%)
> BAmean-95 fault-both-4 1853.49 ( 0.00%) 2038.27 ( -9.97%) 1758.37 ( 5.13%) 1522.69 ( 17.85%)
> BAmean-99 fault-base-1 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%)
> BAmean-99 fault-base-3 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%)
> BAmean-99 fault-base-4 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%) 0.00 ( 0.00%)
> BAmean-99 fault-huge-1 1447.04 ( 0.00%) 1394.38 ( 3.64%) 1188.40 ( 17.87%) 1359.40 ( 6.06%)
> BAmean-99 fault-huge-3 2139.13 ( 0.00%) 2127.28 ( 0.55%) 2500.44 ( -16.89%) 2024.05 ( 5.38%)
> BAmean-99 fault-huge-4 2051.01 ( 0.00%) 2141.53 ( -4.41%) 1959.27 ( 4.47%) 1705.47 ( 16.85%)
> BAmean-99 fault-both-1 1447.04 ( 0.00%) 1394.38 ( 3.64%) 1188.40 ( 17.87%) 1359.40 ( 6.06%)
> BAmean-99 fault-both-3 2139.13 ( 0.00%) 2127.28 ( 0.55%) 2500.44 ( -16.89%) 2024.05 ( 5.38%)
> BAmean-99 fault-both-4 2051.01 ( 0.00%) 2141.53 ( -4.41%) 1959.27 ( 4.47%) 1705.47 ( 16.85%)
>
> thpscale Percentage Faults Huge
> thp thp thp thp
> base1 base2 lru1 lru2
> Percentage huge-1 100.00 ( 0.00%) 100.00 ( 0.00%) 100.00 ( 0.00%) 100.00 ( 0.00%)
> Percentage huge-3 100.00 ( 0.00%) 100.00 ( 0.00%) 100.00 ( 0.00%) 100.00 ( 0.00%)
> Percentage huge-4 100.00 ( 0.00%) 100.00 ( 0.00%) 100.00 ( 0.00%) 100.00 ( 0.00%)
>
> thp thp thp thp
> base1 base2 lru1 lru2
> Duration User 93.79 93.26 93.61 97.02
> Duration System 592.52 601.32 585.31 575.20
> Duration Elapsed 1239.40 1243.79 1232.97 1229.09
>
> thp thp thp thp
> base1 base2 lru1 lru2
> Ops Minor Faults 2532126.00 2377121.00 2410222.00 2414384.00
> Ops Major Faults 836.00 722.00 778.00 610.00
> Ops Swap Ins 676.00 644.00 656.00 760.00
> Ops Swap Outs 15887.00 15705.00 11341.00 6812.00
> Ops Allocation stalls 67986.00 44391.00 48791.00 70025.00
> Ops Fragmentation stalls 0.00 0.00 0.00 0.00
> Ops DMA allocs 0.00 0.00 0.00 0.00
> Ops DMA32 allocs 0.00 0.00 0.00 0.00
> Ops Normal allocs 163203689.00 154651154.00 159354937.00 165237378.00
> Ops Movable allocs 0.00 0.00 0.00 0.00
> Ops Direct pages scanned 832666.00 1385214.00 652686.00 499004.00
> Ops Kswapd pages scanned 808301.00 3517190.00 2310658.00 615450.00
> Ops Kswapd pages reclaimed 28737.00 314122.00 13761.00 21149.00
> Ops Direct pages reclaimed 1542.00 118664.00 278.00 5002.00
> Ops Kswapd efficiency % 3.56 8.93 0.60 3.44
> Ops Kswapd velocity 652.17 2827.80 1874.06 500.74
> Ops Direct efficiency % 0.19 8.57 0.04 1.00
> Ops Direct velocity 671.83 1113.70 529.36 405.99
> Ops Percentage direct scans 50.74 28.26 22.03 44.78
> Ops Page writes by reclaim 15887.00 15705.00 11341.00 6812.00
> Ops Page writes file 0.00 0.00 0.00 0.00
> Ops Page writes anon 15887.00 15705.00 11341.00 6812.00
> Ops Page reclaim immediate 0.00 0.00 0.00 0.00
> Ops Sector Reads 79350969.00 79290528.00 79326339.00 79294800.00
> Ops Sector Writes 105724641.00 105722825.00 105705814.00 105687161.00
> Ops Page rescued immediate 0.00 0.00 0.00 0.00
> Ops Slabs scanned 83935.00 113839.00 100286.00 78442.00
> Ops Direct inode steals 56.00 44.00 86.00 31.00
> Ops Kswapd inode steals 0.00 1.00 2.00 2.00
> Ops Kswapd skipped wait 0.00 0.00 0.00 0.00
> Ops THP fault alloc 232051.00 232036.00 232035.00 232035.00
> Ops THP fault fallback 1.00 0.00 0.00 0.00
> Ops THP collapse alloc 13.00 12.00 15.00 11.00
> Ops THP collapse fail 0.00 0.00 0.00 0.00
> Ops THP split 22665.00 52360.00 35592.00 15363.00
> Ops THP split failed 0.00 0.00 0.00 0.00
> Ops Compaction stalls 144127.00 120784.00 125512.00 144578.00
> Ops Compaction success 8039.00 24423.00 15316.00 3672.00
> Ops Compaction failures 136088.00 96361.00 110196.00 140906.00
> Ops Compaction efficiency 5.58 20.22 12.20 2.54
> Ops Page migrate success 4170185.00 12177428.00 7860810.00 3037053.00
> Ops Page migrate failure 9476.00 9987.00 1987.00 623.00
> Ops Compaction pages isolated 9551164.00 25547480.00 16290756.00 6222921.00
> Ops Compaction migrate scanned 6529450.00 15982527.00 9845328.00 5711509.00
> Ops Compaction free scanned 17198448.00 21663579.00 9531167.00 10790527.00
> Ops Compact scan efficiency 37.97 73.78 103.30 52.93
> Ops Compaction cost 4555.96 13237.59 8537.99 3310.69
> Ops Kcompactd wake 51.00 88.00 48.00 20.00
> Ops Kcompactd migrate scanned 127333.00 325654.00 144656.00 670374.00
> Ops Kcompactd free scanned 100028.00 35658.00 12486.00 127376.00
> Ops NUMA alloc hit 48379083.00 48250045.00 48278578.00 48259808.00
> Ops NUMA alloc miss 68141.00 51994.00 61465.00 70881.00
> Ops NUMA interleave hit 0.00 0.00 0.00 0.00
> Ops NUMA alloc local 48379030.00 48250007.00 48278535.00 48259769.00
> Ops NUMA base-page range updates 19832.00 4340.00 14959.00 5190.00
> Ops NUMA PTE updates 2424.00 756.00 1647.00 1094.00
> Ops NUMA PMD updates 34.00 7.00 26.00 8.00
> Ops NUMA hint faults 2148.00 386.00 1411.00 415.00
> Ops NUMA hint local faults % 1814.00 362.00 1401.00 238.00
> Ops NUMA hint local percent 84.45 93.78 99.29 57.35
> Ops NUMA pages migrated 5392.00 535.00 1030.00 1710.00
> Ops AutoNUMA cost 10.98 1.97 7.18 2.14
>
prev parent reply other threads:[~2020-09-25 0:03 UTC|newest]
Thread overview: 31+ messages / expand[flat|nested] mbox.gz Atom feed top
2020-09-24 3:28 [PATCH v19 00/20] per memcg lru_lock Alex Shi
2020-09-24 3:28 ` [PATCH v19 01/20] mm/memcg: warning on !memcg after readahead page charged Alex Shi
2020-09-24 3:28 ` [PATCH v19 02/20] mm/memcg: bail early from swap accounting if memcg disabled Alex Shi
2020-09-24 3:28 ` [PATCH v19 03/20] mm/thp: move lru_add_page_tail func to huge_memory.c Alex Shi
2020-09-26 6:06 ` Alex Shi
2020-09-24 3:28 ` [PATCH v19 04/20] mm/thp: use head for head page in lru_add_page_tail Alex Shi
2020-09-24 3:28 ` [PATCH v19 05/20] mm/thp: Simplify lru_add_page_tail() Alex Shi
2020-09-24 3:28 ` [PATCH v19 06/20] mm/thp: narrow lru locking Alex Shi
2020-09-26 6:08 ` Alex Shi
2020-09-24 3:28 ` [PATCH v19 07/20] mm/vmscan: remove unnecessary lruvec adding Alex Shi
2020-09-26 6:14 ` Alex Shi
2020-09-24 3:28 ` [PATCH v19 08/20] mm: page_idle_get_page() does not need lru_lock Alex Shi
2020-09-24 3:28 ` [PATCH v19 09/20] mm/memcg: add debug checking in lock_page_memcg Alex Shi
2020-09-24 3:28 ` [PATCH v19 10/20] mm/swap.c: fold vm event PGROTATED into pagevec_move_tail_fn Alex Shi
2020-09-24 3:28 ` [PATCH v19 11/20] mm/lru: move lock into lru_note_cost Alex Shi
2020-09-24 3:28 ` [PATCH v19 12/20] mm/vmscan: remove lruvec reget in move_pages_to_lru Alex Shi
2020-09-24 3:28 ` [PATCH v19 13/20] mm/mlock: remove lru_lock on TestClearPageMlocked Alex Shi
2020-09-24 3:28 ` [PATCH v19 14/20] mm/mlock: remove __munlock_isolate_lru_page Alex Shi
2020-09-24 3:28 ` [PATCH v19 15/20] mm/lru: introduce TestClearPageLRU Alex Shi
2020-09-24 3:28 ` [PATCH v19 16/20] mm/compaction: do page isolation first in compaction Alex Shi
2020-09-24 3:28 ` [PATCH v19 17/20] mm/swap.c: serialize memcg changes in pagevec_lru_move_fn Alex Shi
2020-09-24 3:28 ` [PATCH v19 18/20] mm/lru: replace pgdat lru_lock with lruvec lock Alex Shi
2020-10-21 7:23 ` Alex Shi
2020-10-23 2:00 ` Rong Chen
2020-10-25 21:51 ` Hugh Dickins
2020-10-26 1:41 ` Alex Shi
2020-10-27 0:36 ` Rong Chen
2020-09-24 3:28 ` [PATCH v19 19/20] mm/lru: introduce the relock_page_lruvec function Alex Shi
2020-09-24 3:28 ` [PATCH v19 20/20] mm/lru: revise the comments of lru_lock Alex Shi
2020-09-24 18:06 ` [PATCH v19 00/20] per memcg lru_lock Daniel Jordan
2020-09-25 0:01 ` Alex Shi [this message]
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=c633b107-845a-1c70-b8bb-5d4163bfa274@linux.alibaba.com \
--to=alex.shi@linux.alibaba.com \
--cc=aaron.lwe@gmail.com \
--cc=akpm@linux-foundation.org \
--cc=alexander.duyck@gmail.com \
--cc=cgroups@vger.kernel.org \
--cc=daniel.m.jordan@oracle.com \
--cc=hannes@cmpxchg.org \
--cc=hughd@google.com \
--cc=iamjoonsoo.kim@lge.com \
--cc=khlebnikov@yandex-team.ru \
--cc=kirill@shutemov.name \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=lkp@intel.com \
--cc=mgorman@techsingularity.net \
--cc=mhocko@suse.com \
--cc=richard.weiyang@gmail.com \
--cc=rong.a.chen@intel.com \
--cc=shakeelb@google.com \
--cc=shy828301@gmail.com \
--cc=tj@kernel.org \
--cc=vdavydov.dev@gmail.com \
--cc=willy@infradead.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).