[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] Kernel 3.11 / 3.12 OOM killer and Xen ballooning
On Tue, Dec 10, 2013 at 02:52:40PM +0000, James Dingwall wrote: > Konrad Rzeszutek Wilk wrote: > >On Mon, Dec 09, 2013 at 05:50:29PM +0000, James Dingwall wrote: > >>Hi, > >> > >>Since 3.11 I have noticed that the OOM killer quite frequently > >>triggers in my Xen guest domains which use ballooning to > >>increase/decrease their memory allocation according to their > >>requirements. One example domain I have has a maximum memory > >>setting of ~1.5Gb but it usually idles at ~300Mb, it is also > >>configured with 2Gb swap which is almost 100% free. > >> > >># free > >> total used free shared buffers cached > >>Mem: 272080 248108 23972 0 1448 63064 > >>-/+ buffers/cache: 183596 88484 > >>Swap: 2097148 8 2097140 > >> > >>There is plenty of available free memory in the hypervisor to > >>balloon to the maximum size: > >># xl info | grep free_mem > >>free_memory : 14923 > >> > >>An example trace (they are always the same) from the oom killer in > >>3.12 is added below. So far I have not been able to reproduce this > >>at will so it is difficult to start bisecting it to see if a > >>particular change introduced this. However it does seem that the > >>behaviour is wrong because a) ballooning could give the guest more > >>memory, b) there is lots of swap available which could be used as a > >>fallback. Keep in mind that swap with tmem is actually no more swap. Heh, that sounds odd -but basically pages that are destined for swap end up going in the tmem code which pipes them up to the hypervisor. > >> > >>If other information could help or there are more tests that I could > >>run then please let me know. > >I presume you have enabled 'tmem' both in the hypervisor and in > >the guest right? > Yes, domU and dom0 both have the tmem module loaded and tmem > tmem_dedup=on tmem_compress=on is given on the xen command line. Excellent. The odd thing is that your swap is not used that much, but it should be (as that is part of what the self-balloon is suppose to do). Bob, you had a patch for the logic of how self-balloon is suppose to account for the slab - would this be relevant to this problem? Thank you. > >>Thanks, > >>James > >> > >> > >> > >> > >>[473233.777271] emerge invoked oom-killer: gfp_mask=0x280da, > >>order=0, oom_score_adj=0 > >>[473233.777279] CPU: 0 PID: 22159 Comm: emerge Tainted: G W 3.12.0 #80 > >>[473233.777282] ffff88000599f6f8 ffff8800117bda58 ffffffff81489a80 > >>ffff88004760e8e8 > >>[473233.777286] ffff88000599f1c0 ffff8800117bdaf8 ffffffff81487577 > >>ffff8800117bdaa8 > >>[473233.777289] ffffffff810f8c0f ffff8800117bda88 ffffffff81006dc8 > >>ffff8800117bda98 > >>[473233.777293] Call Trace: > >>[473233.777305] [<ffffffff81489a80>] dump_stack+0x46/0x58 > >>[473233.777310] [<ffffffff81487577>] dump_header.isra.9+0x6d/0x1cc > >>[473233.777315] [<ffffffff810f8c0f>] ? super_cache_count+0xa8/0xb8 > >>[473233.777321] [<ffffffff81006dc8>] ? xen_clocksource_read+0x20/0x22 > >>[473233.777324] [<ffffffff81006ea9>] ? xen_clocksource_get_cycles+0x9/0xb > >>[473233.777328] [<ffffffff8148f336>] ? > >>_raw_spin_unlock_irqrestore+0x47/0x62 > >>[473233.777333] [<ffffffff812915d3>] ? ___ratelimit+0xcb/0xe8 > >>[473233.777338] [<ffffffff810b2aa7>] oom_kill_process+0x70/0x2fd > >>[473233.777343] [<ffffffff81048775>] ? has_ns_capability_noaudit+0x12/0x19 > >>[473233.777346] [<ffffffff8104878e>] ? has_capability_noaudit+0x12/0x14 > >>[473233.777349] [<ffffffff810b31c6>] out_of_memory+0x31b/0x34e > >>[473233.777353] [<ffffffff810b72f0>] __alloc_pages_nodemask+0x65b/0x792 > >>[473233.777358] [<ffffffff810e3c1b>] alloc_pages_vma+0xd0/0x10c > >>[473233.777361] [<ffffffff81003f69>] ? > >>__raw_callee_save_xen_pmd_val+0x11/0x1e > >>[473233.777365] [<ffffffff810cf685>] handle_mm_fault+0x6d4/0xd54 > >>[473233.777371] [<ffffffff81037f40>] __do_page_fault+0x3d8/0x437 > >>[473233.777374] [<ffffffff81006dc8>] ? xen_clocksource_read+0x20/0x22 > >>[473233.777378] [<ffffffff810115d2>] ? sched_clock+0x9/0xd > >>[473233.777382] [<ffffffff810676c7>] ? sched_clock_local+0x12/0x75 > >>[473233.777386] [<ffffffff810a44b4>] ? __acct_update_integrals+0xb4/0xbf > >>[473233.777389] [<ffffffff810a4827>] ? acct_account_cputime+0x17/0x19 > >>[473233.777392] [<ffffffff81067bc0>] ? account_user_time+0x67/0x92 > >>[473233.777395] [<ffffffff810680b3>] ? vtime_account_user+0x4d/0x52 > >>[473233.777398] [<ffffffff81037fd8>] do_page_fault+0x1a/0x5a > >>[473233.777401] [<ffffffff8148f9d8>] page_fault+0x28/0x30 > >>[473233.777403] Mem-Info: > >>[473233.777405] Node 0 DMA per-cpu: > >>[473233.777408] CPU 0: hi: 0, btch: 1 usd: 0 > >>[473233.777409] CPU 1: hi: 0, btch: 1 usd: 0 > >>[473233.777411] CPU 2: hi: 0, btch: 1 usd: 0 > >>[473233.777412] CPU 3: hi: 0, btch: 1 usd: 0 > >>[473233.777413] Node 0 DMA32 per-cpu: > >>[473233.777415] CPU 0: hi: 186, btch: 31 usd: 103 > >>[473233.777417] CPU 1: hi: 186, btch: 31 usd: 110 > >>[473233.777419] CPU 2: hi: 186, btch: 31 usd: 175 > >>[473233.777420] CPU 3: hi: 186, btch: 31 usd: 182 > >>[473233.777421] Node 0 Normal per-cpu: > >>[473233.777423] CPU 0: hi: 0, btch: 1 usd: 0 > >>[473233.777424] CPU 1: hi: 0, btch: 1 usd: 0 > >>[473233.777426] CPU 2: hi: 0, btch: 1 usd: 0 > >>[473233.777427] CPU 3: hi: 0, btch: 1 usd: 0 > >>[473233.777433] active_anon:35740 inactive_anon:33812 isolated_anon:0 > >> active_file:4672 inactive_file:11607 isolated_file:0 > >> unevictable:0 dirty:4 writeback:0 unstable:0 > >> free:2067 slab_reclaimable:3583 slab_unreclaimable:3524 > >> mapped:3329 shmem:324 pagetables:2003 bounce:0 > >> free_cma:0 > >>[473233.777435] Node 0 DMA free:4200kB min:60kB low:72kB high:88kB > >>active_anon:264kB inactive_anon:456kB active_file:140kB > >>inactive_file:340kB unevictable:0kB isolated(anon):0kB > >>isolated(file):0kB present:15996kB managed:6176kB mlocked:0kB > >>dirty:0kB writeback:0kB mapped:100kB shmem:0kB slab_reclaimable:96kB > >>slab_unreclaimable:112kB kernel_stack:24kB pagetables:24kB > >>unstable:0kB bounce:0kB free_cma:0kB writeback_tmp:0kB > >>pages_scanned:33270 all_unreclaimable? yes > >>[473233.777443] lowmem_reserve[]: 0 1036 1036 1036 > >>[473233.777447] Node 0 DMA32 free:4060kB min:4084kB low:5104kB > >>high:6124kB active_anon:41256kB inactive_anon:33128kB > >>active_file:8544kB inactive_file:14312kB unevictable:0kB > >>isolated(anon):0kB isolated(file):0kB present:1163264kB > >>managed:165780kB mlocked:0kB dirty:0kB writeback:0kB mapped:6428kB > >>shmem:604kB slab_reclaimable:9800kB slab_unreclaimable:12908kB > >>kernel_stack:1832kB pagetables:5924kB unstable:0kB bounce:0kB > >>free_cma:0kB writeback_tmp:0kB pages_scanned:152386 > >>all_unreclaimable? yes > >>[473233.777454] lowmem_reserve[]: 0 0 0 0 > >>[473233.777457] Node 0 Normal free:8kB min:0kB low:0kB high:0kB > >>active_anon:101440kB inactive_anon:101664kB active_file:10004kB > >>inactive_file:31776kB unevictable:0kB isolated(anon):0kB > >>isolated(file):0kB present:393216kB managed:256412kB mlocked:0kB > >>dirty:16kB writeback:0kB mapped:6788kB shmem:692kB > >>slab_reclaimable:4436kB slab_unreclaimable:1076kB kernel_stack:136kB > >>pagetables:2064kB unstable:0kB bounce:0kB free_cma:0kB > >>writeback_tmp:0kB pages_scanned:368809 all_unreclaimable? yes > >>[473233.777464] lowmem_reserve[]: 0 0 0 0 > >>[473233.777467] Node 0 DMA: 41*4kB (U) 0*8kB 0*16kB 0*32kB 1*64kB > >>(R) 1*128kB (R) 1*256kB (R) 1*512kB (R) 1*1024kB (R) 1*2048kB (R) > >>0*4096kB = 4196kB > >>[473233.777480] Node 0 DMA32: 1015*4kB (U) 0*8kB 0*16kB 0*32kB > >>0*64kB 0*128kB 0*256kB 0*512kB 0*1024kB 0*2048kB 0*4096kB = 4060kB > >>[473233.777490] Node 0 Normal: 0*4kB 0*8kB 0*16kB 0*32kB 0*64kB > >>0*128kB 0*256kB 0*512kB 0*1024kB 0*2048kB 0*4096kB = 0kB > >>[473233.777498] 5018 total pagecache pages > >>[473233.777500] 16 pages in swap cache > >>[473233.777501] Swap cache stats: add 2829330, delete 2829314, find > >>344059/481859 > >>[473233.777503] Free swap = 2096980kB > >>[473233.777503] Total swap = 2097148kB > >>[473233.794497] 557055 pages RAM > >>[473233.794500] 189326 pages reserved > >>[473233.794501] 544934 pages shared > >>[473233.794502] 358441 pages non-shared > >>[473233.794504] [ pid ] uid tgid total_vm rss nr_ptes > >>swapents oom_score_adj name > >>[473233.794523] [ 6597] 0 6597 8156 252 20 0 > >>-1000 udevd > >>[473233.794530] [ 7194] 0 7194 2232 137 10 0 0 metalog > >>[473233.794534] [ 7195] 0 7195 2223 31 10 3 0 metalog > >>[473233.794537] [ 7211] 0 7211 1064 35 8 0 > >>0 acpid > >>[473233.794546] [ 7227] 702 7227 4922 183 14 0 0 > >>dbus-daemon > >>[473233.794553] [ 7427] 0 7427 13630 179 29 15 0 rpcbind > >>[473233.794560] [ 7442] 0 7442 14743 332 32 0 0 > >>rpc.statd > >>[473233.794569] [ 7472] 0 7472 6365 115 17 0 0 > >>rpc.idmapd > >>[473233.794576] [ 7488] 0 7488 43602 349 40 0 0 cupsd > >>[473233.794583] [ 7512] 0 7512 14856 243 30 0 0 > >>rpc.mountd > >>[473233.794592] [ 7552] 0 7552 148819 940 68 0 0 > >>automount > >>[473233.794595] [ 7592] 0 7592 16006 233 32 0 -1000 sshd > >>[473233.794598] [ 7608] 0 7608 87672 2257 128 6 > >>0 apache2 > >>[473233.794601] [ 7633] 0 7633 521873 631 56 0 0 > >>console-kit-dae > >>[473233.794604] [ 7713] 106 7713 15453 295 34 2 0 nrpe > >>[473233.794607] [ 7719] 986 7719 91303 798 41 0 0 polkitd > >>[473233.794610] [ 7757] 123 7757 7330 259 17 0 0 ntpd > >>[473233.794613] [ 7845] 0 7845 3583 94 12 0 0 master > >>[473233.794616] [ 7847] 207 7847 17745 311 38 0 0 qmgr > >>[473233.794619] [ 7861] 65534 7861 2101 21 9 19 > >>0 rwhod > >>[473233.794622] [ 7864] 65534 7864 2101 99 9 0 > >>0 rwhod > >>[473233.794625] [ 7876] 0 7876 48582 533 47 19 0 smbd > >>[473233.794628] [ 7881] 0 7881 44277 372 38 0 0 nmbd > >>[473233.794631] [ 7895] 0 7895 48646 621 45 18 0 smbd > >>[473233.794634] [ 7902] 2 7902 1078 39 8 4 0 slpd > >>[473233.794637] [ 7917] 0 7917 38452 1073 28 1 0 snmpd > >>[473233.794640] [ 7945] 0 7945 27552 58 9 0 0 cron > >>[473233.794648] [ 7993] 0 7993 201378 5432 63 39 0 nscd > >>[473233.794658] [ 8064] 0 8064 1060 28 7 0 > >>0 agetty > >>[473233.794664] [ 8065] 0 8065 26507 29 9 0 > >>0 agetty > >>[473233.794667] [ 8066] 0 8066 26507 29 9 0 > >>0 agetty > >>[473233.794670] [ 8067] 0 8067 26507 28 9 0 > >>0 agetty > >>[473233.794673] [ 8068] 0 8068 26507 28 8 0 > >>0 agetty > >>[473233.794678] [ 8069] 0 8069 26507 30 9 0 > >>0 agetty > >>[473233.794686] [ 8070] 0 8070 26507 30 9 0 > >>0 agetty > >>[473233.794693] [ 8071] 0 8071 26507 30 9 0 > >>0 agetty > >>[473233.794701] [ 8072] 0 8072 26507 28 9 0 > >>0 agetty > >>[473233.794708] [ 8316] 0 8316 3736 83 11 6 0 > >>ssh-agent > >>[473233.794712] [ 8341] 0 8341 3390 66 12 7 0 > >>gpg-agent > >>[473233.794716] [ 2878] 81 2878 88431 2552 121 5 > >>0 apache2 > >>[473233.794718] [ 2879] 81 2879 88431 2552 121 5 > >>0 apache2 > >>[473233.794721] [ 2880] 81 2880 88431 2552 121 5 > >>0 apache2 > >>[473233.794724] [ 2881] 81 2881 88431 2552 121 5 > >>0 apache2 > >>[473233.794727] [ 2882] 81 2882 88431 2552 121 5 > >>0 apache2 > >>[473233.794734] [ 3523] 81 3523 88431 2552 121 5 > >>0 apache2 > >>[473233.794737] [30259] 1000 30259 3736 118 11 0 0 > >>ssh-agent > >>[473233.794741] [30284] 1000 30284 3390 141 12 0 0 > >>gpg-agent > >>[473233.794745] [21263] 207 21263 17703 771 39 1 0 pickup > >>[473233.794748] [21663] 0 21663 30743 228 16 0 0 cron > >>[473233.794751] [21665] 0 21665 2980 392 12 0 0 > >>gentoosync.sh > >>[473233.794755] [22158] 0 22158 3181 273 12 0 0 sendmail > >>[473233.794757] [22159] 0 22159 77646 54920 158 0 > >>0 emerge > >>[473233.794760] [22160] 0 22160 1068 85 8 0 0 tail > >>[473233.794764] [22161] 0 22161 3173 277 11 0 0 postdrop > >>[473233.794768] Out of memory: Kill process 22159 (emerge) score 57 > >>or sacrifice child > >>[473233.794771] Killed process 22159 (emerge) total-vm:310584kB, > >>anon-rss:215840kB, file-rss:3840kB > >> > >> > >>_______________________________________________ > >>Xen-devel mailing list > >>Xen-devel@xxxxxxxxxxxxx > >>http://lists.xen.org/xen-devel _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |