[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] Kernel 3.11 / 3.12 OOM killer and Xen ballooning
On Mon, Dec 09, 2013 at 05:50:29PM +0000, James Dingwall wrote: > Hi, > > Since 3.11 I have noticed that the OOM killer quite frequently > triggers in my Xen guest domains which use ballooning to > increase/decrease their memory allocation according to their > requirements. One example domain I have has a maximum memory > setting of ~1.5Gb but it usually idles at ~300Mb, it is also > configured with 2Gb swap which is almost 100% free. > > # free > total used free shared buffers cached > Mem: 272080 248108 23972 0 1448 63064 > -/+ buffers/cache: 183596 88484 > Swap: 2097148 8 2097140 > > There is plenty of available free memory in the hypervisor to > balloon to the maximum size: > # xl info | grep free_mem > free_memory : 14923 > > An example trace (they are always the same) from the oom killer in > 3.12 is added below. So far I have not been able to reproduce this > at will so it is difficult to start bisecting it to see if a > particular change introduced this. However it does seem that the > behaviour is wrong because a) ballooning could give the guest more > memory, b) there is lots of swap available which could be used as a > fallback. > > If other information could help or there are more tests that I could > run then please let me know. I presume you have enabled 'tmem' both in the hypervisor and in the guest right? > > Thanks, > James > > > > > [473233.777271] emerge invoked oom-killer: gfp_mask=0x280da, > order=0, oom_score_adj=0 > [473233.777279] CPU: 0 PID: 22159 Comm: emerge Tainted: G W 3.12.0 #80 > [473233.777282] ffff88000599f6f8 ffff8800117bda58 ffffffff81489a80 > ffff88004760e8e8 > [473233.777286] ffff88000599f1c0 ffff8800117bdaf8 ffffffff81487577 > ffff8800117bdaa8 > [473233.777289] ffffffff810f8c0f ffff8800117bda88 ffffffff81006dc8 > ffff8800117bda98 > [473233.777293] Call Trace: > [473233.777305] [<ffffffff81489a80>] dump_stack+0x46/0x58 > [473233.777310] [<ffffffff81487577>] dump_header.isra.9+0x6d/0x1cc > [473233.777315] [<ffffffff810f8c0f>] ? super_cache_count+0xa8/0xb8 > [473233.777321] [<ffffffff81006dc8>] ? xen_clocksource_read+0x20/0x22 > [473233.777324] [<ffffffff81006ea9>] ? xen_clocksource_get_cycles+0x9/0xb > [473233.777328] [<ffffffff8148f336>] ? > _raw_spin_unlock_irqrestore+0x47/0x62 > [473233.777333] [<ffffffff812915d3>] ? ___ratelimit+0xcb/0xe8 > [473233.777338] [<ffffffff810b2aa7>] oom_kill_process+0x70/0x2fd > [473233.777343] [<ffffffff81048775>] ? has_ns_capability_noaudit+0x12/0x19 > [473233.777346] [<ffffffff8104878e>] ? has_capability_noaudit+0x12/0x14 > [473233.777349] [<ffffffff810b31c6>] out_of_memory+0x31b/0x34e > [473233.777353] [<ffffffff810b72f0>] __alloc_pages_nodemask+0x65b/0x792 > [473233.777358] [<ffffffff810e3c1b>] alloc_pages_vma+0xd0/0x10c > [473233.777361] [<ffffffff81003f69>] ? > __raw_callee_save_xen_pmd_val+0x11/0x1e > [473233.777365] [<ffffffff810cf685>] handle_mm_fault+0x6d4/0xd54 > [473233.777371] [<ffffffff81037f40>] __do_page_fault+0x3d8/0x437 > [473233.777374] [<ffffffff81006dc8>] ? xen_clocksource_read+0x20/0x22 > [473233.777378] [<ffffffff810115d2>] ? sched_clock+0x9/0xd > [473233.777382] [<ffffffff810676c7>] ? sched_clock_local+0x12/0x75 > [473233.777386] [<ffffffff810a44b4>] ? __acct_update_integrals+0xb4/0xbf > [473233.777389] [<ffffffff810a4827>] ? acct_account_cputime+0x17/0x19 > [473233.777392] [<ffffffff81067bc0>] ? account_user_time+0x67/0x92 > [473233.777395] [<ffffffff810680b3>] ? vtime_account_user+0x4d/0x52 > [473233.777398] [<ffffffff81037fd8>] do_page_fault+0x1a/0x5a > [473233.777401] [<ffffffff8148f9d8>] page_fault+0x28/0x30 > [473233.777403] Mem-Info: > [473233.777405] Node 0 DMA per-cpu: > [473233.777408] CPU 0: hi: 0, btch: 1 usd: 0 > [473233.777409] CPU 1: hi: 0, btch: 1 usd: 0 > [473233.777411] CPU 2: hi: 0, btch: 1 usd: 0 > [473233.777412] CPU 3: hi: 0, btch: 1 usd: 0 > [473233.777413] Node 0 DMA32 per-cpu: > [473233.777415] CPU 0: hi: 186, btch: 31 usd: 103 > [473233.777417] CPU 1: hi: 186, btch: 31 usd: 110 > [473233.777419] CPU 2: hi: 186, btch: 31 usd: 175 > [473233.777420] CPU 3: hi: 186, btch: 31 usd: 182 > [473233.777421] Node 0 Normal per-cpu: > [473233.777423] CPU 0: hi: 0, btch: 1 usd: 0 > [473233.777424] CPU 1: hi: 0, btch: 1 usd: 0 > [473233.777426] CPU 2: hi: 0, btch: 1 usd: 0 > [473233.777427] CPU 3: hi: 0, btch: 1 usd: 0 > [473233.777433] active_anon:35740 inactive_anon:33812 isolated_anon:0 > active_file:4672 inactive_file:11607 isolated_file:0 > unevictable:0 dirty:4 writeback:0 unstable:0 > free:2067 slab_reclaimable:3583 slab_unreclaimable:3524 > mapped:3329 shmem:324 pagetables:2003 bounce:0 > free_cma:0 > [473233.777435] Node 0 DMA free:4200kB min:60kB low:72kB high:88kB > active_anon:264kB inactive_anon:456kB active_file:140kB > inactive_file:340kB unevictable:0kB isolated(anon):0kB > isolated(file):0kB present:15996kB managed:6176kB mlocked:0kB > dirty:0kB writeback:0kB mapped:100kB shmem:0kB slab_reclaimable:96kB > slab_unreclaimable:112kB kernel_stack:24kB pagetables:24kB > unstable:0kB bounce:0kB free_cma:0kB writeback_tmp:0kB > pages_scanned:33270 all_unreclaimable? yes > [473233.777443] lowmem_reserve[]: 0 1036 1036 1036 > [473233.777447] Node 0 DMA32 free:4060kB min:4084kB low:5104kB > high:6124kB active_anon:41256kB inactive_anon:33128kB > active_file:8544kB inactive_file:14312kB unevictable:0kB > isolated(anon):0kB isolated(file):0kB present:1163264kB > managed:165780kB mlocked:0kB dirty:0kB writeback:0kB mapped:6428kB > shmem:604kB slab_reclaimable:9800kB slab_unreclaimable:12908kB > kernel_stack:1832kB pagetables:5924kB unstable:0kB bounce:0kB > free_cma:0kB writeback_tmp:0kB pages_scanned:152386 > all_unreclaimable? yes > [473233.777454] lowmem_reserve[]: 0 0 0 0 > [473233.777457] Node 0 Normal free:8kB min:0kB low:0kB high:0kB > active_anon:101440kB inactive_anon:101664kB active_file:10004kB > inactive_file:31776kB unevictable:0kB isolated(anon):0kB > isolated(file):0kB present:393216kB managed:256412kB mlocked:0kB > dirty:16kB writeback:0kB mapped:6788kB shmem:692kB > slab_reclaimable:4436kB slab_unreclaimable:1076kB kernel_stack:136kB > pagetables:2064kB unstable:0kB bounce:0kB free_cma:0kB > writeback_tmp:0kB pages_scanned:368809 all_unreclaimable? yes > [473233.777464] lowmem_reserve[]: 0 0 0 0 > [473233.777467] Node 0 DMA: 41*4kB (U) 0*8kB 0*16kB 0*32kB 1*64kB > (R) 1*128kB (R) 1*256kB (R) 1*512kB (R) 1*1024kB (R) 1*2048kB (R) > 0*4096kB = 4196kB > [473233.777480] Node 0 DMA32: 1015*4kB (U) 0*8kB 0*16kB 0*32kB > 0*64kB 0*128kB 0*256kB 0*512kB 0*1024kB 0*2048kB 0*4096kB = 4060kB > [473233.777490] Node 0 Normal: 0*4kB 0*8kB 0*16kB 0*32kB 0*64kB > 0*128kB 0*256kB 0*512kB 0*1024kB 0*2048kB 0*4096kB = 0kB > [473233.777498] 5018 total pagecache pages > [473233.777500] 16 pages in swap cache > [473233.777501] Swap cache stats: add 2829330, delete 2829314, find > 344059/481859 > [473233.777503] Free swap = 2096980kB > [473233.777503] Total swap = 2097148kB > [473233.794497] 557055 pages RAM > [473233.794500] 189326 pages reserved > [473233.794501] 544934 pages shared > [473233.794502] 358441 pages non-shared > [473233.794504] [ pid ] uid tgid total_vm rss nr_ptes > swapents oom_score_adj name > [473233.794523] [ 6597] 0 6597 8156 252 20 0 > -1000 udevd > [473233.794530] [ 7194] 0 7194 2232 137 10 0 0 metalog > [473233.794534] [ 7195] 0 7195 2223 31 10 3 0 metalog > [473233.794537] [ 7211] 0 7211 1064 35 8 0 > 0 acpid > [473233.794546] [ 7227] 702 7227 4922 183 14 0 0 > dbus-daemon > [473233.794553] [ 7427] 0 7427 13630 179 29 15 0 rpcbind > [473233.794560] [ 7442] 0 7442 14743 332 32 0 0 > rpc.statd > [473233.794569] [ 7472] 0 7472 6365 115 17 0 0 > rpc.idmapd > [473233.794576] [ 7488] 0 7488 43602 349 40 0 0 cupsd > [473233.794583] [ 7512] 0 7512 14856 243 30 0 0 > rpc.mountd > [473233.794592] [ 7552] 0 7552 148819 940 68 0 0 > automount > [473233.794595] [ 7592] 0 7592 16006 233 32 0 -1000 sshd > [473233.794598] [ 7608] 0 7608 87672 2257 128 6 > 0 apache2 > [473233.794601] [ 7633] 0 7633 521873 631 56 0 0 > console-kit-dae > [473233.794604] [ 7713] 106 7713 15453 295 34 2 0 nrpe > [473233.794607] [ 7719] 986 7719 91303 798 41 0 0 polkitd > [473233.794610] [ 7757] 123 7757 7330 259 17 0 0 ntpd > [473233.794613] [ 7845] 0 7845 3583 94 12 0 0 master > [473233.794616] [ 7847] 207 7847 17745 311 38 0 0 qmgr > [473233.794619] [ 7861] 65534 7861 2101 21 9 19 > 0 rwhod > [473233.794622] [ 7864] 65534 7864 2101 99 9 0 > 0 rwhod > [473233.794625] [ 7876] 0 7876 48582 533 47 19 0 smbd > [473233.794628] [ 7881] 0 7881 44277 372 38 0 0 nmbd > [473233.794631] [ 7895] 0 7895 48646 621 45 18 0 smbd > [473233.794634] [ 7902] 2 7902 1078 39 8 4 0 slpd > [473233.794637] [ 7917] 0 7917 38452 1073 28 1 0 snmpd > [473233.794640] [ 7945] 0 7945 27552 58 9 0 0 cron > [473233.794648] [ 7993] 0 7993 201378 5432 63 39 0 nscd > [473233.794658] [ 8064] 0 8064 1060 28 7 0 > 0 agetty > [473233.794664] [ 8065] 0 8065 26507 29 9 0 > 0 agetty > [473233.794667] [ 8066] 0 8066 26507 29 9 0 > 0 agetty > [473233.794670] [ 8067] 0 8067 26507 28 9 0 > 0 agetty > [473233.794673] [ 8068] 0 8068 26507 28 8 0 > 0 agetty > [473233.794678] [ 8069] 0 8069 26507 30 9 0 > 0 agetty > [473233.794686] [ 8070] 0 8070 26507 30 9 0 > 0 agetty > [473233.794693] [ 8071] 0 8071 26507 30 9 0 > 0 agetty > [473233.794701] [ 8072] 0 8072 26507 28 9 0 > 0 agetty > [473233.794708] [ 8316] 0 8316 3736 83 11 6 0 > ssh-agent > [473233.794712] [ 8341] 0 8341 3390 66 12 7 0 > gpg-agent > [473233.794716] [ 2878] 81 2878 88431 2552 121 5 > 0 apache2 > [473233.794718] [ 2879] 81 2879 88431 2552 121 5 > 0 apache2 > [473233.794721] [ 2880] 81 2880 88431 2552 121 5 > 0 apache2 > [473233.794724] [ 2881] 81 2881 88431 2552 121 5 > 0 apache2 > [473233.794727] [ 2882] 81 2882 88431 2552 121 5 > 0 apache2 > [473233.794734] [ 3523] 81 3523 88431 2552 121 5 > 0 apache2 > [473233.794737] [30259] 1000 30259 3736 118 11 0 0 > ssh-agent > [473233.794741] [30284] 1000 30284 3390 141 12 0 0 > gpg-agent > [473233.794745] [21263] 207 21263 17703 771 39 1 0 pickup > [473233.794748] [21663] 0 21663 30743 228 16 0 0 cron > [473233.794751] [21665] 0 21665 2980 392 12 0 0 > gentoosync.sh > [473233.794755] [22158] 0 22158 3181 273 12 0 0 sendmail > [473233.794757] [22159] 0 22159 77646 54920 158 0 > 0 emerge > [473233.794760] [22160] 0 22160 1068 85 8 0 0 tail > [473233.794764] [22161] 0 22161 3173 277 11 0 0 postdrop > [473233.794768] Out of memory: Kill process 22159 (emerge) score 57 > or sacrifice child > [473233.794771] Killed process 22159 (emerge) total-vm:310584kB, > anon-rss:215840kB, file-rss:3840kB > > > _______________________________________________ > Xen-devel mailing list > Xen-devel@xxxxxxxxxxxxx > http://lists.xen.org/xen-devel _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |