[lxc-users] CentOS 6.3 kernel-2.6.32-279.el6.x86_64 crash
Shibashish
shib4u at gmail.com
Sat May 3 14:10:19 UTC 2014
Hi,
My server with 4 lxc VM is kernel panicking often. On analyzing the crash
dump, it shows the following. I have 4 VMs, with 3G memory each,
memory+swap set at 4G in the croup settings.
lxc version: 0.9.0.alpha2
I did a hardware swap, but the problem persists. Please let me know what to
do next. Should I upgrade kernel to 2.6.32-431.11.2.el6.centos.plus ?
KERNEL: /usr/lib/debug/lib/modules/2.6.32-279.el6.x86_64/vmlinux
DUMPFILE: /var/crash/127.0.0.1-2014-05-03-06:11:01/vmcore [PARTIAL
DUMP]
CPUS: 8
DATE: Sat May 3 06:09:12 2014
UPTIME: 07:56:53
LOAD AVERAGE: 0.07, 0.06, 0.01
TASKS: 651
NODENAME: myhost
RELEASE: 2.6.32-279.el6.x86_64
VERSION: #1 SMP Fri Jun 22 12:19:21 UTC 2012
MACHINE: x86_64 (1866 Mhz)
MEMORY: 16 GB
PANIC: "kernel BUG at mm/slab.c:533!"
PID: 0
COMMAND: "swapper"
TASK: ffff880426373540 (1 of 8) [THREAD_INFO: ffff880426374000]
CPU: 7
STATE: TASK_RUNNING (PANIC)
------------[ cut here ]------------
kernel BUG at mm/slab.c:533!
invalid opcode: 0000 [#1] SMP
last sysfs file: /sys/devices/system/cpu/online
CPU 7
Modules linked in: veth bridge stp llc ipv6 e1000e(U) sg microcode i2c_i801
iTCO_wdt iTCO_vendor_support i5000_edac edac_core i5k_amb ioatdma dca
shpchp ext3 jbd mbcache sd_mod crc_t10dif aacraid pata_acpi ata_generic
ata_piix radeon ttm drm_kms_helper drm i2c_algo_bit i2c_core dm_mirror
dm_region_hash dm_log dm_mod [last unloaded: scsi_wait_scan]
Pid: 0, comm: swapper Tainted: G I---------------
2.6.32-279.el6.x86_64 #1 Supermicro X7DVL/X7DVL
RIP: 0010:[<ffffffff81163f75>] [<ffffffff81163f75>] free_block+0x165/0x170
RSP: 0018:ffff8800283c32d0 EFLAGS: 00010046
RAX: ffffea0009fd5878 RBX: ffff88042fcf03c0 RCX: 0000000000000010
RDX: 0040000000000000 RSI: ffff8802bba2cec0 RDI: ffff8802daab9800
RBP: ffff8800283c3320 R08: ffffea0009d7b600 R09: 0000000000000000
R10: 0000000000000000 R11: 0000000080042000 R12: 000000000000000c
R13: ffff880426350aa8 R14: 0000000000000002 R15: ffffea0000000000
FS: 0000000000000000(0000) GS:ffff8800283c0000(0000) knlGS:0000000000000000
CS: 0010 DS: 0018 ES: 0018 CR0: 000000008005003b
CR2: 0000003fafe7b3f0 CR3: 00000004240a1000 CR4: 00000000000006e0
DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
Process swapper (pid: 0, threadinfo ffff880426374000, task ffff880426373540)
Stack:
ffff88042fc216c0 ffff8802bba2cec0 000000000000100c ffff8802daab9800
<d> ffff8800283c3360 ffff8802daabc800 ffff880426350a80 0000000000000286
<d> ffff88042fcf03c0 ffff880426350a98 ffff8800283c3390 ffffffff81164500
Call Trace:
<IRQ>
[<ffffffff81164500>] kfree+0x310/0x320
[<ffffffff8143c949>] ? enqueue_to_backlog+0x179/0x210
[<ffffffff8142fef8>] skb_release_data+0xd8/0x110
[<ffffffff8143c949>] ? enqueue_to_backlog+0x179/0x210
[<ffffffff8142fa2e>] __kfree_skb+0x1e/0xa0
[<ffffffff8142fb72>] kfree_skb+0x42/0x90
[<ffffffff8143c949>] enqueue_to_backlog+0x179/0x210
[<ffffffff8143fb20>] netif_rx+0xb0/0x160
[<ffffffff8143fe32>] dev_forward_skb+0x122/0x180
[<ffffffffa02396e6>] veth_xmit+0x86/0xe0 [veth]
[<ffffffff8143b0cc>] dev_hard_start_xmit+0x2bc/0x3f0
[<ffffffff81458c1a>] sch_direct_xmit+0x15a/0x1c0
[<ffffffff8143f878>] dev_queue_xmit+0x4f8/0x6f0
[<ffffffffa032c6bc>] br_dev_queue_push_xmit+0x6c/0xa0 [bridge]
[<ffffffffa0332378>] br_nf_dev_queue_xmit+0x28/0xa0 [bridge]
[<ffffffffa0332e10>] br_nf_post_routing+0x1d0/0x280 [bridge]
[<ffffffff814665e9>] nf_iterate+0x69/0xb0
[<ffffffffa032c650>] ? br_dev_queue_push_xmit+0x0/0xa0 [bridge]
[<ffffffff814667a4>] nf_hook_slow+0x74/0x110
[<ffffffffa032c650>] ? br_dev_queue_push_xmit+0x0/0xa0 [bridge]
[<ffffffffa032c6f0>] ? br_forward_finish+0x0/0x60 [bridge]
[<ffffffffa032c733>] br_forward_finish+0x43/0x60 [bridge]
[<ffffffffa03329b8>] br_nf_forward_finish+0x128/0x140 [bridge]
[<ffffffffa0333ea8>] ? br_nf_forward_ip+0x318/0x3c0 [bridge]
[<ffffffffa0333ea8>] br_nf_forward_ip+0x318/0x3c0 [bridge]
[<ffffffff814665e9>] nf_iterate+0x69/0xb0
[<ffffffffa032c6f0>] ? br_forward_finish+0x0/0x60 [bridge]
[<ffffffff814667a4>] nf_hook_slow+0x74/0x110
[<ffffffffa032c6f0>] ? br_forward_finish+0x0/0x60 [bridge]
[<ffffffffa032c750>] ? __br_forward+0x0/0xc0 [bridge]
[<ffffffffa032c7c2>] __br_forward+0x72/0xc0 [bridge]
[<ffffffffa032c601>] br_flood+0xc1/0xd0 [bridge]
[<ffffffffa032c625>] br_flood_forward+0x15/0x20 [bridge]
[<ffffffffa032d7ae>] br_handle_frame_finish+0x27e/0x2a0 [bridge]
[<ffffffffa0333318>] br_nf_pre_routing_finish+0x228/0x340 [bridge]
[<ffffffffa033388f>] br_nf_pre_routing+0x45f/0x760 [bridge]
[<ffffffff814665e9>] nf_iterate+0x69/0xb0
[<ffffffffa032d530>] ? br_handle_frame_finish+0x0/0x2a0 [bridge]
[<ffffffff814667a4>] nf_hook_slow+0x74/0x110
[<ffffffffa032d530>] ? br_handle_frame_finish+0x0/0x2a0 [bridge]
[<ffffffffa032d95c>] br_handle_frame+0x18c/0x250 [bridge]
[<ffffffff8143a839>] __netif_receive_skb+0x519/0x6f0
[<ffffffff8143ca38>] netif_receive_skb+0x58/0x60
[<ffffffff8143cbe4>] napi_gro_complete+0x84/0xe0
[<ffffffff8143ce0b>] dev_gro_receive+0x1cb/0x290
[<ffffffff8143cf4b>] __napi_gro_receive+0x7b/0x170
[<ffffffff8143f06f>] napi_gro_receive+0x2f/0x50
[<ffffffffa027733b>] e1000_receive_skb+0x5b/0x90 [e1000e]
[<ffffffffa027a601>] e1000_clean_rx_irq+0x241/0x4c0 [e1000e]
[<ffffffffa0281b8d>] e1000e_poll+0x8d/0x380 [e1000e]
[<ffffffff8143aaaa>] ? process_backlog+0x9a/0x100
[<ffffffff8143f193>] net_rx_action+0x103/0x2f0
[<ffffffff81073ec1>] __do_softirq+0xc1/0x1e0
[<ffffffff810db800>] ? handle_IRQ_event+0x60/0x170
[<ffffffff8100c24c>] call_softirq+0x1c/0x30
[<ffffffff8100de85>] do_softirq+0x65/0xa0
[<ffffffff81073ca5>] irq_exit+0x85/0x90
[<ffffffff81505af5>] do_IRQ+0x75/0xf0
[<ffffffff8100ba53>] ret_from_intr+0x0/0x11
<EOI>
[<ffffffff81014877>] ? mwait_idle+0x77/0xd0
[<ffffffff8150338a>] ? atomic_notifier_call_chain+0x1a/0x20
[<ffffffff81009e06>] cpu_idle+0xb6/0x110
[<ffffffff814f6cdf>] start_secondary+0x22a/0x26d
Code: 41 5c 41 5d 41 5e 41 5f c9 c3 0f 1f 40 00 48 8b 72 08 48 89 c7 e8 2c
f0 11 00 e9 07 ff ff ff 48 8b 40 10 48 8b 10 e9 3e ff ff ff <0f> 0b eb fe
0f 1f 80 00 00 00 00 55 48 89 e5 48 83 ec 30 48 89
RIP [<ffffffff81163f75>] free_block+0x165/0x170
RSP <ffff8800283c32d0>
ShiB.
while ( ! ( succeed = try() ) );
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.linuxcontainers.org/pipermail/lxc-users/attachments/20140503/3bd4bf30/attachment.html>
More information about the lxc-users
mailing list