[lxc-users] {Disarmed} [lxc-devel] CentOS 6.3 kernel-2.6.32-279.el6.x86_64 crash
Shibashish
shib4u at gmail.com
Wed May 7 16:41:21 UTC 2014
Upgraded lxc
lxc-libs-1.0.3-1.el6.x86_64
lxc-1.0.3-1.el6.x86_64
CentOS release 6.3 (Final)
uname -a
Linux myhost 2.6.32-279.el6.x86_64 #1 SMP Fri Jun 22 12:19:21 UTC 2012
x86_64 x86_64 x86_64 GNU/Linux
But the problem persists, have had couple of kernel panics.
------------[ cut here ]------------
kernel BUG at mm/slab.c:533!
invalid opcode: 0000 [#1] SMP
last sysfs file: /sys/devices/virtual/dmi/id/sys_vendor
CPU 0
Modules linked in: veth bridge stp llc ipv6 e1000e(U) sg microcode i2c_i801
iTCO_wdt iTCO_vendor_support shpchp i5000_edac edac_core i5k_amb ioatdma
dca ext3 jbd mbcache sd_mod crc_t10dif aacraid pata_acpi ata_generic
ata_piix radeon ttm drm_kms_helper drm i2c_algo_bit i2c_core dm_mirror
dm_region_hash dm_log dm_mod [last unloaded: scsi_wait_scan]
Pid: 0, comm: swapper Tainted: G I---------------
2.6.32-279.el6.x86_64 #1 Supermicro X7DVL/X7DVL
RIP: 0010:[<ffffffff81163f75>] [<ffffffff81163f75>] free_block+0x165/0x170
RSP: 0018:ffff8800282032d0 EFLAGS: 00010046
RAX: ffffea000a54e368 RBX: ffff88042fcf03c0 RCX: 0000000000000010
RDX: 0040000000000000 RSI: ffff8802f3bb6d40 RDI: ffff8802f3aeb000
RBP: ffff880028203320 R08: ffffea000e79b720 R09: 0000000000000000
R10: 0000000000000000 R11: 0000000080042000 R12: 000000000000000c
R13: ffff88042fea13a8 R14: 0000000000000002 R15: ffffea0000000000
FS: 0000000000000000(0000) GS:ffff880028200000(0000) knlGS:0000000000000000
CS: 0010 DS: 0018 ES: 0018 CR0: 000000008005003b
CR2: 00007fc077681000 CR3: 000000042216f000 CR4: 00000000000006f0
DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
Process swapper (pid: 0, threadinfo ffffffff81a00000, task ffffffff81a8d020)
Stack:
ffff88042fc216c0 ffff8802f3bb6d40 000000000000100c ffff8802f3aeb000
<d> ffff880028203360 ffff8802f3bc4000 ffff88042fea1380 0000000000000286
<d> ffff88042fcf03c0 ffff88042fea1398 ffff880028203390 ffffffff81164500
Call Trace:
<IRQ>
[<ffffffff81164500>] kfree+0x310/0x320
[<ffffffff8143c949>] ? enqueue_to_backlog+0x179/0x210
[<ffffffff8142fef8>] skb_release_data+0xd8/0x110
[<ffffffff8143c949>] ? enqueue_to_backlog+0x179/0x210
[<ffffffff8142fa2e>] __kfree_skb+0x1e/0xa0
[<ffffffff8142fb72>] kfree_skb+0x42/0x90
[<ffffffff8143c949>] enqueue_to_backlog+0x179/0x210
[<ffffffff8143fb20>] netif_rx+0xb0/0x160
[<ffffffff8143fe32>] dev_forward_skb+0x122/0x180
[<ffffffffa03446e6>] veth_xmit+0x86/0xe0 [veth]
[<ffffffff8143b0cc>] dev_hard_start_xmit+0x2bc/0x3f0
[<ffffffff81458c1a>] sch_direct_xmit+0x15a/0x1c0
[<ffffffff8143f878>] dev_queue_xmit+0x4f8/0x6f0
[<ffffffffa03276bc>] br_dev_queue_push_xmit+0x6c/0xa0 [bridge]
[<ffffffffa032d378>] br_nf_dev_queue_xmit+0x28/0xa0 [bridge]
[<ffffffffa032de10>] br_nf_post_routing+0x1d0/0x280 [bridge]
[<ffffffff814665e9>] nf_iterate+0x69/0xb0
[<ffffffffa0327650>] ? br_dev_queue_push_xmit+0x0/0xa0 [bridge]
[<ffffffff814667a4>] nf_hook_slow+0x74/0x110
[<ffffffffa0327650>] ? br_dev_queue_push_xmit+0x0/0xa0 [bridge]
[<ffffffffa03276f0>] ? br_forward_finish+0x0/0x60 [bridge]
[<ffffffffa0327733>] br_forward_finish+0x43/0x60 [bridge]
[<ffffffffa032d9b8>] br_nf_forward_finish+0x128/0x140 [bridge]
[<ffffffffa032eea8>] ? br_nf_forward_ip+0x318/0x3c0 [bridge]
[<ffffffffa032eea8>] br_nf_forward_ip+0x318/0x3c0 [bridge]
[<ffffffff814665e9>] nf_iterate+0x69/0xb0
[<ffffffffa03276f0>] ? br_forward_finish+0x0/0x60 [bridge]
[<ffffffff814667a4>] nf_hook_slow+0x74/0x110
[<ffffffffa03276f0>] ? br_forward_finish+0x0/0x60 [bridge]
[<ffffffffa0327750>] ? __br_forward+0x0/0xc0 [bridge]
[<ffffffffa03277c2>] __br_forward+0x72/0xc0 [bridge]
[<ffffffffa0327601>] br_flood+0xc1/0xd0 [bridge]
[<ffffffffa0327625>] br_flood_forward+0x15/0x20 [bridge]
[<ffffffffa03287ae>] br_handle_frame_finish+0x27e/0x2a0 [bridge]
[<ffffffffa032e318>] br_nf_pre_routing_finish+0x228/0x340 [bridge]
[<ffffffffa032e88f>] br_nf_pre_routing+0x45f/0x760 [bridge]
[<ffffffff814665e9>] nf_iterate+0x69/0xb0
[<ffffffffa0328530>] ? br_handle_frame_finish+0x0/0x2a0 [bridge]
[<ffffffff814667a4>] nf_hook_slow+0x74/0x110
[<ffffffffa0328530>] ? br_handle_frame_finish+0x0/0x2a0 [bridge]
[<ffffffffa032895c>] br_handle_frame+0x18c/0x250 [bridge]
[<ffffffff8143a839>] __netif_receive_skb+0x519/0x6f0
[<ffffffff8143ca38>] netif_receive_skb+0x58/0x60
[<ffffffff8143cbe4>] napi_gro_complete+0x84/0xe0
[<ffffffff8143ce0b>] dev_gro_receive+0x1cb/0x290
[<ffffffff8143cf4b>] __napi_gro_receive+0x7b/0x170
[<ffffffff8143f06f>] napi_gro_receive+0x2f/0x50
[<ffffffffa027233b>] e1000_receive_skb+0x5b/0x90 [e1000e]
[<ffffffffa0275601>] e1000_clean_rx_irq+0x241/0x4c0 [e1000e]
[<ffffffffa027cb8d>] e1000e_poll+0x8d/0x380 [e1000e]
[<ffffffff8143aaaa>] ? process_backlog+0x9a/0x100
[<ffffffff8143f193>] net_rx_action+0x103/0x2f0
[<ffffffff81073ec1>] __do_softirq+0xc1/0x1e0
[<ffffffff810db800>] ? handle_IRQ_event+0x60/0x170
[<ffffffff8100c24c>] call_softirq+0x1c/0x30
[<ffffffff8100de85>] do_softirq+0x65/0xa0
[<ffffffff81073ca5>] irq_exit+0x85/0x90
[<ffffffff81505af5>] do_IRQ+0x75/0xf0
[<ffffffff8100ba53>] ret_from_intr+0x0/0x11
<EOI>
[<ffffffff81014877>] ? mwait_idle+0x77/0xd0
[<ffffffff8150338a>] ? atomic_notifier_call_chain+0x1a/0x20
[<ffffffff81009e06>] cpu_idle+0xb6/0x110
[<ffffffff814e433a>] rest_init+0x7a/0x80
[<ffffffff81c21f7b>] start_kernel+0x424/0x430
[<ffffffff81c2133a>] x86_64_start_reservations+0x125/0x129
[<ffffffff81c21438>] x86_64_start_kernel+0xfa/0x109
Code: 41 5c 41 5d 41 5e 41 5f c9 c3 0f 1f 40 00 48 8b 72 08 48 89 c7 e8 2c
f0 11 00 e9 07 ff ff ff 48 8b 40 10 48 8b 10 e9 3e ff ff ff <0f> 0b eb fe
0f 1f 80 00 00 00 00 55 48 89 e5 48 83 ec 30 48 89
RIP [<ffffffff81163f75>] free_block+0x165/0x170
RSP <ffff8800282032d0>
ShiB.
while ( ! ( succeed = try() ) );
On Sat, May 3, 2014 at 8:14 PM, Michael H. Warfield <mhw at wittsend.com>wrote:
> On Sat, 2014-05-03 at 19:40 +0530, Shibashish wrote:
> > Hi,
>
> > My server with 4 lxc VM is kernel panicking often. On analyzing the
> > crash dump, it shows the following. I have 4 VMs, with 3G memory each,
> > memory+swap set at 4G in the croup settings.
>
> > lxc version: 0.9.0.alpha2
>
> You definitely need to upgrade that version of LXC.
>
> 1) It's old.
> 2) It's an alpha version.
> >
>
> > I did a hardware swap, but the problem persists. Please let me know
> > what to do next. Should I upgrade kernel
> > to 2.6.32-431.11.2.el6.centos.plus ?
>
> I would most definitely update the entire system including and
> especially the kernel. Nothing that LXC does should cause a kernel
> panic.
> >
> > KERNEL: /usr/lib/debug/lib/modules/2.6.32-279.el6.x86_64/vmlinux
> > DUMPFILE: /var/crash/MailScanner has detected a possible fraud
> > attempt from "127.0.0" claiming to be MailScanner warning: numerical
> > links are often malicious: 127.0.0.1-2014-05-03-06:11:01/vmcore
> > [PARTIAL DUMP]
> > CPUS: 8
> > DATE: Sat May 3 06:09:12 2014
> > UPTIME: 07:56:53
> > LOAD AVERAGE: 0.07, 0.06, 0.01
> > TASKS: 651
> > NODENAME: myhost
> > RELEASE: 2.6.32-279.el6.x86_64
> > VERSION: #1 SMP Fri Jun 22 12:19:21 UTC 2012
> > MACHINE: x86_64 (1866 Mhz)
> > MEMORY: 16 GB
> > PANIC: "kernel BUG at mm/slab.c:533!"
> > PID: 0
> > COMMAND: "swapper"
> > TASK: ffff880426373540 (1 of 8) [THREAD_INFO:
> > ffff880426374000]
> > CPU: 7
> > STATE: TASK_RUNNING (PANIC)
> >
> >
> I'm not even sure how this relates to LXC. From what I'm seeing below,
> I do see functions in veth and br_*, which could be from an LXC
> container, so the fault is failing somewhere down through the bridging
> code and into e1000 NIC driver and interrupt handlers. That's a kernel
> fault of some sort, and really deep. Definitely upgrade that kernel.
> >
> > ------------[ cut here ]------------
> > kernel BUG at mm/slab.c:533!
> > invalid opcode: 0000 [#1] SMP
> > last sysfs file: /sys/devices/system/cpu/online
> > CPU 7
> > Modules linked in: veth bridge stp llc ipv6 e1000e(U) sg microcode
> > i2c_i801 iTCO_wdt iTCO_vendor_support i5000_edac edac_core i5k_amb
> > ioatdma dca shpchp ext3 jbd mbcache sd_mod crc_t10dif aacraid
> > pata_acpi ata_generic ata_piix radeon ttm drm_kms_helper drm
> > i2c_algo_bit i2c_core dm_mirror dm_region_hash dm_log dm_mod [last
> > unloaded: scsi_wait_scan]
> >
> >
> > Pid: 0, comm: swapper Tainted: G I---------------
> > 2.6.32-279.el6.x86_64 #1 Supermicro X7DVL/X7DVL
> > RIP: 0010:[<ffffffff81163f75>] [<ffffffff81163f75>] free_block
> > +0x165/0x170
> > RSP: 0018:ffff8800283c32d0 EFLAGS: 00010046
> > RAX: ffffea0009fd5878 RBX: ffff88042fcf03c0 RCX: 0000000000000010
> > RDX: 0040000000000000 RSI: ffff8802bba2cec0 RDI: ffff8802daab9800
> > RBP: ffff8800283c3320 R08: ffffea0009d7b600 R09: 0000000000000000
> > R10: 0000000000000000 R11: 0000000080042000 R12: 000000000000000c
> > R13: ffff880426350aa8 R14: 0000000000000002 R15: ffffea0000000000
> > FS: 0000000000000000(0000) GS:ffff8800283c0000(0000)
> > knlGS:0000000000000000
> > CS: 0010 DS: 0018 ES: 0018 CR0: 000000008005003b
> > CR2: 0000003fafe7b3f0 CR3: 00000004240a1000 CR4: 00000000000006e0
> > DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
> > DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
> > Process swapper (pid: 0, threadinfo ffff880426374000, task
> > ffff880426373540)
> > Stack:
> > ffff88042fc216c0 ffff8802bba2cec0 000000000000100c ffff8802daab9800
> > <d> ffff8800283c3360 ffff8802daabc800 ffff880426350a80
> > 0000000000000286
> > <d> ffff88042fcf03c0 ffff880426350a98 ffff8800283c3390
> > ffffffff81164500
> > Call Trace:
> > <IRQ>
> > [<ffffffff81164500>] kfree+0x310/0x320
> > [<ffffffff8143c949>] ? enqueue_to_backlog+0x179/0x210
> > [<ffffffff8142fef8>] skb_release_data+0xd8/0x110
> > [<ffffffff8143c949>] ? enqueue_to_backlog+0x179/0x210
> > [<ffffffff8142fa2e>] __kfree_skb+0x1e/0xa0
> > [<ffffffff8142fb72>] kfree_skb+0x42/0x90
> > [<ffffffff8143c949>] enqueue_to_backlog+0x179/0x210
> > [<ffffffff8143fb20>] netif_rx+0xb0/0x160
> > [<ffffffff8143fe32>] dev_forward_skb+0x122/0x180
> > [<ffffffffa02396e6>] veth_xmit+0x86/0xe0 [veth]
> > [<ffffffff8143b0cc>] dev_hard_start_xmit+0x2bc/0x3f0
> > [<ffffffff81458c1a>] sch_direct_xmit+0x15a/0x1c0
> > [<ffffffff8143f878>] dev_queue_xmit+0x4f8/0x6f0
> > [<ffffffffa032c6bc>] br_dev_queue_push_xmit+0x6c/0xa0 [bridge]
> > [<ffffffffa0332378>] br_nf_dev_queue_xmit+0x28/0xa0 [bridge]
> > [<ffffffffa0332e10>] br_nf_post_routing+0x1d0/0x280 [bridge]
> > [<ffffffff814665e9>] nf_iterate+0x69/0xb0
> > [<ffffffffa032c650>] ? br_dev_queue_push_xmit+0x0/0xa0 [bridge]
> > [<ffffffff814667a4>] nf_hook_slow+0x74/0x110
> > [<ffffffffa032c650>] ? br_dev_queue_push_xmit+0x0/0xa0 [bridge]
> > [<ffffffffa032c6f0>] ? br_forward_finish+0x0/0x60 [bridge]
> > [<ffffffffa032c733>] br_forward_finish+0x43/0x60 [bridge]
> > [<ffffffffa03329b8>] br_nf_forward_finish+0x128/0x140 [bridge]
> > [<ffffffffa0333ea8>] ? br_nf_forward_ip+0x318/0x3c0 [bridge]
> > [<ffffffffa0333ea8>] br_nf_forward_ip+0x318/0x3c0 [bridge]
> > [<ffffffff814665e9>] nf_iterate+0x69/0xb0
> > [<ffffffffa032c6f0>] ? br_forward_finish+0x0/0x60 [bridge]
> > [<ffffffff814667a4>] nf_hook_slow+0x74/0x110
> > [<ffffffffa032c6f0>] ? br_forward_finish+0x0/0x60 [bridge]
> > [<ffffffffa032c750>] ? __br_forward+0x0/0xc0 [bridge]
> > [<ffffffffa032c7c2>] __br_forward+0x72/0xc0 [bridge]
> > [<ffffffffa032c601>] br_flood+0xc1/0xd0 [bridge]
> > [<ffffffffa032c625>] br_flood_forward+0x15/0x20 [bridge]
> > [<ffffffffa032d7ae>] br_handle_frame_finish+0x27e/0x2a0 [bridge]
> > [<ffffffffa0333318>] br_nf_pre_routing_finish+0x228/0x340 [bridge]
> > [<ffffffffa033388f>] br_nf_pre_routing+0x45f/0x760 [bridge]
> > [<ffffffff814665e9>] nf_iterate+0x69/0xb0
> > [<ffffffffa032d530>] ? br_handle_frame_finish+0x0/0x2a0 [bridge]
> > [<ffffffff814667a4>] nf_hook_slow+0x74/0x110
> > [<ffffffffa032d530>] ? br_handle_frame_finish+0x0/0x2a0 [bridge]
> > [<ffffffffa032d95c>] br_handle_frame+0x18c/0x250 [bridge]
> > [<ffffffff8143a839>] __netif_receive_skb+0x519/0x6f0
> > [<ffffffff8143ca38>] netif_receive_skb+0x58/0x60
> > [<ffffffff8143cbe4>] napi_gro_complete+0x84/0xe0
> > [<ffffffff8143ce0b>] dev_gro_receive+0x1cb/0x290
> > [<ffffffff8143cf4b>] __napi_gro_receive+0x7b/0x170
> > [<ffffffff8143f06f>] napi_gro_receive+0x2f/0x50
> > [<ffffffffa027733b>] e1000_receive_skb+0x5b/0x90 [e1000e]
> > [<ffffffffa027a601>] e1000_clean_rx_irq+0x241/0x4c0 [e1000e]
> > [<ffffffffa0281b8d>] e1000e_poll+0x8d/0x380 [e1000e]
> > [<ffffffff8143aaaa>] ? process_backlog+0x9a/0x100
> > [<ffffffff8143f193>] net_rx_action+0x103/0x2f0
> > [<ffffffff81073ec1>] __do_softirq+0xc1/0x1e0
> > [<ffffffff810db800>] ? handle_IRQ_event+0x60/0x170
> > [<ffffffff8100c24c>] call_softirq+0x1c/0x30
> > [<ffffffff8100de85>] do_softirq+0x65/0xa0
> > [<ffffffff81073ca5>] irq_exit+0x85/0x90
> > [<ffffffff81505af5>] do_IRQ+0x75/0xf0
> > [<ffffffff8100ba53>] ret_from_intr+0x0/0x11
> > <EOI>
> > [<ffffffff81014877>] ? mwait_idle+0x77/0xd0
> > [<ffffffff8150338a>] ? atomic_notifier_call_chain+0x1a/0x20
> > [<ffffffff81009e06>] cpu_idle+0xb6/0x110
> > [<ffffffff814f6cdf>] start_secondary+0x22a/0x26d
> > Code: 41 5c 41 5d 41 5e 41 5f c9 c3 0f 1f 40 00 48 8b 72 08 48 89 c7
> > e8 2c f0 11 00 e9 07 ff ff ff 48 8b 40 10 48 8b 10 e9 3e ff ff ff <0f>
> > 0b eb fe 0f 1f 80 00 00 00 00 55 48 89 e5 48 83 ec 30 48 89
> > RIP [<ffffffff81163f75>] free_block+0x165/0x170
> > RSP <ffff8800283c32d0>
> >
> >
> >
> >
> > ShiB.
> > while ( ! ( succeed = try() ) );
> >
> >
>
> Regards,
> Mike
> --
> Michael H. Warfield (AI4NB) | (770) 978-7061 | mhw at WittsEnd.com
> /\/\|=mhw=|\/\/ | (678) 463-0932 |
> http://www.wittsend.com/mhw/
> NIC whois: MHW9 | An optimist believes we live in the best of
> all
> PGP Key: 0x674627FF | possible worlds. A pessimist is sure of it!
>
>
> _______________________________________________
> lxc-users mailing list
> lxc-users at lists.linuxcontainers.org
> http://lists.linuxcontainers.org/listinfo/lxc-users
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.linuxcontainers.org/pipermail/lxc-users/attachments/20140507/3a5332d1/attachment-0001.html>
More information about the lxc-users
mailing list