You are here
ZZRabbit - Fri, 2022/05/06 - 12:37
Hi
Fresh install of Lamp 17.0, Install security updates, reboot and kernel panic. boots old kernel but not the update.
Forum:
Hi
Fresh install of Lamp 17.0, Install security updates, reboot and kernel panic. boots old kernel but not the update.
BUMP
BUMP
I can't be the only one that has had this problem. I have tried on 3 machines Proxmox VE, Virtualbox and bare metal. Every time the same Kernel Panic on fresh install with security updates.
I can't reproduce?!
What platform(s) are you guys running on? I see you note Proxmox Ian, I assume v7?[update: re-reading I see these questions are already answered]. I've tested on Proxmox v6 (pve-qemu-kvm=5.2.0-6) and can't reproduce?FWIW the recent kernel update was a Debian security update, so not something specific to TurnKey. It was released nearly a week ago and covered a number of CVEs.
I guess that there is a chance that some default TurnKey config (that isn't default Debian config) is a compounding factor. But I can't think of anything off the top of my head, and we're certainly not doing anything radical.
I just did a clean install of a v17.0 TKLDev from ISO to KVM (Proxmox v6.x). I installed all security updates on firstboot and rebooted. Everything went as expected and I definately have the new kernel installed and loaded:
I have googled around and can't find any other specific reports related to this kernel. My suspicion is that it may be a regression in a specific kernel driver that you guys are using but I'm not. It still doesn't explain why there aren't more reports of this though.
A little more info... (but still no answer...)
I have done a heap more digging and have so far been unable to find any other reports of kernel panics related to this kernel update?! There are no bugs on the Debian bugtracker specifically related to this kernel. I've also checked the bugs assigned to the kernel meta-package and trolled through all the bugs assigned to the kernel source package and can't find any others specifically related to this kernel update.
Other than official Debian sources, the only thing I could find online specific to this kernel update is news about it e.g. here.
Recap...
Apologies on posting stream of consciousness...
Ok, so I've re-read through the posts here and realised there were some details I have overlooked. So to recap:
I assume from your messages Ian, that you experienced the exact same message on both PVE platforms?
ZZRabbit, I have a few questions for you:
Although on face value it sounds like you may have hit the same issue, we should keep in mind that it is possible that they're different issues (that both stem from the updated kernel).
After more searching online, common causes of the particular kernel panic noted by Ian are:
There may be other causes of that kernel panic message, because I'm certainly no kernel expert. But this certainly seems like a strange one...
Ok I can reproduce on VirtualBox
Ok I can reproduce the issue on VirtualBox.
I'll do a bug report to Debian. For now I can only suggest that you keep running on the old kernel.
[update] I have done a bug report upstream to Debian. Hopefully someone might come through with some ideas?!
TBH, as I've already noted, I'm not sure what more we can do for now other than keep running the old kernel.
FWIW I also have Debian running on my laptop (where VirtualBox is installed; used to recreate the issue) and that is running this latest kernel too. So it must be some side case issue.
Just chiming in to say I
Just chiming in to say I experienced this with the Syncthing 17.0 image a couple days ago, tried 3x to install a VM on Proxmox 7.x, tried with both i440fx and q35 "chipset" but not with EFI yet.
The panic also happens if you install security updates manually with e.g. 'apt dist-upgrade' or your favorite command. Which would make sense if it's being caused by the update, it wouldn't matter whether you installed the updates as part of the TKL initial setup scripts, or deferred it til later like I experimented with doing.
no workaround
After digging a bit deeper it appears that the issue occurs when /boot is inside the LVM. I don't know, but I wonder if this is a kernel regression, in that it can no longer find LVM volumes in early boot?
UPDATE: Scratch all that. I re-ran through these steps again and still get a kernel panic! :( I don't understand why it worked the first time, but not the second. There's obviously something I don't understand going on, perhaps something additional I did that I forgot to note?
I have tested moving /boot into a separate ext4 partition and it seems to boot ok now with the new kernel.
This is a non-trivial operation so I don't recommend trying this on a production server. I have got it to work, but it's not for the faint-hearted. It's also clearly not a great fix so we'll keep looking for something better. In the meantime, follow along if you want to try this yourself:
First add a new drive to your VM. It doesn't have to be huge, but unless you clean up old kernels it will fill up over time. So I suggest 200MB minimum, but 500MB or larger is probably better if free space isn't an object. You may need to stop or reboot your VM to see (or perhaps even add) your new disk (reboot with the old kernel still).
I'll assume that the new disk is /dev/sdb but it may be something different for you.
Press 'n' to make a new partition and accept all the defaults (it will make single partition that spans all the free space). Then press 'a' to make partition '1' bootable. Finally 'w' to write and exit.
Now create the filesystem:
Make a temporary mount point and mount it. Then copy across the contents of /boot:
Now get the UUID of /boot and write it to the fstab:
Now test your fstab:
Check to see /boot is mounted:
You should get something like:
Now I reconfigure the new kernel (just to be on the safe side), install grub to the new disk and also run 'update-grub' for good measure (update-grub probably isn't required).
Now poweroff your VM and ensure that it boots from the new drive (in VirtualBox I had to delete both then re-add them, the new little disk first, then the original disk - on Proxmox you can likely just rearrange the boot order). Reboot and fingers crossed, it will now boot from the new disk with the new kernel.Not Creating Initramfs
The update is not creating the Initramfs. Even if you try and manually it still won't do it.
Keeps deferring update.
Right you are! But not the only issue unfortunately...
Wow I hadn't noticed that, but you are right! All the servers that work have '/boot/initrd.img-5.10.0-14-amd64'. The ones that don't, only have '/boot/initrd.img-5.10.0-13-amd64', with '/boot/initrd.img-5.10.0-14-amd64' missing.
I'm not sure what is wrong with update-initramfs but I reinstalled it and then it would create the initrd:
Then I manually generated the initrd:
So now I do have the initrd, but still - the same kernel panic persists... :(
Update-Grub Works
I noticed that after reinstall initramfs-tools and manually initrd worked, but I checked the grub.cfg and noticed:
linux /boot/vmlinuz-5.10.0-14-amd64 root=/dev/sda1 ro net.ifnames=0 consoleblank=0
linux /boot/vmlinuz-5.10.0-13-amd64 root=UUID=aa513ef5-9aa8-4e8c-ab69-c8ff5e6a802e ro net.ifnames=0 consoleblank=0
ran update-grub and it worked
linux /boot/vmlinuz-5.10.0-14-amd64 root=UUID=aa513ef5-9aa8-4e8c-ab69-c8ff5e6a802e ro net.ifnames=0 consoleblank=0
linux /boot/vmlinuz-5.10.0-13-amd64 root=UUID=aa513ef5-9aa8-4e8c-ab69-c8ff5e6a802e ro net.ifnames=0 consoleblank=0
Hey there, just FYI, the
Hey there, just FYI, the kernel panic was happening on my VMs without LVM set up.
In my case I was creating two drives in Proxmox, both with 'manual' partitioning in the installer, no LVM.
The first drive had a swap partition and /, both were 'primary' partitions.
The second drive had /home as a 'primary' partition.
Thanks everyone, I think we have a work around now!
Thanks everyone for all the info that has been shared. The bug reports and feedback have been invaluable. I am now confident that I have a handle on what the issue is and a vague idea where to start looking to fix the bug (FYI it's in our build process).
Existing v17.0 servers that are affected can work around the issue like this:
As noted by others, there may be other issues, but in all the tests I did (with a clean install of LAMP) the above was all that was required. So that is the minimum required fix.
FYI the dpkg-reconfigure line will both create the initrd and update grub.
I'll share more when I get a chance. Until then, if you haven't already, please share exactly which appliance(s) you have experienced this bug with and/or what other specific issues you had with which appliance(s). Also post back if you continue to have issues.
Thanks again to all of you!
Yes Thank You to All
Hi All
Thank You
I have been using Turnkey Lamp for about 10 years & this is the first problem that I have ever had.
Great team work really shows in the Turnkey Community.
Thank You All
Add new comment