NoxiousPluK
Cadet
- Joined
- Aug 2, 2022
- Messages
- 9
Today I did the following steps (in order):
- Upgrade TrueNAS Scale from 22.02.4 to 22.12.3 (everything seemed OK after the upgrade, my VMs started fine)
- Upgrade my motherboard (AsRock B450 Pro4) BIOS from 1.50 to 1.80
- Upgrade my motherboard BIOS from 1.80 to 8.02
- Upgrade my CPU from Ryzen 7 1700X to Ryzen 7 5800X
- Upgrade my RAM from 4x16GB DDR4-2166 to 2x32GB DDR4-3200
I double-checked that hardware virtualization was enabled and enabled SR-IOV (since my NIC supports it) and IOMMU (just in case I wanted to experiment with hardware mapping in the future).
TrueNAS seemed to boot fine, but the Virtualization page in the GUI got stuck on seemingly infinite loading (I gave it ~5-10 minutes, but no result).
To debug, I disabled SR-IOV and IOMMU again and rebooted, but no result.
I went through a bunch of logfiles, but the only seemingly relevant thing I could find was this in /var/log/messages:
These being my VM names.
After ~10-15 minutes, the page suddenly worked, but all VMs (including those with autostart enabled) were stopped, and hovering over the State-switch shows a tooltip that only states 'ERROR'.
I tried to enable one of the VMs and it took about 10 minutes (with a spinning 'Please wait' prompt) for it to start, but it did end up starting.
At the exact moment that the VM did start, I got this message in my open shell:
I managed to start two more VMs this way (giving similar messages), and on a 4rd try to start another one, I got (after ~5 minutes) this error in the GUI:
I have not been able to start any more VMs since; every suggestion is very welcome!
- Upgrade TrueNAS Scale from 22.02.4 to 22.12.3 (everything seemed OK after the upgrade, my VMs started fine)
- Upgrade my motherboard (AsRock B450 Pro4) BIOS from 1.50 to 1.80
- Upgrade my motherboard BIOS from 1.80 to 8.02
- Upgrade my CPU from Ryzen 7 1700X to Ryzen 7 5800X
- Upgrade my RAM from 4x16GB DDR4-2166 to 2x32GB DDR4-3200
I double-checked that hardware virtualization was enabled and enabled SR-IOV (since my NIC supports it) and IOMMU (just in case I wanted to experiment with hardware mapping in the future).
TrueNAS seemed to boot fine, but the Virtualization page in the GUI got stuck on seemingly infinite loading (I gave it ~5-10 minutes, but no result).
To debug, I disabled SR-IOV and IOMMU again and rebooted, but no result.
I went through a bunch of logfiles, but the only seemingly relevant thing I could find was this in /var/log/messages:
Code:
Jun 18 09:59:27 truenas middlewared[2202]: libvirt: QEMU Driver error : Domain not found: no domain with matching name '1_gateway' Jun 18 09:59:28 truenas middlewared[2202]: libvirt: QEMU Driver error : Domain not found: no domain with matching name '2_dc01' Jun 18 09:59:29 truenas middlewared[2202]: libvirt: QEMU Driver error : Domain not found: no domain with matching name '3_akpublicrust' Jun 18 09:59:29 truenas middlewared[2202]: libvirt: QEMU Driver error : Domain not found: no domain with matching name '6_hass' Jun 18 09:59:30 truenas middlewared[2202]: libvirt: QEMU Driver error : Domain not found: no domain with matching name '7_pufferpanel' Jun 18 09:59:30 truenas middlewared[2202]: libvirt: QEMU Driver error : Domain not found: no domain with matching name '8_webserver' Jun 18 09:59:31 truenas middlewared[2202]: libvirt: QEMU Driver error : Domain not found: no domain with matching name '10_pihole2' Jun 18 09:59:31 truenas middlewared[2202]: libvirt: QEMU Driver error : Domain not found: no domain with matching name '11_bothost'
These being my VM names.
After ~10-15 minutes, the page suddenly worked, but all VMs (including those with autostart enabled) were stopped, and hovering over the State-switch shows a tooltip that only states 'ERROR'.
I tried to enable one of the VMs and it took about 10 minutes (with a spinning 'Please wait' prompt) for it to start, but it did end up starting.
At the exact moment that the VM did start, I got this message in my open shell:
Code:
Broadcast message from systemd-journald@truenas (Sun 2023-06-18 12:53:09 CEST): haproxy[149277]: backend be_13 has no server available! Broadcast message from systemd-journald@truenas (Sun 2023-06-18 12:53:10 CEST): haproxy[149277]: backend be_29 has no server available! 2023 Jun 18 12:53:09 truenas backend be_13 has no server available! 2023 Jun 18 12:53:10 truenas backend be_29 has no server available! Broadcast message from systemd-journald@truenas (Sun 2023-06-18 12:53:10 CEST): haproxy[149277]: backend be_33 has no server available! 2023 Jun 18 12:53:10 truenas backend be_33 has no server available! Broadcast message from systemd-journald@truenas (Sun 2023-06-18 12:53:10 CEST): haproxy[149277]: backend be_38 has no server available! 2023 Jun 18 12:53:10 truenas backend be_38 has no server available! Broadcast message from systemd-journald@truenas (Sun 2023-06-18 12:53:10 CEST): haproxy[149277]: backend be_47 has no server available! Broadcast message from systemd-journald@truenas (Sun 2023-06-18 12:53:11 CEST): haproxy[149277]: backend be_52 has no server available! 2023 Jun 18 12:53:10 truenas backend be_47 has no server available! 2023 Jun 18 12:53:11 truenas backend be_52 has no server available!
I managed to start two more VMs this way (giving similar messages), and on a 4rd try to start another one, I got (after ~5 minutes) this error in the GUI:
Code:
CallError [EFAULT] Failed to connect to libvirt More info... Error: Traceback (most recent call last): File "/usr/lib/python3/dist-packages/middlewared/plugins/vm/supervisor/supervisor.py", line 172, in start if self.domain.create() < 0: File "/usr/lib/python3/dist-packages/libvirt.py", line 1353, in create raise libvirtError('virDomainCreate() failed') libvirt.libvirtError: Cannot recv data: Connection reset by peer During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/lib/python3/dist-packages/middlewared/plugins/vm/vm_lifecycle.py", line 46, in start await self.middleware.run_in_thread(self._start, vm['name']) File "/usr/lib/python3/dist-packages/middlewared/main.py", line 1261, in run_in_thread return await self.run_in_executor(self.thread_pool_executor, method, *args, **kwargs) File "/usr/lib/python3/dist-packages/middlewared/main.py", line 1258, in run_in_executor return await loop.run_in_executor(pool, functools.partial(method, *args, **kwargs)) File "/usr/lib/python3.9/concurrent/futures/thread.py", line 52, in run result = self.fn(*self.args, **self.kwargs) File "/usr/lib/python3/dist-packages/middlewared/plugins/vm/vm_supervisor.py", line 68, in _start self.vms[vm_name].start(vm_data=self._vm_from_name(vm_name)) File "/usr/lib/python3/dist-packages/middlewared/plugins/vm/supervisor/supervisor.py", line 181, in start raise CallError('\n'.join(errors)) middlewared.service_exception.CallError: [EFAULT] Cannot recv data: Connection reset by peer During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/lib/python3/dist-packages/middlewared/main.py", line 204, in call_method result = await self.middleware._call(message['method'], serviceobj, methodobj, params, app=self) File "/usr/lib/python3/dist-packages/middlewared/main.py", line 1344, in _call return await methodobj(*prepared_call.args) File "/usr/lib/python3/dist-packages/middlewared/schema.py", line 1378, in nf return await func(*args, **kwargs) File "/usr/lib/python3/dist-packages/middlewared/schema.py", line 1246, in nf res = await f(*args, **kwargs) File "/usr/lib/python3/dist-packages/middlewared/plugins/vm/vm_lifecycle.py", line 48, in start if (await self.middleware.call('vm.get_instance', id))['status']['state'] != 'RUNNING': File "/usr/lib/python3.9/concurrent/futures/thread.py", line 52, in run result = self.fn(*self.args, **self.kwargs) File "/usr/lib/python3/dist-packages/middlewared/plugins/vm/vms.py", line 100, in extend_context self._check_setup_connection() File "/usr/lib/python3/dist-packages/middlewared/plugins/vm/connection.py", line 71, in _check_setup_connection self._check_connection_alive() File "/usr/lib/python3/dist-packages/middlewared/plugins/vm/connection.py", line 66, in _check_connection_alive raise CallError('Failed to connect to libvirt') middlewared.service_exception.CallError: [EFAULT] Failed to connect to libvirt
I have not been able to start any more VMs since; every suggestion is very welcome!
Last edited: