PBS kernel trace problems

frankz

Active Member
Nov 16, 2020
364
24
38
That's all, I noticed that my PBs goes into kernel trace:



error.png
 
Last edited:
Hi,

That's not the full message, please post more from the dmesg output, ideally as text - images of text are unecesarrily hard to handle.

There's on small hint in the middle, namely a process stuck in D state, that means uninterruptible IO and, for example, can happen if NFS is in use and network goes down. Can you post also a bit more about that system, version in use, storage in use, ...?
 
Hi,

That's not the full message, please post more from the dmesg output, ideally as text - images of text are unecesarrily hard to handle.

There's on small hint in the middle, namely a process stuck in D state, that means uninterruptible IO and, for example, can happen if NFS is in use and network goes down. Can you post also a bit more about that system, version in use, storage in use, ...?
Si scusami hai ragione , Comunque per quello poco che mi hai scritto effettivamente c'e' un riscontro in merito ad NFS . Ecco altre informazioni :






Code:
proxmox-backup: 2.1-1 (running kernel: 5.13.19-4-pve)
proxmox-backup-server: 2.1.5-1 (running version: 2.1.5)
pve-kernel-helper: 7.1-10
pve-kernel-5.13: 7.1-7
pve-kernel-5.11: 7.0-10
pve-kernel-5.4: 6.4-4
pve-kernel-5.13.19-4-pve: 5.13.19-9
pve-kernel-5.13.19-3-pve: 5.13.19-7
pve-kernel-5.13.19-2-pve: 5.13.19-4
pve-kernel-5.11.22-7-pve: 5.11.22-12
pve-kernel-5.4.124-1-pve: 5.4.124-1
pve-kernel-5.4.65-1-pve: 5.4.65-1
ifupdown2: 3.1.0-1+pmx3
libjs-extjs: 7.0.0-1
proxmox-backup-docs: 2.1.5-1
proxmox-backup-client: 2.1.5-1
proxmox-mini-journalreader: 1.2-1
proxmox-widget-toolkit: 3.4-5
pve-xtermjs: 4.16.0-1
smartmontools: 7.2-pve2
zfsutils-linux: 2.1.2-pve1

The Pbs is configured on two NICs, one LAN and the other for the backup network, where there are servers running in NFS on the latter network.
I restarted the server and unfortunately in Messages I can't find anything from the previous track.
Datastores have always worked from the one with an internal sata disk that the 500 GB USB one, although not exceptionally with great performance, all in all I have never had problems, even if in recent times at booting the PBS freezes after checking the LVM of the disk and goes as if it were stalled.
 

Attachments

  • disk.png
    disk.png
    75.7 KB · Views: 2
Hi,

That's not the full message, please post more from the dmesg output, ideally as text - images of text are unecesarrily hard to handle.

There's on small hint in the middle, namely a process stuck in D state, that means uninterruptible IO and, for example, can happen if NFS is in use and network goes down. Can you post also a bit more about that system, version in use, storage in use, ...?
Ciao ecco altri dati :





Code:
[231272.015730] INFO: task txg_sync:720 blocked for more than 120 seconds.
[231272.015745]       Tainted: P           O      5.13.19-4-pve #1
[231272.015748] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
[231272.015752] task:txg_sync        state:D stack:    0 pid:  720 ppid:     2 flags:0x00004000
[231272.015758] Call Trace:
[231272.015761]  <TASK>
[231272.015765]  __schedule+0x2fa/0x910
[231272.015775]  schedule+0x4f/0xc0
[231272.015778]  schedule_timeout+0x8a/0x140
[231272.015781]  ? __bpf_trace_tick_stop+0x10/0x10
[231272.015787]  io_schedule_timeout+0x51/0x80
[231272.015792]  __cv_timedwait_common+0x131/0x170 [spl]
[231272.015807]  ? wait_woken+0x80/0x80
[231272.015810]  __cv_timedwait_io+0x19/0x20 [spl]
[231272.015823]  zio_wait+0x133/0x2c0 [zfs]
[231272.016002]  dsl_pool_sync+0xcc/0x4f0 [zfs]
[231272.016146]  spa_sync+0x55a/0xff0 [zfs]
[231272.016292]  ? spa_txg_history_init_io+0x106/0x110 [zfs]
[231272.016439]  txg_sync_thread+0x2d3/0x460 [zfs]
[231272.016586]  ? txg_init+0x260/0x260 [zfs]
[231272.016733]  thread_generic_wrapper+0x79/0x90 [spl]
[231272.016745]  ? __thread_exit+0x20/0x20 [spl]
[231272.016756]  kthread+0x128/0x150
[231272.016762]  ? set_kthread_struct+0x50/0x50
[231272.016765]  ret_from_fork+0x1f/0x30
[231272.016771]  </TASK>
[231755.341390] INFO: task txg_sync:720 blocked for more than 120 seconds.
[231755.341413]       Tainted: P           O      5.13.19-4-pve #1
[231755.341421] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
[231755.341427] task:txg_sync        state:D stack:    0 pid:  720 ppid:     2 flags:0x00004000
[231755.341438] Call Trace:
[231755.341442]  <TASK>
[231755.341450]  __schedule+0x2fa/0x910
[231755.341467]  schedule+0x4f/0xc0
[231755.341473]  schedule_timeout+0x8a/0x140
[231755.341479]  ? __bpf_trace_tick_stop+0x10/0x10
[231755.341488]  io_schedule_timeout+0x51/0x80
[231755.341498]  __cv_timedwait_common+0x131/0x170 [spl]
[231755.341526]  ? wait_woken+0x80/0x80
[231755.341533]  __cv_timedwait_io+0x19/0x20 [spl]
[231755.341562]  zio_wait+0x133/0x2c0 [zfs]
[231755.341735]  dsl_pool_sync+0xcc/0x4f0 [zfs]
[231755.341871]  spa_sync+0x55a/0xff0 [zfs]
[231755.342013]  ? spa_txg_history_init_io+0x106/0x110 [zfs]
[231755.342154]  txg_sync_thread+0x2d3/0x460 [zfs]
[231755.342296]  ? txg_init+0x260/0x260 [zfs]
[231755.342437]  thread_generic_wrapper+0x79/0x90 [spl]
[231755.342449]  ? __thread_exit+0x20/0x20 [spl]
[231755.342460]  kthread+0x128/0x150
[231755.342464]  ? set_kthread_struct+0x50/0x50
[231755.342467]  ret_from_fork+0x1f/0x30
[231755.342473]  </TASK>
 

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!