Hello guys,
I have a dell 2950 server with one 120 ssd drive and a 1Tb 8 drives array in raid50 for local storage 64GB of RAM and 2xX5460 . I have just installed the latest version of pve in the attempt to run a test environment for client api calls emulation on a software development project.
There is no outside storage, everything is being run from the local storage I just specify and this is no cluster install, just a plain server setup.
So, for this test we need to run around 140-200 CT on this setup, with the exact same configuration, using the bionic LTS latest install we have downloaded from the repository.
I have created one template, which after I have set it up I have cloned it up to 139 CT time/instances.
The issue really arises after i try to start all CT, which fail after just about 105~111 CT running with the following errors while trying to start the rest up to 140:
Syslog says:
Tried tunning the kernel params, performed a reboot but couldn't make it eat more than 111 CT without making the base plain PVE install start complaining of no space left.
Below is the useful information for the current env.
Anybody that might have an idea on what should be changed/adapted feel free to knock-in, as help is appreciated.
chroot inside a container:
Thank you.
I have a dell 2950 server with one 120 ssd drive and a 1Tb 8 drives array in raid50 for local storage 64GB of RAM and 2xX5460 . I have just installed the latest version of pve in the attempt to run a test environment for client api calls emulation on a software development project.
There is no outside storage, everything is being run from the local storage I just specify and this is no cluster install, just a plain server setup.
So, for this test we need to run around 140-200 CT on this setup, with the exact same configuration, using the bionic LTS latest install we have downloaded from the repository.
I have created one template, which after I have set it up I have cloned it up to 139 CT time/instances.
The issue really arises after i try to start all CT, which fail after just about 105~111 CT running with the following errors while trying to start the rest up to 140:
unable to fork worker - No space left on device at /usr/share/perl5/PVE/RESTEnvironment.pm line 504.
starting 213
command 'systemctl start pve-container@213' failed: open3: fork failed: No space left on device at /usr/share/perl5/PVE/Tools.pm line 429.
starting 214
Job for pve-container@214.service failed because the control process exited with error code.
See "systemctl status pve-container@214.service" and "journalctl -xe" for details.
command 'systemctl start pve-container@214' failed: exit code 1
starting 215
Job for pve-container@215.service failed because of unavailable resources or another system error.
See "systemctl status pve-container@215.service" and "journalctl -xe" for details.
command 'systemctl start pve-container@215' failed: exit code 1
starting 216
Job for pve-container@216.service failed because of unavailable resources or another system error.
See "systemctl status pve-container@216.service" and "journalctl -xe" for details.
command 'systemctl start pve-container@216' failed: exit code 1
starting 217
Job for pve-container@217.service failed because the control process exited with error code.
See "systemctl status pve-container@217.service" and "journalctl -xe" for details.
command 'systemctl start pve-container@217' failed: exit code 1
starting 218
Job for pve-container@218.service failed because the control process exited with error code.
See "systemctl status pve-container@218.service" and "journalctl -xe" for details.
command 'systemctl start pve-container@218' failed: exit code 1
starting 219
Job for pve-container@219.service failed because the control process exited with error code.
See "systemctl status pve-container@219.service" and "journalctl -xe" for details.
command 'systemctl start pve-container@219' failed: exit code 1
starting 220
Job for pve-container@220.service failed because the control process exited with error code.
See "systemctl status pve-container@220.service" and "journalctl -xe" for details.
command 'systemctl start pve-container@220' failed: exit code 1
starting 221
Job for pve-container@221.service failed because the control process exited with error code.
See "systemctl status pve-container@221.service" and "journalctl -xe" for details.
command 'systemctl start pve-container@221' failed: exit code 1
starting 222
Job for pve-container@222.service failed because the control process exited with error code.
See "systemctl status pve-container@222.service" and "journalctl -xe" for details.
command 'systemctl start pve-container@222' failed: exit code 1
starting 223
Job for pve-container@223.service failed because the control process exited with error code.
See "systemctl status pve-container@223.service" and "journalctl -xe" for details.
command 'systemctl start pve-container@223' failed: exit code 1
starting 224
Job for pve-container@224.service failed because the control process exited with error code.
See "systemctl status pve-container@224.service" and "journalctl -xe" for details.
command 'systemctl start pve-container@224' failed: exit code 1
starting 225
Job for pve-container@225.service failed because the control process exited with error code.
See "systemctl status pve-container@225.service" and "journalctl -xe" for details.
command 'systemctl start pve-container@225' failed: exit code 1
starting 226
Job for pve-container@226.service failed because the control process exited with error code.
See "systemctl status pve-container@226.service" and "journalctl -xe" for details.
command 'systemctl start pve-container@226' failed: exit code 1
starting 227
unable to fork worker - No space left on device at /usr/share/perl5/PVE/RESTEnvironment.pm line 504.
starting 228
unable to fork worker - No space left on device at /usr/share/perl5/PVE/RESTEnvironment.pm line 504.
starting 229
unable to fork worker - No space left on device at /usr/share/perl5/PVE/RESTEnvironment.pm line 504.
starting 230
unable to fork worker - No space left on device at /usr/share/perl5/PVE/RESTEnvironment.pm line 504.
starting 231
command 'systemctl start pve-container@231' failed: open3: fork failed: No space left on device at /usr/share/perl5/PVE/Tools.pm line 429.
starting 232
unable to fork worker - No space left on device at /usr/share/perl5/PVE/RESTEnvironment.pm line 504.
starting 233
Job for pve-container@233.service failed because of unavailable resources or another system error.
See "systemctl status pve-container@233.service" and "journalctl -xe" for details.
command 'systemctl start pve-container@233' failed: exit code 1
starting 234
unable to fork worker - No space left on device at /usr/share/perl5/PVE/RESTEnvironment.pm line 504.
starting 235
Job for pve-container@235.service failed because the control process exited with error code.
See "systemctl status pve-container@235.service" and "journalctl -xe" for details.
command 'systemctl start pve-container@235' failed: exit code 1
starting 236
unable to fork worker - No space left on device at /usr/share/perl5/PVE/RESTEnvironment.pm line 504.
starting 237
Job for pve-container@237.service failed because the control process exited with error code.
See "systemctl status pve-container@237.service" and "journalctl -xe" for details.
command 'systemctl start pve-container@237' failed: exit code 1
starting 238
Job for pve-container@238.service failed because the control process exited with error code.
See "systemctl status pve-container@238.service" and "journalctl -xe" for details.
command 'systemctl start pve-container@238' failed: exit code 1
starting 239
command 'systemctl start pve-container@239' failed: open3: fork failed: No space left on device at /usr/share/perl5/PVE/Tools.pm line 429.
starting 240
unable to fork worker - No space left on device at /usr/share/perl5/PVE/RESTEnvironment.pm line 504.
root@pveds01:/etc#
root@pveds01:/etc# systemctl status pve-container@218.service
● pve-container@218.service - PVE LXC Container: 218
Loaded: loaded (/lib/systemd/system/pve-container@.service; static; vendor preset: enabled)
Active: failed (Result: exit-code) since Wed 2019-06-19 23:02:52 EEST; 9s ago
Docs: man:lxc-start
man:lxc
manct
Process: 21134 ExecStart=/usr/bin/lxc-start -n 218 (code=exited, status=1/FAILURE)
Jun 19 23:02:52 pveds01 systemd[1]: Starting PVE LXC Container: 218...
Jun 19 23:02:52 pveds01 lxc-start[21134]: lxc-start: 218: tools/lxc_start.c: main: 330 The container failed to start
Jun 19 23:02:52 pveds01 lxc-start[21134]: lxc-start: 218: tools/lxc_start.c: main: 333 To get more details, run the container in foreground mode
Jun 19 23:02:52 pveds01 lxc-start[21134]: lxc-start: 218: tools/lxc_start.c: main: 336 Additional information can be obtained by setting the --logfile and --logpriority options
Jun 19 23:02:52 pveds01 systemd[1]: pve-container@218.service: Control process exited, code=exited status=1
Jun 19 23:02:52 pveds01 systemd[1]: Failed to start PVE LXC Container: 218.
Jun 19 23:02:52 pveds01 systemd[1]: pve-container@218.service: Unit entered failed state.
Jun 19 23:02:52 pveds01 systemd[1]: pve-container@218.service: Failed with result 'exit-code'.
root@pveds01:/etc#
Syslog says:
Jun 19 22:02:27 pveds01 lxcfs[914]: fuse: error creating thread: No space left on device
Jun 19 22:02:27 pveds01 lxcfs[914]: fuse: error creating thread: No space left on device
Jun 19 22:02:27 pveds01 lxcfs[914]: fuse: error creating thread: No space left on device
Jun 19 22:02:27 pveds01 lxcfs[914]: fuse: error creating thread: No space left on device
Jun 19 22:02:28 pveds01 pvestatd[1895]: fork failed: No space left on device
Jun 19 22:02:28 pveds01 pct[32359]: <root@pam> starting task UPIDveds01:00002529:00076FAE:5D0A86C4:vzstart:239:root@pam:
Jun 19 22:02:28 pveds01 pct[9513]: starting CT 239: UPIDveds01:00002529:00076FAE:5D0A86C4:vzstart:239:root@pam:
Jun 19 22:02:28 pveds01 pvestatd[1895]: command 'lxc-info -n 202 -p' failed: open3: fork failed: No space left on device at /usr/share/perl5/PVE/Tools.pm line 429.
Jun 19 22:02:28 pveds01 pvestatd[1895]: command 'lxc-info -n 134 -p' failed: open3: fork failed: No space left on device at /usr/share/perl5/PVE/Tools.pm line 429.
Jun 19 22:02:28 pveds01 pvestatd[1895]: command 'lxc-info -n 110 -p' failed: open3: fork failed: No space left on device at /usr/share/perl5/PVE/Tools.pm line 429.
Jun 19 22:02:28 pveds01 pvestatd[1895]: command 'lxc-info -n 129 -p' failed: open3: fork failed: No space left on device at /usr/share/perl5/PVE/Tools.pm line 429.
Jun 19 22:02:28 pveds01 pct[9513]: command 'systemctl start pve-container@239' failed: open3: fork failed: No space left on device at /usr/share/perl5/PVE/Tools.pm line 429.
Jun 19 22:02:28 pveds01 pvestatd[1895]: command 'lxc-info -n 208 -p' failed: open3: fork failed: No space left on device at /usr/share/perl5/PVE/Tools.pm line 429.
Jun 19 22:02:28 pveds01 pct[32359]: <root@pam> end task UPIDveds01:00002529:00076FAE:5D0A86C4:vzstart:239:root@pam: command 'systemctl start pve-container@239' failed: open3: fork failed: No space left on device at /usr/share/perl5/PVE/Tools.pm line 429.
Jun 19 22:02:28 pveds01 pvedaemon[1915]: command 'lxc-info -n 114 -p' failed: open3: fork failed: No space left on device at /usr/share/perl5/PVE/Tools.pm line 429.
Jun 19 22:02:29 pveds01 lxcfs[914]: bindings.c: 2473: recv_creds: Timed out waiting for scm_cred: Success
Jun 19 22:02:29 pveds01 lxcfs[914]: fuse: error creating thread: No space left on device
Jun 19 22:02:29 pveds01 lxcfs[914]: fuse: error creating thread: No space left on device
Jun 19 22:02:29 pveds01 lxcfs[914]: fuse: error creating thread: No space left on device
Jun 19 22:02:29 pveds01 lxcfs[914]: fuse: error creating thread: No space left on device
Jun 19 22:02:30 pveds01 pvestatd[1895]: fork failed: No space left on device
Jun 19 22:02:31 pveds01 lxcfs[914]: bindings.c: 2473: recv_creds: Timed out waiting for scm_cred: No such file or directory
Jun 19 22:02:31 pveds01 lxcfs[914]: fuse: error creating thread: No space left on device
Jun 19 22:02:32 pveds01 pvestatd[1895]: command 'lxc-info -n 178 -p' failed: open3: fork failed: No space left on device at /usr/share/perl5/PVE/Tools.pm line 429.
Jun 19 22:02:33 pveds01 pvestatd[1895]: status update time (5.680 seconds)
Jun 19 22:02:33 pveds01 lxcfs[914]: bindings.c: 2473: recv_creds: Timed out waiting for scm_cred: No such file or directory
Jun 19 22:02:33 pveds01 lxcfs[914]: fuse: error creating thread: No space left on device
Jun 19 22:02:34 pveds01 snmpd[1420]: error on subcontainer 'ia_addr' insert (-1)
Jun 19 22:02:35 pveds01 lxcfs[914]: bindings.c: 2473: recv_creds: Timed out waiting for scm_cred: Success
Jun 19 22:02:35 pveds01 lxcfs[914]: fuse: error creating thread: No space left on device
Jun 19 22:02:37 pveds01 pvestatd[1895]: command 'lxc-info -n 181 -p' failed: open3: fork failed: No space left on device at /usr/share/perl5/PVE/Tools.pm line 429.
Jun 19 22:02:37 pveds01 pvestatd[1895]: command 'lxc-info -n 195 -p' failed: open3: fork failed: No space left on device at /usr/share/perl5/PVE/Tools.pm line 429.
Jun 19 22:02:38 pveds01 pvestatd[1895]: command 'lxc-info -n 125 -p' failed: open3: fork failed: No space left on device at /usr/share/perl5/PVE/Tools.pm line 429.
Jun 19 22:02:38 pveds01 pvedaemon[1914]: <root@pam> successful auth for user 'root@pam'
Jun 19 22:02:38 pveds01 pvestatd[1895]: command 'lxc-info -n 206 -p' failed: open3: fork failed: No space left on device at /usr/share/perl5/PVE/Tools.pm line 429.
Jun 19 22:02:38 pveds01 pvestatd[1895]: command 'lxc-info -n 149 -p' failed: open3: fork failed: No space left on device at /usr/share/perl5/PVE/Tools.pm line 429.
Jun 19 22:02:38 pveds01 pvestatd[1895]: command 'lxc-info -n 193 -p' failed: open3: fork failed: No space left on device at /usr/share/perl5/PVE/Tools.pm line 429.
Jun 19 22:02:38 pveds01 pvestatd[1895]: command 'lxc-info -n 170 -p' failed: open3: fork failed: No space left on device at /usr/share/perl5/PVE/Tools.pm line 429.
Jun 19 22:02:38 pveds01 pvestatd[1895]: command 'lxc-info -n 198 -p' failed: open3: fork failed: No space left on device at /usr/share/perl5/PVE/Tools.pm line 429.
Jun 19 22:02:38 pveds01 pvestatd[1895]: command 'lxc-info -n 158 -p' failed: open3: fork failed: No space left on device at /usr/share/perl5/PVE/Tools.pm line 429.
Jun 19 22:02:38 pveds01 pvestatd[1895]: command 'lxc-info -n 166 -p' failed: open3: fork failed: No space left on device at /usr/share/perl5/PVE/Tools.pm line 429.
Jun 19 22:02:38 pveds01 pvestatd[1895]: command 'lxc-info -n 173 -p' failed: open3: fork failed: No space left on device at /usr/share/perl5/PVE/Tools.pm line 429.
Jun 19 22:02:38 pveds01 pvestatd[1895]: command 'lxc-info -n 124 -p' failed: open3: fork failed: No space left on device at /usr/share/perl5/PVE/Tools.pm line 429.
Jun 19 22:02:38 pveds01 pvestatd[1895]: command 'lxc-info -n 122 -p' failed: open3: fork failed: No space left on device at /usr/share/perl5/PVE/Tools.pm line 429.
Jun 19 22:02:38 pveds01 pvestatd[1895]: command 'lxc-info -n 152 -p' failed: open3: fork failed: No space left on device at /usr/share/perl5/PVE/Tools.pm line 429.
Jun 19 22:02:38 pveds01 pvestatd[1895]: command 'lxc-info -n 138 -p' failed: open3: fork failed: No space left on device at /usr/share/perl5/PVE/Tools.pm line 429.
Jun 19 22:02:38 pveds01 pvestatd[1895]: command 'lxc-info -n 144 -p' failed: open3: fork failed: No space left on device at /usr/share/perl5/PVE/Tools.pm line 429.
Jun 19 22:02:38 pveds01 pvestatd[1895]: command 'lxc-info -n 197 -p' failed: open3: fork failed: No space left on device at /usr/share/perl5/PVE/Tools.pm line 429.
Jun 19 22:02:38 pveds01 pvestatd[1895]: command 'lxc-info -n 186 -p' failed: open3: fork failed: No space left on device at /usr/share/perl5/PVE/Tools.pm line 429.
Jun 19 22:02:39 pveds01 pvedaemon[1915]: command 'lxc-info -n 114 -p' failed: open3: fork failed: No space left on device at /usr/share/perl5/PVE/Tools.pm line 429.
Jun 19 22:02:39 pveds01 lxcfs[914]: bindings.c: 2473: recv_creds: Timed out waiting for scm_cred: Success
Jun 19 22:02:39 pveds01 lxcfs[914]: fuse: error creating thread: No space left on device
Jun 19 22:02:39 pveds01 lxcfs[914]: fuse: error creating thread: No space left on device
Jun 19 22:02:39 pveds01 lxcfs[914]: fuse: error creating thread: No space left on device
Jun 19 22:02:39 pveds01 lxcfs[914]: fuse: error creating thread: No space left on device
Jun 19 22:02:39 pveds01 lxcfs[914]: fuse: error creating thread: No space left on device
Jun 19 22:02:39 pveds01 lxcfs[914]: fuse: error creating thread: No space left on device
Jun 19 22:02:39 pveds01 lxcfs[914]: fuse: error creating thread: No space left on device
Jun 19 22:02:39 pveds01 lxcfs[914]: fuse: error creating thread: No space left on device
Jun 19 22:02:39 pveds01 lxcfs[914]: fuse: error creating thread: No space left on device
Jun 19 22:02:39 pveds01 lxcfs[914]: fuse: error creating thread: No space left on device
Jun 19 22:02:39 pveds01 lxcfs[914]: fuse: error creating thread: No space left on device
Jun 19 22:02:39 pveds01 lxcfs[914]: fuse: error creating thread: No space left on device
Jun 19 22:02:39 pveds01 lxcfs[914]: fuse: error creating thread: No space left on device
Jun 19 22:02:39 pveds01 lxcfs[914]: fuse: error creating thread: No space left on device
Tried tunning the kernel params, performed a reboot but couldn't make it eat more than 111 CT without making the base plain PVE install start complaining of no space left.
Below is the useful information for the current env.
Anybody that might have an idea on what should be changed/adapted feel free to knock-in, as help is appreciated.
proxmox-ve: 5.4-1 (running kernel: 4.15.18-16-pve)
pve-manager: 5.4-6 (running version: 5.4-6/aa7856c5)
pve-kernel-4.15: 5.4-4
pve-kernel-4.15.18-16-pve: 4.15.18-41
pve-kernel-4.13.13-2-pve: 4.13.13-33
corosync: 2.4.4-pve1
criu: 2.11.1-1~bpo90
glusterfs-client: 3.8.8-1
ksm-control-daemon: 1.2-2
libjs-extjs: 6.0.1-2
libpve-access-control: 5.1-10
libpve-apiclient-perl: 2.0-5
libpve-common-perl: 5.0-52
libpve-guest-common-perl: 2.0-20
libpve-http-server-perl: 2.0-13
libpve-storage-perl: 5.0-43
libqb0: 1.0.3-1~bpo9
lvm2: 2.02.168-pve6
lxc-pve: 3.1.0-3
lxcfs: 3.0.3-pve1
novnc-pve: 1.0.0-3
openvswitch-switch: 2.7.0-3
proxmox-widget-toolkit: 1.0-28
pve-cluster: 5.0-37
pve-container: 2.0-39
pve-docs: 5.4-2
pve-edk2-firmware: 1.20190312-1
pve-firewall: 3.0-22
pve-firmware: 2.0-6
pve-ha-manager: 2.0-9
pve-i18n: 1.1-4
pve-libspice-server1: 0.14.1-2
pve-qemu-kvm: 3.0.1-2
pve-xtermjs: 3.12.0-1
qemu-server: 5.0-52
smartmontools: 6.5+svn4324-1
spiceterm: 3.0-5
vncterm: 1.5-3
zfsutils-linux: 0.7.13-pve1~bpo2
root@pveds01:/etc# df -h
Filesystem Size Used Avail Use% Mounted on
udev 31G 0 31G 0% /dev
tmpfs 6.2G 26M 6.2G 1% /run
/dev/mapper/pve-root 89G 6.6G 83G 8% /
tmpfs 31G 43M 31G 1% /dev/shm
tmpfs 5.0M 0 5.0M 0% /run/lock
tmpfs 31G 0 31G 0% /sys/fs/cgroup
/dev/sdb1 5.5T 221G 5.3T 4% /storage
/dev/fuse 30M 48K 30M 1% /etc/pve
tmpfs 6.2G 0 6.2G 0% /run/user/0
root@pveds01:/etc# df -hi
Filesystem Inodes IUsed IFree IUse% Mounted on
udev 7.8M 860 7.8M 1% /dev
tmpfs 7.8M 1.7K 7.8M 1% /run
/dev/mapper/pve-root 45M 71K 45M 1% /
tmpfs 7.8M 85 7.8M 1% /dev/shm
tmpfs 7.8M 155 7.8M 1% /run/lock
tmpfs 7.8M 17 7.8M 1% /sys/fs/cgroup
/dev/sdb1 559M 6.0K 559M 1% /storage
/dev/fuse 9.8K 165 9.7K 2% /etc/pve
tmpfs 7.8M 10 7.8M 1% /run/user/0
root@pveds01:/etc#
root@pveds01:/etc# lsblk
NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT
loop0 7:0 0 20G 0 loop
loop1 7:1 0 20G 0 loop
loop2 7:2 0 20G 0 loop
loop3 7:3 0 20G 0 loop
loop4 7:4 0 20G 0 loop
loop5 7:5 0 20G 0 loop
loop6 7:6 0 20G 0 loop
loop7 7:7 0 20G 0 loop
loop8 7:8 0 20G 0 loop
loop9 7:9 0 20G 0 loop
loop10 7:10 0 20G 0 loop
loop11 7:11 0 20G 0 loop
loop12 7:12 0 20G 0 loop
loop13 7:13 0 20G 0 loop
loop14 7:14 0 20G 0 loop
loop15 7:15 0 20G 0 loop
loop16 7:16 0 20G 0 loop
loop17 7:17 0 20G 0 loop
loop18 7:18 0 20G 0 loop
loop19 7:19 0 20G 0 loop
loop20 7:20 0 20G 0 loop
loop21 7:21 0 20G 0 loop
loop22 7:22 0 20G 0 loop
loop23 7:23 0 20G 0 loop
loop24 7:24 0 20G 0 loop
loop25 7:25 0 20G 0 loop
loop26 7:26 0 20G 0 loop
loop27 7:27 0 20G 0 loop
loop28 7:28 0 20G 0 loop
loop29 7:29 0 20G 0 loop
loop30 7:30 0 20G 0 loop
loop31 7:31 0 20G 0 loop
loop32 7:32 0 20G 0 loop
loop33 7:33 0 20G 0 loop
loop34 7:34 0 20G 0 loop
loop35 7:35 0 20G 0 loop
loop36 7:36 0 20G 0 loop
loop37 7:37 0 20G 0 loop
loop38 7:38 0 20G 0 loop
loop39 7:39 0 20G 0 loop
loop40 7:40 0 20G 0 loop
loop41 7:41 0 20G 0 loop
loop42 7:42 0 20G 0 loop
loop43 7:43 0 20G 0 loop
loop44 7:44 0 20G 0 loop
loop45 7:45 0 20G 0 loop
loop46 7:46 0 20G 0 loop
loop47 7:47 0 20G 0 loop
loop48 7:48 0 20G 0 loop
loop49 7:49 0 20G 0 loop
loop50 7:50 0 20G 0 loop
loop51 7:51 0 20G 0 loop
loop52 7:52 0 20G 0 loop
loop53 7:53 0 20G 0 loop
loop54 7:54 0 20G 0 loop
loop55 7:55 0 20G 0 loop
loop56 7:56 0 20G 0 loop
loop57 7:57 0 20G 0 loop
loop58 7:58 0 20G 0 loop
loop59 7:59 0 20G 0 loop
loop60 7:60 0 20G 0 loop
loop61 7:61 0 20G 0 loop
loop62 7:62 0 20G 0 loop
loop63 7:63 0 20G 0 loop
loop64 7:64 0 20G 0 loop
loop65 7:65 0 20G 0 loop
loop66 7:66 0 20G 0 loop
loop67 7:67 0 20G 0 loop
loop68 7:68 0 20G 0 loop
loop69 7:69 0 20G 0 loop
loop70 7:70 0 20G 0 loop
loop71 7:71 0 20G 0 loop
loop72 7:72 0 20G 0 loop
loop73 7:73 0 20G 0 loop
loop74 7:74 0 20G 0 loop
loop75 7:75 0 20G 0 loop
loop76 7:76 0 20G 0 loop
loop77 7:77 0 20G 0 loop
loop78 7:78 0 20G 0 loop
loop79 7:79 0 20G 0 loop
loop80 7:80 0 20G 0 loop
loop81 7:81 0 20G 0 loop
loop82 7:82 0 20G 0 loop
loop83 7:83 0 20G 0 loop
loop84 7:84 0 20G 0 loop
loop85 7:85 0 20G 0 loop
loop86 7:86 0 20G 0 loop
loop87 7:87 0 20G 0 loop
loop88 7:88 0 20G 0 loop
loop89 7:89 0 20G 0 loop
loop90 7:90 0 20G 0 loop
loop91 7:91 0 20G 0 loop
loop92 7:92 0 20G 0 loop
loop93 7:93 0 20G 0 loop
loop94 7:94 0 20G 0 loop
loop95 7:95 0 20G 0 loop
loop96 7:96 0 20G 0 loop
loop97 7:97 0 20G 0 loop
loop98 7:98 0 20G 0 loop
loop99 7:99 0 20G 0 loop
loop100 7:100 0 20G 0 loop
loop101 7:101 0 20G 0 loop
loop102 7:102 0 20G 0 loop
loop103 7:103 0 20G 0 loop
loop104 7:104 0 20G 0 loop
loop105 7:105 0 20G 0 loop
loop106 7:106 0 20G 0 loop
loop107 7:107 0 20G 0 loop
loop108 7:108 0 20G 0 loop
loop110 7:110 0 20G 0 loop
sda 8:0 0 111.8G 0 disk
├─sda1 8:1 0 1M 0 part
├─sda2 8:2 0 256M 0 part
└─sda3 8:3 0 111.6G 0 part
├─pve-swap 253:0 0 8G 0 lvm [SWAP]
└─pve-root 253:1 0 88.8G 0 lvm /
sdb 8:16 0 5.5T 0 disk
└─sdb1 8:17 0 5.5T 0 part /storage
sdc 8:32 1 16M 0 disk
sr0 11:0 1 1024M 0 rom
sr1 11:1 1 590M 0 rom
sr2 11:2 1 1024M 0 rom
root@pveds01:/etc#
root@pveds01:/etc# pct config 101
arch: amd64
cores: 2
hostname: 101.int.hosthub.ro
memory: 4096
net0: name=eth0,bridge=vmbr0,hwaddr=76:30:A1:30:03:4B,ip=dhcp,tag=106,type=veth
ostype: ubuntu
rootfs: vol_containers:101/vm-101-disk-0.raw,size=20G
swap: 128
unprivileged: 1
root@pveds01:/etc# free -m
total used free shared buff/cache available
Mem: 63411 10341 25594 295 27476 52064
Swap: 8191 0 8191
root@pveds01:/etc# ulimit -a
core file size (blocks, -c) 0
data seg size (kbytes, -d) unlimited
scheduling priority (-e) 0
file size (blocks, -f) unlimited
pending signals (-i) 253482
max locked memory (kbytes, -l) 64
max memory size (kbytes, -m) unlimited
open files (-n) 1048576
pipe size (512 bytes, -p) 8
POSIX message queues (bytes, -q) 819200
real-time priority (-r) 0
stack size (kbytes, -s) 8192
cpu time (seconds, -t) unlimited
max user processes (-u) 253482
virtual memory (kbytes, -v) unlimited
file locks (-x) unlimited
root@pveds01:/etc# ulimit -Hn
1048576
root@pveds01:/etc# ulimit -Sn
1048576
root@pveds01:/etc# cat /etc/sysctl.conf | grep -v "#" | awk 'NF >0'
fs.inotify.max_queued_events = 1048576
fs.inotify.max_user_instances = 1048576
fs.inotify.max_user_watches = 1048576
vm.max_map_count = 262144
chroot inside a container:
root@pveds01:/etc# pct enter 101
root@101:/etc#
root@101:/etc# df -h
Filesystem Size Used Avail Use% Mounted on
/dev/loop0 20G 1.5G 18G 8% /
none 492K 0 492K 0% /dev
udev 31G 0 31G 0% /dev/tty
tmpfs 31G 0 31G 0% /dev/shm
tmpfs 31G 112K 31G 1% /run
tmpfs 5.0M 0 5.0M 0% /run/lock
tmpfs 31G 0 31G 0% /sys/fs/cgroup
root@101:/etc# df -i
Filesystem Inodes IUsed IFree IUse% Mounted on
/dev/loop0 1310720 40362 1270358 4% /
none 8116654 21 8116633 1% /dev
udev 8111439 860 8110579 1% /dev/tty
tmpfs 8116654 1 8116653 1% /dev/shm
tmpfs 8116654 129 8116525 1% /run
tmpfs 8116654 2 8116652 1% /run/lock
tmpfs 8116654 17 8116637 1% /sys/fs/cgroup
root@101:/etc#
Thank you.