Hi gentelmen:
I'm running a tiny cluster with 4 nodes. On Saturday there was a failure and one node started to behave weird. To make ceph chug away I decided to delete monitor from it and just let everything go with its way. Unfortunately I've deleted monitor from wrong node ... and the I've...
Can somebody provide some explanation here:
When using pveceph command to create osd with --bluestore = true. It does create 2 partitions on HDD. One of them is 100MB XFS, other one is reminder of the disk.
Now without the
it seems to work, but I'm not sure whenever it's using an actual...
fyi I've got same problem and I've got NO capital leters in host names (capital leters in host names don't exist anyway so why somebody would put them there)
Edit:
in my case for some reason /etc/ssh/ssh_known_hosts was NOT a symlink to /etc/pve/priv/known_hosts as it should.
Hi.
Before somebody calls me silly ... I know that this trick shall not be done on production systems ... but I needed some of the features of 5.0 so I jumped the ship early.
Anyway I've got 3 systems with proxmox 5 beta 1 & ceph runing as a small cluster and everything was ok.
After some time...
Thanks @FastLaneJB, I was slowly thinking that nobody here understands real issues with backups :/
Anyway check out the backy2 I'm intrigued with it because I run everything on top of ceph and ceph -> S3 seems like a logical move. That's why I wanted somebody from team to look at it and maybe...
@remark thanks for you remark about being able to backup to anything that is mounted - if you would carefully read few words further you would understand that I had problem with stuff that is mounted because people tent to mess it up. Anyway thanks for you remark @remark
If using setup as...
Hi,
You may treat this it as a kinda feature request :)
So I've got most of stuff at my workplace running on proxmox, but I've noticed that backup is somehow lacking. This is not meant as bitching topic - so no flames please.
So main problem I've got is disaster recovery, right now backup only...
So, thanks for all the replies - it got me going the right way.
I anybody ever looks for stuff like that, here is my dirty script:
/etc/systemd/system/multi-user.target.wants/mount-cephfs.service
[Unit]
Description=Mount cephsFS info /cephfs
ConditionPathExists=/cephfs...
Thing is that @reboot will run at the point where console is initiated (level 6 when login prompt shows up) - while ceph services are actually still starting up. I've added ceph-mds@0 and you can see that one still being in state starting up (along with whole load of ceph-osd@0 services)
I'll...
Hi, this is kinda trivial question, but I found usual scripts that should run at the end of level 6 init process but since proxmox and the rest of the world is mowing towards systemd it's all a bit messy to me and it seems that usual suspects don't get run as last ...
Also I would like the run...
I was not aware of fixes that you guys are making (or those were not communicated properly).
From your statement I can gather that fix is essentially making sure that I won't use virtio_scsi - but more inept and performance hitting virtio_blk (I know that i's still virtio_scsi, but it will mask...
@fabian, unfortunately I can't do that for you. All proxmox installations I've had to manually revert to 4.3 and I'm waiting for issue to be resolved to push them up to 4.4 ... I can't afford any more data loss and time consuming re-installations.
I thought that you guys have managed to...
@fabian, do you want to have an output under 4.3 ? (also sas -> sas_disk setup has had proxmox removed from for safety sake, so this will take some time to get proxmox back on it)
I've got setup with
SATA controller - SATA disk
SAS controller - SATA disk
SAS controller - SAS disk
What will be a net effect for my setups vs. 4.3 proxmox ?
@e100 I think this might be a unrelated issue that you have there. Problems that people have is that without any evidence on host the virtio_scsi will cause corruption of data between guest-disk ...
Also logs will "just end" if you have a kernel panic within guest, I think the best way for...
Just FYI I can reproduce this without any FS, just by having DD write data to passed through disk in VM.
ps. I've received an email few days ago from proxmox to root at my company domain that is diverted to my mailbox ... and it let me know that there was a smart discovered bad sector on...
Yep - I discovered it on btrfs and then was able to replicate it with dd directly to disk device :/ But it seems here that if it's not following "the ceph of lvm way" it's not worth the bother - even devs can't commit to test it as described in original post ... they use network iscsi :/
Can...
Just FYI,
(since I've had few hours free in new year)
when downgrading kernel to version used on proxmox 4.3 it's all the same :/
uname -a
Linux proxmox-dl180-14bay-2 4.4.19-1-pve #1 SMP Wed Sep 14 14:33:50 CEST 2016 x86_64 GNU/Linux
So bottom line is that irrespective of hardware...
In terms of logs - nothing on host :/
In terms of your passthrough logic - you are slightly wrong, if you use virtio_scsi_pci with image backing it will emulate SCSI for you (actually using a separate thread for it OR thread per device), if you actually connect to real device qemu will pass...
@manu first of all you are not passing a real disk (at least it does not look like that from your example), second of all have you got that:
cpu: host
scsihw: virtio-scsi-pci
S.M.A.R.T. is continuously monitored on all drives and there is no problem with any of them (even level of helium on...
This site uses cookies to help personalise content, tailor your experience and to keep you logged in if you register.
By continuing to use this site, you are consenting to our use of cookies.