-
Notifications
You must be signed in to change notification settings - Fork 123
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
snapshotter aborting while taking snapshot from a 5 TB 3 node cluster #120
Comments
I've just run the tool on a 2 node C* cluster with around 25 GB of data which worked fine. I was running it with user root. |
I found out the issue: even with setting all sudoers options to NOPASSWD my issues was not solved, because I'm using ssh agent forwarding and was runnint cassandra-snapshotter in a screen, which by default is not knowing about the underlying ssh session. so if you use ssh with agent key forwarding in a screen you'll be by default prompted for a password. following this gist https://gist.github.com/martijnvermaat/8070533 fixed my issue of being able to connect without password to the other hosts. |
I thought I've fixed it, but it again aborted at the very end with this message:
I don't get, it dumps 4,8 TB and at the very end it aborts with the above message. I don't see any ring or manfiest.json files backed up. the command i was using:
fab and paramiko are in the following version
any idea what's going wrong? I'm inside the screen session and can ssh to every host in the cluster with the user ramo without password. I can also edit and save files like `sudo vi /etc/hosts" on every host without entering a password any ideas? |
I was able to fix the error: This is done by copying the id_rsa.pub of your HOST to the authorized_keys file of each C_NODE in CNODES. Note that my HOST machine is an instance within the same VPC of the C_NODES. |
Thanks @markediez I am able to access all nodes on the host and in a screen via SSH without being prompted for a passowrd. I did it once with root and created also a seperate user that was listed in sudoers file with NOPASSWD set on all hosts. Again, manually it worked but while running snapshotter it didn't work. The strange issue here is: it's throwing the error after streaming 4,8 TB out of 5 TB , so at the very end. I'm assuming that snapshotter also tries to login all hosts at the very beginning, before it streams. So that seems to work fine. I've upgraded now my DSE version and will try to run an update with Opscenter. I'm not sure if this ticket should be closed or not, since there was no solution found for my case |
Hi,
I'm running into trouble when using snapshotter on a 3 node DSE cluster with a volume of 5 TB.
I'm running DSE 5.2.4 in the cluster which as C* 2.1 as a version
Furthermore Python 2.7.6
I've tried to run snapshotter with in two versions:
with user
ramo
which has the following configuration invisudo
ramo ALL=(ALL:ALL) NOPASSWD:ALL
and performing this command
and
with user
root
user ramo has sudo access with NOPASSWD:ALL and can also login on each node of the cluster.
Furtermore user ramo has access to the /tmp folder and owns the backupmanifest file.
Any ideas what's going wrong here?
The text was updated successfully, but these errors were encountered: