i will soon migrate cupani AKA git-rw.torproject.org to the new ganeti cluster. this will involve an IP address change which might affect the service.
please let me know if there are any problems you can think of. in particular, do let me know if any internal (inside the server) or external (outside the server) services hardcodes the IP address of cupani.
thanks!
To upload designs, you'll need to enable LFS and have an admin enable hashed storage. More information
Child items ...
Show closed items
Linked items 0
Link issues together to show that they're related.
Learn more.
We should test that IRC notifications still work after the move. Other notifications and hooks seem to be authenticated in some way, but the IRC ones aren't so I suspect IP address based auth.
managed to boot the box. turns out disk order matters so i'll have to improve my inventory routine to list disks in the right order.
the other problem is that it's not finding /srv, but that might not be the migration script's fault, because the disk is available in the right order in the vm now. crypto?
the ip has been changed and the new host is available for testing at 116.202.120.182 and 2a01:4f8:fff0:4f:266:37ff:fe5f:c6c6. All changes will be lost in the final sync.
remote: ssh: connect to host vineale.torproject.org port 22: Connection refusedremote: rsync: connection unexpectedly closed (0 bytes received so far) [sender]remote: rsync error: error in rsync protocol data stream (code 12) at io.c(235) [sender=3.1.2]
git-rw needs to be able to SSH to vineale.torproject.org directly, which I guess is an iptables thing but maybe it's more complicated.
Trac: Status: accepted to needs_revision Reviewer: N/Ato irl
ah yes, that's because the jumphost configuration (which allows all hosts to talk to all others) is defined in puppet and exported based on the current ip address (which hasn't been updated in puppet yet). I expect this will fix itself after the migration is done for real.
for now, I fixed this by hand with this on vineale:
iptables -I INPUT -j ACCEPT -s 116.202.120.182
can you try again please?
even better: could you provide a quick few commands I could run to test this myself?
Somehow the SSH key is missing to push to vineale, and the IRC message hook is broken (and at a guess this may also be affecting the email hook).
You can't test this unless you are part of the git team, because you won't have permission to push to the config repo. While this host may not be a production host, all the hosts that the triggers are talking to are production hosts.
I'm afk today at a meeting, and then it's the weekend, so this may be all the progress we make this week. Note that next week I only have 3 days, and then I'm back on the 23rd.
Somehow the SSH key is missing to push to vineale, and the IRC message hook is broken (and at a guess this may also be affecting the email hook).
I don't understand that diagnostic. What do you mean "missing SSH key" - the actual file is gone? Which path exactly? What error message are you getting?
What is the IRC message hook? How does it get fired?
It's hard for me to help you in those conditions, as I have no way of reproducing this problem in any form and you are providing very little information describing the actual issues you are having.
You can't test this unless you are part of the git team, because you won't have permission to push to the config repo.
Couldn't I add myself to the git team, at least at the technical level?
While this host may not be a production host, all the hosts that the triggers are talking to are production hosts.
I understand that we need to be careful. :)
I'm afk today at a meeting, and then it's the weekend, so this may be all the progress we make this week. Note that next week I only have 3 days, and then I'm back on the 23rd.
Okay well, I'll leave this on my desk until monday, but after that I'm tempted to just resync the VM and change the DNS and deal with the problems that result after... Otherwise we'll never get through with this. :)
And sorry you have to go through with those things, you're kind of my guinea pig of this procedure. I underestimated the trouble involved with the renumbering and server move and I am starting to see how much trouble that might entail. I will try to proceed differently for the next migrations.
i redid a sync without problems today, but i've removed the cloned machine. i'll finalize the migration tomorrow morning (UTC-4) and just fix the problems as they come along.
revocation procedures problems were discussed in #33587 (moved). i recreated a new cert by moving /var/lib/puppet/ssl aside on the client and rebootstrapping puppet. all done.
I solved it for hetzner-nbg1-02.torproject.org as following:
While trying to just regenerate the certificate on the client I noticed puppet was throwing an error:
To fix this, remove the CSR from both the master and the agent and then start a puppet run, which will automatically regenerate a CSR.On the master: puppet cert clean hetzner-nbg1-02.torproject.orgOn the agent: 1a. On most platforms: find /var/lib/puppet/ssl -name hetzner-nbg1-02.torproject.org.pem -delete 1b. On Windows: del "\var\lib\puppet\ssl\certs\hetzner-nbg1-02.torproject.org.pem" /f 2. puppet agent -t