Furud and Flerovium are two special hadoop client nodes, since they were created to host data backups that we still care/need.
We can avoid to copy data to another host before the reimage if we use the reuse-parts script that Stevie created for partman (and that we used for all the Hadoop workers). This will be a good use case for @razzi to learn how to reimage without loosing data :)
High level plan:
- Check what we do/use for other use cases, like for the hadoop workers.
- Check analytics105[1-9]|analytics106[0-9]|analytics107[0-9]|an-worker*) echo reuse-parts.cfg partman/custom/reuse-analytics-hadoop-worker-12dev.cfg ;; \ in netboot.cfg (in puppet)
- More specifically, the above points to the partman/custom/reuse-analytics-hadoop-worker-12dev.cfg partman recipe.
- https://gerrit.wikimedia.org/r/c/operations/puppet/+/664788 for an example about how the above recipe was created.
- Create the partman recipe to keep /srv for furud/flerovium, and set it in netboot.cfg.
- The first time we should use echo reuse-parts-test.cfg ..., since it stops the Debian Install process right before confirming to proceed with the creation/destruction of partitions. It is extremely helpful to be sure that the recipe is right and that we'll not wipe everything.
- Use T231067#6863800 (if needed) before reimaging to set fixed uid/gids properly.