Page MenuHomePhabricator

Decommission prometheus4001
Closed, ResolvedPublicRequest

Description

This task will track the decommission-hardware of server prometheus4001.

With the launch of updates to the decom cookbook, the majority of these steps can be handled by the service owners directly. The DC Ops team only gets involved once the system has been fully removed from service and powered down by the decommission cookbook.

Steps for service owner:

  • - all system services confirmed offline from production use
  • - set all icinga checks to maint mode/disabled while reclaim/decommmission takes place. (likely done by script)
  • - remove system from all lvs/pybal active configuration
  • - any service group puppet/hiera/dsh config removed
  • - remove site.pp, replace with role(spare::system) recommended to ensure services offline but not 100% required as long as the decom script is IMMEDIATELY run below.
  • - login to cumin host and run the decom cookbook: cookbook sre.hosts.decommission <host fqdn> -t <phab task>. This does: bootloader wipe, host power down, netbox update to decommissioning status, puppet node clean, puppet node deactivate, debmonitor removal, and run homer.
  • - remove all remaining puppet references and all host entries in the puppet repo
  • - reassign task from service owner to DC ops team member and site project (ops-sitename) depending on site of server

End service owner steps / Begin DC-Ops team steps:

  • - system disks removed (by onsite)
  • - determine system age, under 5 years are reclaimed to spare, over 5 years are decommissioned.
  • - IF DECOM: system unracked and decommissioned (by onsite), update netbox with result and set state to offline
  • - IF DECOM: mgmt dns entries removed.
  • - IF RECLAIM: set netbox state to 'inventory' and hostname to asset tag

Event Timeline

andrea.denisse renamed this task from decommission prometheus4001 to Decommission prometheus4001.Apr 28 2023, 2:47 PM
andrea.denisse changed the task status from Open to In Progress.
andrea.denisse created this task.

Change 913250 had a related patch set uploaded (by Andrea Denisse; author: Andrea Denisse):

[operations/puppet@production] prometheus: Decommission prometheus4001 in ulsfo

https://gerrit.wikimedia.org/r/913250

Change 913250 merged by Andrea Denisse:

[operations/puppet@production] prometheus: Decommission prometheus4001 in ulsfo

https://gerrit.wikimedia.org/r/913250

cookbooks.sre.hosts.decommission executed by denisse@cumin1001 for hosts: prometheus4001.ulsfo.wment

  • prometheus4001.ulsfo.wment (FAIL)
    • Downtimed host on Icinga/Alertmanager
    • Found Ganeti VM
    • Failed to shutdown VM, manually run gnt-instance remove on the Ganeti master for the ulsfo cluster: Cumin execution failed (exit_code=2)
    • Started forced sync of VMs in Ganeti cluster ulsfo to Netbox
    • Removed from DebMonitor
    • Removed from Puppet master and PuppetDB
    • Failed to remove VM, manually run gnt-instance remove on the Ganeti master for the ulsfo cluster: Cumin execution failed (exit_code=2)
    • Started forced sync of VMs in Ganeti cluster ulsfo to Netbox

ERROR: some step on some host failed, check the bolded items above

Mentioned in SAL (#wikimedia-operations) [2023-05-12T00:32:21Z] <denisse> manually removing prometheus4001.ulsfo.wmnet from the Ganeti master after a failed step in the decommission cookbook - T335585

andrea.denisse changed the task status from In Progress to Open.May 12 2023, 12:39 AM
andrea.denisse removed andrea.denisse as the assignee of this task.
andrea.denisse updated the task description. (Show Details)
andrea.denisse added projects: DC-Ops, ops-ulsfo.

Cookbook cookbooks.sre.debmonitor.remove-hosts run by jmm: for 1 hosts: prometheus4001.ulsfo.wmnet

So VMs don't need/warrant a hardware decom ticket, resolving.