Page MenuHomePhabricator

decommission db1097.eqiad.wmnet
Closed, ResolvedPublicRequest

Description

This task will track the decommission-hardware of server db1097.eqiad.wmnet

With the launch of updates to the decom cookbook, the majority of these steps can be handled by the service owners directly. The DC Ops team only gets involved once the system has been fully removed from service and powered down by the decommission cookbook.

db1097

Steps for service owner:

  • - all system services confirmed offline from production use
  • - set all icinga checks to maint mode/disabled while reclaim/decommmission takes place. https://gerrit.wikimedia.org/r/c/operations/puppet/+/610236
  • - remove system from all lvs/pybal active configuration
  • - any service group puppet/hiera/dsh config removed
  • - remove site.pp, replace with role(spare::system) recommended to ensure services offline but not 100% required as long as the decom script is IMMEDIATELY run below.
  • - login to cumin host and run the decom cookbook: cookbook sre.hosts.decommission <host fqdn> -t <phab task>. This does: bootloader wipe, host power down, netbox update to decommissioning status, puppet node clean, puppet node deactivate, debmonitor removal.
  • - remove all remaining puppet references (include role::spare) and all host entries in the puppet repo
  • - remove ALL dns entries except the asset tag mgmt entries.
  • - reassign task from service owner to DC ops team member depending on site of servee.

End service owner steps / Begin DC-Ops team steps:

  • - disable switch port / set to asset tag if host isn't being unracked / remove from switch if being unracked.
  • - system disks wiped (by onsite) will be shreded with disposal
  • - determine system age, under 5 years are reclaimed to spare, over 5 years are decommissioned.
  • - IF DECOM: system unracked and decommissioned (by onsite), update netbox with result and set state to offline
  • - IF DECOM: switch port configration removed from switch once system is unracked.
  • - IF DECOM: mgmt dns entries removed.

Event Timeline

Change 612135 had a related patch set uploaded (by Marostegui; owner: Marostegui):
[operations/puppet@production] mariadb: Remove puppet references for db1097

https://gerrit.wikimedia.org/r/612135

cookbooks.sre.hosts.decommission executed by marostegui@cumin1001 for hosts: db1097.eqiad.wmnet

  • db1097.eqiad.wmnet (FAIL)
    • Downtimed host on Icinga
    • Found physical host
    • Downtimed management interface on Icinga
    • Wiped bootloaders
    • Failed to power off, manual intervention required: Remote IPMI for db1097.mgmt.eqiad.wmnet failed (exit=1): b''
    • Set Netbox status to Decommissioning
    • Removed from DebMonitor
    • Removed from Puppet master and PuppetDB

ERROR: some step on some host failed, check the bolded items above

Change 612135 merged by Marostegui:
[operations/puppet@production] mariadb: Remove puppet references for db1097

https://gerrit.wikimedia.org/r/612135

cookbooks.sre.hosts.decommission executed by marostegui@cumin1001 for hosts: db1097.eqiad.wmnet

  • db1097.eqiad.wmnet (FAIL)
    • Downtimed host on Icinga
    • Found physical host
    • Downtimed management interface on Icinga
    • Wiped bootloaders
    • Failed to power off, manual intervention required: Remote IPMI for db1097.mgmt.eqiad.wmnet failed (exit=1): b''
    • Set Netbox status to Decommissioning
    • Removed from DebMonitor
    • Removed from Puppet master and PuppetDB

ERROR: some step on some host failed, check the bolded items above

I have powered off the host manually, the IPMI connection was failing

Change 612136 had a related patch set uploaded (by Marostegui; owner: Marostegui):
[operations/dns@master] wmnet: Remove db1097 DNS

https://gerrit.wikimedia.org/r/612136

Change 612136 merged by Marostegui:
[operations/dns@master] wmnet: Remove db1097 DNS

https://gerrit.wikimedia.org/r/612136

Marostegui added subscribers: Jclark-ctr, wiki_willy.

@Jclark-ctr please note that this host has mainboard/memory issues, so let's label it as such. However, the disks and the BBU should be usable as spare if any other host like this requires it, as they are perfectly usable. Can we set them aside somewhere so we have spare pieces for those hosts which are a similar model and are not under warranty anymore?

Change 616884 had a related patch set uploaded (by Cmjohnson; owner: Cmjohnson):
[operations/dns@master] Removing mgmt dns for decom host db1097

https://gerrit.wikimedia.org/r/616884

Change 616884 merged by Cmjohnson:
[operations/dns@master] Removing mgmt dns for decom host db1097

https://gerrit.wikimedia.org/r/616884

Cmjohnson updated the task description. (Show Details)