When doing file export, we source the list of public, active wikis by querying noc.wikimedia.org.
For brand new wikis though, we will not have the site information available on wmf_raw.mediawiki_project_namespace_map.
In these cases, we fail that particular wiki like in this example:
In this task we should figure what we can do so that this doesn't become a burden on OpsWeek person, as we do expect to have new wikis created on a semi-monthly basis.
An idea: we could do a SQL check on wmf_raw.mediawiki_project_namespace_map, and if we find that we do not have data for a particular wiki, we can put it in a 'skip list'. Later, we can mark that particular wiki as skipped.
Looks like we can use an AirflowSkipException top accomplish this. In a task group, the first @task should check if we should skip and if so, throw that exception.