==== **User Story**
> ==== As a platform engineer, I need to design, implement and deploy a streaming job that produces event streams of mediawiki page changes with raw content.
==== The service willl:
* Call MW API to get the wikitext for the article
* Format the input stream data and wikitext into the new topic format
* Output the formatted data to a new Kafka topic
==== Expected Spikes:
* Data modeling exercise for new consolidated stream - T308017
==== Why are we doing this?
- Simplify event stream consumption. Consumers can listen to a single stream that represent the state of a page rather than a page action (current design)
- Adding content to streams to make them usable by consumers without having to enrich themselves
==== What is needed for GA internal release
[] {T309699}
[] {T338169}
[] {T338233}
[] Quality / consistency metrics. Might have to do this in Hive/Airflow? (Phab task TODO)
[] Alerting on SLIs (uptime, latency, and maybe quality/consistency?)
[] Announcement