Per T393038#10810826, some subset of MediaWiki Logstash logs seem to be capped at 100 / sec, with further log events discarded. (Not whole channels, and certainly not the total log volume which is around 400 / sec; seems to be based on the normalized_message field?)
The dropped events are not in the dead letter queue, which is much lower volume. I think they make it through udp2log, so this is specific to the Logstash-only part of the logging stack. The cap seems to be applied with a few-minute window, e.g. here is a high-volume log event (normalized_message:"Central autologin attempt"):
exactly 90K events in 15 minutes (ie. 100/sec) but distributed in an uneven, sawtooth pattern of exactly 15K events in a roughly 1-minute window and then zero events for a slightly less than 1-minute or 2-minute interval (those two lengths seem to be alternating). So maybe a combination of a 15K/60s cap and a 30K/300s cap?
While this is probably a good thing (the log stream where I noticed this ended up much higher volume than I thought, and without such filtering would have tripled the total MediaWiki log volume), it doesn't seem to be documented anywhere and it can cause a lot of confusion. (Docs I checked: wikitech:Logstash, especially the Configuration, Gotchas and Troubleshooting sections; wikitech:Logs#mwlog1002:/srv/mw-log/; wikitech:Rsyslog; wikitech:Kafka#logging (eqiad and codfw)). Nor did I find any guidance on what would be an acceptable level of log volume for a new event or MediaWiki logs in total.




