this post was submitted on 02 Mar 2026
21 points (95.7% liked)
Tech
3205 readers
1 users here now
A community for high quality news and discussion around technological advancements and changes
Things that fit:
- New tech releases
- Major tech changes
- Major milestones for tech
- Major tech news such as data breaches, discontinuation
Things that don't fit
- Minor app updates
- Government legislation
- Company news
- Opinion pieces
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Does this mean for the (ab)users, or for the repo? If it's for the bandwidth hogs, then the brownouts are properly a good thing, as it'll force people to pay attention to these otherwise unmonitored systems.
Also, if it makes the upstream service seem flaky and unreliable, it could convince users to set up the proper caching proxy just for self-interested availability reasons.
I can see some companies happily paying for access, as they'll think it's easier than paying someone internally to manage a proxy/mirror, especially as on-prem is unfashionable lately.
for context, the full sentence is "Throttling efforts led to “brownouts” via 429 errors, but patterns mutated, forcing a "Whack-a-Mole" game, especially since most consumption is headless and unnoticed." so they tried to use brownouts to control it but it couldn' t stop them
Yeah ok, I guess that's what's meant.
I'd be interested to know how the patterns changed - perhaps requests moved to IPv6 which made grouping request origins harder, or maybe too many unconnected users were coming from a single IP and getting false positives (leading to bad UX and support requests).