If there’s no connection or you’re in airplane mode, some apps will let you access locally stored/cached data, but as soon as there’s a bad connection, they’ll wipe that data by trying to unsuccessfully refresh it from the server.
So you're stuck in a loop of not being able to use the web because the websites keep downloading stuff you don't need.
The obvious easy fix is to give them unlimited data. If the intent is to give them internet, they should give them internet that functions for the modern web.
Cell phone systems already have some tiering built in, at least based on the fine print I've read about my plans. Once I run out of "official data" I fall back to low-priority usage, but the cell system is generally so well-provisioned nowadays that I hardly notice. In 2026, one must take explicit action to force people back to 2G. Nothing would stop these plans from, say, simply always being "low priority usage" but at full speed, and for the most part this would satisfy everyone.
This sort of clause reeks of "it was written into a contract 15 years ago and nobody has even so much as thought about it since then" rather than some sort of choice.
Data caps are to an extent “fake”, in that telcos’ costs aren’t measured in how many bytes their customers download/upload. Telcos’ costs come from renting bandwidth from tier 1 and tier 2 ISPs. This bandwidth is constant.
And for popular websites, they will cache lots of content on their own network or peer directly with data centers so they don’t have to pay for the bandwidth there. The routers will continue routing and the switches will continue switching whether you download 5GBs or 5TBs.
One more way to understand how much of a scam mobile data caps are, is that the same ISP will sell you unlimited fiber plans even though essentially your traffic goes through the same backbone.
Data caps may help lessen congestion on their cell towers, but they don’t need to be as low as they are today.
In the long run, all costs are variable. Phone companies lack the bandwidth to provide all their customers unlimited data all the time. Most of them can’t even provide full speeds to their existing customers at peak times. If they gave everyone unlimited data they’d have to get more bandwidth, and they’d pass on every penny of the cost.
Data caps make congestion worse, because you are more likely to restrict where you use data and people are predictable. You'll no longer use bits everywhere because you care less, you'll use it where everyone else does.
In the case of TFA, PC Gamer isn't directly consuming the bandwidth with their own servers on their own domain name. It's an ad distribution network doing that, and odds are reasonable they're already colocated someplace with your ISP and the bandwidth consumed by ads is totally irrelevant to everyone except the poor sap at the home end of the last mile.
Just wait 15 years when the middle class has been struggling with easily accessible gambling and it can't be explained as problem of character. There will be laws passed and people prosecuted or successfully sued.
Those folks that did fall to it, then became (often) lower class while failing to it.
The thing to realize, is that the upper classes ‘eat their own’ just like any other. It’s why Trump is as frantic as he is, he knows what will happen when he stops being ‘useful’/necessary.
But even if that wasn't a thing - the way it's ruined watching sports now, with the constant odds flashing, etc, I'd ban it JUST for that, on top of all the detrimental effects on society.
I don't remember "I can't throw money away on this football play" being a massive society wide problem that needed fixing in 2010, pretty sure everyone could bet with their friends already.
They can also just go to the local library or Starbucks for the WiFi if they need more.
You can barely even use FB messenger (you need to get messenger-lite).
I only know this cuz tmobile would give you free 2g all over europe. it was JUST BARELY helpful. mostly just sms and email.
google maps was unusable etc. This only got worse over the years.
They now give you free 3G and it's bearable. 2G is insanely slow in the 2020+ world.
2G ~= 5 KB/s. That means 40 seconds just to download a properly optimized react bundle.
5MB site? 16+ minutes.
Note that many European countries have already got rid of their 3G networks completely [0]. So it's either "you have 4G/5G" or "the internet is pretty much unusable", nothing in between.
As someone living in a European country with no 3G network, my experience with mobile data is that when my phone fails to find a 4G signal and switches to 2G (pretty much only happens in remote areas, thankfully), I can as well send my packets using a pigeon carrier, they're going to arrive to the destination sooner.
My experience with 2G speeds is:
1. Open job application site
2. Upload resume pdf
3. Upload required picture of ID
4. Server's nginx config has a hard-coded timeout after 1 minute. Connection error
5. Try to upload again
6. Connection error
A huge number of pieces of the web have hardcoded timeouts and limits designed to stop slowloris style attacks, and if your connection is slow enough, those will prevent you from ever being able to complete some tasks.
I've paid for 2GB/mo for years now. I think I ran out once.
2G is really from a different age. Does anyone remember WAP and i-mode? I was certainly not able to afford data back then, but that is what all the business’y types were raving about.
2G EDGE was 384 kbit/s (48 kB/s) (https://en.wikipedia.org/wiki/2G). That means 21 seconds to download a 1 MB page.
I just loaded the careers page at my employer, and the page weighed in at 3.6 MB, so you're talking 75 seconds.
I guess you could configure it not to do that, or write your own imap client with better behavior -- on your 2G smartphone.
"I mean, it's one banana, Michael. What could it cost? 10 dollars?"
In reality this is like 16kb with speeds similar to 56k modem. No modern JS website works. Ssh, irc works and thats about it. Modern stuff like certificates also slows it down.
I'm talking here about real life - not some emulated 2G in your browser.
Guess how I know you've never actually tried this.
Part of my job is testing the web sites I build in the terrible real-world conditions where our customers are. Places like machine rooms, deep basements, and small towns with only municipal or small-carrier 3G cell service. (In spite of what HN believes, there are plenty of places in America with 3G or even zero cell service.)
2G speeds will not work. The device or one of the essential thousands of processes in it will time out because they were designed by tech bubble tech bros who never use their apps in the real world.
https://pgealerts.alerts.pge.com/outage-tools/outage-map/
When the power's out and broadband is down, if you are lucky the cell network is still up. However, everyone fails over to it simultaneously, so there's no way you'll get 2G speeds out of your 5G plan with 5 bars.
I throttled my browser dev tools to 2G and clicked reload. After 2+ minutes, it popped up a "It looks like you are on a slow connection" modal (sometimes this loads on top of the outage info, obscuring the data you want!), so I clicked "Use low bandwidth version". After 51 seconds, the debugger says page load is finished (for the low bandwidth one), but the page is just a white background, and the browser "loading" animation is still running.
After 2.17 more minutes, I get a form where you can type in an address + a bunch of irrelevant info. I typed an address, waited a minute, then typed this paragraph. After 1.5 minutes, it showed my address in the autocomplete menu, and I clicked it.
33 seconds later, it started loading a google maps ajax. 47 seconds after that, it displayed a header, with a white body. 1.33 minutes later, ignoring styling, it returned the string "power is on".
(No map, etc, at this point.)
Had this been a real outage in a storm, I'd be standing outside in the rain, or in a situation where getting cell coverage for more than 30 seconds at a time is impossible if you're holding the phone. I usually just put the phone on a carefully placed chair, then back away slowly.
Anyway it takes 9 minutes under ideal scenarios (no drops) to send 30 bytes of address, and an 11 byte response. 44 bytes / 540 seconds = 0.08 bytes per second good put.
For reference, voyager can send 20 bytes per second, so it's 250x faster than this. Morse code / telegraphs are typically sent at 1.5 words per minute, where a word is 15 characters, so 0.375 bytes per second = 4.6x faster.
Remember, these are simulated conditions for the PG&E site. My record best time getting it to load during a major outage is 45 minutes, not 9.
But I can promise you from sitting with them dozens of times things like Google Maps are unusable once the connection is throttled. It might load some of the map, some of the time. But it never loads all of it and it is just plain unworkable. Even if it loads some of it it takes so long that the busses have gone past by the time they've tried to figure out what direction they need to go.
The worst thing is load balancers with a 10 or 20 second timeout, because there's almost nothing you can do other than use Opera Mini or something.
At that point, why would you use Google Maps at all? Osmand will do the same thing, and requires no connection.
If you assume that mapping services on a low-to-no bandwidth connection are important to them, they'll hear about it through word of mouth. Anything that solves a real problem will spread that way.
Contrapositively, we can conclude that mapping isn't much of a problem for these people.
The signal is terrible, but it’s there. You can talk on the phone or send texts.
Surfing is horrible. At times you get great speeds. Two seconds later it feels like slow dial up. Really that’s what it feels like most of the time, any kind of speed is the anomaly.
As said in other comments, very few apps actually handle this well. They seem to expect that you either have a good connection or nothing.
It’s been like that for a decade plus. I assume it’s just overloaded and will never be fixed.
0 of course, but wasn't 3G all shut down in the US in 2022 to open up the airspace?
One of those HN myths that comes from only being willing to Google (or ChatGPT) information, rather than encountering it in the real world.
3G still exists in rural and remote areas that no major carrier wants to serve, at least as of April, 2025 — the last time I did a round of real-world web testing. Next round is in September. Maybe with 5G in the cities, some hand-me-down 4G equipment has made it to the places where I test.
https://www.fcc.gov/consumers/guides/plan-ahead-phase-out-3g...
The FCC page you linked is talking about major carriers decommissioning 3G.
The grandparent comment is talking about rural/remote areas that no major carrier served in the first place.
This doesn’t jibe with my experience trying to make phone calls on rural highways, where it seems there is no signal whatsoever more often than not.
I suppose this could be because ATT-Verizon-T-Mobile used to have 2G in that area (which was discontinued — 900Mhz analog voice band, also decommissioned) has moved on and left swathes of the US without signal, whereas, certain areas (commenter omits an example) never were served by major telecoms and have “evolved” their tech more slowly, so 3G is not decommissioned in those places. In that sense yes there is no contradiction. It still feels like we’ve gone backwards since there are places I used to be able to make a phone call that are now considered remote area with satellite SOS being you’re only way to reach someone
But even beside that, AFAICT USCellular shut down 3G in January 2024, Appalachian Wireless in Dec 2022, Cellcom in Dec 2023, and C Spire sometime in 2022.
I'm interested to know where exactly public 3G still exists in the USA.
Thank you for proving my point, that people on HN falsely think they know more than others because they can Google a link, even though what's happening on the ground is entirely different.
Reality ≠ policy papers, press releases, or web links.
You'd be surprised how far you can get with that. IRC works just fine (as long as you use Quassel w/ Quasseldroid), HN works well, so does reddit (via redreader). RSS readers and wikipedia work as well, and for general web browsing you can set up a readability proxy (basically Firefox' Reader Mode, but server-side). And of course email works really well, too.
The issue isn't really living with 500 MB/month of data. For most people, it will simply be knowing that you can do that. The next issue they will face is having the technical ability to actually do so. Then, once you've done all of that, the question will remain: will they be interested in the stripped down Internet. A lot of us who frequent HN may be since the results will still reflect our interests. There are people on IRC who we would want to talk to. There is a slant towards tech sites with RSS. And so on. That isn't going to be reflected in sites targeted at a general audience.
Some of these guys have been locked up for 40 years straight. They're not doing all that extra stuff. They want to go on Indeed or Monster or YouTube. One job site I had to load on my desktop to find out why it wasn't working for them, only to discover the pages had a 250MB payload of random crap downloading, including videos.
But as a practical matter, what people rely on phones for are services that are app-based. Good luck completing a Venmo transaction or any amount of banking.