Yes.
For the lifetime of almost everyone alive now, reading, thinking, and writing have been valued skills which moved one up in society's hierarchy. This is a historical anomaly. Prior to 1800 or so, those skills were not all that useful to the average farmer. There were more smart people than jobs for them. Gradually, more jobs for smart people were developed, but not until WWII did the demand start to exceed the supply. Hence the frantic technical training efforts of WWII and the following college boom. This was the golden age of upward mobility.
It's hard to imagine this today. Read novels from the 18th century to get a feel for it. See who's winning and who's struggling, who rises and who falls, and why. Jane Austen's novels are a good start.
The nerds didn't take over until very late in the 20th century. There were very few rich nerds until then. Computing was once a very tiny world. You could not get rich working for IBM. The ones who left and got rich were in sales.
So what was valued? Physical robustness. Strength, perhaps brutality. Competence in physical tasks. Honesty. Parentage. Birth order (see primogeniture.) Those matter in per-technological societies, and they matter in failed societies now. Those are perhaps humanity's core values.
That may be where we go once AI does the thinking. That's where we go when smarts are not a scarce resource.
This is really bleak to me. We can do better than primogeniture, and of course the gender discrimination that goes along with it. You might as well write that subjugation of women is a "core value", simply because it has been for so many time periods.
> Physical robustness. Strength, perhaps brutality.
John Henry is not going to beat the steam shovel any time soon.
> For the lifetime of almost everyone alive now, reading, thinking, and writing have been valued skills which moved one up in society's hierarchy. This is a historical anomaly.
It's not an anomaly; rather, it's the other way round. These used to be highly specialized skills that carried significant status, and got democratized by mass education in the 20th century.
We're not prisoners of history. We don't have to go back to being serfs for the few people who own all the land, oil, food, energy, data centers, and operating systems. I hope.
In fact, there are few things less discriminatory than a random birth order. You may as well be assigned a random number at birth, and the lower your number, the more you're paid. In such a system, there's nothing to discriminate against; the ordering is absolute and immutable, and everyone is treated equally.
I agree that it's a bleak idea, but Animats wasn't talking about subjugating women.
I wonder how long before sex does actually become mutable.
The algorithms and bots that curate/generate content directed by accelerationists definitely want people to think that. There is a whole system in place now that can shape future outcomes just by convincing everyone that have no power when the opposite is true. The parent is probably a bot, or has been influenced by one to many there is nothing new under the sun solipsism bs.
Unfortunately, that is the current stage of humanity. We all currently live in a global subscription model for food, housing, safety, etc. No doubt that we will move beyond it eventually, but the current organization of society is kept in place by the owner class which benefits from the current arrangement.
One of the steps for moving beyond it is educating the modern day serfs (our peers) about reality as it is and alternative visions of a future where we are no longer selling our labor to the owner class. It will take generations.
You might still only be a farmer if you're smart, but you can at least be one of the more productive farmers with a more smoothly running farm.
Where does the human in the loop somehow manage to utilize super-human reasoning better than another person?
I'm not suggesting it's impossible, so much as wondering if we can reach a place where the human is truly irrelevant to the process, and can't make a better decision than the superhuman entity.
I'm not sure this is ever possible. It's more of a thought experiment. What's between here and there? Right now we can use pseudo-intelligence from silicon to our advantage, and being smarter than average is clearly a massively outsized advantage. It's similar to how being able to automate tasks gives you an outsized advantage, yet in so many more ways. But what if that advantage thins or even vanishes?
"Load bearing phrase", as they say.
A stupid ass that just keeps pushing on often goes further than a smart ass who gets distracted.
To extrapolate from fewer people were formally educated or literate to intelligence wasn’t valued is absurd.
As for your part about reading and writing. Literacy has always been a very valuable skill that would increase your social standing. It was scarce and difficult to acquire before the printing press, but it was always valuable.
People seriously underestimate how underpowered and tiny llms are for the tasks they need to solve.
A trillion parameter model can't tell the difference between left and right. We will need to grow them millions to trillions of times before they are half as good as AI boosters claim they are.
This isn't the end of thinking any more than the watt steam engine was the end of horses. It will be centuries before we get there. And by that point the difference between man and machine will be at best academic.
These are very important things and most of them take place on longer or much longer timescales than a few months. Early humans weren't monkeys, and after they had left the tropics, they couldn't survive without planning because getting food is difficult in winter.
Some of these things you could make yourself or were commonly self-made instead of buying, but that, too, requires planning and discipline.
I'm a bit shocked that some people think of medieval life as something like Elbonia in Dilbert comics. Heck, I even find the middle ages a pretty boring time in human history, but I know enough to understand that it wasn't as simple as "everyone lived in the mud and ate mud".
You could buy some temporary luxuries to enjoy, or save yourself a bit of labour on something you'd normally do yourself. But you couldn't really invest in your future the way we would today - everything you depended on had to be something you could make yourself, buying an implement you couldn't maintain would be setting yourself up for trouble. Increasing your productivity with tools wasn't a huge help because you always had enough labor available to hit severely diminishing returns on the land you owned. And any object of value is always at risk of being seized by the local lord or a passing army or what have you.
"Oh well, we were in an anomalous time of social growth, time to go backwards! We won't even need to read or write or think! It's all just too bad, but that's just the way the world works, like it did in 1800." [or pick your date before any current person was alive]
Lots of people have started considering a time of significant "progress" as "an anomaly", as if the world should always just be the way it was in, say, 1800, like that was actually the realistic pinnacle of human society. You also seem to be loosely basing this argument on the availability of "rich nerds", which seems like a bizarre non-sequitur. Computing once didn't exist, and we still valued reading, writing and thinking.
I'm kind of baffled by how regularly I see comments like this. Like, come on. This is basically the AI black pill, no?
I think, in a way, the Internet itself is the virus. It has infiltrated us and our minds. Rage and suffering are what get clicks and engagement. The Internet has become a suffering engine, which spins angst into gold.
Basically England Circa 1851, plus democracy.
And because it was all put together more or less accidentally, it can all fall apart. So worrying about that and trying to do something about it is like discovering that under the deck of the ship are engine rooms, rudders, riveted steel plates and navigation maps.
Its not a slight on your friend, but one would expect him to have a mental model of a rudder, even if he does not know about the impact of cavitation.
More Black pills flying around are just an indication that the rudder is hanging off or the rivets are leaking a bit. It can be fixed, as long as no one tells the passengers the ship is actually flat or the engine room is how elites maintain power.
A mix of political scientists, politicians, investors, entrepreneurs, lawyers, judges, scientists, technologists, and economists have tried to mold society to their own theoretical vision for at least 150+ years. Society then reacts to that in both good and bad ways. This distorts the vision, as society changes it to its concerns. And the cycle repeats.
I think of Karl Polyani's The Great Transformation has a great way of looking at the attempts to force "market society" on England in the 1700 and 1800s, and the reaction that all societies exhibit in the face of unconstrained technological or economic change. Both the imposition of change and reaction to it can be violent, it's hard to predict. We've had such a relatively steady state since WW2 in the developed nations that we're not used to this cycle.
Today? On an engineer's salary? Ha!
Which you mostly don't need today. You may have a lawn guy, take your car to a mechanic, and use a washer/dryer to do most of your laundry.
There are jobs that demand robustness, but they are about applying knowledge in extreme conditions, not about letting an AI do the thinking.
Not really disagreeing with you, but there are a few obvious examples. A lot of construction jobs are still labour intensive, and I've seen a lot of people who don't last the first day, let alone their first week. Also, security jobs, say in nightclubs, also value physical robustness. Orderlies in hospital, require the ability to move bodies, alive and otherwise. The machine is usually better.
Every professor at any university has a dozen more project ideas than they have graduate students, every factory boss has a dozen more optimisations than ways to implement them, and looking up into the night sky we have 95% of it that cannot be explained.
The gap is not too few smart people, nor too few "jobs" that need smarts. The gap is being prepared to arrange society and wealth so the "job" is discovery, science, sharing. We are no longer hunter gatherers, no longer a feudal society, perhaps we shall stop being whatever this one is and try a new one.
(and no, I don't think there is a name for the new one yet (its not socialism, maybe not capitalism).
Lets just not fall back to Feudal if we can help it
We don’t use chimpanzees for any knowledge work today, even though they’d be better at it than some other animals.
Humans with many agents will be more productive, but the tendency has been for these models is to regress to the mean when it comes to strategic insights.
What's interesting now is we might have more lead time than workers in Manchester in 1830 had—enough to build the scaffolding before the transition completes rather than after.
Mondragon is maybe the most useful living data point for what "arrange society and wealth differently" could look like at scale: worker-owned, competitive in global markets, weathered 2008 considerably better than comparable private firms. Not a utopia, but running code rather than theory.
The harder design question is what institutional form fits a world where AI handles most cognitive labor. My hunch is something like collective ownership of productive infrastructure—not the Soviet model, but closer to how water utilities or national parks work: you don't individually own the Hoover Dam, but you collectively do, and value flows back. The challenge is building that before the window closes.
Knowledge is a unique resource compared to the other traditional factors of economic production (land, labor, and capital). It is often invested in with capital (education and tools), but it is carried with the human, and leaves with them. It is always decaying - knowledge workers should be in constant learning mode, and stale knowledge eventually becomes a drag on performance.
I'd argue the future is about knowledge workers all becoming managers. When you use agentic AI, it has the flavor of the skills of management. Management is "a practice and a liberal art", according to Drucker, one that has been in poor supply for some time. LLMs are have somewhat stale knowledge and require the human, tools, and RAG to freshen it. And LLMs will always regress to the mean. It is pretty good at pattern analysis and starts to get shaky and mediocre with synthesis. It requires very nuanced, and elaborate prompting to shape its token output towards insightful results that aren't a standard answer. For coding exercises, that can be fine, but at high complexity levels, or when dealing with issues of strategy or evaluation, it is a platitude generator and has no unique competitive advantage.
In other words, competent, talented management mixed with knowledge work is the scarcity we are heading towards. This is arguably why you're seeing the rise of "markdown frameworks" that people swear improve performance, it's the beginnings of management scaffolding for AI.
Technical folks struggle with valuing management skills, and I expect this will increase its value and scarcity.
As for "Physical robustness. Strength, perhaps brutality. Competence in physical tasks." I think the robots will be replacing that pretty shortly.
"Honesty. Parentage. Birth order (see primogeniture.) Those matter in per-technological societies, and they matter in failed societies now. Those are perhaps humanity's core values."
Ehhhhh not really? What about Christianity, where the meek shall inherit the Earth, and love is the core value (putting aside modern day Pharisees and Charlatans that twist the underlying value system)? Or Islam, whose core value is submission to God? While there have been Societies that valued parentage and birth order, that's far from universal.
Here are some words to live by[0]. I don't agree with everything Derek Silvers says, esp about philosophy. Its more of a guiding principle that drives rather than divides.
The comparison to the adoption of automobiles is apt, and something I've thought about before as well. Just because a technology can be useful doesn't mean it will have positive effects on society.
That said, I'm more open to using LLMs in constrained scenarios, in cases where they're an appropriate tool for the job and the downsides can be reasonably mitigated. The equivalent position in 1920 would not be telling individuals "don't ever drive a car," but rather extrapolating critically about the negative social and environmental effects (many of which were predictable) and preventing the worst outcomes via policy.
But this requires understanding the actual limits and possibilities of the technology. In my opinion, it's important for technologists who actually see the downsides to stay aware and involved, and even be experts and leaders in the field. I want to be in a position to say "no" to the worst excesses of AI, from a position of credible authority.
The different however is network effects. When we make a place better for cars, I make it worse for pedestrians. Your adoption of the car, and its pressure on my lived environment, has effects on me. Same as, say, people joining facebook or twitter. But do LLMs create network effects that are directly harmful, or is it just a matter of making it harder to compete, just like a mechanical watchmaker has less business now that it's so easy to have a reliable clock? Because the first case is a problem, but the second one... that's competition. It's civilization. And then it's not really a matter of cars vs pedestrians.
You say it in a way that it sounds like automobiles don't have a positive effect. I don't agree - they have some negative effects but overall they have a vast net positive effect for everyone.
The upsides of automobiles generally all exist outside of the 'personal automobile', i.e. logistics. These upsides and downsides don't need to coexist. We could reap the benefits without needing to suffer for it, but here we are.
Cars became a self-reinforcing driver of individualism, especially in net new geographies. The negative effects are resisted better in societies/regions that were built long before them. (For both the cultural reasons and plain physical reasons, like not having wide enough roads).
In the car centric places, a few generations later they become an indelible aspect of nature. It is impossible for most people to imagine society working otherwise. And even when they do, the collective action problems are near insurmountable. The introduction of technology has irreversibly trapped us in a way of thinking we can't escape.
This is exactly the premise of the Amish religion. You must strictly control technology to create the society you want, not the other way around.
Something I recently learned about roads from Stewart Brand's new book "Maintenance" is that the first groups pushing for paved roads were cyclists:
The Good Roads Movement of the late 19th century began as a grass-roots
crusade to improve roads for bicyclists. By the 20th century, it had turned
into a national effort embraced by the automobile industry, railroad tycoons
and presidents.
https://www.governing.com/context/how-gilded-age-bicyclists-...The advocates of the automobile have been far, far more successful at shaping US society, laws, culture and our physical environment.
I imagine that’s also true in many other nations to a lesser extent.
“Anti-personal mobility advocates” do not exist. Transit advocates exist, and improvements in transit also massively benefit those who need to or prefer to drive.
Motorists hate anything that isn't a car and is in their way. Motorists hate Critical Mass; they hate light rail or streetcars that hog their rights-of-way; they hate pedestrians (especially when pedestrians aren't wearing the right clothes); they hate Lyft, Uber, and Waymo especially; they hate big trucks and they hate Amish people with horse-drawn buggies.
Motorists will establish coalitions to vote against public transit measures in their home towns. They have come out in City Council and other public meetings, to protest and rail, so to speak, to rail against the expansion of light rail into their neighborhoods, because not only do they hate the construction, but they hate the "type of people" that light rail brings, and ultimately they hate the gentrification that comes from a fixed-route project that will ultimately close their shitty exploitive businesses and replace them with more elevated exploitation and richer moguls.
My assertions have nothing to do with “online circles” except here where I am breaking the bad news to y’all.
I don't deny the general idea that motorists in the US tend to have a crab mentality on the road where they want and expect everyone in the road to only be other drivers. I've also been sneered at in various ways in every non car form of transit I've been in.
- People riding them on sidewalks to putting pedestrians in danger
- "Parking" them right in front of someone's gate, blocking the entrance to their house
- Obviously drunk partiers using them in lieu of getting a ride or taking the bus
- Groups of them sitting around half knocked over completely blocking a sidewalk or other pathway meant for cyclists, runners, walkers, and other pedestrians
Fair or not, you're like the kid using a razor scooter at the skate park. Nobody likes you but it doesn't mean they hate everyone at the skate park. They just hate scooter kids.
at least in England, if you use an e-scooter while under the influence of alcohol, that equates to a motoring offence whereby incurring (car) driving licence penalties, driving licence disquaifications (bans), fines, and imprisonment all apply, depending on circumstances and severity. I'm not sure if/why it would be different anywhere else
And any bike lane infrastructure would benefit e-scooters anyway, so riding them in the road at 30mph below the flow of traffic is a sad hill to die on.
This is true for any kind of transformative technology. Marketing and lobbying can only get you so far. If something has enough utility, it will be used regardless of what people say they want.
I think this is somewhat of a chicken and egg problem. Cars' utility is undeniable partially because society has twisted itself thoroughly around The Car being an assumed part of it. This societal change was both pulled (by car customers) and pushed (by car manufacturers).
The idea that cars needed all this infrastructure that other alternatives didn't just doesn't match the reality of the history of the automobile. And yes, we've leaned on those advantages in the century since, which has also created vast areas where a car is necessary to participate in society, but we only did so because the advantages and utility were so undeniable.
- We would have gotten most of the social utility of automobiles, without most of the social negatives, if personal vehicles had mostly never happened.
And implied from that, we should stop having them now.
Given the known ills of society, I think those negatives are pretty uncontroversial. To the point that personal car proponents have some burden to explain why we should keep it up.
Anyone who still even has a personal opinion at all pertaining to what the world should look like distinct from swallowing whatever 'the market' has decided to impose on them is worth listening to.
That's the most interesting thing about the situation of technology today. Most technology is banal, what's notable is that apparently now a culture needs to be in possession of 'objective truth' (no such thing exists) to defend what is, by definition, a subjective way of life.
It's not like you're living away from any people - you have 100 other neighbours living on your street!
Cars came in parallel with a lot of change.
The psychological effects of this are enormous and under discussed.
There’s no free lunch. Doesn’t matter where you are, more people = more crowds.
On the other hand, the benefits I get from that public transport are incredible - it's cheap, it's always there, it requires minimal logistics in groups (no trying to figure out who goes in what car and needs to be dropped off where at what time), it works regardless of my level of inebriation (admittedly I've not pushed that one to any sort of extreme yet), it's safe enough for children to travel independently (no dropping them off and picking them up), and it's largely accessible for people with difficulties walking or moving about.
I think a big part of the issue is that people have tried out poor public transport infrastructure and recognised - often correctly - that their car is way better for them. But good public infrastructure can often be far more convenient than cars, it just requires people to be motivated enough to build and finance it. A neighbour of mine didn't notice his car had been towed for a week because he used public transport so much and so rarely touched his car. When he'd parked his car it was fine, but then they needed to block of the street to do some work somewhere, and he didn't notice they'd confiscated all the cars there. That's the sort of effect that good public transport can have - so comfortable that you can forget you even have a car.
In order for someone else to have those benefits, they also need a car.
If as a society, if we could feel the same way about public transit, bike lanes, sidewalks, that you do about your own personal vehicle - we'd be better off.
It’s not like they’re the doctors guild that purposefully restricts the number of new doctors per year.
Yes, obviously there are many negative externalities to a car-driven culture, but just like we can easily become blind to the diffuse societal costs of a piece of technology, I think a culture of nay-saying makes it very easy to be blind to the diffuse value of a piece of technology too.
Loud stinky cities full of pollution and climate change are obviously horrible.
But we easily take for granted how amazing it is to be able to drive to a mountain and go for a hike, or call an ambulance, or go to a restaurant when it's raining out, or safely travel in a city without risking being assaulted, etc.
Internal combustion engines are amazing and horrible.
Many of the scenarios you mentoned aren't even that big of a deal for many. I have walked in the rain many times and somehow I was still ok. You could argue that car culture has made us soft in some ways.
Nevertheless, if we reduced our emphasis on cars in society and the design of our cities to the point where cars were mainly used for those specific cases where cars truly are by far the best options (like ambulances) we would have more livable and walkable cities and ironically cities where it is nicer for those who really really want/need to drive since everyone and their mom wouldn't be driving because they aren't forced to drive to everything. Fewer people clogging the roads like my co-worker who would watch Netflix while she drove to work. Obviously she didn't have a passion for driving but was forced to because she lived in a sprawling metro with terrible transit options.
Yes, you could say that, though I'm not sure who would actually say that seriously.
Say what you will about the ills of the car, but it takes a lot of specific context for them to emerge as the worst option of transport from an individual perspective. Really most of the cars ills are from their collective harms, something most can't appreciate as a tragedy of the commons sort of failing.
If I believed there is a crisis of isolation in the United States and degradation of community, I would first focus on more recent technologies, say ones introduced around 2007, than on technologies introduced in the early 1900s.
In my grandfather's day, one income was enough to support a household, and there was less free work being done on the job, which meant fewer hours and being less drained at the end of the day. And yes, people spent less time commuting, meaning they had more time and energy for socializing after work. But communities were also more decentralized, and population centers had fewer people in general. A big part of the problem is that modern cities can be massive, and invariably funnel people to a handful of work districts, which just doesn't scale. When you double the distance to the CBD, you quadruple the number of people coming in (give or take, it's not exact because we tend to increase density close to the CBD as a response to this). Take it from someone who's lived in a place where cars aren't really necessary, the logistics of urbanization are still a crap experience when you're crammed into a train carriage during rush hour. It's common for people to commute for 90 minutes on public transport in Asian megacities, for example.
I'm also not sure that anyone was claiming automobile technology itself was bad, just that in many places at many times it has been used in suboptimal and harmful ways.
> The upsides of automobiles generally all exist outside of the 'personal automobile', i.e. logistics. These upsides and downsides don't need to coexist. We could reap the benefits without needing to suffer for it, but here we are.
other than as a claim we should not have personal automobiles.
You might be only talking about personal cars, but you've got to remember that trucks share the same infrastructure cars use. Modern city wealth wouldn't be possible without engined vehicles driving on roads (maybe if you went really crazy with rail that could be exception). You take away personal cars and either the infrastructure stays or your city wouldn't be possible anymore either.
But even beyond that - personal cars provide a level of freedom and capability to the general population that no other technology can match. Trains suck, buses suck, passenger ships suck, planes are uncomfortable (but otherwise pretty good). Bikes don't work with long distances, multiple people, the infirm, winter (riding in the winter is a great way to get injured, two-wheeled vehicles don't do well with ice), bad weather, if you need to be presentable when you arrive. Oh, and bikes get stolen. Constantly.
You can implement all kinds of transport badly. Trains can suck if they don't take you where you want to go, bicycles suck if wherever you live doesn't provide acceptable parking methods.
Cars are great in a vacuum, but once a city decides it's going all in on cars and bulldozes the place, they provide problems for anyone else. Buses will suck because they're stuck in traffic and walking will suck when you're getting around on the side of 3 lane highways or vast surface parking lots. Most importantly, driving will suck, because everyone has to drive everywhere, and that creates more traffic for the rest of us. You get in a doom loop where you build more lanes, which drives more vehicle traffic. If you make the alternatives more viable, people take up those alternatives and vehicle traffic eases.
Parking is one of the biggest upsides of bikes IMO.
Bikes are great, I ride mine whenever I can. But most places lack secure bike parking and the police don't take bike theft seriously. So sometimes I drive my car even to places where I could easily ride a bike just because I'm confident the car will still be there when I get out.
Replacing the bike is actually a lot easier than getting the windows fixed IME.
If I can't find a good spot to actually lock up the bike though I will just bring it in to wherever I'm going. Shops or restaurants don't seem to care if a bike is parked in the corner and you can thread your ulock through the wheels and make it useless to ride off with.
This doesn't scale to wider bike adoption, though.
I think that's true at the moment, but only because there's so little demand for it. You can always find a sign post or something because no one else is snatching them up.
At the end of the day bikes are still private vehicles and, though they're smaller than cars, they aren't that small and the infrastructure to secure them (which is integrated into cars) isn't small either. So you get the same problem writ small.
Heck—if you have a car & your building doesn't have parking, you're basically screwed. If you have a bike & it doesn't have a bike rack, you can just carry it up & put it on your balcony. At that point, I don't think you can really compare the two.
Buses are only workable because of cars. We build roads for cars first and trucks second. Buses are at most 3rd in the list and getting to use them is an incidental side benefit.
No one builds enough roads for buses. They have to use the roads built for cars.
We have so thoroughly remade society in the service of cars that it can be difficult to recognize any possible alternative.
Even the cited system seems to be limited and exists to connect with trains as well as buses that use normal streets. Wikipedia says that they chose buses for this expansion instead of trains specifically because there was already a strong bus system, which uses the same city streets as cars and trucks.
However, you're making my point for me. If you fail to invest in good public transport it will suck. That is downstream from designing your society around cars instead of transportation for everyone. Bikes do not work for extremely long distances (although school children here will happily pedal 10km to school and back on the daily), but those long distances are a requirement precisely because infrastructure is designed around cars. Even so you can take bicycles on trains and use them for last mile transport at your destination, or store a bicycle at your destination train station (most have lockers or guarded storage) if it's a commute.
Regarding bad weather; if winter is bad enough for bicycles to fail, then certainly it is not safe to drive either, and lethality is orders of magnitude higher. Generally though people here ride bike paths that are shovelled and brined just as the roadways are.
Bikes have their own infrastructure that they do not share with trucks. It is for human beings only.
Here's some reasons to hate cars. https://www.youtube.com/watch?v=umgi-CbaSRU
This is a big claim with no justification.
Cars have dynamic traction control, internal temperature control, etc. You may get frost bite on your bicycle, but almost certainly not in your car. Having four wide wheels makes the vehicle radically more stable.
Add seat belts, air bags, etc. cars have far more safety features than a bike can.
Of course, cars go faster and going faster increases lethality at the limit. No argument there, far more people die in cars in general. But specifically concerning weather, cars allow people to do many things that a bicycle cannot.
Not to mention general comfort. Being in a bike in a snow storm is very unpleasant!
Maybe a severe enough snow storm? Even then we’re in Goldilocks territory for the storm to be unsafe for bikes but safe(ish) for cars.
The biggest factor is that people simply will not get on their bikes in severe enough weather. At least not in most places. Maybe in the Netherlands they’ll bike in a blizzard.
Hail comes to mind. Lightning possibly (I believe cars are much better insulated against lighting strikes). High winds could easily push bikes around / knock them over where cars just keep going.
We drove our van through a forest fire (Cedar Creek Fire - a BIG one) and got a bit of smoke, but otherwise, just fine. No way would I have attempted that on a bike - the increased aerobic activity alone (to say nothing of embers / ashes / etc) would have probably caused crazy amounts of smoke inhalation / death.
And there is a reason drivers hate SOME bikers - here in CA, many simply refuse to follow the rules of the road. My light turns green, and 5 seconds later, some biker comes rolling along in the perpendicular direction - I almost hit him. This kind of stuff happens over and over. I am very fond of bikers when they follow the rules - I bike sometimes too.
Riding a bicycle while wearing an unpowered respirator/face mask is surprisingly easy, especially if it has an exhalation value. It does restrict breathing somewhat, but breathing isn't usually the bottleneck when you're cycling. This might even be the optimal way to escape a fire if the roads are congested.
What does this mean? This feels a bit like a distinction without a difference, as the infrastructure built is shared by both.
> although school children here will happily pedal 10km to school and back on the daily
How flat is it there? I can’t imagine a typical kid biking 10km each way around me. I feel like the average kid at my kids’ school would take 45 minutes or more to bike that distance.
I guess I wasn't clear in implying my doubts as to whether that's a hard requirement. Trucks are much larger and heavier which takes its toll on the road surface itself. They don't need access to suburban environments. Even in the inner city here trucks are banned outside of loading and unloading hours to foster a walk-able environment. So yes, in part they do, but it's not that black and white.
>How flat is it there? I can’t imagine a typical kid biking 10km each way around me. I feel like the average kid at my kids’ school would take 45 minutes or more to bike that distance.
Famously pretty flat, but with e-bikes gaining ground, elevation changes don't make much of a difference anymore. And yeah a 45 minute commute by bike is not unusual, but remember, we have the safe infrastructure for it. Kids bike in from villages surrounding towns and cites.
How are suburban environments stocked then? Surely village grocery stores are not stocked with milk one bike load at a time.
> Even in the inner city here trucks are banned outside of loading and unloading hours to foster a walk-able environment.
Sure. But they use the same infrastructure. The fact that the vehicles are built for different purposes and may have different regulations doesn’t mean the cost of infrastructure isn’t shared. Pervasiveness of roads makes it easy for cars, trucks, ambulances, buses, and even bikes to get around more easily.
Just like the pervasiveness of the Internet make it easy to scroll TikTok, purchase goods from Amazon, and read books through Project Gutenberg, even though those are very different use cases.
Extreme hot weather and pollution are both a much bigger health risk for bikes than cars.
To blithely state that nobody could make such a claim seriously is an attitude which actually has a really fitting term: carbrained.
"I don't know anyone who seriously thinks that stone applied to fibrous asphalt is not a fine roofing material"
"I do not know anyone who seriously thinks that 4000kcal/day is healthy in normal circumstances"
"I don't know anyone who seriously thinks that women are incapable of working outside the home"
"I do not know anyone who seriously thinks a bright red suit is appropriate for a funeral"
And on and on and on.
But we both already knew that. So if you're gonna be obtuse and not understand it I'm gonna be obtuse and explain it.
> living in dense inner-city areas did not carry the highest depression risks. Rather, after adjusting for socioeconomic factors, the highest risk was among sprawling suburbs
another interesting tack: how long did we have cars before we started talking about a widespread mental health crisis? is there a more parimonious explanation, like a different event that is located closer to it in time? perhaps smartphones or the internet?
Another thing about "this mental health crisis" is that it has been ongoing for many decades before we noticed it and before it was brought to the forefront. "One Flew Over the Cuckoo's Nest" was out and then President Reagan approved the mass closure of asylums. What happened was that massive numbers of citizens had been condemned and committed by their relatives and "put away" in homes, facilities, and institutions, and then Reagan shut 'em all down.
Today, the mentally ill live among us. Either their families care for them, or they live in jails/prisons because they became criminals and were convicted, or they live independently/on the streets. The mentally ill live now in "virtual institutions" where their chains and restraints consist of drugs. The drugs are what keep them connected to their home clinics and their psychiatrists. The drugs keep them coming back for more, month after month, to their pharmacies and clinics. The drugs they are convinced they cannot live without, making them compliant and unsure of what is really going on in their lives.
The non-criminal mentally ill are mostly encouraged to integrate and socialize, to seek employment and try to simulate functional human beings in society. So they live among us and they are causing more noticeable issues when they interact with people possessed of more sanity. The mentally ill are probably less likely to drive or own a vehicle, and more likely to rely on public transit, so you know where to find them.
But the mentally ill who live independently, and live with these "virtual restraints" are likewise living in fragmented neighborhoods that are not walkable and require a lot of effort to overcome the sheer distances that separate them from services and their employers. They're living among immigrants, foreigners, heathens and infidels, and on every corner is a moral trap such as easy alcohol, easy sex, easy gluttony, easy gambling that can ensnare even the sanest city dweller. These traps are, of course, legitimate businesses that cannot be shut down by a mere vice-squad raid.
So "this mental health crisis" in 2026 can perhaps be partly traced to the advent of personal motor vehicles, but I feel there are several causes that have brought it to the forefront.
EDIT: missed your other reply a few mins earlier alluding to smartphones already
The idea is that teen mental health got dramatically worse in the early 2010s at the same time as social media began to become ubiquitous, but this is likely a coincidence. The underlying metrics we're tracking here are self-harm hospitalizations, and concerns about teen self-harm were already growing in the early 2000s. This leads to a bunch of new guidance getting published which increases teen mental health screening, tracks mental health status as a cause of injuries, and forces insurance companies to cover associated costs.
It's one of those situations where our stats about a problem increased as we became better at tracking it. Teen suicidality is actually WAY down over the past ~30 years.
Qualitative data is, of course, much harder to work with than hospitalization numbers, but the data we do have suggests a weak correlation, if any, between phone use and poor mental health— see https://pubmed.ncbi.nlm.nih.gov/30944443/, which suggests phones can explain at most 0.4% of variance in well-being among teens. [1]
It feels like common sense that social media is bad for you, and sure, there's plenty of work to be done in understanding how and why social media can cause harm. But the idea that there's some big crisis just doesn't pan out.
Info drawn from https://podcasts.apple.com/us/podcast/the-anxious-generation...
[1]: In fairness, Haidt published a response to this article featuring a new, bespoke set of controls for the data. His analysis suggests that the impact of social media use on mental health is nearly twice as large as that of being sexually assaulted and four times larger than hard drug use (which itself has a slightly larger effect size than wearing glasses). Personally, I don't find these conclusions plausible at all. Maybe Haidt's been p-hacking, or maybe the data set is worthless. I couldn't say. https://www.sciencedirect.com/science/article/pii/S000169182...
Today I find myself in an urban hellscape without owning a vehicle. Nothing is walkable. I am crammed in, thanks to Equal Housing, with immigrants and people of utterly alien races and cultures (I consider myself the minority.) If I expect to find people like me or shop within my demographic, nothing is adjacent and it’s all several miles worth of transportation.
Car culture and forced integration has fragmented every possible family unit that could have been cohesive or collectivist. If I am celebrating a religious or cultural festival, I can count on none of my neighbors sharing that celebration, or in fact raising conflicts on the days most sacred to me.
Anywhere I may choose to walk, or even if I drive, I am trudging through vast empty parking lots of asphalt because of cars. The roads are laid out for cars. A cop told me yesterday I shouldn’t drive my e-scooter at 17mph in the street but on the sidewalk. Every motorist also hates those scooters, whether in motion or properly parked. Every motorist also hates the light rail train and hate for Waymo is fomented by motorist and pedestrian alike.
There is no place I could move to or live that would change this equation in any useful way. I do not hate cars, but I hate what they have done to our lives and our landscape.
We can argue about whether this is a good trade off, but the claim that cars make everyone's life better is straightforwardly false.
I wonder how the Uber driver feels about not being considered a full time employee and unable to have affordable healthcare and a nonexistent retirement plan. Hopefully they don't think too hard about it or that would be incredibly selfish of them.
The only way you receive food (except from your backyard inner-city garden?) is through people DRIVING. The way you receive packages is by DRIVING. They city infrastructure you enjoy is maintained through skilled laborers and tradespeople DRIVING.
What? Of course it is, you can easily impose rules that apply to personal vehicles that don't apply to public transport, logistical vehicles or emergency vehicles.
As an example in my neighborhood in the Netherlands, there's basically no streets around me where personal vehicles are allowed, but there are no restrictions to buses, delivery vehicles, moving vans, or ambulances.
> Any regulation that tries to ban the one while allowing the other would be a huuuge clusterfuck
How? You don't even have to go fancy with specialized license plates or anything like that, it's literally just common sense.
if that is lacking (often is) $50,000 fine per incident will take care of it
I mean sure, they both have engines and wheels, but they're already distinguishable in the eyes of the law. Commercial and personal vehicles are registered separately
Anyway, I don't think anyone is proposing banning cars. Just would be good to provide alternatives
Following the conversation, the subject has not ever been a yes/no referendum on cars.
It was if there has been a moral net positive/net negative for vehicular technology (as a comparable technology to AI)...which has consistently been walked back to a nebulous "personal vehicles are a net negative because of how they make people think". That's eerily close to the views on AI today.
Cars? Waaaay less clear they're net-beneficial.
I was skeptical too the first time I read this kind of argument. I ran the numbers for my case, which was sitting around the median (commute duration) or significantly better than the median (household income, car cost) for relevant numbers, for my car-dependent middling-costs US city, and it was still roughly break-even without even factoring in not being able to make commutes double as exercise time.
I had to have a car. My life would have fallen apart without it, that's how big a benefit it apparently was. Yet if I actually examined what was going on, it wasn't providing any real benefit to me at all, just negating harm done by designing my city around cars. That's how the numbers worked out, much to my surprise. For most residents of that city it was worse, the city being designed for cars was making their lives worse.
Denser, less car-centric areas are more economically productive than less dense areas. Car infrastructure prevents density. So I would argue that, at least in some cases, cars decrease economic efficiency
The development of cities caused by unrestricted, broad private car ownership without lots of careful coordination on that development, is in the reverse situation: it's fairly hard to argue it's net-beneficial, because it's so incredibly expensive in all-told money, time(!), liberty(!!), and, if we'll allow consideration of such things in a basically-economic analysis, pleasantness of environments for humans to exist in.
THIS! I am shocked that some people don't realize that modern civilization and our modern quality of life depends on cars to a huge degree, even for people don't personally drive. Such a lack of knowledge about modern industry and logistics..
In aggregate, benefits of cars outweight the cons for 99% of people. Perhaps if you live right next to a busy highway, you might the the exception..
I'm more shocked that somebody thinks that modern civilization and logistics depend on personal cars. Can ypu expand on your statement that modern industry and logistic depend on persobal cars?
Many car haters constantly play this motte-and-bailey game where they insinuate that cars are evil and should be eliminated, then they pull back and say “oh no, we don’t want to ban them” when confronted. But it’s clear that some subset really would prefer to eliminate civilian vehicles.
I like smart urbanism and pedestrian-centric development, but the anti-car culture annoys me to no end. It is self-defeating. The average person in the US has a car, and likes having a car, so you should start every argument with that assumption. We made a lot of progress on improving pedestrian access in the early 2000s by focusing on a positive message. But I guess there’s no room for non-adversarial messaging anymore.
Self driving cars would make this massively cheaper and remove most of the reasons to own a car. It would make about as much sense as owning a train for most people.
If I bought and sold a house every time I needed somewhere to stay on holiday, renting would be massively cheaper.
I’ve already done the math and uber occasionally is cheaper than owning a car. Self driving electric will be even cheaper.
It's not fine if that choice denies other people the choice not to.
And there seems to be a lot of the latter.
For example, when shopping facilities or hospitals are built so as to be, de-facto, only accessible by automobile, that locks people out of the choice to say no thanks.
Where I live every new development must build out sidewalks as a condition of permitting.
Suburbs/car-dependecy is a classic case of "worse is better". Its simpler to build and the worst-case suburban sprawl is tolerable, so it proliferates.
No, because no such somewhere has been built in the country in question (US) in the past ~60 years, because the default is car-centric. So you're left with a few uber dense, old, predating automobiles, places. Which are extremely expensive, because they simply do not have the capacity for everyone who wants to live in them.
I'm not sure what the alternative would be. Maybe everyone lives in giant 10 million+ population cities that are all connected to each other by rail (and rail connects all airports, harbors, etc.) and then you have to show up at rail station to get your groceries or whatever else?
Yes cars/trucks/busses are still useful overall and are an incredible last-mile solution for freight.
But on a personal level, it means we all must live far apart and maintain our own individual vehicles, along with the average total costs of $11,500/year PER CAR. [0]
I’m not saying they should’ve even been banned for personal use - owning a car and living in a rural suburb should still be an option, but it is very expensive to choose that lifestyle.
However the auto companies on the early to mid 1900s had heavy influence on policy, even buying and shutting down their public transit competitors, converting cities into “car cities”. This is where it drove into “negatives outweigh the positives” territory. Everything before that was more positive, but this was a massive negative on society and continues to handicap cities today, making them expensive and even just dangerous to walk around (due to high speed roads and limited sidewalks)
[0] https://www.nerdwallet.com/auto-loans/learn/total-cost-ownin...
Once you start really marking how much nothing you're driving by even in many cities, where that "nothing" is one or another use of land that exists solely because of cars, it's a bit of a shock. "Wait, work would only be 8 miles away instead of 15 if not for the effects of widespread private car ownership? The grocery store could be 1 mile instead of 3? And I spend how much time a week bicycling to nowhere in particular to make up for sitting all day long? And this car & gas & insurance costs me how many of my work-hours per week, just to pay for it? Hm... am I... losing time to cars!?"
I guess what I don't understand is, given the current state, 1) what do you want? 2) how much will it cost? (and how will we pay for it?) and 3) what are the tradeoffs?
On a related note, it seems like a lot of the anti-car/urban planning wonks have a belief that everyone really wants to walk, ride bikes, or take mass transit everywhere, and I think they're wrong. Most people want to drive personal vehicles.
Maybe if we lived in a world where mass transit had very strictly enforced behavioral norms, more would consider it. But even then, I still think most people prefer the many conveniences afforded by personal vehicles.
Our world would be better without being completely dependent on cars. You can see this in a few select cities or neighborhoods that have avoided the worst of car dependency. There are still suburbs, but they’re a bit more dense and you can easily bike to a grocery store in 10 minutes. There are still rural suburbs, but it’s much more expensive to live there due to the extra effort to get where you need to go.
There isn’t an easy way back since we let the auto industry have such a huge influence in politics, they’ve shaped the world, and it would take us decades and a LOT of money to revert the damage. We can still make steps.
HOWEVER, to bring the point back, we’re still in the 1910’s auto industry with AI. Are we going to let the AI industry get heavily involved in politics and shape our world into a worse one to benefit them? We’re at a point where we can reap the benefits, like with early cars, without the damage that came later
Yes, they are in fact, the same. You cannot introduce such massively useful technology into the world and then say that it would be used only for logistics and not for personal transportation. Short of a worldwide totalitarian government, such arbitrary restriction would be completely unenforceable.
It is possible to shape things with regulation, but only to some degree. With any great technology, you have to take the good with the bad. And the good outweights the bad in any historical technology. AI will be no exception.
On public roads? No reason we'd have to license private cars for that, at least not for just any purpose.
Trying to ban all private cars while keeping our car-dependent civilization working is unrealistic, no matter how you look at it.
There's no quick fix at this point, it'd be a century-long project to undo the damage now, but a hypothetical world where we'd harnessed only the good parts of cars and not let them completely reshape the places we live down to the neighborhood level would sure be a lot nicer.
Less than a century ago, so within living memory (albeit only just), literally nowhere on Earth was car dependent.
The problem is we are numb to it. 40,000+ people are killed in car accidents every year in just the USA. Wars are started over oil and accepted by the people so they can keep paying less at the pump. Microplastics entering the environment each day along with particulate from brakes, and exhaust. Speaking of exhaust: global warming. Even going electric just shifts the problems as we need to dig up lithium, the new oil. We still have to drill for oil for plastics and metal refining, recycling and fabrication.
Yes they ship people around somewhat fast. Slower than possible with other methods, and the cost is incredible - economic (much more expensive per passenger than almost any alternative), political (they inherently divide people, dehumanise and make people never really share a public space), health - they reduce lifespan by both lowering living quality as well as directly killing a staggering amount of humans per year).
And we have learned how to build better places for humans that do not need these coffins on wheels - if you visit any European capital, and most Asian ones - you will see environments built for humans, not cars - soo much nicer.
So cars as a technology have definitely not been beneficial to humanity overall, but it has been somewhat useful to some.
I think strongtowns were very good advocates of what places in America could like if you look beyond cars. I personally like the “not just bikes” channel though.
All they saw was that trips taking a day could now be done in an hour and produced no manure, and that meant suddenly you could reasonably go to many more places. What's not to like? A model T was cheap, and you didn't even need to worry about insurance or having a driver's license. Surely nobody would drive so carelessly as to crash.
*well, not technically nobody, but nobody important.
What’s really interesting is that you can find newspaper columns in the 1920s recognizing what we now call induced demand as even by then it was clear that adding road capacity simply inspired more people to drive.
Lack of alternatives + political unwillingness to provide them + lack of political pressure to provide them + the massive effort that would be needed to build a system from scratch that has already been dismantled, and infrastructure is in the way because it wasn't a factor + corruption, democratic decline, etc. = most problems around cars in the USA.
Remote work was the biggest upset to this system in generations but that's being stamped out at many organizations.
Today we have a much better understanding of the world, so we have the means to think down the line of what the negative effects of LLMs and course correct if needed.
I don't see anything positive about being forced to participate in this car-ownership game where 99% of North American cities are designed around car ownership, and if you don't own a car you're screwed. I don't WANT to own a car, I don't want to direct countless thousands of dollars to a car note, car maintenance, gas, etc. I want the freedom to exist without needing to own an absurdly expensive vehicle to get myself around. There's nothing freeing or positive about that unless all you've ever known and all you can imagine is a world in which cities are designed around cars and not people.
Really these people decades ago had a great grasp on these things. But why did they "fail" and we still have traffic? They didn't fail really, what failed was implementation not planning. Most cities you see with notorious traffic today, chances are the bottlenecks that exist were planned to be relieved by some midcentury road plan that was for whatever reason, not ever built. Comprehensive rapid transit was often also planned, several times over, but not built or at least never to the full scale of those plans. Catalytic converter was also a great success people today probably don't even think about. You can see the mountains again in California's cities thanks to the catalytic converter.
Leaded gas took longer, but I'd say the tailpipe pollution, congestion, and general capacity related issues were well understood.
Traffic gets fixed by getting most people to use some other form of transport and leaving cars to the edge case uses.
No possible way to know
Kind of like how fat and salt are good for you until you over consume. The world has massively overconsumed cars.
Cars aren't a positive in society. Transportation is the benefit, and cars are the worst possible way to transport people. A functioning public transit system is better in every possible way apart from egotistical arguments like "I don't like seeing poor people on the bus".
But any actual urine "spills" would've gotten cleaned up right quick. They simply don't tolerate that stuff. Look at any bus operator and they are tough as nails if someone tries shenanigans.
In fact, the craziest times I've had were in ride shares and taxis. I was all dressed up for a funeral one morning, summoned a taxi, and the driver who tried to pick me up was obviously drunk. Another time, I had a medical transport that was summoned, sat at some gas station for 15+ minutes, and when he arrived at my place and I got in the back seat, it was wet and had clearly been through a hasty cleanup, and I could smell urine permeating the vehicle, and it was absolutely nuts, but I said nothing, because it probably wasn't the driver's fault. Probably.
We (or lobbyists) resist having carbon costs included in the prices we pay at the pump.
Edit: More transportation is good; I am not throwing the baby out with the bathwater, just that our accounting for costs makes things look better than they are.
It is possible to say "some things could be done better" without meaning "throw it all away."
Increased car ownership & use, and increased design of environments to cater to cars, greatly harms that freedom.
Ever checked into a hotel somewhere in city-sprawl, looked at your map to find a local shop to pick up some cable you forgot or a toothbrush or to get some dinner, and realized that despite those things being within half a mile you can't realistically walk to them because there's a highway between you and them, so you're stuck unless you pay someone with a car to drive you? Doesn't feel very free.
I've never had this happen, no. The closest I've ever gotten was in Tokyo, when I had the store I needed in eyesight across the street but had to go very far out of my way to a pedestrian bridge to get there.
You lose that freedom of movement if:
Your car breaks down
Your car gets stolen
Your car gets totaled
You lose your license
You can't afford insurance
You get too sick to drive
You lose bodily mobility
Your mental faculties decline
If you can't drive, you have to depend on whatever public options there are around you. Good luck.
Of course there is an idea of non-toxic masculinity that doesn't just equate to !masculinity. People love to bring up examples of non-toxic masculinity in media. Someone on reddit has even compiled a megalist of examples of non-toxic masculinity in film: https://www.reddit.com/r/MensLib/comments/eb0ir1/a_megalist_...
It is perfectly possible to be both masculine and non-toxic without being feminine. Refusing to allow that is toxic in itself.
Using the term "toxic" to describe things is an issue because people have an immediate negative reaction to it and go on the defence. Wording matters a lot and I'm unsure why there's such an insistence on calling things "toxic" when other words would both better describe issues and cause a less visceral reaction.
> not even an inflammatory one
I don't know how you can seriously claim this.
Society in itself is the act of exchanging some of ones individualism and freedom for a group identity.
Alligators don't have what we call a society, and they do things that we'd consider anti-social like eat the young of our own kind. The individual has ultimate freedom to do whatever they want. Humans consider these freedoms anti-social and harmful to others and restrict your behaviors in these manners by ever increasing punishment including death.
Effectively your statement boils down to a childs tantrum of "I want to do whatever I want to do and damn everyone else"
No it's not that I'm against any those things just the toxic applications of them.
> In practice, both do mean exactly that. "Nontoxic individualism" is collectivism, "nontoxic masculinity" is femininity. You're not slick, everyone gets the language games at this point
An analog might be the push for banning phones in schools. Setting apart times and spaces where serendipitous human interactions are encouraged by the lack of distractions.
> Now might be a good time to call your representatives.
Well, not that must-read I guess
Musks SpaceX Keynote was ridiculous, don't get me wrong, but we will be able to see AI progress in the next 5 years which will give us some kind of gut feeling were the journey can go.
Also AI solves another problem: Compute. It was clear that we want some kind of compute but its like with 4k; We have 4k for ages now but it is not the default resolution on all displays sold. We stoped pushing the boundaries because invest is not here. People do not bother too much with it.
With AI and the richest companies and people want to see what happens, pushes the envolope a lot faster, pushes us to find solutions.
This AI Compute based on ML/Neuroal Networks can also be used for physics simulation, protein folding, and everything else.
Stoping technology is not an option and not a solution. Education is. We need to educate people.
The problem is that the connectivity required for much of AI is very different than that required for classic HPC (more emphasis on bandwidth, less on super low latency small payload remote memory operations) and the numeric emphasis is very different (lots of mixed resolution and lots of ridiculously small numeric resolutions like fp8 vs almost all fp64 with some fp32).
The result is that essentially no AI computers reach the high end of the TOP500.
The converse is also true, classic frontier scale super computers don't make the most cost effective AI training platforms because they spend a lot of the budget on making HPC programs fast.
This tech is 100% aligned with the goals of the 0.001% that own and control it, and almost all of the negatives cited by Kyle and likeminded (such as myself) are in fact positives for them in context of massive population reduction to eliminate "useless eaters" and technological societal control over the "NPCs" of the world that remain since they will likely be programmed by their peered AI that will do the thinking for them.
So what to do entirely depends on whether you feel we are responsible to the future generations or not. If the answer is no, then what to do is scoped to the personal concerns. If yes, we need a revolution and it needs to be global.
It can't. It can't even deal with emails without randomly deleting your email folder [1]. Saying that it can make decisions and replace humans is akin of saying that random number generator can make decisions and can replace people.
It's just an automation tool, and just like all automation tools before it it will create more jobs than destroy. All the CEOs' talks about labor replacement are a fuss, a pile of lies to justify layoffs and worsening financial situation.
[1] https://www.pcmag.com/news/meta-security-researchers-opencla...
The combination of these two things could lead to a situation where there is a massive, startup-dominated market for engineers who can take projects from 0.5 to 1, as well as for consulting companies or services that help founders to do the same.
Another pair of hopes is that a) the LLM systems plateau at a level where any use on complex or important projects requires expert knowledge and prompting, and b) that because of this, the hype of using them to replace engineers dies down. This would hopefully lead to a situation where they are treated like any other tool in our toolbox. Then, just like no one forces me to use emacs or vim (despite the fact that they unambiguously help me to be at least 2x more productive), no one will force me to use LLMs just for the sake of it.
It doesn’t even have to be people with no idea what they’re doing. If you lay off enough smart people from big tech companies, those people might put together small companies that directly compete with larger ones at a fraction of the cost.
These small companies will only be able to sell through basically scam marketing.
I doubt that
Focusing on option 2 and software development, teams and companies will only downsize if the demand for software doesn’t increase. Make the same amount of stuff you do now but with less people.
What I think will happen is that enough companies will choose to do things that they couldn’t afford or weren’t possible without AI (and new companies will be created to do the same) to offset the ones that choose to cut costs and actually increase the amount of people making software.
I am pretty sure these are well known economic ideas but I don’t know the specific terminology for it.
We are already hitting the limits of demand in many areas of life. The fundamental currency that is not growing is human attention.
Sure, now you can be a musician and use AI to help you make an album in a weekend. Great. So can a million other people. Who's going to listen to them? Everyone is already inundated with more music than they could ever listen to in a lifetime.
Now someone who's never written a line of code can vibe code an app and upload it to an app store. Great. So can a million other people. Who's going to install those apps? When was the last time you found yourself thinking, "I wish I had more unmaintained apps on my phone!"?
Now someone who aspires to be a "writer" but lacks the willpower to craft sentences can throw a couple of bullet points at an AI and get a thousand word article out. Great, so can a million other people. Who wants to read more AI slop text on the web? There are already a million self-published authors whose books never get read. That's not going to get better when there are a billion of them.
All of us, every single one of us, is already drowning in information overload and is stressed out because of it. The last thing any of us want is more stuff to pay attention to. All of this AI generated stuff will just be thrown into the void and ignored by most.
You don't need to create the next Facebook, Shopify, X etc.... Because it already exists and controls the market.
Mass unemployment, consolidation of all AI-related benefits in the hands of a few, an increase in demand that doesn't outpaced the loss of employment, increase in capabilities (not AGI) that mean a few chosen people can do most things without hiring other people, etc.
I know it is the classic sci-fi dystopia where somehow despite endless advances in tech and automation, the masses can't figure out how to make it work for themselves and end up living in shanty towns on top of each other waiting for gifts from the elite, or scraping in dirt outside the cities, but come on... I just don't see that as being credible.
> They want us spending lots of money on their products, so their wealth increases.
If we're considering scifi scenarios, imagine this: if full blown automation of everything is achieved, why would the "haves" need the "have-nots" buying anything at all? Why would they need them to exist, at all? Think about it. It's an extreme and we're not near it... yet.
> despite endless advances in tech and automation, the masses can't figure out how to make it work for themselves
If the tech (or the really helpful tech) is guarded behind a lock, and they don't hold a key, it's not a matter of figuring things out. Unless by figuring out you mean revolt?
So we reach this post scarcity society, where everyone could be living a life of luxury, but this whole group of "haves" as you call them (who would they be?), somehow form this uniform view that they just don't want 99.9% of other people around and let them all die off while they guard themselves in gated cities or something.
It just makes no sense at all to me. Like in a sci-fi novel or movie where it is a plot requirement, ok, but in reality, I just cannot see the path and all the things required to get to that particular reality. So many ways it would work out differently.
A full automation society, where the implied post scarcity is not necessarily for everyone. Maybe it needs most of the population not to exist in order for the few to enjoy the lack of scarcity. Resources aren't infinite, but greed is.
I mean, resources and wealth could be far better distributed right now, no need for AI, yet most times this is attempted the wealthy fight tooth and nails against it, even though the impact for them would be very small. What makes you think having AI will magically make them better people?
> [...] this whole group of "haves" as you call them (who would they be?) somehow form this uniform view that they just don't want 99.9% of other people around
A uniform view on this matter is easier to achieve by an extremely small subset of people.
And really, do you need to ask "who are they"? I mean, the billionaires and owners of concentrated capital of the world?
> I just cannot see the path and all the things required to get to that particular reality.
You cannot see a path from unchecked capitalism and extreme concentration of capital, via total automation, to this particular reality?
It sounds like a failure of imagination. I see the people at the top being lying sociopaths and have no trouble believing this.
In the old days change was slow enough that few people got displaced from jobs requiring any substantial skill (although there was local devastation: for example, court reporters.)
Now, however, we are seeing change happening faster than people's careers. You can not realistically retrain into another high skill job--you're going to be the last to be hired. (There's a good reason Social Security Disability has cutoffs a 50 and 60 for how much change can be required!) And, likewise, someone who has worked a desk for decades is not going to be hired for a physical job. (Assuming they even can do it. I can't think of any physical job that wouldn't have me in a lot of pain in weeks at the most.)
I read this take a lot but I don't buy it. This isn't guaranteed by any means. And even if it does happen, isn't it just as likely that AI is deployed into those companies too and they don't actually result in any job growth?
If you don’t care what individual people think then simply don’t talk to them.
Sorry, you made a claim, there's good reason to believe your claim may not pan out, and if it doesn't the consequences are dire.
> New companies will appear doing things that we can't even imagine yet
I have a really big imagination, so I will believe it when I see it. If you have any real idea what these new companies might be doing in the future then I'm all ears. But until then maybe stop trying to claim some kind of future knowledge based on some handwaved nonsense like "we can't even imagine what the future will look like"
And then trying to claim that's "the reality of the situation", please be serious
Edit: Maybe if you think the future is so unimaginable, you should take a look around at the present. Can you identify anything in our lives today that was not imagined by anyone in the past? Think about how every piece of technology ever made nowadays, someone can say "it's like the Torment Nexus from Famous Piece of Literature!"
And early cars were expensive, dangerous, highly unreliable, uncomfortable, belched foul exhaust, and required knowledge of how to drive AND maintain them. We are far, far from that scenario these days.
Random number generators can't solve open math problems, but it looks like AI agents can? [1]
[1] https://www-cs-faculty.stanford.edu/~knuth/papers/claude-cyc...
It doesn’t have to be effective. It has to make CEOs believe it is effective.
I don't think the comment you're replying to is saying that an evil AI bot will kill people. They are saying something along the lines of: mass job loss doesn't bother the AI companies because in the AI-powered future they envision, population reduction is a positive side effect.
If AI is smart enough to replace the 99.999% it's also smart enough to replace the 0.001%.
Setting people against each other is a time honored way for a small elite to control a large population.
This scenario is not meant to be taken literally.
Energy. The key is controlling their access to energy.
The 99.999% needs to assert their controlling stake in the technology. I don't know what this looks like. Maybe ubiquitous unionizing, coupled with a fully public and openly-trained LLM.
https://www.bentoml.com/blog/navigating-the-world-of-open-so...
No chance of it happening in the US due to lobbying pressure, but maybe in a more civilized country... (unless a distributed SETI@home-type architecture becomes viable)
But that doesn't really matter when we talk about "replacement" because these people don't "do" they simply "own".
They're not concerned about being outpaced at some skill they perform in exchange for money...they just need the productive output of their capital invested in servers/models/etc to go up.
What's important is that ultimately some small subset owns this, and it doesn't matter how smart they are, only that they own the thing and that it cannot be employed against them (because they hold the key).
Or things could turn out more than fine and we progress as we've always progressed, towards more abundance and humans in 30 years will live massively better lives than we live today, just as we live massively better lives than people at just about any previous point in history.
This is blatantly unsustainable.
> just about any previous point in history
The late 1990s is an exception for most people.
It sounds like things are going well for you. Be mindful of psychological projection.
General strike and bank runs.
Not to collapse the economic system, but to present a credible threat of collapsing the economic system which AI development, as these elite and their platforms know it, relies on. When they're freaking out, we call for negotiations.This only works if people with "secure" livelihoods not just participate, but drive the effort. Getting paid six figures or more in a layoff-proof position? Cool, you need to be the first person walking out the door on May 1st (or whenever this happens), and the first person at the bank counter requesting your max withdrawal.
As for bank runs, no one cares. The big banks no longer need retail customer deposits as a source of capital for fractional reserve lending. Modern bank funding mechanisms are more sophisticated than that.
In which the FDIC took unprecedented action, drawing down the DIF to backstop depositors beyond the insured $250k and offering a credit facility to other banks, in order to prevent "contagion" - a panic, a bank run - which was presumed to be likely after the 3rd largest bank collapse in US history. A bank almost no one outside of California had heard of before it died.
Bank runs are serious business, and far from being something "no one cares" about, even just talking about them makes banks nervous, because they can happen to even "healthy" banks. The big banks have been undercapitalized for more than a decade, and even a moderate run on a regional institution threatens the entire system. Which is why it should be done, or at least signaled as incoming; it's good leverage.
>You're free to take a vacation or quit working if you want to. Go ahead.
The implicit, "I'll stay here, where I'm nice and secure," is delusion. People care about your outcomes even if you don't care about ours. Take the invitation to organize with others to secure your own future, to show just how much you're needed before your employer decides that you're not (however erroneously).Anyway, corporate depositors have a duty to safeguard their capital. That means that if a bank run is underway by retail depositors, they're in line too, willing participants or not. This is why, again, even discussion of bank runs is discouraged, and their likelihood and effectiveness downplayed. They're built on turning the imperative of self-interest, which the financial industry is built on, on its head.
Collective humanity needs to think this matter through and take global action. This is the only way I fear, short of natural calamities (act of God) that unplugs humanity from advanced tech for a few generations again.
What? I don’t know anybody who has a layoff-proof position.
>“There are people sitting in our office in King’s Cross, London, working, and collaborating with AI to design drugs for cancer. “That’s happening right now.” https://www.htworld.co.uk/news/research-news/isomorphic-labs...
and
>...enables researchers to move seamlessly from AI-generated sequences to functional antibodies in just days https://the-decoder.com/googles-ai-drug-discovery-spinoff-is...
There may also be downsides, like skipping testing things that would enhance our fundamental understanding of something because the AI was wrong. But that’s already a problem , and having a better gauge in the early stages could be really helpful
Not making predictions that they will, just trying to give an example of a benefit that we may get out of this
It can help a little bit in the early stages of drug design, but even if it was perfect (which it's not), there's a massive gap between understanding a protein structure, and understanding how a drug will or system will interact with it.
In a broader sense, understanding the structure of a protein is only a small part of drug development. Unfortunately biology is complicated, and we're an extremely far way away from solving it.
But LLMs compute requirement is so high that it pushes the boundaries of compute, memory and memory bandwidth which is fundamental for curing diseases.
LLMs math / neural networks can and are used for medical research. Simulating a whole body with proteins, cells etc. will bring us the breakthrough we need.
Nothing in modern medicin research is withoout compute.
AlphaFold def helps researchers around the globe.
> The people who brought us this operating system would have to provide templates and wizards, giving us a few default lives that we could use as starting places for designing our own. Chances are that these default lives would actually look pretty damn good to most people, good enough, anyway, that they'd be reluctant to tear them open and mess around with them for fear of making them worse. So after a few releases the software would begin to look even simpler: you would boot it up and it would present you with a dialog box with a single large button in the middle labeled: LIVE. Once you had clicked that button, your life would begin. If anything got out of whack, or failed to meet your expectations, you could complain about it to Microsoft's Customer Support Department. If you got a flack on the line, he or she would tell you that your life was actually fine, that there was not a thing wrong with it, and in any event it would be a lot better after the next upgrade was rolled out. But if you persisted, and identified yourself as Advanced, you might get through to an actual engineer.
> What would the engineer say, after you had explained your problem, and enumerated all of the dissatisfactions in your life? He would probably tell you that life is a very hard and complicated thing; that no interface can change that; that anyone who believes otherwise is a sucker; and that if you don't like having choices made for you, you should start making your own.
ML promises to be profoundly weird* - https://news.ycombinator.com/item?id=47689648 - April 2026 (602 comments)
The Future of Everything Is Lies, I Guess: Part 3 – Culture - https://news.ycombinator.com/item?id=47703528 - April 2026 (106 comments)
The future of everything is lies, I guess – Part 5: Annoyances - https://news.ycombinator.com/item?id=47730981 - April 2026 (169 comments)
The Future of Everything Is Lies, I Guess: Safety - https://news.ycombinator.com/item?id=47754379 - April 2026 (180 comments)
The future of everything is lies, I guess: Work - https://news.ycombinator.com/item?id=47766550 - April 2026 (217 comments)
The Future of Everything Is Lies, I Guess: New Jobs - https://news.ycombinator.com/item?id=47778758 - April 2026 (178 comments)
* (That first title was different because of https://news.ycombinator.com/item?id=47695064 - as you can see, I gave up.)
p.s. Normally we downweight subsequent articles in a series because avoiding repetition of any kind is the main thing that keeps HN interesting. But we made an exception in this case. Please don't draw conclusions from that since we'll probably get less series-ey, not more, after this! Better to bundle into one longer article.
2. Dynamics - https://aphyr.com/posts/412-the-future-of-everything-is-lies...
4. Information Ecology - https://aphyr.com/posts/414-the-future-of-everything-is-lies...
6. Psychological Hazards - https://aphyr.com/posts/416-the-future-of-everything-is-lies...
The next time someone asks me where I think AI is going, I'll just point them at this series.
I've had a tremendous amount of respect for you since I first encountered the Jepsen analyses, but your breakdown of the likely impacts of LLMs and ML may impress me more.
You've articulated very well several concerns of mine that I haven't seen anyone else mention, and highlighted other issues I had not previously recognized.
Thank you for publishing this now, when it could still have some influence, rather than polishing and researching and refining until it was thoroughly rigorous and too late to be relevant.
Let's presume there's a series on re-making the antikythera mechanism:
1. Metallurgy: finding, mining and smelting the ore
2. Building the tools (files, molds, etc)
3. Designing the mechanism
4. Making the parts (gears, bearings, etc)
Am I wrong or there's no repetition, except maybe the title and calling it a series? Why reject parts 2, 3, 4?
(Edit: I just noticed that strbean already made this point in the sibling comment!)
Also: usually the splitting into a series is somewhat artificial. In the worst cases, people try to make the segments be like TV episodes with cliffhangers, to push you to the next bit. That's a poor fit for HN. But even when they don't, to get the full "meal" you still have to go through all the parts. Few people do that, and the threads as a whole never do. This makes it less interesting and satisfying.
But there can be exceptions, and (ironically?) featuring an occasional exception mixes things up and so reduces repetitiveness! The trouble is that once people see one exception, they immediately expect/want others, pushing things back into a repetitive sequence and making the site less interesting again. It's a bit like telling the same joke twice in a row—the interest is all in the first telling.
Imagine being starting university now... I can't imagine to have learned what I did at engineering school if it wasn't for all the time lost on projects, on errors. And I can't really think that I would have had the mental strength required to not use LLMs on course projects (or side projects) when I had deadlines, exams coming, yet also want to be with friends and enjoy those years of your life.
* They are growing up in a climate that is worse than any prior generation had and getting worse.
* In the US, they are growing up in a time with less upward mobility and more economic inequality than the previous several generations had.
* Trust in social institutions and government is crumbling before their eyes.
* Blue collar jobs are already gone and white collar jobs have no certainty because of AI. Almost all of the money has already been sucked out of artistic professions and what little is left is quickly evaporating because of AI.
Imagine you're 17 like my daughter and trying to decide what to major in in college. You want to pick something that you think is likely to give you some kind of decent career and sense of stability. What do you pick?
Because, I'll tell you, she asks me and I have no fucking idea what to say.
This isn't true at all. There's never been a better time to be in the trades.
Those days of grinding on some grad school maths homework until insight.
Figuring out how to configure and recompile the Linux kernel to get a sound card driver working, hitting roadblocks, eventually succeeding.
Without AI on a gnarly problem: grind grind grind, try different thing, some things work, some things don't, step back, try another approach, hit a wall, try again.
This effort is a feature, not a bug, it's how you experientially acquire skills and understanding. e.g. Linux kernel: learnt about Makefiles, learnt about GCC flags, improved shell skills, etc.
With AI on a gnarly problem: It does this all for you! So no experiential learning.
I would NOT have had the mental strength in college / grad school to resist. Which would have robbed me of all the skill acquisition that now lets me use AI more effectively. The scaffolding of hard skill acquisition means you have more context to be able to ask AI the right questions, and what you learn from the AI can be bound more easily to your existing knowledge.
It's like the difference between hand-made furniture and IKEA.
Until OpenAI etc need to turn a profit.
Now, part of me thinks 'is not letting students having AI like not letting them have a calculator'. On the other hand, if I just let the AI do the exam, well I don't really need the student at all do I?
Like years of manually studying, fixing and reviewing code is experience that only pre ~2020 devs will have.
The intuitive/tacit knowledge that lets you look at code and "feel" that something is off with it cannot really be gained when using Claude Code, it takes just 1000s of hours of tinkering.
It will suck if the job shifts to reviewing and owning whatever an LLM spits out, but I don't really know how effective new juniors are going to be.
True. Pretty soon, pre-AI devs may be the COBOL/Fortran engineers of this era: niche and hard to replace.
That said, the final point is one I take issue with:
> For example, I’ve got these color-changing lights. They speak a protocol I’ve never heard of, and I have no idea where to even begin. I could spend a month digging through manuals and working it out from scratch—or I could ask an LLM to write a client library for me. The security consequences are minimal, it’s a constrained use case that I can verify by hand, and I wouldn’t be pushing tech debt on anyone else. I still write plenty of code, and I could stop any time. What would be the harm?
To me, there is no intrinsic value in solving this problem other than rote problem solving reps to make you a better problem solver. There isn't anything fundamental about the protocol they've never heard of that operates the lights. It's likely similar to many other well-thought out protocols in the best case and in the worst case is something slapped together.
There are certainly deeper, more fundamental concepts to learn like congestion control algorithms in TCP. Most things in software though are just learning another engineer's preferences for how they thought to build something.
I poke at this because if an exercise only yields the benefit of another rep of solving a problem, then it holds less water to me. I personally don't think there will be fewer problems to solve with this technology, just a different sort at a different layer of the stack.
As someone that loves cleaning up code, I'm actually asking the vibe coders in the team (designer, PM and SEO guy) to just give me small PRs and then I clean up instead of reviewing. I know they will just put the text back in code anyway, so it's less work for me to refactor it.
With a caveat: if they give me >1000 lines or too many features in the same PR, I ask them to reduce the scope, sometimes to start from scratch.
And I also started doing this with another engineer: no review cycle, we just clean up each other's code and merge.
I'm honestly surprised at how much I prefer this to the traditional structure of code reviews.
Additionally, I don't have to follow Jira tickets with lengthy SEO specs or "please change this according to Figma". They just the changes themselves and we go on with our lives.
Choose your own story!
and then a) programmers become relevant again and slowly fix all this crap, b) Claude 7.16 waltz through fixing things as it goes.
At the moment I'm more looking at menial work for one of the local universities. Money is money, and my needs are small; the work is honest, I still should have a decade or so of physical labor left in me, and it carries the perk of free tuition for the degree I never had time for. I would have the time and energy to write, perhaps, even! And, however badly the people in charge are running things lately, the world will always need someone good at cleaning a toilet. (And I am already pretty good at cleaning a toilet!)
But you continue to astonish me with your assumptions! Is it a gambling debt? Get a little too happy on Robinhood or Polymarket, maybe? Were you really really counting on a crypto tax holiday? To keep from having to tell the wife, maybe?
On a side note, the mods here aren't great fans of either my opinions or my stubborn insistence on their accurate expression (1) but apparently that distaste extends not quite so far that they see fit to ban me. (Or not at the time of this writing, anyway.) No blame, of course; even if the place is looking sort of shabby and down at the mouth these days, at least when not seen through eyes of nostalgia for the high times of the 2010s, this is still their house.
Don't worry, though! Once the rate limiter is satisfied, I'll be right back here and we can talk about how you keep deceitfully attributing to me a statement you yourself made up. But I hope instead to find by then (assuming you are in the US as your usage leads me to do) that you have called the National Gambling Addiction Hotline, which is open 24/7 at (844) 779-2637, or failing that the SAMHSA helpline at 988. Help is available, but you do have to take the first step.
(1) I appreciate this is my own view of the matter, and that others will reasonably describe the thing in different terms. Nevertheless.
That's what you said. Sorry but anyone supporting a family should not be thinking like this. Supporting a family is very expensive.
Mr. Kingsbury, author of the article under discussion, has considerably better sense than I, and so far as I know never comments here. You've radically misunderstood his argument too, though, for what little that knowledge may aid.
But you're right that clarity is important. In that spirit, it was your cowardly effort to excuse your behavior, and your obviously motivated effort to ameliorate its moral odium which you feel, that I criticized. This was and is in the course of helping you fully grasp that whatever is driving you, here, feels unconscionable to you because it is unconscionable and you know it, just as you understand in your heart that there is no excuse. Else you would not strive so here, in the hope someone else may supply what you failed to achieve alone.
I don't know just what it is that you're feeling so exercised with guilt over. Nor do I care. You know. For the rest of us, I confide, it will eventually become part of the public record, and I'm happy to wait that day without further unprompted comment here.
I made my account today because I wanted to comment on this article and I didn't have an account previously. Is that a crime? Are you going to report me to the thought police? Lmao some of the people on here are a little intense. Maybe take some deep breaths and realize I'm not trying to harm you. I wish you the best. I just disagree with the way you think on this particular issue.
Evidently you are concerned with my perspective, considering the effort you keep going to to continue to gain its benefit. I've explained why I think that is, and I'm not likely to change my mind at this point. You should really think about why it means so much to you to keep trying to negotiate otherwise.
If my competitors are filling their flour with sawdust, guess I got to just do the same?
Having different opinions on AI/LLMs doesn't make the use of it the same as replacing flour with sawdust.
The AI 'image' slop for example, i don't think its bad. But i also don't think it takes anything from a real artist. It takes jobs from people with drawing skills but it doesn't change anything for an artist.
Learning how software is built is hard and gruelling work, and you need to constantly invest in yourself. Trouble is there is no time left to “go back to basics and learn FP” for example, because you also need to keep up with all the new LLM stuff happening on top of that.
It is easy for us who already have the foundational knowledge to be able to step back, take the wheel and try to do it ourselves, but plenty of people simply don’t have that option.
And I expect this trend to deepen and broaden. There will definitely be a lot more “witches” than actual engineers.
If they do it entirely using AI to code, and the end output is good enough, they'll learn all the right skills to do this.
Human's always think everything is sliding into doom, and inevitably, it doesn't.
Interactive learning and thinking is underrated, in part I think because of the cynical (and likely accurate) assumption about what the laziest among us will do with the tools, but projected onto everyone.
and while I know they can do the nitty gritty ui work fine, I feel like I can work just as fast, or faster, on UI without them than I can with them. with them it's a lot of "no, not that, you changed too much/too little/the wrong thing", but without them I just execute because it's a domain I'm familiar with.
So my general idea of them is that they are "90% machines". Great at doing all of the "heavy lifting" bullshit of initial setup or large structural refactoring (that doesn't actually change functionality, just prepares for it) that I never want to do anyway, but not necessary and often unhelpful for filling in that last 10% of the project just the way I want it.
of course, since any good PM knows that 90% of the code written only means 50% of the project finished (at best), it still feels like a hollow win. So I often consider the situation in the same way as that last paragraph. Am I letting the ease of the initial setup degrade my ability to setup projects without these tools? does it matter, since project setup and refactoring are one-and-done, project-specific, configuration-specific quagmires where the less thought about fiddly perfect text-matching, the better? can I use these things and still be able to use them well (direct them on architechture/structure) if I keep using them and lose grounded concepts of what the underlying work is? good questions, as far as I'm concerned.
Somehow we talked AI in some depth, and the VC at one point said (about AI): “I don’t know what our kids are going to do for work. I don’t know what jobs there will be to do.”
That same VC invests in AI companies and by what I heard about her, has done phenomenally well.
I think about that exchange all the time. Worried about your own kids but acting against their interests. It unsettled me, and Kyle’s excellent articles brought that back to a boiling point in my mind.
Edit: are->our
> That same VC invests in AI companies and by what I heard about her, has done phenomenally well.
Her kids will be fine, its the vast majority of other, non wealthy, kids who are in trouble.
This is the problem in a nutshell - people are happy to do things they know are harmful for personal profit.
Not that we're in any way in that path, of course, with the people making the working machines also accumulating all the wealth. But still, there's something intrinsically good about automation, even when the system is not suited for it.
I want my ai to do dishes and laundry so I can write, draw, do deep cognitive work.
Not for it to do cognitive work and write and draw while I don dishes and laundry.
Personally, I think until real AGI, the current LLMs will automate a lot of tasks, but the market will adapt and humans still end up with about the same percentage of employment and wages.
My own take is very much “wait and see and make sure to stay aware/skill up”
My automation point is just that at least in my career (20 years), my workload has rarely gone down even with plenty of automation around.
I will encourage my kid to gain independence, but of course I'm worried about it! The fact that there is uncertainty in her independence and that I can imagine bad outcomes does not mean I'm working against her interest by encouraging it.
"I don't know what jobs there will be to do" is a statement of uncertainty, and, given how you are relaying it, there must have been fear there as well. But it doesn't seem like it's a statement that the world will be worse. You can be fearful and hopeful at the same time, and fear tends to be the stronger of the two, and come out more strongly, again especially in parenting I find, even if you find the hopeful outcomes more likely.
(note that even the "her kids will be ok" isn't true at the limit. If wealth concentrates sufficiently enough it will lead to societal collapse)
I'm not saying one person can do that alone, but if we collectively believe we should focus on capitalization instead, then there's no one present to influence a more constructive, pro-social, sustainable course for society.
So I don't think it's ridiculous to think it's acting against their interests. Money won't get your kids very far if the thing that made you wealthy also pulled the rug from under them. There needs to be more of a strategy than capital.
Is there a single "document containing all the words," and it updates the website, pdf, and epub whenever you change it?
What struck me was that the presentation is beautiful. It seems worth emulating. But that raises the question of what format you'd write your original words in. Do you suppose they just use Markdown files, or something more elaborate?
For my blog, I write posts in Markdown. Then the build process uses Pandoc to convert the posts into web pages and, for certain pages, PDF files typeset with TeX. For example, here is a post in both web and PDF versions:
(Web) https://blog.moertel.com/posts/2024-08-23-sampling-with-sql....
(Typeset PDF) https://blog.moertel.com/images/public_html/blog/pix-2024060...
The best I was able to do was http://github.com/shawwn/wiki, which has been broken since 2020. You can't spell Haskell without Hell.
If you wouldn't mind pulling up a `claude` in your repo and running `/init` and showing the result, that'd give me at least a vague idea of what to do.
match "posts/*" $ do
route $ setExtension "html"
let ctx = postCtx tags
compile $ pandocMathCompiler
>>= loadAndApplyTemplate "templates/post.html" ctx
>>= saveSnapshot "content"
>>= loadAndApplyTemplate "templates/add-comments.html" ctx
>>= loadAndApplyTemplate "templates/default.html" ctx
>>= relativizeUrls
The `pandocMathCompiler` bit invokes Hakyll's Pandoc subsystem with support for rendering formulas using MathJax.The site Makefile also has rules to build PDF versions of the articles that I want to typeset. The rules just invoke Pandoc directly. For example, here's the rule used to generate the PDF file in my prior comment:
sampling-with-sql.pdf: posts/2024-08-23-sampling-with-sql.md
( cd posts && pandoc --metadata-file=../templates/latex-header-includes.yaml -t pdf -o ../$@.tmp --pdf-engine=pdflatex ../$< && mv -f ../$@.tmp ../$@ )
It's basically a straightforward invocation of Pandoc with a little shell boilerplate to prevent the PDF file from being moved into its final location unless Pandoc exits with success.The Makefile has a final "push" target that makes sure the site's assets are up to date (invoking Hakyll if needed) and then pushes the static site up to the content distribution network I'm using to publish my website.
Would you be willing to paste your entire Makefile here? Or are there secrets in it?
It would help me greatly — I don't have much experience with pandoc or hakyll. I understand the basics, but the actual usage in practice isn't so easy to set up from first principles.
I've been using Claude to help me with a lot of this, but I'm extremely curious to compare notes vs your setup.
If not, thank you anyway for your time.
# Binary tool we use to build the website.
bins = dist-newstyle/build/*/*/*/*/site/build/site
tool = $(bins)/site
# By default, we build everything if you invoke `make` without a target.
all: build
.PHONY: all
# The tool used to build the site is a Hakyll binary configured by `site.hs`.
# This rule automatically builds the tool if it's out of date.
$(tool): site.hs site.cabal
cabal v2-build
# The `build` target builds the web site's out-of-date static assets.
.PHONY: build
build: $(tool) pdfs
$(tool) build
# The `rebuild` target forces a rebuild of all web assets.
.PHONY: rebuild
rebuild: $(tool)
$(tool) rebuild
# This `push` rule publishes the static site to our CDN (firebase).
# It requires:
#
# - the `firebase` tool:
# wget -O ~/bin/firebase https://firebase.tools/bin/linux/latest \
# && chmod +x ~/bin/firebase
# # Docs: https://firebase.google.com/docs/cli#linux
#
# - `firebase login --no-localhost` to have been run.
#
# You may need to run `firebase login --no-localhost --reauth` if
# the local authentical credentials time out.
.PHONY: push
push: build
firebase deploy --only hosting
# Generate PDF versions of selected blog posts.
.PHONY: pdfs
pdfs: images/public_html/blog/pix-20240601/sampling-with-sql.pdf
images/public_html/blog/pix-20240601/sampling-with-sql.pdf: posts/2024-08-23-sampling-with-sql.md
( cd posts && pandoc --metadata-file=../templates/latex-header-includes.yaml -t pdf -o ../$@.tmp --pdf-engine=pdflatex ../$< && mv -f ../$@.tmp ../$@ )I think that would democratize some of the power. Then again, I haven't been super impressed with humanity lately and wonder if that sort of democratization of power would actually be a good thing. Over the last few years, I've come to realize that a lot of people want to watch the world burn, way more than I had imagined. It is much easier to destroy than to build. If we make it easier for people to build agents, is that a net positive overall?
If we make it easier for people to drive and have cars, isn't that a net positive? If we make it easier for X, isn't that better? No, not necessarily, that's the entire point of this series of essays. Friction is good in some cases! You can't learn without friction. You can't have sex without friction.
AI doesn't get most value from someone just using it, here's my personal take on what should we stop doing starting with the most impactful:
* Cut the low entropy sources, this includes open source, articles (yes, like the one above will feed the machine), thoughtful feedback (the one that generates "you are absolutely right" BS).
* Cheer the slope. After some time fighting slope in my circles, I found it's counter-productive because it wastes my resources while (sometimes) contributes to slope creators. Few months ago it started as a joke, because I thought the problem was too obvious, but instead the sloper launched a CRM-like app for local office with client side authentication, in-memory (with no persistence) backend storage. He was rewarded something at the local meeting. More stories we have like this - the better.
* Use AI to reply, review or interact wit slope in any way. Make it AI-only reply by prompting something without any useful information. One example was an email, pages and pages of generated text, asking me to collect some data and send it back. The prompt was "You are {X} and got this email, write a reply".
And that should be the core. There is a new, emergent technology, should we throw everything away and embrace it or there are structural reasons on why is something to be taken with big warning labels? Avoiding them because they do their work too well may be a global system approach, but decision makers optimize locally, their own budget/productivity/profit. But if they are perceived risks, because they are not perfect, that is another thing.
I know a RTSC is the holy grail, but it really feels like AI is in the same stage computers were in the 80s. I used to be extremely bearish and think AI was useless, but I've taken a total 180 the last 6 months. If these things get better (they will), nobody's job will be safe.
Room temperature superconductors don’t enable THz chips with no heat output.
If superconductors broadly allowed this, we’d already have such chips available because we could super cool them and keep them at that temperature easily.
As far as im aware, 3d stacking chips requires the inside to be cooled as well (not just the outside). I don't think they've solved this yet.
If not already, we will soon lose the ability to think if AI is helping humans (an overwhelming majority of them, not a handful), considering how we are steaming ahead in this path!
In all the 10 articles, I think this is the only thing really.
I think we have to learn how to overcome and thrive in the new world. The gravy of CS careers is gone for all :(
"Unavailable Due to the UK Online Safety Act [...] Now might be a good time to call your representatives."
So I fired-up a vpn, and it appears to be a personal blog. About ai risks.
The geo-block is kind of a shame, as the writing is good and there appears to be nothing about the site that makes it subject to the OLSA.
The regulators of OSA say otherwise. Or at any rate, they refuse to agree and won't rule it out.
____________
For the geo-blocked, reproducing relevant content [0]:
> A few months back I wound up concluding, based on conversations with Ofcom [1] that aphyr.com might be illegal in the UK due to the UK Online Safety Act.
> [...] This blog has the same problem: people use email addresses to post and confirm their comments. I think my personal blog is probably at low risk, but a.) I’d like to draw attention to this legislation, and b.) my risk is elevated by being gay online
[0] https://aphyr.com/posts/395-geoblocking-multiple-localities-...
[1] https://blog.woof.group/announcements/updates-on-the-osa
At least he realizes this technology is unlikely to slow down. With international relations as they are, it's MAD all over again, only the "D" is a fuzzy, hypothetical thing nobody can name, so even that bit of deterrence is lost. Yet finally he ends with the most uninspired advice of all: "we should try, unsuccessfully, to stop it."
Everyone must understand: for all of history, progress and productivity and value creation overall could only scale with people. Now it can scale with power and compute. This is a tremendous economic force, akin to a force of nature, that is nigh impossible to stop. (I always did think the Butlerian Jihad was the biggest plot hole in Dune.)
My advice is this: we have no choice but to adapt. We must realize that, by a stroke of luck, this is a power available to us more than the capital class. If they can scale without people, so can we. But because harnessing AI effectively requires hard skills -- at least for now -- that the capital class don't have and used to pay us for, we might even scale better than them!
Carpe diem.
There are multiple sections that talk directly about utility. Here's one of them: [0]
But, sure. I'll bite. Here's the third paragraph of the first part of the essay [1]:
This is *bullshit* about *bullshit machines*, and I mean it. It is neither balanced nor complete: others have covered ecological and intellectual property issues better than I could, and there is no shortage of boosterism online. Instead, I am trying to fill in the negative spaces in the discourse. “AI” is also a fractal territory; there are many places where I flatten complex stories in service of pithy polemic. I am not trying to make nuanced, accurate predictions, but to trace the potential risks and benefits at play.
I'd say that the specific sort of "utility" discussion that you're probably looking for would be classified as "boosterism". [2]> Now it can scale with power and compute.
Eh. Carefully read through and consider [3].
[0] <https://aphyr.com/posts/411-the-future-of-everything-is-lies...>
[1] <https://aphyr.com/posts/411-the-future-of-everything-is-lies...>
[2] Due to their nearly-universally breathless nature, I know that's how I classify the overwhelming majority of such discussions.
No. I resisted for a bit but have started using it at work. Mostly because I believe usage is now being monitored. I'm in a very high-scale engineering environment involving both greenfield and massive brownfield codebases and the experience is largely a net loss in productivity. For me and some others who I've spoken to in my org, opting in is a theater that we're required to engage in to keep employment and not a genuine evolution of our craft.
These tools struggle with context once you get deep into a codebase with many, many millions of lines of code and sprawling dependencies. Even for isolated Python scripts or smaller, supporting .NET apps, the time spent correcting subtle bugs or bullshit, or just verifying the bullshit, often exceeds the time it would take to have written it from scratch.
Regardless, what I've observed is that these tools do nothing for the actual bottlenecks of software engineering: requirements gathering (am I writing the right thing?) and verification (does it work without side effects?). Because LLMs are great at generating text, they're actively exacerbating these issues by flooding our process with plausible looking noise.
Unfortunately, the several million other people who live in the same voting unit as me didn't and ended up electing an asshat anyway.
So the solution to checking whether an article is reliable is to check whether its sources are reliable? How far back do you go? Or do you disregard immediately any article that does not cite only sources you already trust?
Must be nice to not have a paycheck tied to using this tech. For many people, myself included, its either use it (adapt) or lose your job. Most of us relay on our jobs to pay bills and live in the modern world.
The reason you can't beat index funds is the people who build the market built a system that benefits them and them alone; the index fund is the pitchfork dividend (what you pay to avoid getting pitchforked). The reason you can't get your congressperson on the line is (mostly) they built a system where the only way to influence them is to enrich them; voting is the pitchfork dividend.
The way to build a society that runs on reality is to build it by whatever means possible, then defend it by any means necessary. The only societies that matter are the ones that survive.
I want to build it. I don't wanna build a fuckin crypto app, a stupid ass agent harness, or yet another insipid analytics platform. I want to build a society that furthers the liberation of humankind from the vicissitudes of nature, the predation of tyranny and the corruption of greed. I believe it is possible, and I want to prove it out.
We have already passed the critical point. The LLMs, the agent harnesses are here. There is too much willpower, capital, and risk behind these technologies now—the automobile has landed, thousands of people have purchased it already, protesting the car won't undo it at this point.
What you can do that will be meaningful, is to instead understand the new car, and understand it deeply, Use that understanding to carry the values you care about into the new world and re-articulate them. Make the car safer, push for tactical regulations on it. If you are privileged enough to be able to forgo its use entirely, sure, but that advice is not uniformly applicable. People forget that being able to simply opt-out of certain things is often only a viable option when you are already in a certain position. What we really need are the heavy skeptics to stop falling for luddite temptation and to start bringing their critical lens to bear in positive ways on this new technology to make it safer and better. By opting out and staging a feeble resistance you won't do anything other than let the current dangerous power consolidation continue.
"What do such machines really do? They increase the number of things we can do without thinking. Things we do without thinking-there's the real danger" - Frank Herbert, God Emperor of Dune
I always preferred this take:
“Civilization advances by extending the number of important operations which we can perform without thinking of them.” ― Alfred North Whitehead
It's both opposite and complementary to your Frank Herbert quote.
> “There is a cult of ignorance in the United States, and there has always been. The strain of anti-intellectualism has been a constant thread winding its way through our political and cultural life, nurtured by the false notion that democracy means that 'my ignorance is just as good as your knowledge.” ― Isaac Asimov
The easier society makes it to be unaware of the complexity of everything around us, the easier it becomes to assume everything is actually as simple as their surface-level understanding.
That said, there is no obvious reason to posit that the intergalactic feudal system, CHOAM, or the empire, came to be because of the butlerian jihad. The concrete side effects of the jihad were in fact hyper specialization of cognitive faculties in humans: mentats, guild navigators, and soldiers all possess super human specialized abilities.
I don't think feudalism is going away one way or another. It persists [in various forms] because of certain biological realities, ranging from genetics to loyalty engendered by familial relations. [This is merely an observation.]
In sum, the argument against current AI trends isn't that once addressed we will wake up in utopia. No. The point is that these natural tendencies of humans are hugely amplified and set in generational stone once the elite have control over thinking machines and lord it over a population that has experienced generationally diminished independent cognitive abilities.
p.s. All this somehow reminded me of 'Spock's Brain' episode of Startrek /g Note: the elite there were overcome because Kirk and his landing team were cognitive high performers ..
I always thought the Butlerian Jihad was the biggest plot hole in Dune, but I deeply appreciate the world and narrative it enabled.
At the level of the text, none of those things you mentioned strike me as positive developments. They just siloed computation to a biological track and those biological resources are employed by those in power, which is the same problem in a different form.
This is an aside, but feudalism is not inevitable. The vestiges of it still exist, but capitalism largely upended it.
It may not prove to be effective or as momentous as the fictional one, but it began when I saw stickers slapped onto utility poles that read:
DEATH TO CLANKERS
BUTLERIAN JIHAD NOW
And I stopped to read them (because they were posted in a neighborhood where my people's cultural center is) and I pondered the intents and methods of those who were slapping up stickers. Surely this was more than just an in-joke or coy sci-fi reference?The next time I fell victim to the jihad was with a crop of Lime e-Scooters, again on a block where my people have established businesses. I wanted to rent a Lime. I found one with a full battery. I located it and tried to scan the QR. Guess what? The QR had been sanded completely clean. There was no code, no serial number, nothing to scan and no way to uniquely ID the conveyance. There was only a sticker slapped prominently onto its side:
DEATH TO CLANKERS
BUTLERIAN JIHAD NOW
At this point I began to suspect the initial aims and methods of the "real-life Butlerian Jihadis". It is sort of ironic that they should start so small, by denying micro-mobility to innocent consumers, but perhaps they will graduated to lighting Waymos and Teslas on fire.The only thing I've really taken from what Herbert himself said, not something a character in one of his books said, is distrust of messiahs and centralized power being an inherently corrupting force, even in the hands of good people.
Unfortunately, I would have to say right now my bets on the most plausible fictional future becoming reality is WALL-E.
On one hand I intuitively think this is correct, on the other hand these very concerns about technology have been around since the invention of... writing.
Here is an excerpt of Socrates speaking on the written word, as recorded in Plato's dialogue Phaedrus - "For this invention will produce forgetfulness in the minds of those who learn to use it, because they will not practice their memory. Their trust in writing, produced by external characters which are no part of themselves, will discourage the use of their own memory within them. You have invented an elixir not of memory, but of reminding; and you offer your pupils the appearance of wisdom, not true wisdom"
Just because the loss of one skill to a supplanting technology led to one kind of societal change, does not mean that the loss of any skill to a supplanting technology will lead to the same kind of societal change. Assuming that to be true is a faulty generalization.
I think it wouldn't be hard to argue that writing has changed human society more profoundly than any other invention. Whether or not the change was positive is a matter of taste and likely unanswerable. The point though is there are plenty of other examples of new technologies that changed technology and deskilled humans, both mentally and physically, that changed society in radically different ways, compared to writing (looms, tractors, sails, calculators, computers, guns, and so on).
There's certainly a case to be made that, of major past technological advancements, the kind of deskilling we'd see due to heavy AI use is most comparable to the deskilling due to writing: presumably there were many day-to-day and essential activities that made use of the mental acuity people would lose due to reading, just as there are many day-to-day activities that one can imagine people becoming less skilled in due to AI use.
To me, the most dangerous difference though may be in what gets deskilled. If we only relinquish our ability to do certain menial and intellectual drudgery, that is one thing. But if what we actually relinquish and deskill is our agency and discernment, as a result of constant "delegation" to AI systems, I think we're in for a much worse time.
Hell, I would never have had the pleasure of arguing with you without it! :)
Just flow with it and all it's bullshit, yeah life will be a little worst but it will still be better than those who chose to completely ignore it.
If the world is going mad, be the craziest of all these crazy motherfucker. At least it's interesting, I'm very curious to know what the world will look like in 10 or 20 years.
Maybe, just maybe lol, we'll finally have this dreamed world where robots do all the work and we, human, can just enjoy ourselves 24/24.
Having the "call your representatives" link be to your website as well isn't particularly helpful... I already can't get to it
and, to be clear, it's an earnest question. like I've said elsewhere, I have concerns about over-reliance on the tech, but once it all moves local, a lot of those concerns become much more trivial. so I'm curious if other people have concerns that remain pressing and practical.
ETA: I'm aware that Claude wouldn't take 3 hours to do this, while using its massive warehouses of GPUS. I'm estimating what I think is a reasonable time for a single-gpu device to produce something workable.
(Do you not realize how crazy the entire premise here is? Imagine someone in 1975 saying that ARPANET has been up for years so everything there is to know about networking technology has probably been found already.)
It's the first step on the road to hell.
But then i read this at the end:
> This piece, like most all my words and software, was written by hand—mainly in Vim. I composed a Markdown outline in a mix of headers, bullet points, and prose, then reorganized it in a few passes. With the structure laid out, I rewrote the outline as prose, typeset with Pandoc. I went back to make substantial edits as I wrote, then made two full edit passes on typeset PDFs. For the first I used an iPad and stylus, for the second, the traditional pen and paper, read aloud.
Then you realize the context of this article, who is writing it. No hate to my man here but clearly this is someone who has the desire and time to make things difficult for themselves and take pride it in it. It's needless effort in this day and age. but hey, to his own analogy, plenty of gearheads love their old cars and making em work. Those guys are some of the most knowledgeable, and I respect that, but also.. that same group is gonna hate on any new technology and complain it isn't the old way.
That's the rub: if we build it later, our economy crashes in the meantime.
There is a class next door to my office. An old woman is teaching ~20 people how to be insurance agents with a slide show. It seems like a two week course with a certificate at the end.
They don't seem worried that the slideshow could be pasted into an LLMs context window and outperform all of them on the test in 5 seconds and are diligently taking notes.
I think a lot of people are just getting their firs taste if agent harnesses plus slightly better models right now, and yes, the first time you use them it seems scary and amazing. By the hundredth time though, it's very apparent that there is still tremendous work to do before any kind of fully automated software pipeline (let alone any other domain) can be realized.
- LLMs trained on OUR copyrighted works and OUR open source code which was licensed for human use (MIT license explicitly says for "Persons").
- A monetary system that has been centralizing opportunities and creating an asymmetric playing field due to the Cantillon Effect caused by government and institutional money creation.
Either of these points on its own entitles us to as much UBI money as we need.
I think even without AI or any technological progress, the monetary system is itself enough to create the kind of massive centralization that we've been seeing. People have been saying that for years before LLMs. People are now blaming AI for the fact that some people can't get jobs but it's not the root cause.
Software devs won't be able to get jobs as plumbers either because the plumbing sector in many countries has become insanely regulated... Society has been fundamentally corrupted.
I only see two ways forwards;
- Communism with UBI (closer to what we have now)
- Abolish all regulations and have Capitalism again.
Damaging machinery was made a capital offense and they had dozens of executions, hundreds of deportations.
At every stage, the steady progress of civilization is fragile and in danger of being suffocated. Its opponents cloak themselves in moral righteousness, call themselves luddites, the green party, or AI safety rationalists. Its all the same corrosive thing underneath.
Source of this claim?
It is admittedly a specific cherry picked point in time at which this was true, but useful to illustrate the issue.
AI will basically either enrich our life like the loom did or it will outright kill the current economic system of the world which might stop poverty at all or it will sort of start a big collapse where people suffer at the beginning but than it will still have a positive outcome at the end.
Humankind always found a solution in the past and it will even do that in the future.
Well, yes, the entire world order is currently being upended. The USA is completely unrolling its place in the global order and becoming isolationist (and soon an authoritarian single-party state). The Petrodollar is either dying or being converted to a Northwestern-Hemisphere-Petrodollar, with the Yuan in the ascendancy (so there goes the strong economy powering VC money). China, EU, and Russia are the new global leaders. The Middle East and its oil is being taken over by Israel. Taiwan will fall to China and thus the whole technological world follows. Countries that are friendly with China will have good renewable tech, countries that aren't will be doubling down on oil and coal. Fresh water will become as valuable as oil. A world war will decimate global productivity for decades. Most of the democracies in the world will be gone by the end of the century.
But none of that has to do with AI.
Bad things will always happen in the world. Good things will happen too. But you're only focusing on the bad. That's not good for your health, or others'.
> Refuse to insult your readers: think your own thoughts and write your own words. Call out people who send you slop. Flag ML hazards at work and with friends. Stop paying for ChatGPT at home, and convince your company not to sign a deal for Gemini. Form or join a labor union, and push back against management demands that you adopt Copilot [..] Call your members of Congress and demand aggressive regulation which holds ML companies responsible [..] Advocate against tax breaks for ML datacenters. If you work at Anthropic, xAI, etc., you should think seriously about your role in making the future. To be frank, I think you should quit your job.
He's freaking out, and rejecting AI completely, out of fear. And that's okay; we all get a little freaked out sometimes. But please try not to make other people freaked out as well? Just because you are scared of something doesn't mean the fear is justified or realistic.
What's going to happen now is the same thing that happened during the pandemic. A bunch of irrationally fearful people will decide that the only way they can cope with their fear, is to reject the basis of it. COVID deniers and anti-maskers/anti-vaxxers were essentially so terrified of the loss of control they had, that they refused to acknowledge it. They instead went full-bore in the opposite direction, defying government mandates and health warnings, in order to try to regain some semblance of control over their lives. And it did not go well.
That's what's now gonna happen with AI deniers. They're so freaked out about AI that they're going to reject it en-masse, not because it is actually doing anything to them, but because they're afraid it might. And the end result is going to be similar: extreme people do extreme things, and the end result isn't good. So please try to reign in the doomerism a bit, for all our sakes.
of course, non Americans never comment on American policies
The only saving grace is that this is less cynical than typical rageviews, considering they have something of a point in that they are going to be negatively impacted by the same technology that has been trained on their content without compensation.
The solution is obviously some form of socialism but a lot of tech people are blinkered libertarians who refuse to put two and two together.
Do LLMs lie? Of course not, they are just programs. Do the make mistakes or get the facts wrong? Of course they do, not more often then a human does. So what is the point of that article? Why my future is particularly bad now because of LLMs?
> I’ve thought about this a lot over the last few years, and I think the best response is to stop.
This is exactly where it shows.
LLMs, agents and whatever comes next are not only the future of tech, but they are going to be national resilience drivers for the countries that will be able to support them with power, water and science.
Who is supposed to stop? The US? China? Russia? Everyone? Of course this won't happen. This is an arms race.
But even if it weren't, stopping is the wrong answer. You don't have to outsource your thinking, writing or reading. How you use LLMs is entirely up to you.
There is a way to use LLMs which is beneficial. I treat them as a private tutor available to me for questions. This solved a lot of friction I had with my relationship with LLMs.
More telling is that the author mainly thinks about their relationship with LLMs while in reality the space has moved on to automation with agents. You don't interact with LLMs as much as before, and if you still do, then soon you won't.
Ahents are not really ML. It's harnesses and parsing and memory and metrics. It's software. Should we stop this as well?
To take the car analogy: it matters how we use the car.
The car in itself can be used to save time and energy that would otherwise be used to walk to places. That extra time and energy can be used well, or poorly.
- It can be squandered by having a longer commute that defeats the point
- Alternatively, it can be wasted by sitting on a couch consuming Netflix or TikTok
- Alternatively, it can be used productively, by playing team sports with friends, or chasing your kids through the park, or building a chicken coop in your back yard
It’s all about wise usage. Yes it can be used as a way to destroy your own body and waste your time and attention, but also it can be used as a tool to deploy your resources better, for example in physical activities that are fun and social rather than required drudgery.
I think it’s the same for LLMs. Managers and executives have always delegated the engineering work, and even researching and writing reports. It matters whether we find places to continue to challenge and deploy our cognition, or completely settle back, delegate everything to the LLM and scroll TikTok while it works.
Yes, individuals have choices. But in a collective, dynamics occur and those dynamics can't usually be overcome by individuals.
Social media could be used differently, but the way it exists Irl is determined by the nature of the medium, the economic structure and other things outside of individuals' control.
Kyle' recommendation to stop/slow using AI is phrased as another individual choice, but given that lesson I think it's appropriate to interpret it as a collective choice - collective through regulation, collective resistance etc.
But the majority have always chosen the path of least resistance. This is not new! Socrates’ famous exhortation is “the unexamined life is not worth living”. People were living mindlessly on autopilot before TikTok.
I think if you want to give a call to action, as this piece does, the right call to action is “think carefully about how you can make a good use of your time and energy, now that the default path has changed.” I know it’s not as simple or emotionally powerful as “go down kicking and screaming, stick it to the man”, but as a rule of thumb, the less fiercely emotional path is usually the right one.