upvote
It's possible I miss something, but are you saying that the author should relax and she should leaves this to smarter people?
reply
Rereading my comment I don’t see anything about the author’s intelligence or anxiety level.
reply
I take that non response as a yes.
reply
I call the author out for not being widely read in areas where they are attempting to coin neologisms. If that seems like insulting someone’s intelligence and telling them to calm down to you then I’m not sure this will be a productive conversation.

Coining new terms for established ideas is not intellectual contribution — it is, in Wiredu’s language, the opposite of conceptual decolonization. The topic does not need new vocabulary. The author needs to read more widely, cite more justly, and recognize that the scholars they overlook understood these problems first, understood them more deeply, and in some cases paid significant professional costs for doing so.

reply
She didn't claim to be an expert, so she doesn't have to satisfy any "standards".

You consider yourself intellectual but you don't apply benefit of the doubt? You neither offered reasonable correction, only a thin bun with a lot of beef.

reply
deleted
reply
> But what actually happens is we have formalized processes and can externalize them.

Even if I believe that is what happens in 10% of uses of AI, it doesn't excuse what happens with the rest.

Many people can not do mental math anymore and still more question why we need to learn math at all in the first place when we have simple calculators. "When will I ever use XYZ?" is a common refrain.

AI is currently developed and owned by billionaires who also happen to own news sources. If that correlation doesn't spark questions about why we shouldn't externalize processes to AI, you have likely been using AI too much already.

reply
I think we're excluding from this analysis the probability that these "AI" products will remain truly unbiased and free from external (corporate) influences.

When AI gains true marketshare in the "think-space", I have zero trust that the corporate overlords controlling these machines will use them in the fairest interests of humanity.

reply
You're absolutely right! But Brawndo has what plants crave!
reply
When I read pieces like this all I think is, resistance to change is a helluva drug.

I've been working on a project and using LLMs heavily to inform my design decisions. There's already a long list of cases where it has taught me things I wasn't familiar with, alerted me to possibilities I didn't consider, shown me how to do things that I was struggling with. In those cases I ask for references, and it delivers.

This is not "endangering human development". If anything, it's the exact opposite - allowing human knowledge to be transmitted to other humans in an accessible way that otherwise, usually simply would not have happened.

Of course, this all depends on using AI to enhance cognition and access to knowledge, as opposed to just letting a machine write all your code for you without review, Yegge-style.

I'm not saying there isn't a moral dimension to all this, and areas of serious concern. But the one about "endangering human development" is wholly in our individual hands. You can use AI to help you learn, or to replace the need to learn. The former will be better for human development.

One real lesson from this is perhaps that we need to teach people how to use AI in ways that benefit their development, not just their output.

reply
Usually when things are "in our individual hands" it ends very poorly.

This is because humans are actually extremely easy to exploit. Our biology is very stupid and also dumb, so even basic attacks can cause us to self-destruct.

And that's how we get obesity, smoking, war, I mean... you name it.

LLMs are basically perfect. While I'm sure some people, somewhere, can theoretically exist attacks from LLMs, on the whole I'm not sure that will be the case.

reply
> When I read pieces like this all I think is, resistance to change is a helluva drug.

When I read comments like yours, I’m reminded of (though I’m not comparing you to—I believe you are arguing in good faith) the cryptocurrency shills saying anyone who is against cryptocurrencies is just jealous they didn’t get in on the gold rush; they are incapable of imagining or accepting other people have their own reasons beyond what the author can themselves conceptualise.

When people criticise cryptocurrencies, NFTs, the Metaverse, LLMs, they’re not just stubbornly “resisting change”. Those technologies have important issues and repercussions which should be addressed, we shouldn’t just accept change unquestionably.

> Of course, this all depends on using AI to enhance cognition and access to knowledge, as opposed to just letting a machine write all your code for you without review, Yegge-style.

And the latter is exactly what is going to happen and is already happening in large enough quantity that it’s going to be a serious problem.

> But the one about "endangering human development" is wholly in our individual hands. You can use AI to help you learn, or to replace the need to learn.

That completely ignores the loss of skill that happens without you realising, as you lean more on a tool.

https://www.thelancet.com/journals/langas/article/PIIS2468-1...

https://arxiv.org/abs/2506.08872

This is nothing new. We already know that e.g. heavy GPS use makes us weaker at navigating on our own.

https://www.nature.com/articles/s41598-020-62877-0

> One real lesson from this is perhaps that we need to teach people how to use AI in ways that benefit their development, not just their output.

Yes, that is a good goal. But good luck achieving it.

reply
I think it depends on the person. As a teacher, I see this. Some kids (the gifted ones) use AI to multiply their efforts. Most kids use to just get by and are actually coming out of the class with less knowledge than they would have without one.
reply
Hence my final sentence:

> One real lesson from this is perhaps that we need to teach people how to use AI in ways that benefit their development, not just their output.

reply
As I see it, LLM’s require far more self-discipline and introspection than people expect or generally engage in.

It’s a corner cutting machine that allows people to shift the burden of their work on to others either in the form of more slop we have to wade through OR more work we have to correct because they couldn’t bother to vet the results.

It’s like writing a paper, running spellcheck, then sending it to some less to look over for you without ever taking a pass yourself. It’s selfish.

reply
>This is a benefit if you can use your newfound capacity and free time for something better, which I think most people ultimately will.

I think for a lot of of us the problem is that this is not a given. It’s often promised and rarely occurs, especially in the modern era. Increased productivity usually just means increased demands in the workplace.

reply
Similar to What people were saying about television. They imagined it would bring Shakespeare to the masses, but what ultimately ended up happening is television met people where they are hence we got reality TV. There’s no reason to believe AI won’t be similar.
reply
If you want to watch Shakespeare on your TV you are welcome to. But also I don't think that's the point at all. If it's my job to hammer 100 nails a day, and then my boss gives me a hammer that is 10x more effective. I don't get to go home early, I'm just expected to hammer 1000 nails now. Maybe the work becomes more pleasant, maybe it doesn't. But either way it definitely benefits the boss and the person who makes the hammer.
reply
Yeah I can’t disagree I think it is really disruptive in a corporate-job-expectation setting where you don’t have the same level of agency.
reply