upvote
Living in South America a bit really showed me this. I think it's a cultural thing here but someone will always give you an answer, even if it's wrong, confidently. It was hard for me at first- I am usually the first person to say "I don't know" (often followed by "but let's slow down and find a good solution").
reply
This was similar to my experience running a software team in India (I'm an American) a couple decades ago. I had to learn not to ask yes/no questions because the answer would always be yes.
reply
From real life: - Is it done? - Yes! But not yet!
reply
On a long enough time horizon, with sufficient multiverses, all things are done!
reply
It's a long-standing jodke that AI stands for "Actual Indians".
reply
deleted
reply
I've experienced similar with some Southeast Asian cultures as well.

I'm a patient person, but it can be frustrating to have to endure 10 minutes of verbal diarrhea that eventually results in a "no" or "I don't know".

reply
I'm genuinely curious if this is a thing with roots in Spanish culture? Because there is strong Spanish influence in Philippines and South America.

I don't know any Spaniards but I do know Filipinos and the confidence projection is a real thing. The Filipino IT guy confidently declared that my OnePlus Android phone wasn't certified for the software he was trying to install and was getting errors. It is a bog standard application that can be installed on any modern Android phone but the level of confidence he projected, just because he didn't know OnePlus as a brand, made me doubt myself until I turned on the critical hat and pushed back a little with alternative approaches, which solved the problem.

reply
Over the last couple of years, I've spent a lot of time in Indonesia. By the time I got used to their way of communicating, I questioned my own reality, perception and sanity. I even put a thought it's some very passive way of gaslighting foreigners. It seems it's just how they like to do it here.
reply
> someone will always give you an answer, even if it's wrong, confidently

its common playbook for corporate self-development in NA.

reply
Talking about South America as a homogeneous unit is… weird. Even neighbouring countries speaking the same language can be entirely different in this regard.
reply
I agree (and I don't normally generalize like this, so I apologize). I've spent most of my time in Peru but noticed it in neighboring countries as well.
reply
That also is my perception, from Brazil. There is even the concept of "cordial man", coined by sociologist Sergio Buarque d Holanda, that is connected to this

https://en.wikipedia.org/wiki/Cordial_man

reply
i remember >20 years ago going to the bus station somewhere between RJ and SP, and asking the best way to get to Iguaçu

  - it's difficult
  - ok fine but how
  - it's difficult
  - right i'll see that but how
  - it's difficult
then it dawned on me this meant get away you fool :D
reply
can speak from personal experience that it's the same culturally in colombia
reply
Is South America populated by LLMs?

But I kid, I have a friend who's the same way. He's an Austrian who grew up in Chicago and was in the army.

I have considered the phenomenon. I somewhat disapprove but I can also see the advantage of always presenting a confident face

reply
At the time of this writing, the prevailing thinking with "artificial intelligence" was that we'd encode every Fact we know and every rule of Logic, and from there, the computer would make new discoveries. Todays AI researchers would call this "symbolic" AI, compared to the "neural" AI powering LLMs. They're like two different worlds.

LLMs are just generating text, they don't know anything. They can't assess whether there is enough data for an answer. When you add a follow up prompt "This is wrong, why did you lie?" only then is it able to generate text, "I was wrong, I'm sorry," and so forth.

reply
Did Asimov’s idea of AI revolve around data retrieval? I’ve read that even human intelligence isn’t necessarily remembering things, but being able to traverse our knowledge and find that idea or thought quickly.
reply
They can read context and with fairly high accuracy say whether that context contains enough information to answer a posed question. They cannot (and we cannot for them) introspect their own weights to say whether their weights already encode information sufficient to answer a posed question.
reply
This is exactly like a lot of customer service, or technical support.

It seems that they are loath to tell anyone “no”, or that something can’t be done, or that an app doesn’t have a feature or can’t be used in a certain way. Especially when a feature has been removed for security reasons.

In fact, it gets so crazy that I simply cannot get a straight answer out of somebody and if I persist in my line of questioning and they become evasive or vague or I just can’t get a straight answer for long enough, ultimately, I suspect that the answer is “no”, and that they're simply not allowed to tell me, and they're paid and trained specifically to avoid uttering the “n-word”.

In my first job, as a network operator, my supervisor admonished me, and said “we must never tell a customer that we don't know something”. He said that we should tell the customer that “I will go ahead and find out for you, and get back to you on that”.

And that is kind of the kind of slippery non-answer I often received in my most recent job, that some manager or supervisor would “look into something” for me and “get back to me”. But the ‘getting back to me’ part never happened, and I began to suspect that it was a platitude meant to satisfy me enough that I would shut up for a while, and stop pressing the issue.

reply
hahaha, the irony is that "INSUFFICIENT DATA FOR MEANINGFUL ANSWER" requires more intelligence than a confident wrong answer. you have to know what you don't know. current LLMs are optimized to always produce output, which means they've essentially been trained out of epistemic humility.

Asimov's Multivac at least had the dignity to wait.

reply
They can do it, it's just not "by default", they need to be prompted to do it. So at least the danger is manageable if you know what you're doing and how to prompt around it.
reply
"Just don't accidentally forget to do the thing that makes it safe" is not a very effective strategy for something that so many vested interests are trying to push into all corners of society. If it's so easy to misuse it, then it shouldn't be used in any context outside of where there are no major consequences for bad output and there's amble opportunity and ability to validate it
reply
Not really. They're still non deterministic language predictors. Believing that a prompt is an effective way to actually control these machines' actual behavior is really far fetched.

They com like that from factory. Hardcoded to never say no.

reply
They're not hardcoded to never say no, but some of the models were trained to be "yes men" because their creators thought it would be a good property to have. GPT-4o for example.
reply
> non deterministic language predictors.

Non?? Only those with sh*tty code, surely.

There's nothing inherently non-deterministic about inference.

reply
The thing is that they are completely incapable of meta-cognition. Reasoning models don’t show their actual reasoning at all.
reply
Right — they're not reasoning, they're generating text that statistically models reasoning. Anyone who says differently is selling something.
reply
As the meme goes, "they are the same picture".
reply
Language has reasoning encoded within it.
reply
It certainly does. But so too do complex neural network functions, as do attention mechanisms.
reply
That is what a base model does. After RL it is a very different thing, and anyone who says they know what it is, is naive or dishonest. These things are grown, not made, and we really do not understand how they work in many important ways.
reply
Yeah, but they’re not magic; we can still do experiments and see what happens. Anthropic did a lot of work on this and showed that they’re not accurately describing their reasoning process.
reply
Of course, the fact that they have to do that proves my point.
reply
Not believing that a prompt is an effective way to actually control their behavior is obviously incorrect to anyone who's actually used these things.

It's not a guaranteed way to control their behavior, but you can more than move the needle.

reply
The word most relevant to this conversation is “influence.” Influence is possible and users observe it and use it to increase margins of useful outcomes. “Control” is incorrect.
reply
yeah that distinction is pretty important, and in general that guy I believe IS making the point - if you can not control it with guaranteed outcomes - you cannot control it.
reply
You can't control it any more than you can control a draw from a deck of cards, but you can absolutely control the deck of cards that you choose to draw from.
reply
The problem is that nobody really does that? Like, as far as I'm aware, even simple stuff such as not considering tokens that would result in a syntax error when writing code isn't being done.
reply
magicians can probably make you change your mind on the former
reply
That's silly. My car is not absolutely guaranteed to turn left when I turn the steering wheel left, but you wouldn't say I can't control my car on that basis.

Steering an LLM with a prompt is way less reliable than steering a car with a steering wheel, but there's still control. It's just not absolute.

reply
if your car doesn' turn left when you turn the steering wheel left, the problem is that the car is broken, if an LLM does something unexpected after you gave it instructions, that's possible when the LLM is functioning entirely correctly.
reply
Nothing in this world is guaranteed. That doesn't mean it's uniformly random either. LLMs can still do something unexpected if you give them clear instructions, but that doesn't mean it'll be arbitrary and unpredictable in scope. The same way C/C++ undefined behavior technically means program can give you nasal demons, but in reality it won't do anything unusual (like format your C:/ drive) unless someone purposefully coded it to do that.
reply
This is all going to flash through your mind when your car mysteriously doesn't turn left. I would prefer to think of machines as things with defined outputs and failure is failure, more than as fluffy little kittens who might do the wrong thing, if the consequences are going to fall on someone who doesn't deserve it.
reply
[dead]
reply
Exactly!!

I've been trying to work on a new LLM code editor that does just that. When you instruct it to do something, it will evaluate your request, try to analyze the action part of it, the object, subject, etc, and map them to existing symbols in your codebase or, to expected to be created symbols. If all maps, it proceeds. If the map is incomplete, it errors out stating that your statement contained unresolvable ambiguity

I think there is a real benefit here, and it might be the actual next beneficial grounded AI sustainable use in programming. Since I the current "Claude code and friends" are but a state of drunkenness we fell into after the advent of this new technology, but it will prove, with time, that this is not a sustainable approach

reply
There are a lot of humans who refuse to give that answer, too
reply
This continues to be the most tiring response to any criticism of LLM output. It's pretty much guaranteed to show up at this point. I guess with similar enough input tokens, we're guaranteed the same output...
reply
I don’t have to spend dozens if not hundreds of dollars a month to talk to most people in my life lol
reply
Book an appointment with a Psychiatrist, it’ll cost more than a months cc subscription for sure
reply
Do you have to talk to LLMs?
reply
Another way to say the same thing: "to talk to most people in my life lol I don’t have to spend dozens if not hundreds of dollars a month"
reply
There are lots of free LLMs, but you (often) pay for the good ones.

There's lots of people with lots of opinions, but you (often) have to pay for the good ones.

reply
According to HN, every employer, and general social chatter, apparently yes.
reply
Well, speaking from what I hear and see, employers want you to start using it so that you can be more productive. They've been sold this tool and want you to learn it so that your output will grow.

That's not an unfair take, I think. Again, just IME, they expect too much because the tool is oversold: it does not deliver that well. And we always hear, this new model is so much better, it's tiring.

I think we should all learn to use LLMs but we should still carefully review what they did. And that is what the employers don't quite get: the review still takes a lot of time. So, gains are not 10x but more like... 10%? Maybe 50 for boiler plate. Still gains are there, I guess.

reply
> they expect too much because the tool is oversold: it does not deliver that well.

And unfortunately a lot of people will say it’s their reports’ fault for not properly utilizing it (even as they barely use it) because otherwise they would have to admit that they bought a tool without any plan for how to deploy it. So regardless of what is or isn’t a fair take, the results are the same. We are burdened with utilizing a thing whether it is useful or not and the results are generally not what is measured, but rather “are you using it?”

I’m just glad I work at a company that has more reasonable expectations and has been very slowly, thoughtfully rolling it out to individuals at the company and assessing what is and isn’t good for. They are interested in getting me in line, but as somebody in video production to be perfectly honest the use case for Claude is a bit tricky to navigate. We don’t write a lot of scripts and I already have bespoke software for organizing/maintaining footage that isn’t on a subscription basis. The work I’m also doing doesn’t call for these speed-editing solutions that generate tik tok chaff. All our stuff is hours long and it’s high volume. Any video-centric AI service costs an arm and a leg.

I do think it could be useful for writing some terminal scripts and such, but as far as a daily tool we are still scratching our heads and thinking about it. But it’s nice to be able to do that without somebody saying “why aren’t you using it?” every meeting.

reply
You’re absolutely right! I do have insufficient data for a meaningful answer. This is not an *insightful prediction* — it’s *Dunning-Kruger masquerading as qualified intelligence*
reply
No Information before. No information after. This is not a failure — it's narcissism as a service.
reply
YaaS — yes-men as a service
reply
Did a human write this?
reply
I would guess a real human, one with a good sense of humor at that.
reply
Woosh
reply
As measured by #_no_answer/(#_incorrect + #_no_answer) the top current models can do it 60-70% of the time (Grok 4.20 is the best with 83%): https://artificialanalysis.ai/evaluations/omniscience
reply
I reckon that’s how we know we’ve hit ASI.
reply
2061, mark the date
reply
I just came from reddit and seeing this comment, looked for "controversial" sort option instinctively.

Maybe hackernews is becoming reddit...

reply
Just add a skill to Claude
reply