Its a bit like choosing JS / python -- of course performance is inferior to a compiled language with highly tailored code, but they are flexible and have an ecosystem that might do 99% of the lifting for you.
But in isolation, I agree with your idea that specialized robots with form fitted specifically to task will likely outperform a more generalized solution in a specific domain of behavior, the more generalized will likely outperform in flexibility and reusability (e.g. capable of reusing the human ecosystem).
You don’t need a human-like hand to hold a tool made for humans. As an extreme example, you can make a robot operate a power drill with strap to hold it and a servo with a small bit of wood to operate the trigger mechanism.
But for a robot operating in a space made for humans there certainly are some physical requirements which are based on the human form: maximum volume and clearances, stairs, fragile fixtures that can’t be operated with too much force, etc.
Ever walk through some over-crowded antique shop where you need to twist and lean your body to avoid knocking into thing?
What makes human hands especially suitable for e.g. assembling a phone or installing a door handle onto a car?
yes. do you think it's safe to just plug usb into some hole and type? the safest option for a robot is typing with fingers
I personally am not bullish on 1:1 human hands either, but IMO the question shouldn't be $100k 2 ton Kuka arm vs biped with hands, it's overactuated robotics (build it from the floor with hard coded operations) vs underactuated (build it from the contact point of the work backwards with ML and sensors). We shall see which form factors prevail, but the type of robotics development posted here seems like the way forwards regardless, an ecosystem of small, power dense, reliable, accurate QDD actuators will lead to many general purpose robot applications. I recognize I am not using underactuated vs overactuated in their strict definition here but if you are familiar with robots I think you'll understand where I am coming from as far as a robot design ethos.
I will say though in designing robots of this type without necessarily being bound by trying to make a robot look like a human, I have often found myself accidentally recreating human arm DOF in a round trip way, it does just end up being well packaged beyond the "world designed for humans" talking point. Maybe hands will end up being a similar situation.
Not to dismiss the value of LLMs in those cases as an interface/interpretation layer.
If grandma goes into the windowless surgery factory, I just want the best bots working on her. There is value in having Dr. Bot the replicant give me the face-to-face status updates. We are not breaking out those layers as much, anymore, as the focus becomes minimizing FOMO.
Similar to how claude code gained so much traction in terminal by just leveraging the command line interface that already exists for humans, no need to invent a domain specific MCP to just run shell commands.
I agree with you that it's far from the most efficient approach for specific tasks. But the analogy would be that you also generally don't want to use LLMs to do something you can "just" write a script for... that doesn't make LLMs useless though.