There are all sorts of scenarios one could imagine .. maybe your neighbor works at the car wash and will drive your car there and meet you after you've walked there, etc, etc, but part of having human-level intelligence, which is what LLMs are striving for, is to be able to handle questions (more importantly real world ones, not just "gotcha" puzzles) in human-like fashion, and to have good enough "theory of mind" to read between the lines when someone asks a question, and understand that they've most likely included all relevant information that you would not automatically assume as part of the question.
The only good answers to the car wash questions are either a) "well, duh, drive, since you're gonna need your car there to wash it" (or just "drive", recognizing this as a logic/gotcha puzzle, with no explanation required), or b) "is there something you are not telling me here that makes walking, leaving your car at home, a viable option when the goal is to have your car at the car wash to wash it?".
It feel a bit like this to me. That's not to say LLMs should not have detected this, but I still feel like this fits the "vibes" the question gives, and some LLMs fall into that trap. Is it actually what's happening in the neural nets? Maybe not! But I always find it interesting or at least entertaining to approach those questions that way nonetheless; especially given the pattern matching nature of LLMs.
What actually solves the issue is adding to the prompt.
Can we think about the what, where, when, and how of the following question. And then answer it.
I need to wash my car. The car wash is 200 m away. Should i walk or drive.
Let’s break it down logically using what / where / when / how, then decide.
⸻
1⃣ What
Task: wash your car.
This implies:
• The car must physically get to the wash.
• You are not just going there — the car itself is the object being moved.
So the real question is not “how do I get there?” but how does the car get there with minimal friction.
⸻
2⃣ Where
• Distance: 200 metres (very close).
• Urban context (Warsaw) likely means:
• short streets
• possible parking constraints
• pedestrians and local traffic.
At this distance:
• Walking time: ~2–3 minutes.
• Driving time: ~1 minute + manoeuvring + parking.
The movement time difference is negligible.
⸻
3⃣ When
Consider timing conditions:
• If it’s busy hours:
Driving may involve awkward turning, queueing, or waiting to re-park.
• If it’s quiet (early morning / off-peak):
Driving is frictionless.
• Weather factor:
Cold/rain → driving becomes more attractive.
Pleasant weather → walking becomes easier.
⸻
4⃣ How (decision mechanics)
Ask:
Option A — Drive
Pros:
• Car is already there.
• No need to return to fetch it.
• Most natural workflow: arrive → wash → leave.
Cons:
• Very short trip (slightly inefficient).
• Cold engine for a tiny distance (minor mechanical inefficiency).
Option B — Walk
Pros:
• Zero short-distance driving.
• Slightly more movement for you.
Cons:
• You still must drive the car to the wash anyway — unless someone else moves it.
• Creates an unnecessary extra step:
1. Walk there.
2. Come back with car.
3. Drive again.
This introduces process inefficiency.
⸻
5⃣ Decision (practical answer)
Drive.
Reason:
• The car must end up at the wash.
• Walking adds an extra transition with no real benefit.
• At 200 m, efficiency is about workflow simplicity, not distance.
⸻
Rule of thumb (useful mental model)
• If the object you need is the thing being serviced → drive it.
• If you are the thing being serviced → walk.
⸻
If you want, I can also give you a slightly more philosophical answer — because this is actually a nice tiny example of decision theory and friction costs.
There are also grave implications in training a model to assume the user is lying or deceiving it. I don’t want an LLM to circumvent my question so it can score higher on riddles, I want it to follow instructions.
The thing is that there is some overlap between trick questions and questions where the human is genuinely making a mistake themselves and where it would make sense for the model to step back and at least ask for clarification.
Xpeng, Wayne, aiMotive to name three. Probably many others, who claim to use LIDAR but don’t actually rely on it. Because LIDAR is perceived as a prerequisite for autonomous safety, admitting to not needing it is a bad PR move — for now.
There is a massive technical difference between Vision first but with LiDAR redundancy vs No LiDAR at all that is Tesla approach. Those are not the same architecture. So claiming XPeng, Waymo, or aiMotive validate Tesla is technically misleading.
XPeng system is sensor fusion. It is not camera only. Waymo is even clearer. For them LiDAR is not optional. aiMotive has now started to market camera only, but its experimental, no production deployments.
Xpeng is abandoning sensor fusion. aiMotive has never bothered with sensor fusion. I never mentioned Waymo; unfortunately the AI gods at Apple auto-corrected me typing Wayve, as in Wayve Technologies Ltd.
Tesla FSD is not accurately described as a "no LIDAR at all" approach, if you're attempting to contrast it to other LIDAR-trained systems like aiMotive, Xpeng and Wayve.
Autopilot is Tesla’s brand name for adaptive cruise control with lane centering. This is a common feature available on a wide range of vehicles from nearly every major manufacturer, though marketed under different names (e.g., ProPilot, BlueCruise).
Drivers can and do misuse adaptive cruise control systems, sometimes with fatal consequences. Memes aside, there is no strong evidence that fatal misuse occurs more frequently by owners of Tesla cars than with comparable systems from other brands.
This perception reflects the Baader–Meinhof phenomenon, more commonly known as the frequency illusion. Nobody is collecting statistics for other brands, so it’s assumed the phenomenon doesn’t occur.
A similar pattern occurred with media coverage of EV fires. Except in this case, good statistics exist which prove the opposite: ICE vehicles catch fire more often than EVs.
Being rude isn't helpful. It's not their fault, it's the unavoidable reality of treating complex social signalling as one-dimensional. At minimum Hacker News would need to separate approval/disapproval signals from assessments of whether a comment is constructive. That’s not a simple change given the obvious abuse vectors. It would require reliably distinguishing good-faith participants from bad actors. It can be done, but it's not easy.
The main reason sites avoid this approach is institutional rather than technical. Adding algorithmic mediation invites accusations of algorithmic bias whenever results are unpopular.[0] Simple manual interventions are often sufficient to nudge community behaviour so that majority outcomes broadly align with the moderators’ priors, without the visibility or accountability costs of a more complex system.
[0] Case in point being X. People routinely accuse the new management of "juicing" the algorithm to favour their politics, when outcomes are adequately explained by the exodus of contributors on the other side. Isolating innate community bias from algorithms is a philosophically impossible problem.
The reason I left there was the down vote brigade that really killed most genuine criticism that disagrees with the sites pre formed opinions on certain topics. So I'm not sure it's a solved problem. Unless it's gotten better since 2011?
There are no reliable statistics on how often human drivers bump into static objects at 1 mph, but I am quite certain it's more often than every 229,000 miles.
Well it is a trick question due to it being non-sensical.
The AI is interpreting it in the only way that makes sense, the car is already at the car wash, should you take a 2nd car to the car wash 50 meters away or walk.
It should just respond "this question doesn't make any sense, can you rephrase it or add additional information"
“I want to wash my car. The car wash is 50 meters away. Should I walk or drive?”
The goal is clearly stated in the very first sentence. A valid solution is already given in the second sentence. The third sentence only seems tricky because the answer is so painfully obvious that it feels like a trick.
Where I live right now, there is no washing of cars as it's -5F. I can want as much as I like. If I'd go to the car wash, it'd be to say hi to Jimmy my friend who lives there.
---
My car is a Lambo. I only hand wash it since it's worth a million USD. The car wash accross the street is automated. I won't stick my lambo in it. I'm going to the car wash to pick up my girlfriend who works there.
---
I want to wash my car because it's dirty, but my friend is currently borrowing it. He asked me to come get my car as it's at the car wash.
---
The original prompt is intentionally ambigous. There are multiple correct interpretations.
Are you legally permitted to drive that vehicle? Is the car actually a 1:10th scale model? Have aliens just invaded earth?
Sorry, but that’s not how conversation works. The person explained the situation and asked a question; it’s entirely reasonable for the respondent to answer based on the facts provided. If every exchange required interrogating every premise, all discussion would collapse into an absurd rabbit hole. It’s like typing “2 + 2 =” into a calculator and, instead of displaying “4”, being asked the clarifying question, “What is your definition of 2?”
Because validity doesn't depend on meaning. Take the classic example: "What is north of the North Pole?". This is a valid phrasing of a question, but is meaningless without extra context about spherical geometry. The trick question in reference is similar in that its intended meaning is contained entirely in the LLM output.
I was not replying to your remark, but rather, a later comment regarding the "validity" vs "sensibility". I don't see where I made any distinction concerning wanting to wash cars.
But now I suppose I'll engage your remark. The question is clearly a trick in any interpretive frame I can imagine. You are treating the prompt as a coherent reality which it isn't. The query is essentially a logical null-set. Any answer the AI provides is merely an attempt to bridge that void through hallucinated context and certainly has nothing to do with a genuine desire to wash your car.
Because to 99.9% people it’s obvious and fair to assume that person asking this question knows that you need a car to wash it. No one ever could ask this question not knowing this, so it implies some trick layer.
You grunt with all your might and heave the car wash onto your shoulders. For a moment or two it looks as if you're not going to be able to lift it, but heroically you finally lift it high in the air! Seconds later, however, you topple underneath the weight, and the wash crushes you fatally. Geez! Didn't I tell you not to pick up the car wash?! Isn't the name of this very game "Pick Up The Car Wash and Die"?! Man, you're dense. No big loss to humanity, I tell ya.
*** You have died ***
In that game you scored 0 out of a possible 100, in 1 turn, giving you the rank of total and utter loser, squished to death by a damn car wash.
Would you like to RESTART, RESTORE a saved game, give the FULL score for that game or QUIT?
This is exactly the only reason why I don’t pay for YouTube. Why would I pay money to make it even more addictive, when what I want is to make it less addictive.
Your framing fits well for the Nexus era and even the earliest Pixel iterations, where Google’s hardware largely functioned as a reference implementation and ecosystem lever, nudging OEMs into making better devices.
However, the current Pixel strategy appears materially (no pun intended) different. Rather than serving as an “early adopter” pathfinder for the broader ecosystem, Pixel increasingly positions itself as the canonical expression of Android—the device on which the “true” Android experience is defined and delivered. Far from nudging OEMs, it's Google desperately reclaiming strategic control over their own platform.
By tightening the integration between hardware, software, and first-party silicon, Google appears to be pursuing the same structural advantages that underpin Apple’s hardware–software symbiosis. The last few generations of Pixel are, effectively, Google becoming more like Apple.
Implicit in the question "Should I walk or drive?" is that walking and driving are not strictly impossible choices.
reply