Hacker Newsnew | past | comments | ask | show | jobs | submit | Azkron's commentslogin

This reminds me of the CEO of a cyber security company that challenged Anonimous https://en.wikipedia.org/wiki/HBGary. If you work for any kind of security company, do not ever ever ever challenge any kind penetration specialist. Everything is hackable, it is only a matter of cost vs reward, but when you challenge someone that goes out of the window.


Generalizing the advice..

Don't challenge people. Don't insult people. Don't humiliate people. Don't threaten people. Allow them to maintain their self-respect even when they lose. Don't rub it in. Give them a face-saving exit.

Plenty of violence and aggression is caused by violation of the above rules. They seem simple but they're broken on a daily basis. Famous last words: "you don't have the guts".


It would be nice if it had the option to indicate you which finger you should be using to press the key ideally. Like the mechanigraphy games.


| "Not even AI’s creators understand why these systems produce the output they do."

I am so tired of this "NoBody kNows hoW LLMs WoRk". It fucking software. Sophisticated probability tables with self correction. Not magic. Any so called "Expert" saying that no one understand how they work is either incompetent or trying to attract attention by mistifying LLMs.


You are assuming there is no such thing as emergent complexity. I would argue the opposite. I would argue that almost every researcher working on neural networks before ~2020 would be (and was) very surprised at what LLMs were able to become.

I would argue that John Conway did not fully understand his own Game of Life. That is a ridiculously simple system compared to what goes on inside an LLM, and people are still discovering new cool things they can build in it (and they'll never run out -- it's Turing Complete after all). It turns out those few rules allow infinite emergent complexity.

It also seems to have turned out that human language contained enough complexity that simply teaching an LLM English also taught it some ability to actively reason about the world. I find that surprising. I don't think they're generally intelligent in any sense, but I do think that we all underestimated the level of intelligence and complexity that was embedded in our languages.

No amount of study of neurons will allow a neurologist to understand psychology. Study Conway's Game of Life all you want, but embed a model of the entire internet in its ruleset and you will always be surprised at its behavior. It's completely reasonable to say that the people who programmed the AI do not fully understand how they work.


Whatever comes out of any LLM will directly depend on the data you feed it and which answers you reinforce as correct. There is nothing unknown or mystical about it. I honestly think that the main reason big tech claims they “don’t understand how they work” is either to avoid responsibility for what comes out of them or as a marketing strategy to impress the public.

EDIT: By the way, I definitely think LLMs are intelligent and could even be considered “synthetic minds.” That’s not to say they are sentient, but they will definitely be subject to all kinds of psychological phenomena, which is very interesting. However, this is outside the scope of my initial comment.


> Whatever comes out of any LLM will directly depend on the data you feed it

Right, and whatever comes out of Conway's Game of Life will directly depend on its initial setup as well. Show me a configuration of Conway's Game of Life that is tailored to emulate human speech and trained on the entire internet and then tell me your prediction of how it will evolve. You will get it completely wrong. Emergent behavior is a real thing.

> There is nothing unknown or mystical about it.

Almost all researchers and practitioners in the field seem to disagree with you on this. It is surprising that teaching a system to be extremely good at auto-completing English text is enough for it to develop an ability to reason. I happen to believe that this is more of an emergent property of our language than of neural networks, but it was definitely not predicted by almost anyone, not easily explainable, and maybe even a bit mystical-feeling.

Ph.D. dissertations have been published about trying to understand what is happening inside large neural networks. It's not as simple and obvious as you make it out to be.


This isn't suggesting no one understands how these models are architected, nor is anyone saying that SDPA / matrix multiplication isn't understood by those who create these systems.

What's being said is that the result of training and the way in which information is processed in latent space is opaque.

There are strategies to dissect a models inner workings, but this is an active field of research and incomplete.


Whatever comes out of any LLM will directly depend upon the data you fed it and which answers your reinforced as correct. There is nothing unknown or mystical about it.


The same could be said of people, revealing the emptiness of this idea. Knowing the process at a mechanism level says nothing about the outcome. Some people output German, some English. It’s sub-mechanisms are plastic and emergent


So many words there carrying too much weight. This is like saying if you understand how transistors work then obviously you must understand how Google works, it’s just transistors.


I guarantee you that whoever designed Google understands how Google works.


The relevant research field is known as mechanistic interpretability. See:

https://arxiv.org/abs/2404.14082

https://www.anthropic.com/research/mapping-mind-language-mod...


This is a bit like saying a computer engineer who wrote and understands a simple RISC machine in college thereby automatically understands all programs that could be compiled for it.


No this is like saying that whovever writes a piece of software understands how it works. Unless one forgot about it or stumbled upon it out of sheer luck. And neither of those are the case with LLMs.


I agree. This is a very dangerous marketing and legal stragety than can end up costing us very dearly.


"Guerrero" is a common last name in Spain.


For many people "wealthy = evil". And "poor = good". It is easier to demonize someone that is doing better than you than to admit that maybe he is just making better choices.


The poor can be anything. A wealthy person could have worked hard for it. A wealthy person could have also exploited others in order to get wealthy.

Virtually all (meaning systemically) very wealthy people had to exploit others to get to their very wealthy state.


Something that has helped me a lot when dealing with noise or other distractions is present time meditation. Just sit there and focus on actively hearing and feeling everything, I would just do 5 min sessions but quite often. After practicing this for while I became less distracted and more comfortable wherever I was.


Did you follow some guide or just "made it up" this particular method? Interesting!


Visual schemas and sketches help greatly when dealing with complex deep abstractions. There is no need to keep it all in your head all the time. Just summarise the architecture in a visual way which can be fast to look up.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: