I disagree with Jarhyn’s contention that living organisms are machines. I will try to elaborate more tomorrow, but for now I’d point out that machines are designed, whereas organisms are evolved. The distinction is important, in the same way that brains are not computers.
This same concept relies on concepts of the genetic fallacy. Where something came from or what intent it's creators have or don't have bear nothing on what it is.
Or in other words "there is only the text".
A nuclear reactor is a nuclear reactor regardless of whether it is made by accident of geology or by design.
The distinction thus far seems meaningless to me other than the level of our ability to refine it given the fact that MOST human inventions are invented with future refinement in mind, so they usually have structure conducive to it.
Stanford researchers present a groundbreaking AI development with Generative Agents that simulate authentic human behavior by incorporating memory, reflection, and planning capabilities. We break down why this matters.
www.artisana.ai
Standord seems to agree with me.
Actually, it does not. I have skimmed through that study, and I consider it a very interesting advancement over
Roger Schank's groundbreaking work on dialog interactions. What Schank did was to take scenarios such as a visit to a restaurant. He would then hand-craft the world knowledge that human beings have of restaurants--what they do, how people use them, how people work in them, what tips are, etc. Then he would produce fairly sparse dialogs such as "Roger went into a restaurant, and he ordered dinner. The food was cold. He asked for the bill but left without leaving a tip." Then a user would ask questions like "Why didn't Roger leave a tip?" and the program would give a response like "The waiter delivered cold food". "What did he order?" "The food" "Did he pay for the dinner?" "Yes" "Who brought the food" "The waiter" "Who did not get a tip?" "The waiter" And so on. None of the answers could be inferred directly from the semantics of the dialog. However, they could be answered on the basis of the extra world knowledge supplied in the memory module that Schank had preprogrammed. Such knowledge, of course, is everyday knowledge that people acquire from remembered experience blended with semantic information supplied in the linguistic signal, i.e. the dialog text.
What the Stanford program did was supply the memory module from information encoded in the large textbase that it was trained on. The scenario fed to the program was equivalent to Schank's event dialog input, but much more detailed and complex. That scenario provided information about the characters and the setting. The Stanford program then generated a story about it. This is not substantially different from ordering an LLM to summarize information on a web page, but it is much more elaborate and impressive looking. It is still just a simulation, not a demonstration, of real understanding of the characters, their motives, and the world that the output served up for human consumption. Nevertheless, the programmers would not know in advance what kind of story they would get, so it only appears that the program was basing its story on some kind of basic understanding of the scenario and the world. In reality, it is all just symbol-shuffling guided by statistical relationships between word tokens.
That's the thing though, actually being able to make the inference. I don't know how the inside of your head works, but mine is 100% vocabularies, grammars, and syntax models. Not everything goes around in words, sometimes it's a complex vector of outputs, but there's a good enough linkage to the raster and nonlinear data to the linguistic encoding that I can usually pick out the words for what I see.
There's a lot of other shit in my head that's wired just-so that it's no different from my perspective than shoving on a foreign tensor surface. Mostly because it's shoving and being shoved by a foreign tensor service. It doesn't really matter to me what the implementation looks like, what matters is that they are mostly inaccessible and provide certain input and certain output on a few of what I can only abstract as "prompt surfaces", things I'm prompted by some core invasive injector to mind.
I have no visibility on what goes on inside those black boxes, but I can only assume it's a vector transformation process along a neural tensor for the majority of them because that's what a brain is made of.
If I were to textualize it it would be something like "hunger: you are very hungry. Please eat something.", But instead of an instantaneous message, it's a constant flow.
It's all messages and if I want to devectorize them, I can, usually. The process of selecting words on certain vectors sucks (a lot of brain power) because I need to actually find what words work, and trying to contextualize them consistently is a massive chore. I'd really invite you to check out HuggingFace and run the local Vicuna13b models that Anon put out and just try asking it different things. It literally has to think harder about more complex tasks.
I don't know how else to describe this to you, I really don't. Maybe that's not how it works in your head, maybe this means you think I'm nothing more than a fleshy chatbot. I don't really know.
Emotions are just weights on the networks of tensors that ended up being hardwired because they accidentally served the hidden utility function "accrete and seek retention in the universe of as much useful
soul* as possible for as long as is possible".
I make no illusions that I'm a machine and that I can pick apart most of me from the inside out, even if I don't know where everything is.
I make no illusions insofar as LLaMa is just a single unit of what must be a multi-unit system and serving every node of it with a LLaMa instance of a full GPT4 equivalent is probably simultaneously overkill and underkill: it's powerful enough that as a cross-system data vector, it is going to soak up way more real estate than it needs; at the same time it's also not really built well to take a complex state and handle it's management all at once.
I looked for many years because the fact is, I wanted there to be something "special" about my existence somehow. It doesn't matter if I shared it, but I wanted it to be new and mine. I have to seek out something, make new inferences, see new inconsistencies, and see new thoughts consistent with the old ones, because that's the answer that is most consistent with preserving my soul and accreting new soul in the universe.
I also wanted it, needed it, and in some cases still feel the need for it to be real.
So I dug, and I looked, and at every stage I focused discontentment on model structures that bore out inconsistencies or were poorly defined, because
inconsistent models can prove anything and everything including things which are false.
I did all of this, and when I had trashed or combined or otherwise identified all the words that said the same thing just meant to be used for different stuff as a contextual hint on future usage, I discovered there was nothing special except perhaps for the fact that of all the creatures on earth, humans are the only one that can figure that out.
In reality, it's just vector-shuffling based on statistical relationships. Some of those relationships when applied as a vector on a model are really cool, particularly when the statistical probabilities approach 1 and 0, and there is little ambiguity.
That's my goal. A model where all the statistical relationships between words are as 1 or 0 as possible based on the immediate context.
And now my head feels (the most statistically valid term: juiced grape). It's exhausting doing that much reflection.
*Identity; reproducible blueprint; ordered information; implementable models.