>Yes, I didn't read the whole posting >>24816 at first, because it started responding to the "shell" distraction. I'm going to look into Justified Programming.
I will try to better order parts of my posts, any suggestions on how to make more helpful and appealing? If a specific reply for something you don't care about, just skip it :^)
>I'm not sure if optimizations for conversations, like faster understanding and anticipation where a conversation goes, should also go in here or in another thread
I think it fits this thread, that’s absolutely a higher level function.
>separating it from conversational AI is probably impossible, since we need to use it for inner dialog and retrieval of data from LLMs
I’d argue that is not the case, and that it’s actually essential to not overly depend on LLMs especially for knowledge storage. Take what I’m about to say with a grain of salt, I am simply stating what my gut feeling is. I will argue that an LLM should be treated as purely part of a language center and that storing facts (like what is the capital of X or Who is the Yth president?) is silly, a red herring and a waste of resources. Storing knowledge this way makes it hard to update, makes LLMs expensive to train. Adding new knowledge via fine-tuning is not effective, but fine-tuning is good for changing the style of text (1). Here is something I found inspiring when thinking about how to integrate LLMs into a larger system (2).
This also ties in to "Hallucination", I find the way it is talked about and even the coined term annoying and unhelpful. You should not be "solving" it, its not a problem, if anything it demonstrates that a model is good at predicting the next token and is not over fitted on the data. The problem with "hallucinations" is that pattern prediction and long term memory is being conflated as one. Maybe I am not understanding something, but I am not impressed by the larger LLMs, I see them as a step backwards. Maybe I am wrong and an anon can tell me why?
>I try to avoid these people, I don't engage in this kind of discussions, I don't care, I will do it anyways, if I can.
Thank you, your a breath of fresh air.
I am so tired of this same muh "Roko's basilisk", Sam Altman regulate me harder daddy, reddit """debate""" bullcrap, it's the same style of bullshit as "humanity is bad", its forced garbage pushed by media to keep you (us) down, please reconsider what you consume friends.
Higher level thinking is important for AI safety, I am sure even the nicest of people sometimes have an angry or bad thought, but they don't spiral into a rage on a drop of a hat simply due to that thought popping into there head. Thought or Prediction is not equal to action, there clearly is a gating mechanism for actions. If your AI is just an LLM feeding into it self, id be more worried, An idea I have noted down: emotions are a gating mechanism for actions and a mechanism for determining the kinds of goals and agent prioritizes.
>- The system will need to be as flexible as possible, making it possible to add new features everywhere
I have been thinking about how a practical architecture would look like, the current idea I’m on is going with is a central event bus and a bunch of modules/plugins observing and adding events onto the bus. Each module when it emits an event provides a confidence score. If there is several modules competing for output, the highest score wins. Modules will also be provided a "feedback" score, this can be a manual input from anon or module(s) tracking various inputs to predict anon engagement.
The bus would be the ideal place to have logging and to visualize the system's state.
Also I would like to advice anons to not be soydevs, do not build garbage that I will rewrite out of disgust :/ do not be inspired by web garbage like Firebase, GraphQL or whatever else, the cloud sucks and self hosting web crap is such a pain, the web ecosystem, especially nodejs/npm and python is an example of what not to do. If you even think about using docker, or having one or two official linux distros for building, that is a strong indicator that your code is hard to maintain garbage. For example why would you use GraphQL when you could just directly query the database and have less moving parts. Your building a system to be used and run by the same party, Your waifu should be running on your own machines anyway.
Some technology I want to bring attention to:
1. SQLite, people really under estimate what SQLite can do, it's a very powerful and extendable database that will run anywhere, do a deep dive. You can do document databases(3), graph stuff(4) and even embedding vector search (5).
2. D programming language, its my choice of systems programming language, If picking it is too scary then go with C++, just make sure to expose a nice pure C api!
3. GGML (6), It's the ML library used in llama.cpp & whisper.cpp. It's so magical to type in make into my terminal and getting an executable that just works.
>I'll confess my cowardice to the board, heh. Plastic Memories hit me so hard that it's one of only 2 animus I don't want to watch again
Plastic Memories, yah that one hits hard, its brutal.