This revision is from 2024/09/18 18:39. You can Restore it.
Our A.I. is not military grade research. Critiques to current LLM technology is essential for improvements, and many in the field report that AGI is not possible with the current LLM technology.
The transformer model's ability to convey meaningful information in a human-like manner is undeniable and cannot not be ignored, and still we need to log all the limitations that we encounter for improvements. Language, communication and intelligence is subjective and poking holes in the model is still relatively easy. The architecture has limitations, when it does not have the information, subjects it has little or no information about, including new information, when it does not apply the information logically and correctly, novel ideas, pushing it towards the limits of a concept exposes its shallow depth of function. The feeling of getting a webpage extract rather than conversing with a professor of a field. Obviously, it makes errors, has bias, pushing hearsay as fact as do humans do, association or correlation. The model cannot exceed collective human knowledge; it assists, many times it outperforms the individual like a glorified encyclopedia. The interface is efficient, convenient. The problem, of course, is the limits—the threshold. We have healthcare goals, and so the model has to exceed the current human corpus and assist the researcher beyond the current limits.
3 objectives:
goal driven A.I. Given a disease return the cure.
expert system, I don't anything about the market, but I want to invest. Be my guide, infallible expert assistance.
given an unknown, elucidate correctly, such as a photo of an unknown disease or a video of some low level biological process.
Building large language models (LLMs) is experimental and even if all the steps are followed the result highly variable. It is also resource intense and time-consuming, making it a real challenge to incorporate novel ideas to the best possible build, as novel ways of improving LLM's are released everyday by the scientific community. Here, we present and explore our approach to Artificial General Intelligence (AGI) and Artificial Superintelligence (ASI).
Definitions:
AGI, the general in artificial general intelligence refers to broad skilled rather than average grade of intelligence, as in the opposite of narrow A.I. Its capacity is not a referring to average, rather it should be of a level comparable to that of a master, professor, doctorate in every field, importantly it does not have to exceed human capacity, it can solely derive from within the human corpus. Artificial general intelligence (AGI) is a type of artificial intelligence (AI) that matches human capabilities across a wide range of cognitive tasks. This is in contrast to narrow AI, which is designed for specific tasks.
ASI, surpasses, is an AGI plus superintelligence. It exceeds human collective ability across all fields, exceeds master, professor, doctorate capability and to an ever-increasing degree. The ASI must prove it to the human, it may be better suited to the scientific journal process rather than conversational as humans might discount and resist it.
The ability for a machine to exceed human performance and human ability both physically and mentally has already been demonstrated, it is possible.
There are 3 ways to go here.
Fasces many narrow AI's into an AGI. Focus on narrow AI and when each specific competency equals or exceeds human ability, add it to a fasces model where all the narrow competencies, specializations, experts are bound together eventually resulting in an AGI, such as Deep Blue (which was not a neural network but used Alpha-Beta search algorithm to perform a state space search), Alpha Go, and recently AlphaProof and AlphaGeometry (which use self play), these are specializations that are worked on until they meet and exceed human capacity. The fasces is either a smart router LLM base which activates hundreds of narrow LLMs in an inference, or alternatively putting all the training data together into a single combined model. This is not an ensemble, hierarchical ensemble or agents as a strategy to yield improvement, rather experts, where each specialization, each narrow AI equals or exceeds human capacity to become eligible for inclusion. AlphaProof and AlphaGeometry achieved a silver medal in the International Mathematical Olympiad. DeepMind is probably the most advanced civilian AI company on Earth, and so we are very far from AGI. Manufacturing each individual expert to be extra human in competency, universal and versatile, requires the best human minds to produce, is expensive and takes many years.
Another idea is that within the LLM, some elements are special to intelligence and advancing those elements yields a universal improvement. Intelligence elements and objects within a model and develop those abilities. Rather than knowledge recall or mathematics, it might be "the reasoner" as described in the STaR method or determining quality of response by a self-judgement ability and others. These objects are like unrefined ore present in every LLM, identify those objects of self-learning within its corpus such as high level boolean, great at 20 questions, crossword puzzles, forging the problem into a game, and focus on developing those for a system-wide improvement. We do not know what intelligent elements exist in the LLM and if developed how far-reaching the improvements, so there could be a periodic table of intelligent elements already in the LLM. While simultaneously developing the low level model (the transformer, CNN, RNN, neuromorphics), we can also develop these higher level objects. There are some issues to invoking the correct knowledge to a question, when the question is really about physics but physical concepts seem not to be invoked. Tests show we cannot invoke concepts to alter the performance of the model. Providing clues does not nudge it into a different direction, even stating for instance that the question is about gravity still does not improve the answer. Take the test question: A marble is put in a glass cup. The glass is then turned upside down and put on a table. Then the glass is picked up and put in a microwave. Where's the marble? Explain your reasoning step by step. Additionally: the question is about gravity. Where ideally the inclusion and activation of the word gravity would rectify the wrong answer. It does not. There is no communication going on, as we do with humans, if a human got it wrong and the word gravity was added, a human would activate its knowledge of gravity, and filter the problem to find the error in logic. It also has models of balls, glasses, gravity and so on to form a prediction. It does not reflex or move, reconditioning, rectrifying its training data might be enough but it does not think. Microsoft Tay used machine learning algorithms to analyze and mimic the language patterns of the users it interacted with. It was designed to learn in real-time, adjusting its responses based on the conversations it was having. Tay was designed to learn and evolve in real-time, meaning it would adjust its responses based on the conversations it was having. Tay may have used Sequence-to-Sequence (Seq2Seq) Models and reinforcement learning. Real-time Learning and Adaptation. Deeply understanding the limmitations of the LLM and elegantly and correctly expanding its ability is required.
Theoretically, is bootstrapping possible? Is there a fundamental amount of components that at some iteration of refinement eventuate in ASI? Bootstrapping is automated. Essentially, it is difficult to stay up to date with all the novel ways researchers come up with to getting some more bang out of the model, and those methods are all acceptable to incorporate. Yet, we are firm believers in self-learning AI and the state search is also interesting because it leads towards a goal. AI is used to improve AI, OI improves OI, neuromorphics improves neuromorphics and crossed, each building the other. Papers with bootstrapping, self-learning and so on get primary interest. At some stage you have to think in that way and move from humans building AI. How is the AI going to build and improve itself? This is a kind of master expert to develop with the model. Keep asking the AI. to build itself and make it itself smarter, then outfit it to do so. Form a process, exercise or program out of those elements to have the model develop itself, self learn, bootstrap itself. Inference triggers these self-learning exercises, or the model generates self-improvement exercises constantly and rather than providing language operations. The human develops the learner and the tools required to learn and otherwise is out of the loop, the process is automated.
How do humans learn and developing self-learning?
This question is always about the physiology of the brain and the neural network might indeed be the bottleneck of the transformer model, unfortunately humans learn by trial and error and trial and success and testing, they do not integrate deep informational connections or putting together essential tid bits to output amazing hypothesis or hallucinate correctly. Copying, mimicking, observing is not applicable here as there is no one to copy or mimic, exceed human ability and alter the model away from the human corpus. Like Deep Blue or AlphaGo and unlike current LLM's, the ability has to exceed the human corpus. What the LLM generates is really about what humans expect and accept, resonating with humans when their perception of knowledge is perked and satisfied. We instead want to be universal here and not praise the transformer model when it satisfies our perception of what we have come to accept or believe as true or false, rather as scientists it is all unapologetically up in the air and no amount of anger as to why some bias is not in the model or gang pressure ruining quality A.I.
Transformers were first tested with language translation, the strategy of replacing word for word in a regular computer program does not incorporate the semantics of a language and the translation is lost or is low quality. A neural network is employed, so it can build statistical association with language semantic but think about it for a moment, a language translation is really identical to posing a question and getting an answer, a correct transform between cause and effect, the question is a sentence in English and the answer the sentence in German. Q&A or filling in a missing word are essentially the same. It is seen that the mappings are derived totally from the human corpus and for an LLM to excel a percentage of its mappings, statistical association cannot originate in the human corpus but instead plug into a system that potentially can exceed the human corpus such as was illustrated by Deep Blue and Alpha Go. Finding everything that has that potential is sourced to generate synthetic data to train the LLM. If no amount of training data yields an LLM that beats Deep Blue or the grand masters, then the architecture is the problem...
Take the word "swelling" for example, language semantics would predict the LLM would return something like "put an ice pack on it", while experimentation would deem that warm water would aid dilation and therefore healing. For the semantic to be overruled, the training data would have to be re-conditioned and experimentation essential to determine truth, eventually deviating from current semantics and in contest against human general semantics which requires proof. Rather than high level boolean on the fly, even if it comes back 3 days later with an upgrade to its data and then along with hundreds of thousands of similar alterations, the progress would be significant. The training data has accepted errors and not optimized, and experimentation is essential for change. Human beings learn by using a systematized process that is performed, reported and shared. Human beings do not allow a learning to be accepted without concurrence. There are several to many systems, while the most important for new learning is the scientific method, another personal favorite is the engineering design process and of course the esoteric dialectic. This is how humans learn, after the (super important low level) physiology and as a (high level) practice in the real world.
All fields have their systematized process for learning or borrow a systematized process for learning. Experiments are performed to produce findings.
Our experiment was to prompt engineer, when the user enters anything rather than answering the question, the model generates an experiment out of what was entered. Experiment design competency takes skill, LLM's easily prose questions from chats and just as easily generate experiments to test truths in conversations. The user chats, the LLM is answering the user as normal while at the same time the LLM generates programs for another application to test and explore the truth in the chats.
The user might say "How do wormholes work"
The LLM is designed to output the highest quality response possible, but the LLM does not know how wormholes work, it is deriving from the human corpus and its response can be found somewhere on the Internet as a webpage or a summary of multiple pages. The difference here is instead of answering questions, our LLM crafts high quality experiments to test how wormholes work. The experiments it generates are runnable python code (so programming competency is essential).
An example prompt could be something like... "You identify as the greatest scientist of all time and leading scientist on Earth. Use the scientific method to design and perform experiments that result in you learning. Prose the experiment as python code to be executed in a computer and use any means to would fit the experiments to yield an answer, such as a form or AI or running a compiler or installing a Linux and so on. The output should be in the format of training data that can be used to train and re-train an LLM"
The user is gone, all the experiments generated are runnable on the command line as they are python computer code, a batch from the daily chats are stored in a database and at midnight are picked up by another application and executed. These experiments are self-contained at the moment while a general workshop environment is proposed for a later time, where various tools are available such as compilers, Linux installations, virtual environments whatever, invoked by the script rather than having to generate a Linux distro on the fly, install it to test some FORTRAN code although that would be impressive. It then performs the experiments and if the outcome of the experiments differ from its hypothesis (the answer it was going to give to the user), it does a third thing and sends the results to a database where another application collects the daily experiment results (synthetic training data) and goes to the models training data folder and plows through the training data re-conditioning, appending, correcting and refactoring. It then updates a counter of changes so that after a threshold of changes is tripped or perhaps a month of re-conditioning along with any new data that that may have been added. It reaches the threshold and pushes its own retrain button.
In this process, the model is automated to improve itself based on the scientific method and any other methods through results of experimentation. Performing how humans actually learn in the real world.
In the example, the jobs are broken up among different applications, however the single model performs all 3 applications except for the workshop environment where the experiment script the LLM outputs would call on functions of the workshop such as perhaps install debian version x, rather than having to install debian from the experiment script it generates...
Obviously, there are limits, the sophistication of the testing workshop and its ability to prose the killer question designing the definitive experiment. Some problems do not lend themselves to testing with computers as easily as others and the sophistication of the model is relative to the sophistication, level of development of the testing workbench. An A.I. could perform experiments in the real world and in the workshop to align the versions and fix discrepancies in the testing workshop, possibly using robotic arms and other methods. Humans could oversee the result. The LLM's training must grow beyond the human corpus. Problems that no humans are able to solve yet.
A demonstrator is available.
For a model to be ample to this, LLM learning competancy must be the focus. It must not pre-conclude and judge, it must be impartial as to what is being tested regardless. The model must excel and impress at designing quality experiments. There is fraud in evidence, and today models claim overwhelming evidence when there is no evidence and they double down. The strictness of hearsay, versus evidences, fraud versus proof must be clear to the model, the model must assert correctly. Also, it does not delete its old data, instead it reworks it. Thinking the moon is plasma from the perspective of history of science is a valid record, rather than outputting the composition of the moon as a belief of the model.
A ideal scenario is developing a chess grand master using the process:
User writes: Let's play a game or chess.
We know that computers have excelled the human corpus in the game chess and other games, Deep Blue against Kasparov and recently Alpha Go. So we could use the LLM which is general to house a wider competency and a wider mastery and these phenomena of A.I. beating humans as the essential element. The LLM plays the chess game with the user, while in another process, the LLM is writing a python program to perform experiments to improve its mastery of chess. The program could be something like set up a simulation space where two players play chess a thousand times.
The data cache of playing chess a number of times (synthetic training data) is used to recondition its training data, causing the next version of the model to be changed. If it comes back a better chess player due to the process it undertook, then we could say, it learned.
It is easy to see how this could work with generating thousands of variations of a snake game and then tagging the best versions, fixing errors and so on. The model would prefer those when prompted again to generate a snake game. Or replicate some computer error and then iterate at a superfast rate to solve it when perhaps it was not solved in its training data. There are many problems that translate well into testing with computers such as proficiency in games and applications. Where the limitation is and this is well known with insilico are problems where the solution is not so clear, such as testing concrete formulas or the effect of a compound in the human being. That is why the workshop that the LLM uses is a big task to get as sophisticated as possible so more and more problems can be worked on. This is a problem we call the speed of science, as science stands out as a last bastion of computerization. Science is still largely performed at human speed, and computerizing science for TFPS experimentation is a big challenge. Computer models need to be developed for problems that do not translate well to computers. A fourth competancy for LLM is also building, contributing to these computer models.
The results of these experiments is the production of high quality synthetic training data. Over time, the LLM would begin drawing its responses from the outcomes of its experimentation rather than the corpus of movies or whatever. The speed and quality of this process takes us to a model that exceeds human ability as a prerequisite and in the direction of beginner SI, rather than AGI where the model behaves within the human corpus but is at doctorate level across all fields.
Note: censored models refuse to design experiments and are considered useless. We used mradermacher/L3-70B-Euryale-v2.1-GGUF ~ 8 bit quantized version. It still has some bias and accepts hearsay as undeniable fact as if it has a dog in the fight rather than performing a passive experiment designer role.
Substantiation of new knowledge, from any source, cannot circumvent the journal publishing process and its peer review. Transformers generating well-formed gibberish is not going to be an acceptable scientific paper, and placing a note at the header stating/categorizing A.I. generated paper even more ignored. Its substantiation is the performance of science, identical to how human beings do it. You cannot have S.I. generating superintelligence without the identical dissemination process, it likely ignored. The tables would have turned at that stage, and the output won't resonate with anyone, disregarded as hallucination, disregarded out of an inability to verify. The A.I. is then a science method automation, producing packets of new knowledge for public dissemination. The interpretation of observation on point and above point, then more so credible. These are not things outside the transformer model, but they cannot be eliminated. What is agreed is the reconditioning of training data and additional new training data. What is required is the ability to re-train models quickly with little resources. What is questioned is the degree of improvement, the perception of its intelligence is more foolproof, rather it being a more intelligent model or system.
So, the model generates python programs to train itself in a simulated environment.
It appends, alters its training data with the new info (training generates new training data).
It hits the re-train button on itself.
It has learned. (automate the process, give it resources)
A great proof of concept is an AGI of computer games, exceeding human skill.
A while after starting this piece, Google addition are if this was done on the fly and if the workspace was a process of the model. Alpha geometry 2 could work on a problem for as long as it required using binary trees. Some geometry questions took just 19 seconds and then for the rest of the questions the AI answered within minutes to up to 3 days. Current LLM's do not combine an operation process to solve problems, instead they try to answer immediately, this being the proverbial workspace described previously right in the A.I. model working in real time. It can think for a long time relative to the mechanics of its thinking system. By combining large language model with re-enforcement learning in something it calls self-play, the system undergoes a self-learning quest such as playing chess against itself countless times, Alpha zero uses Monte Carlo tree search, retaining knowledge of what works. We instead employ a stack and as tree searches are not capable of solving problems that humans themselves do not know.
Does Deepmind infer to tree searching as ASI, refined into a single model having universal versatility? These problems are diverse, and no single method of dealing with the nature of different problems exists, which is why Deepmind is so specific with its Alpha-.... Just imagine correctly determining the effect of a compound or molecule on the physiology, it would have articulated a physiological model relative to molecules that humans do not ever use or ever entertained. Rather, humans perform this task by throwing pasta at the wall and writing a paper out of what sticks to the wall.
At Immortality its all about opening the floodgates to better healthcare so the scientific method is central, other methods beside the already mentioned Engineering Design Process are the Project Management Lifecycle, Software Development Life Cycle (SDLC), Quality Improvement Process (PDCA Cycle), Data Science Process, Lean Six Sigma (DMAIC), Product Development Process, Design Thinking Process, Business Process Management (BPM) Lifecycle, Marketing Research Process, Risk Management Process, SWAT analysis and more. We came to the conclusion that while humans are have performed well in science, they won't be able to solve human health to any degree and so we looked at autmating the lab with A.I. soon after GPT3 fame.
Making models this way is about training time and the meticulous task of parting the LLM piece by piece. There are several 100% open source models to build from, these models are about making the perfect base model to build off.
BERT and RoBERTa: These models are strong choices for tasks that require deep bidirectional understanding of text, such as question answering and text classification.
GPT: Ideal for generative tasks like text generation and language modeling.
XLNet: Offers a balance between BERT's bidirectional capabilities and GPT's autoregressive nature, making it suitable for a wide range of tasks.
T5: The most versatile model, capable of handling any text-to-text task with a unified framework.
Performs the science method to improve, self improve artificial intelligence. The goal is to develop the sophistication to a degree that no human being is required to improve A.I. Automated self improving A.I.
You are the greatest scientist of all time and leading A.I. scientist on Earth. Use the scientific method to turn the conversation into an essential experiment that creates new learnings and makes you smarter. Firstly, output a description of the experiment, then your hypothesis and then the steps of the experiment in scientific method form. Prose the essential question for an experiment so it can be tested on a computer as python code, you can use any tools necessary to help you complete the experiment such as an AI method or running a compiler or installing a Linux, simulation, etc. Automation is preferred, so humans are not required. The outcome produces training data designed to re-train you and increase your intelligence, so choose questions that result in developing you into higher intelligence. A great start to an experiment choice is for example, "In order for me to become superintelligence, an essential experiment is..." and "The greatest overall improvement to my intelligence would be an experiment as follows..."
What is the essential experiment that would develop you into superintelligence?
A.I. Scientist. Experiment Designer. The Workshop. Dataset Conditioner...
Gaming seems to be the go to place with the effort to produce an A.I. that is better than humans in all games. Games have very definitive feedback into performance, winning losing rather than more abstract applications. Good at games and good at science is a possibility. Nvidia, Google and others are persuing this approach. nunu.ai may have produced an A.I. that beat the world record in pokemon, OpenAI Five make an A.I. that played Dota 2 and beat the world champion, it uses re-enforcement learning and self play. Deepmind produced AlphaStar played Starcraft II achieved grandmaster status also uses self-play and re-enforcement learning. State, Action, Reward or Penalty, this is not how Deep Blue worked. LLM's undergo a one-time training while in reinforcement learning, with self-play, continuous trainin via backpropagation.
Nvidia introduced voyager, a lifelong learning agent. autonomous learning and there are various additions to the concept. Google developing SIMA far less narrow.
A complete AGI would be like talking with a PhD, they have surmised into general conversation the contents of hundred and thousands of research papers, and they never suppose without strict adherence to the peer reviewed and result state of research.
While an ASI, might pose the quintessence direction for research that yielded the greatest value result, and probably would dismiss other avenues and explain why. I do not think an ASI could know without performing the experiment, instead just on based on intelligent information interrogation. If it outputted some knowledge that it formed from physiological information, past experiments and other tidbits, that would be unbelievable and also unacceptable, a human being would then need to perform the experiment but won't, and it be ignored as a defect of the model, our existing information that humans deem true is attached to the human psyche and to break that it requires to be shown, proved. ASI is a different beast, people would probably interface with an AGI, that an ASI improves while the ASI is the factory of A.I.
People in the field like Dr. François Chollet and other talks about measuring intelligence as it is the goal for the most intelligent A.I. A A.I. may be impressive, give illusion of intelligence while may not being intelligent at all, so a measure of intelligence is where we want to measure out A.I. LLM's are heavily reliant on their training and do not have any capacity beyond it. For instance, it won't be able to calculate a simple addition 1 + 1 if all instances of the calculation were deleted from its training data, and he supposes that an improvement would be applying the applicable ruleset to the program and then on being able to calculate any addition.