- Parameters: The key measure for artificial neural networks is the number of trainable parameters. Each parameter represents a connection between simulated “neurons” in the network. More parameters allow the model to learn more complex patterns.
- Billions of parameters: Models like GPT-3 have 175 billion parameters. Other large models like Tencent’s Wu Dao 2.0 and Anthropic’s Claude have similar magnitudes of parameters (100s of billions).
- Trillions of parameters: In 2022, Meta and DeepMind announced models with trillions of parameters – Meta’s OPT-175B has 175 trillion, DeepMind’s Gopher has 280 trillion. These are the largest models to date.
- Neurons: However, these artificial neural networks do not contain actual neurons. The “neurons” are just mathematical functions, and the parameters represent connections between them.
- Biological neurons: In contrast, the human brain contains around 86 billion biological neurons. Each neuron is a real cell that connects and transmits signals to other neurons via synapses. The complexity arises from the trillions of intricate biological connections.
So in summary, while the largest AI models may have trillions of parameters, equating these directly to biological neurons is not accurate. The advanced AI capabilities arise from massively layered mathematical functions, not real brain-like biological complexity. The goal is to mimic aspects of cognition, not precisely replicate the brain.
How many billions of parameters is Claude2?
- Claude 1.0 had 12 billion parameters when launched in April 2022.
- In June 2022, Anthropic mentioned they were training Claude 2.0 which would be significantly larger than Claude 1.0.
- In recent presentations, Anthropic has referenced benchmarks using a model with 35 billion parameters, implying this could be the size of Claude 2.0.
- Anthropic has stated Claude 2.0 contains “tens of billions” of parameters.
- Based on their compute usage, some AI researchers estimate Claude 2.0 could have between 30-60 billion parameters.
So while the exact number is not public, it seems reasonable to estimate that Claude 2.0 likely contains 35 to 60 billion parameters based on these clues.
This makes it larger than GPT-3 (175 billion), but smaller than models like Meta’s OPT-175B (175 trillion) or DeepMind’s Gopher (280 trillion). But Claude 2.0 is designed specifically for conversational AI rather than general capabilities. The number of parameters alone does not determine the performance. But having billions of parameters allows Claude 2.0 to learn more complex language patterns and skills.
Comparision chart of Claude 2.0 35-60 billion parameters vs human 86 billion biological neurons, average general intelligence
| Metric | Claude 2.0 | Human Brain |
|---|---|---|
| Architecture | Artificial neural network | Biological neural network |
| Processing Units | 35-60 billion parameters | 86 billion neurons |
| Connections | Trillions of synaptic connections between parameters | 100 trillion synapses between neurons |
| Computing Power | Massive parallel processing on GPUs | Slower firing of biological neurons |
| Training Data | Trained on diverse text/dialogue datasets | Learns from life experiences |
| Reasoning | Narrow AI, limited logic/reasoning | General intelligence, complex reasoning |
| Knowledge | Fixed training data, no common sense | Dynamic real world knowledge |
| Language | Advanced natural language processing | Human-level language understanding |
| Creativity | Limited creative capabilities | Highly creative and imaginative |
| Self-Awareness | No consciousness/self-awareness | Consciousness, self-identity |
In summary, while Claude 2.0 has some impressive language processing capabilities from its training, it still lacks the general intelligence, reasoning, contextual knowledge, self-awareness, and creative abilities that come naturally to humans from our vastly more complex neural architecture evolved over millions of years. Contemporary AI systems are still narrowly focused on specific tasks compared to the generality of biological human cognition. But advances in deep learning are rapidly pushing AI to new frontiers.
SPAUN
“Spaun is the world’s largest functional brain model. It was created using specialized simulation software for modelling neural networks called Nengo”
Spaun (Semantic Pointer Architecture Unified Network) is an artificial neural network model of the brain created by researchers at the University of Waterloo in Canada. It does not contain real biological neurons, but rather uses simulated “neurons” and connections called semantic pointers.
Some key details about Spaun’s architecture:
- It is composed of around 2.5 million simulated spiking neuron models. These are mathematical approximations of biological neurons.
- The neurons are arranged in various interconnected modules to mimic brain function. For example, modules for vision processing, short-term memory, action selection, etc.
- The connections between the simulated neurons are semantic pointers – vectors that represent knowledge using their directionality and overlap with other pointers.
- By adjusting the strengths of these semantic pointer connections, Spaun can learn associations and patterns in data.
- In total, Spaun has around 137 million trainable parameters. These represent the connection weights between semantic pointers that are adjusted during learning.
So in summary, while Spaun’s architecture is inspired by the brain, it does not use real biological neurons. The “neurons” are equations modeled on the spiking behavior of actual neurons. And its knowledge is encoded in the mathematical semantic pointers that connect these simulated neurons rather than biological synapses. The parameters represent the trainable connection weights rather than a count of distinct neurons.
prompt engineered: claude.ai
stay connected with Inner I Network;
