What we will do next?
Some thoughts about future research questions and our final goals towards the unknown.
What works are scientifically meaningful for computational neuroscience? Specially, what can we do for network neuroscience?
Why are we doing spiking neural networks or neuromorphic computing? What will it finally lead to?
What can we do to integrate artificial intelligence and neuroscience? What are the critical problems in neuroscience that can be solved by artificial intelligence? What are the phenomenon and mechanism in artificial neuroscience that requires the integration of neuroscience?
What makes human beings in the current stages? In the previous research path, our paradigm is to investigate a specific task through the reductionism. However, the LLM’s motivation may be the opposite. It is not because of the complex infrastructure of the brain so that it can execute complex tasks, but the complex tasks that shapes the high intelligence.
Some big thoughts?
- The AI develops from decision-making to generative AI. Reasoning from the first-principle, what can be the next step?
- The physics and AI development has several key time points. Regarding the physics, the emergence of Newton’s three principles for mechanics, the quantum mechanics and relative theory permanently change the landscape of the physics. Regarding the AI, the prosperity of DNN and associated hardware boost its development waves by waves. What can be the changing point for neuroscience?
- The next-word prediction serves as a good pre-train task for the LLMs. What types of properties does this task capture so that it is valid and efficient ? Regarding the general artificial intelligence, what kinds of pre-train tasks are appropriate? For the vision-LM or multimodal-LM, what kinds of pre-train tasks are appropriate?
- Transformers are good for predicting the next word. Yet is it the proper design for intelligence? Relatively, what type of data structure matches the network infrastructure?
Suppose we have a model that can predict brain’s response to the given stimulus. What kind of questions about encoding and decoding can we ask? What can serve as a fundamental model for neural encoding?
If we take the SNN as a special type of quantization, can we derive its advantages on the representation and parallelism? For the representation, we mean that compared to the quantization of the low-bit setup, what benefit it brings when we divide the 1-time expression into multiple steps? For the parallelism, we mean that what kind of computational benefit it brings with the fixed size spikes?
The brain uses simple and imprecise mechanism to support the learning. Moreover, the underline infrastructure appears to be dominated by the genes rather than the data, which suggests that they might be a generic initialization for deep neural networks. Or, the pre-training part of a large ANN might be realized through simple learning rules.
Quick Thoughts:
- Combine Transformer pretraining and quantization approaches in a spike manner. Use the spiking mechanism to control masks.
- Regarding the behavioral experiments, we can provide detailed description of the subject and then has the character to perform the task. Can we have an example here?
- What are the common space for vision-language? It should be not only defined by similarity in the representation (CLIP), but also by certain tasks. The CLIP maps an image into a space aligned with the caption, which is typically a subspace of the information that can be represented by the image. So our question can be: how to generate questions/dialogue for the given images? What are the questions that can expands the image space?
What are good research works in AI?
- It established new learning paradigm, like RL, generative AI, CLIP, etc…
- It proposes new fundamental learning infrastructures, like CNN, ResNet, Transformer, stableDiff, etc…
The bitter lesson: what should we believe to generate intelligence.
- Our brain is a trajectory full of randomness in its current stage. Not all mechanisms in the brain are useful. Many mechanisms appear only because of the compromise to the underline infrastructure and only works in that way.
- Considering the evolution theory, maybe it is random for the brain to appear in its current state, but it could be of high probability to show up under some scalable rules. Designing scalable infrastructure and identifying scalable principles may be more effective in building more intelligent agents.
- The human’s utilization of information and energy is efficient. However, we should interpret it as a distinct scalable solution rather than a perfect model to follow. It makes no sense in the computation to say something is brain-inspired without digging into the underline computational rules.
- What’s the difference between the scalability of current AIs and those in the previous decades? How should be design research plan? Or more broadly, how should we design our intelligent agent? Is there any scalable plan that works for all tasks?
- The evolution leads to diversities in the function. For example, there are different types of eyes. These eyes are represented by genes. How can we build the similar “design” or “encode” system? Or even more, can we talk whether such a system is efficient? If it is, how can we design one with similar efficiency?
- We can either build AI through connection and symbols. BUT, can we do both?
How to realize the scaling law for brain mapping?
- Scan time, resolution, task?
- What does the brain mapping tell us from the view of machine learning?
- What will be the more effective paradigm for neural intervention?
An interesting idea of abstraction for AI: we can have AI to draw a human body or other figures, then have some professional painters to judge the body-science and highlight-shade.