Menu Close

The thoughts about designing a brain-like model

This is a great discussion point. If I am right, we aim to build a model that can have similar computational mechanisms underlying various cognitions and behaviors. These mechanisms are formed through anatomical structures in the brain and altered through learning (e.g., visual experience). When speaking a model (e.g., ANNs), it starts from its structure and ends by learning (e.g., training on data). We have input, internal representation, and output parts to consider for building a model. We also need to consider the spatial and temporal dimensions of information flow formed in the internal model. I believe that this question may be too big to talk about here. I want to highlight the following points I have been considering:

  1. Some low-level features are still missing in the early layers of current artificial neuralnetworks. For example, some study has suggested eye position information in the V1 brain area, which probably has not been seen in current ANNs.
  2. It seems that the “neurons” in current ANNs do not have one unified principle to follow, such as the free-energy principle says that any self-organizing system that is at equilibrium with its environment must minimize its free energy (by Karl Friston). Here I like to know why the brain has inhibitory and excitatory neurons.
  3. It is related to both visual perception and decision-making when we speak image classification. It may be tightly associated with the prefrontal cortex (PFC) from the brain’s perspective. My current emotion perception study demonstrates that the emotion valence rating is associated with PFC. We may consider PFC into the brain-like model for visual perception. In addition, we still have many questions about how the concepts (e.g., “knife,” “dog,” or “party”) are stored in the brain. A recent study from Leonardo Fernandino indicates that conceptual knowledge is stored as patterns of neural activity that encode sensory-motor and affective information about each concept. Here we may argue that PFC is not necessary to allow animals to see. I believe that it is likely true. However, is the “see” the same as the “see” in the human brain? Probably not. Another recent study (by Schafroth, J.L., Basile, B.M., Martin, A. et al. ) indicates that monkeys just know that a triangle is just a triangle, but humans can spontaneously ascribe mental states to animated shapes.

It would be great to discuss more ideas to build a model that can match the brain better.

Casual thoughts about deep neural network design

Ai, Artificial Intelligence, Intelligence, Network

Inspired by A ConvNet for the 2020s by Zhuang Liu, Hanzi Mao, Chao-Yuan Wu, Christoph Feichtenhofer, Trevor Darrell, Saining Xie from Facebook AI Research (FAIR) and UC Berkeley.

Convolution and transformer are two approaches to design a deep neural network. Recently, the transformer seems to be becoming dominant in developing AI techniques. I have been asking myself: is that true? What makes the transformer special? I do not know the answer and hope to figure it out sooner or later. What I like to deliver in this blog is all about my experience with developing deep neural networks during my Ph.D. study.

What’s the secret to designing a state-of-the-art artificial deep neural network?

  • Learning Structure – You need to tell the network how to extract features from the input layer by layer. For example,

Hierarchical representation by starting from small-sized patches and gradually increasing the size through merging to achieve scale-invariance

By Sieun Park
  • Block Design – Play with the internal representation. For example,

Achieves efficient, linear computational complexity by computing self-attention locally. (shifted window approach)

By Sieun Park
  • Size of Convolution Kernel. For example,

The researchers observed that the benefit of larger convolution kernels and the saturation point is reached at 7 × 7

from A ConvNet for the 2020s
  • Pay Attention to Temporal Learning – Play with the representation embedding timely. For example,

The position of the spatial depth-wise Conv layer is moved up.

A ConvNet for the 2020s
  • Training with Fine Hyper-parameters: Optimizer, Learning Rate, Batch-size, Activation functions, and so on. No secrets, but try.

Final thoughts

Thinking about the followings:

  1. Pay attention to each of them: Input->Representation->Output
  2. Spatial and Temporal Learning

Keep them in mind whenever designing a deep neural network for any task.

Understand your data statistically before developing your model-Chapter I

Inspired by “Understanding 8 types of Cross-Validation” by Satyam Kumar

Cross-Validation (CV) is one critical way for evaluating our machine learning models. However, it should be applied correctly to your own data. You need to check your data and understand it statistically before developing your model on your data.

There are more than eight types of Cross-Validation variants you may use to develop your model. However, which one you should use largely depends on your data. We should check the data at least by looking at (1) sample size: small or large? (2) class balance or not? (3) whether it is time-series data?

The following 8 types of CV are explained in the Understanding 8 types of Cross-Validation” by Satyam Kumar.

  1. Leave p out cross-validation
  2. Leave one out cross-validation
  3. Holdout cross-validation
  4. Repeated random subsampling validation
  5. k-fold cross-validation
  6. Stratified k-fold cross-validation
  7. Time Series cross-validation
  8. Nested cross-validation

You may find the Pros vs. Cons for each one in the article. Here, I just put my key idea:

  1. Make sure you have balanced data and not time-series one. The fast and safe way is to up-sample or down-sample your data. After balancing the data, we can easily apply Nest Cross-Validation (why? check my previous blog)
  2. If it is time-series data, there are few CV options for you, and you need to use time Series cross-validation.

I will keep this topic updated.