Human action recognition has become one of the hot areas today. The skeleton-based human action recognition method has been widely noticed for its ability to show human action clearly. In response to the problem that the manually designed human topology graph cannot obtain global information and has a lot of redundant information when extracting features. A Dual-mode Attention Spatio-Temporal Graph Convolutional Network is proposed to achieve the full utilization of node information that plays a key role in action recognition. First, the SGSAE module is proposed to model the relationship between all nodes using a self-attention mechanism to achieve the extraction of global features of node information, and to optimize the topology of the graph during the training process of the network to finally obtain a graph topology that adapts to various data samples. Second, the global and local features of the nodes are fused by different weights. Finally, the channel attention mechanism is introduced into the network, and the MCA module is proposed to fuse channel features to reduce a large amount of redundant information and improve action recognition accuracy. The experimental results show that the proposed Dual-mode Attention Spatio-Temporal Graph Convolutional Network achieves better performance on NTU-RGB+D and Kinetics datasets.