Video content analysis and understanding are active research topics in modern visual computing and communication. In this context, a particular challenging problem that attracts much attention is human action recognition. In this chapter, we propose a new methodology to solve the problem using geometric statistical information. Two new approaches, Differential Luminance Field Trajectory (DLFT) and Luminance Aligned Projection Distance (LAPD), are proposed. Instead of extracting the object or using interest points as a representation, we treat each video clip as a trajectory in a very high dimensionality space and extract the useful statistical geometric information for action recognition. For DLFT, we take advantage of the differential signals which preserve both the temporal and spatial information, and then classify the action by supervised learning. For the LAPD approach, we generate a trajectory for each video clip and compute a distance metric to describe the similarity for classification. Decision is made by applying a K-Nearest Neighbor classifier. Since DLFT is more sensitive in the temporal domain while the LAPD approach can handle more variance in appearance luminance field, a potential fusion of the two methods would yield more desirable properties. Experimental results demonstrate that the methods work effectively and efficiently. The performance is comparable or better and more robust than conventional methods.