Gait datasets are often limited by a lack of diversity in terms of the participants, appearance, viewing angle, environments, annotations, and availability. We present a primary gait dataset comprising 1,560 annotated casual walks from 64 participants, in both indoor and outdoor real-world environments. We used two digital cameras and a wearable digital goniometer to capture visual as well as motion signal gait-data respectively. Traditional methods of gait identification are often affected by the viewing angle and appearance of the participant therefore, this dataset mainly considers the diversity in various aspects (e.g., participants’ attributes, background variations, and view angles). The dataset is captured from 8 viewing angles in 45° increments along-with alternative appearances for each participant, for example, via a change of clothing. The dataset provides 3,120 videos, containing approximately 748,800 image frames with detailed annotations including approximately 56,160,000 bodily keypoint annotations, identifying 75 keypoints per video frame, and approximately 1,026,480 motion data points captured from a digital goniometer for three limb segments (thigh, upper arm, and head).