View on GitHub

Penn Action

Action Dataset with human body joints for all frames!

Download this project as a .zip file Download this project as a tar.gz file

Introducing the Penn Action Dataset

Penn Action Dataset (University of Pennsylvania) contains 2326 video sequences of 15 different actions and human joint annotations for each sequence. The dataset is available for download via the following link:


Dataset At a Glance


If you use our dataset, please cite the following paper:

Weiyu Zhang, Menglong Zhu and Konstantinos Derpanis, "From Actemes to Action: 
A Strongly-supervised Representation for Detailed Action Understanding"
International Conference on Computer Vision (ICCV). Dec 2013.

Dataset Content

The dataset is organized in the following format:

/frames  ( all image sequences )
/labels  ( all annotations )
/tools   ( visualization scripts )

The image frames are located in the /frames folder. All frames are in RGB. The resolution of the frames are within the size of 640x480.

The annotations are in the /labels folder. The sequence annotations include class label, coarse viewpoint, human body joints (2D locations and visibility), 2D bounding boxes and training/testing label. Each annotation is a separate MATLAB .mat file under /labels.

An example annotation looks as follows in MATLAB:

annotation = 

      action: 'tennis_serve'
        pose: 'back'
           x: [46x13 double]
           y: [46x13 double]
  visibility: [46x13 logical]
       train: 1
        bbox: [46x4 double]
  dimensions: [272 481 46]
     nframes: 46

List of Actions

baseball_pitch  clean_and_jerk  pull_ups  strumming_guitar  
baseball_swing  golf_swing      push_ups  tennis_forehand   
bench_press     jumping_jacks   sit_ups   tennis_serve
bowling         jump_rope       squats    

Annotation Tools

The annotation tool used in creating this dataset is also available. Please refer to for more details.


Please direct any questions regarding the dataset to

Menglong Zhu