Synthesising 2D Video from 3D Motion Data for Machine Learning Applications

Marion Mundt, Henrike Oberlack, Molly Goldacre, Julia Powles, Johannes Funken, Corey Morris, Wolfgang Potthast, Jacqueline Alderson

Research output: Contribution to journalArticlepeer-review

4 Citations (Scopus)


To increase the utility of legacy, gold-standard, three-dimensional (3D) motion capture datasets for computer vision-based machine learning applications, this study proposed and validated a method to synthesise two-dimensional (2D) video image frames from historic 3D motion data. We applied the video-based human pose estimation model OpenPose to real (in situ) and synthesised 2D videos and compared anatomical landmark keypoint outputs, with trivial observed differences (2.11–3.49 mm). We further demonstrated the utility of the method in a downstream machine learning use-case in which we trained and then tested the validity of an artificial neural network (ANN) to estimate ground reaction forces (GRFs) using synthesised and real 2D videos. Training an ANN to estimate GRFs using eight OpenPose keypoints derived from synthesised 2D videos resulted in accurate waveform GRF estimations (r > 0.9; nRMSE < 14%). When compared with using the smaller number of real videos only, accuracy was improved by adding the synthetic views and enlarging the dataset. The results highlight the utility of the developed approach to enlarge small 2D video datasets, or to create 2D video images to accompany 3D motion capture datasets to make them accessible for machine learning applications.
Original languageEnglish
Article number6522
Issue number17
Publication statusPublished - Sept 2022


Dive into the research topics of 'Synthesising 2D Video from 3D Motion Data for Machine Learning Applications'. Together they form a unique fingerprint.

Cite this