Recently, there is a growing interest in utilizing wireless signals for human gesture recognition and activity recognition. At the same time, the scarcity and lack of diversity of radar echo signature datasets of human gestures and activities is well recognized. This work demonstrates a framework for synthetically generating a vast and diverse set of radar echo signatures starting from a small set of optical motion capture (MoCap) trajectories. The captured trajectories are perturbed using a pool of composable spatial and temporal transformation functions assembled by a data augmentation pipeline builder. The transformed trajectories, combined with a simple radar cross-section (RCS) modeling process, are used to simulate radar CIR signals. Features extracted from this synthetic dataset show a strong correlation with the features obtained from simultaneously collected real radar data. Furthermore, we demonstrate that the synthetically generated radar echo signals can improve the performance of ML-based wireless gesture and activity recognition systems especially where the availability of real data is limited.
|