Abstract
Multi-animal pose estimation is a critical technique for enabling fine-grained quantification of group animal behaviors, which holds significant scientific value for uncovering behavioral changes under space environmental factors such as microgravity and radiation. Currently, the China Space Station has conducted a series of space biology experiments involving typical model organisms, including Caenorhabditis elegans (C. elegans), zebrafish, and Drosophila. However, substantial differences in species types, body scales, and posture dynamics among these animals pose serious challenges to the generalization and robustness of traditional pose estimation methods. To address this, we propose a novel, flexible, and general single-stage multi-animal pose estimation method. The method constructs species-specific pose group representations based on anatomical priors, incorporates a multi-scale feature-sampling module to integrate shallow and deep visual cues, and employs a structure-guided learning mechanism to enhance keypoint localization robustness under occlusion and overlap. We evaluate our method on the SpaceAnimal dataset-the first public benchmark for pose estimation and tracking of model organisms in space-containing multi-species samples from both spaceflight and ground-based experiments. Our method achieves AP scores of 72.8%, 62.1%, and 67.1% on C. elegans, zebrafish, and Drosophila, respectively, surpassing the state-of-the-art performance. These findings demonstrate the effectiveness and robustness of the proposed method across species and imaging conditions, offering strong technical support for on-orbit behavior modeling and large-scale quantitative analysis.