AI Tech Revives Virtual Pets

Abstract

We tackle animatable 3D dog reconstruction from a single image, noting the overlooked potential of animals. Particularly, we focus on dogs, emphasizing their intrinsic characteristics that complicate 3D observation. First, the considerable variation in shapes across breeds presents a complexity for modeling. Additionally, the nature of quadrupeds leads to frequent joint occlusions compared to humans. These challenges make 3D reconstruction from 2D observations difficult, and it becomes dramatically harder when constrained to a single image. To address these challenges, our insight is to combine the acquisition of appearance from generative models, without additional data, with geometric guidance provided by a parametric representation, aiming to achieve complete geometry. To this end, we present DogRecon, our framework consists of two key components: Canine-centric novel view synthesis with canine prior for multi-view generation of dog and a reliable sampling weight strategy with Gaussian Splatting for animatable 3D dog reconstruction. Extensive experiments on the GART, DFA, and internet-sourced datasets confirm our framework has state-of-the-art performance in image-to-3D generation and comparable performance in animatable 3D reconstruction. Additionally, we demonstrate novel pose animation and text-to-3D dog reconstruction as applications. Project page: https://vision3d-lab.github.io/dogrecon/

Researchers at UNIST have developed an innovative AI technology capable of reconstructing highly detailed three-dimensional (3D) models of companion animals from a single photograph, enabling realistic animations. This breakthrough allows users to experience lifelike digital avatars of their companion animals in virtual reality (VR), augmented reality (AR), and metaverse environments.

Led by Professor Kyungdon Joo at the Artificial Intelligence Graduate School of UNIST, the research team announced the development of DogRecon, a novel AI framework that can reconstruct an animatable 3D dog Gaussian from a single dog image.

Due to their diverse breeds, varying body shapes, and frequent occlusion of joints caused by their quadrupedal stance, reconstructing 3D models of dogs presents unique challenges. Moreover, creating accurate 3D structures from a single 2D photo is inherently difficult, often resulting in distorted or unrealistic representations.

DogRecon overcomes these challenges by utilizing breed-specific statistical models to capture variations in body shape and posture. It also employs advanced generative AI to produce multiple viewpoints, effectively reconstructing occluded areas with high fidelity. Additionally, the application of Gaussian Splatting techniques enables the model to accurately reproduce the curvilinear body contours and fur textures characteristic of dogs.

Performance evaluations using various datasets demonstrated that DogRecon can generate natural, precise 3D dog avatars comparable to those produced by existing video-based methods, but from only a single image. Unlike prior models, which often rendered dogs with unnatural postures-such as stretched bodies with bent joints, or bundled ears, tails, and fur-especially when dogs are in relaxed or crouched positions, DogRecon delivers more realistic results.

Teaser of DogRecon. Figure 1. Teaser of DogRecon. Left: The proposed DogRecon takes a single image as input and reconstructs 3D Gaussian, including texture and shape. Right: With the obtained 3D dogs, we can retarget motion to an existing video or even create new animations within pre-trained scenes by editing. In addition, DogRecon seamlessly applies to text-to-animatable 3D dogs.

Furthermore, due to its scalable architecture, DogRecon holds significant promise for applications in text-driven animation generation, as well as AR/VR environments.

This research was led by first author Gyeongsu Cho, with contributions from Changwoo Kang (UNIST) and Donghyeon Soon (DGIST). Gyeongsu Cho remarked, "With over a quarter of households owning pets, expanding 3D reconstruction technology-traditionally focused on humans-to include companion animals has been a goal," adding, "DogRecon offers a tool that enables anyone to create and animate a digital version of their companion animals."

Professor Joo added, "This study represents a meaningful step forward by integrating generative AI with 3D reconstruction techniques to produce realistic models of companion animals." He further added, "We look forward to expanding this approach to include other animals and personalized avatars in the future."

The findings of this research were published in the International Journal of Computer Vision (IJCV), one of the most prestigious journals in the field of computer vision on June 2, 2025. It was carried out with the support from the Ministry of Science and ICT (MSIT) and the Institute for Information & Communications Technology Planning & Evaluation (IITP), through initiatives aimed at developing AI technologies capable of understanding and reasoning about everyday life based on common sense, as well as by the UNIST Artificial Intelligence Graduate School.

Journal Reference

Gyeongsu Cho, Changwoo Kang, Donghyeon Soon, and Kyungdon Joo, "DogRecon: Canine Prior-Guided Animatable 3D Gaussian Dog Reconstruction From A Single Image," IJCV, (2025).

/Public Release. This material from the originating organization/author(s) might be of the point-in-time nature, and edited for clarity, style and length. Mirage.News does not take institutional positions or sides, and all views, positions, and conclusions expressed herein are solely those of the author(s).View in full here.