![]() Garcia Sanchez, Albert ![]() ![]() ![]() in Proceedings of Conference on Computer Vision and Pattern Recognition Workshops (2021, June) Being capable of estimating the pose of uncooperative objects in space has been proposed as a key asset for enabling safe close-proximity operations such as space rendezvous, in-orbit servicing and active ... [more ▼] Being capable of estimating the pose of uncooperative objects in space has been proposed as a key asset for enabling safe close-proximity operations such as space rendezvous, in-orbit servicing and active debris removal. Usual approaches for pose estimation involve classical computer vision-based solutions or the application of Deep Learning (DL) techniques. This work explores a novel DL-based methodology, using Convolutional Neural Networks (CNNs), for estimating the pose of uncooperative spacecrafts. Contrary to other approaches, the proposed CNN directly regresses poses without needing any prior 3D information. Moreover, bounding boxes of the spacecraft in the image are predicted in a simple, yet efficient manner. The performed experiments show how this work competes with the state-of-the-art in uncooperative spacecraft pose estimation, including works which require 3D information as well as works which predict bounding boxes through sophisticated CNNs. [less ▲] Detailed reference viewed: 263 (35 UL)![]() ; Mohamed Ali, Mohamed Adel ![]() ![]() in European Conference on Space Debris (2021), 8(1), Detailed reference viewed: 127 (14 UL)![]() Mohamed Ali, Mohamed Adel ![]() ![]() ![]() in 2021 IEEE International Conference on Image Processing (ICIP) (2021) Detailed reference viewed: 90 (16 UL)![]() Papadopoulos, Konstantinos ![]() ![]() ![]() in International Conference on Pattern Recognition, Milan 10-15 January 2021 (2021) Detailed reference viewed: 155 (28 UL)![]() Oyedotun, Oyebade ![]() ![]() ![]() Poster (2021) Detailed reference viewed: 109 (12 UL)![]() Shabayek, Abd El Rahman ![]() ![]() in IEEE Access (2021), 9 Detailed reference viewed: 99 (4 UL)![]() Oyedotun, Oyebade ![]() ![]() Poster (2020, November 18) Detailed reference viewed: 134 (12 UL)![]() Ghorbel, Enjie ![]() ![]() ![]() in Journal of medical systems (2020) Detailed reference viewed: 110 (13 UL)![]() Shabayek, Abd El Rahman ![]() ![]() ![]() in 27th IEEE International Conference on Image Processing (ICIP 2020), Abu Dhabi 25-28 October 2020 (2020, October) Detailed reference viewed: 93 (2 UL)![]() Cherenkova, Kseniya ![]() ![]() Scientific Conference (2020, October) We propose a Point-Voxel DeConvolution (PVDeConv) mod- ule for 3D data autoencoder. To demonstrate its efficiency we learn to synthesize high-resolution point clouds of 10k points that densely describe ... [more ▼] We propose a Point-Voxel DeConvolution (PVDeConv) mod- ule for 3D data autoencoder. To demonstrate its efficiency we learn to synthesize high-resolution point clouds of 10k points that densely describe the underlying geometry of Computer Aided Design (CAD) models. Scanning artifacts, such as pro- trusions, missing parts, smoothed edges and holes, inevitably appear in real 3D scans of fabricated CAD objects. Learning the original CAD model construction from a 3D scan requires a ground truth to be available together with the corresponding 3D scan of an object. To solve the gap, we introduce a new dedicated dataset, the CC3D, containing 50k+ pairs of CAD models and their corresponding 3D meshes. This dataset is used to learn a convolutional autoencoder for point clouds sampled from the pairs of 3D scans - CAD models. The chal- lenges of this new dataset are demonstrated in comparison with other generative point cloud sampling models trained on ShapeNet. The CC3D autoencoder is efficient with respect to memory consumption and training time as compared to state- of-the-art models for 3D data generation. [less ▲] Detailed reference viewed: 194 (9 UL)![]() Oyedotun, Oyebade ![]() ![]() ![]() in IEEE Access (2020) Detailed reference viewed: 151 (16 UL)![]() Oyedotun, Oyebade ![]() ![]() ![]() in Applied Intelligence (2020) Detailed reference viewed: 133 (18 UL)![]() Saint, Alexandre Fabian A ![]() ![]() ![]() Scientific Conference (2020, August 23) We propose 3DBooSTeR, a novel method to recover a textured 3D body mesh from a textured partial 3D scan. With the advent of virtual and augmented reality, there is a demand for creating realistic and high ... [more ▼] We propose 3DBooSTeR, a novel method to recover a textured 3D body mesh from a textured partial 3D scan. With the advent of virtual and augmented reality, there is a demand for creating realistic and high-fidelity digital 3D human representations. However, 3D scanning systems can only capture the 3D human body shape up to some level of defects due to its complexity, including occlusion between bodyparts, varying levels of details, shape deformations and the articulated skeleton. Textured 3D mesh completion is thus important to enhance3D acquisitions. The proposed approach decouples the shape and texture completion into two sequential tasks. The shape is recovered by an encoder-decoder network deforming a template body mesh. The texture is subsequently obtained by projecting the partial texture onto the template mesh before inpainting the corresponding texture map with a novel approach. The approach is validated on the 3DBodyTex.v2 dataset [less ▲] Detailed reference viewed: 198 (8 UL)![]() ![]() Saint, Alexandre Fabian A ![]() ![]() ![]() Scientific Conference (2020, August 23) The SHApe Recovery from Partial textured 3D scans challenge, SHARP 2020, is the first edition of a challenge fostering and benchmarking methods for recovering complete textured 3D scans from raw ... [more ▼] The SHApe Recovery from Partial textured 3D scans challenge, SHARP 2020, is the first edition of a challenge fostering and benchmarking methods for recovering complete textured 3D scans from raw incomplete data. SHARP 2020 is organized as a workshop in conjunction with ECCV 2020. There are two complementary challenges, the first one on 3D human scans, and the second one on generic objects. Challenge 1 is further split into two tracks, focusing, first, on large body and clothing regions, and, second, on fine body details. A novel evaluation metric is proposed to quantify jointly the shape reconstruction, the texture reconstruction, and the amount of completed data. Additionally, two unique datasets of 3D scans are proposed, to provide raw ground-truth data for the benchmarks. The datasets are released to the scientific community. Moreover, an accompanying custom library of software routines is also released to the scientific community. It allows for processing 3D scans, generating partial data and performing the evaluation. Results of the competition, analyzed in comparison to baselines, show the validity of the proposed evaluation metrics and highlight the challenging aspects of the task and of the datasets. Details on the SHARP 2020 challenge can be found at https://cvi2.uni.lu/sharp2020/ [less ▲] Detailed reference viewed: 174 (8 UL)![]() Oyedotun, Oyebade ![]() ![]() ![]() in IEEE International Conference on Image Processing (ICIP 2020), Abu Dhabi, UAE, Oct 25–28, 2020 (2020, May 30) Detailed reference viewed: 157 (9 UL)![]() Shabayek, Abd El Rahman ![]() ![]() ![]() in 45th International Conference on Acoustics, Speech, and Signal Processing (ICASSP 2020), Barcelona 4-8 May 2020 (2020, May) Detailed reference viewed: 93 (0 UL)![]() Oyedotun, Oyebade ![]() ![]() ![]() in IEEE 2020 Winter Conference on Applications of Computer Vision (WACV 20), Aspen, Colorado, US, March 2–5, 2020 (2020, March 01) Detailed reference viewed: 156 (16 UL)![]() Shabayek, Abd El Rahman ![]() ![]() ![]() in Proceedings of the 15th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications (VISIGRAPP 2020), GRAPP (2020, February) Detailed reference viewed: 154 (13 UL)![]() Papadopoulos, Konstantinos ![]() ![]() ![]() in IEEE International Conference on Automatic Face and Gesture Recognition, Buenos Aires 18-22 May 2020 (2020) Detailed reference viewed: 145 (19 UL)![]() Baptista, Renato ![]() ![]() ![]() in International Conference on Pattern Recognition (ICPR) Workshop on 3D Human Understanding, Milan 10-15 January 2021 (2020) In this paper, we propose 3DBodyTex.Pose, a dataset that addresses the task of 3D human pose estimation in-the-wild. Generalization to in-the-wild images remains limited due to the lack of adequate ... [more ▼] In this paper, we propose 3DBodyTex.Pose, a dataset that addresses the task of 3D human pose estimation in-the-wild. Generalization to in-the-wild images remains limited due to the lack of adequate datasets. Existent ones are usually collected in indoor controlled environments where motion capture systems are used to obtain the 3D ground-truth annotations of humans. 3DBodyTex.Pose offers high quality and rich data containing 405 different real subjects in various clothing and poses, and 81k image samples with ground-truth 2D and 3D pose annotations. These images are generated from 200 viewpoints among which 70 challenging extreme viewpoints. This data was created starting from high resolution textured 3D body scans and by incorporating various realistic backgrounds. Retraining a state-of-the-art 3D pose estimation approach using data augmented with 3DBodyTex.Pose showed promising improvement in the overall performance, and a sensible decrease in the per joint position error when testing on challenging viewpoints. The 3DBodyTex.Pose is expected to offer the research community with new possibilities for generalizing 3D pose estimation from monocular in-the-wild images. [less ▲] Detailed reference viewed: 179 (16 UL) |
||