Results 1-20 of 27.
((uid:50031397))
![]() Mejri, Nesryne ![]() ![]() ![]() in IEEE International Conference on Acoustics, Speech and Signal Processing. Proceedings (2023) This paper introduces a novel framework for unsupervised type-agnostic deepfake detection called UNTAG. Existing methods are generally trained in a supervised manner at the classification level, focusing ... [more ▼] This paper introduces a novel framework for unsupervised type-agnostic deepfake detection called UNTAG. Existing methods are generally trained in a supervised manner at the classification level, focusing on detecting at most two types of forgeries; thus, limiting their generalization capability across different deepfake types. To handle that, we reformulate the deepfake detection problem as a one-class classification supported by a self-supervision mechanism. Our intuition is that by estimating the distribution of real data in a discriminative feature space, deepfakes can be detected as outliers regardless of their type. UNTAG involves two sequential steps. First, deep representations are learned based on a self-supervised pretext task focusing on manipulated regions. Second, a one-class classifier fitted on authentic image embeddings is used to detect deepfakes. The results reported on several datasets show the effectiveness of UNTAG and the relevance of the proposed new paradigm. The code is publicly available. [less ▲] Detailed reference viewed: 54 (3 UL)![]() Dupont, Elona ![]() ![]() Scientific Conference (2023) Detailed reference viewed: 38 (0 UL)![]() Singh, Inder Pal ![]() ![]() ![]() in AAAI-22 Workshop Program-Deep Learning on Graphs: Methods and Applications (2022, February) In this paper, we propose the Improved Multi-Label Graph Convolutional Network (IML-GCN) as a precise and efficient framework for multi-label image classification. Although previous approaches have shown ... [more ▼] In this paper, we propose the Improved Multi-Label Graph Convolutional Network (IML-GCN) as a precise and efficient framework for multi-label image classification. Although previous approaches have shown great performance, they usually make use of very large architectures. To handle this, we propose to combine the small version of a newly introduced network called TResNet with an extended version of Multi-label Graph Convolution Networks (ML-GCN); therefore ensuring the learning of label correlation while reducing the size of the overall network. The proposed approach considers a novel image feature embedding instead of using word embeddings. In fact, the latter are learned from words and not images making them inadequate for the task of multi-label image classification. Experimental results show that our framework competes with the state-of-the-art on two multi-label image benchmarks in terms of both precision and memory requirements. [less ▲] Detailed reference viewed: 279 (22 UL)![]() Singh, Inder Pal ![]() ![]() ![]() in IEEE International Conference on Image Processing (2022) In this paper, a novel graph-based approach for multi-label image classification called Multi-Label Adaptive Graph Convolutional Network (ML-AGCN) is introduced. Graph-based methods have shown great ... [more ▼] In this paper, a novel graph-based approach for multi-label image classification called Multi-Label Adaptive Graph Convolutional Network (ML-AGCN) is introduced. Graph-based methods have shown great potential in the field of multi-label classification. However, these approaches heuristically fix the graph topology for modeling label dependencies, which might be not optimal. To handle that, we propose to learn the topology in an end-to-end manner. Specifically, we incorporate an attention-based mechanism for estimating the pairwise importance between graph nodes and a similarity-based mechanism for conserving the feature similarity between different nodes. This offers a more flexible way for adaptively modeling the graph. Experimental results are reported on two well-known datasets, namely, MS-COCO and VG-500. Results show that ML-AGCN outperforms state-of-the-art methods while reducing the number of model parameters. [less ▲] Detailed reference viewed: 81 (5 UL)![]() Garcia Sanchez, Albert ![]() ![]() ![]() in Proceedings of Conference on Computer Vision and Pattern Recognition Workshops (2021, June) Being capable of estimating the pose of uncooperative objects in space has been proposed as a key asset for enabling safe close-proximity operations such as space rendezvous, in-orbit servicing and active ... [more ▼] Being capable of estimating the pose of uncooperative objects in space has been proposed as a key asset for enabling safe close-proximity operations such as space rendezvous, in-orbit servicing and active debris removal. Usual approaches for pose estimation involve classical computer vision-based solutions or the application of Deep Learning (DL) techniques. This work explores a novel DL-based methodology, using Convolutional Neural Networks (CNNs), for estimating the pose of uncooperative spacecrafts. Contrary to other approaches, the proposed CNN directly regresses poses without needing any prior 3D information. Moreover, bounding boxes of the spacecraft in the image are predicted in a simple, yet efficient manner. The performed experiments show how this work competes with the state-of-the-art in uncooperative spacecraft pose estimation, including works which require 3D information as well as works which predict bounding boxes through sophisticated CNNs. [less ▲] Detailed reference viewed: 256 (35 UL)![]() Papadopoulos, Konstantinos ![]() ![]() ![]() in International Conference on Pattern Recognition, Milan 10-15 January 2021 (2021) Detailed reference viewed: 150 (28 UL)![]() ; Mohamed Ali, Mohamed Adel ![]() ![]() in European Conference on Space Debris (2021), 8(1), Detailed reference viewed: 122 (12 UL)![]() Mohamed Ali, Mohamed Adel ![]() ![]() ![]() in 2021 IEEE International Conference on Image Processing (ICIP) (2021) Detailed reference viewed: 84 (16 UL)![]() Ghorbel, Enjie ![]() ![]() ![]() in Journal of medical systems (2020) Detailed reference viewed: 102 (11 UL)![]() Ghorbel, Enjie ![]() ![]() ![]() in IEEE Signal Processing Letters (2020) In this paper, a fast approach for curve reparametrization, called Fast Adaptive Reparamterization (FAR), is introduced. Instead of computing an optimal matching between two curves such as Dynamic Time ... [more ▼] In this paper, a fast approach for curve reparametrization, called Fast Adaptive Reparamterization (FAR), is introduced. Instead of computing an optimal matching between two curves such as Dynamic Time Warping (DTW) and elastic distance-based approaches, our method is applied to each curve independently, leading to linear computational complexity. It is based on a simple replacement of the curve parameter by a variable invariant under specific variations of reparametrization. The choice of this variable is heuristically made according to the application of interest. In addition to being fast, the proposed reparametrization can be applied not only to curves observed in Euclidean spaces but also to feature curves living in Riemannian spaces. To validate our approach, we apply it to the scenario of human action recognition using curves living in the Riemannian product Special Euclidean space SE(3) n. The obtained results on three benchmarks for human action recognition (MSRAction3D, Florence3D, and UTKinect) show that our approach competes with state-of-the-art methods in terms of accuracy and computational cost. [less ▲] Detailed reference viewed: 283 (9 UL)![]() Papadopoulos, Konstantinos ![]() ![]() ![]() in IEEE International Conference on Automatic Face and Gesture Recognition, Buenos Aires 18-22 May 2020 (2020) Detailed reference viewed: 141 (19 UL)![]() Baptista, Renato ![]() ![]() ![]() in IEEE International Conference on Acoustics, Speech and Signal Processing, Brighton, UK, 12–17 May 2019 (2019, May) In this paper, we propose a novel view-invariant action recognition method using a single monocular RGB camera. View-invariance remains a very challenging topic in 2D action recognition due to the lack of ... [more ▼] In this paper, we propose a novel view-invariant action recognition method using a single monocular RGB camera. View-invariance remains a very challenging topic in 2D action recognition due to the lack of 3D information in RGB images. Most successful approaches make use of the concept of knowledge transfer by projecting 3D synthetic data to multiple viewpoints. Instead of relying on knowledge transfer, we propose to augment the RGB data by a third dimension by means of 3D skeleton estimation from 2D images using a CNN-based pose estimator. In order to ensure view-invariance, a pre-processing for alignment is applied followed by data expansion as a way for denoising. Finally, a Long-Short Term Memory (LSTM) architecture is used to model the temporal dependency between skeletons. The proposed network is trained to directly recognize actions from aligned 3D skeletons. The experiments performed on the challenging Northwestern-UCLA dataset show the superiority of our approach as compared to state-of-the-art ones. [less ▲] Detailed reference viewed: 293 (32 UL)![]() Ghorbel, Enjie ![]() ![]() ![]() in 14th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications, Prague, 25-27 February 2018 (2019, February) View-invariant action recognition using a single RGB camera represents a very challenging topic due to the lack of 3D information in RGB images. Lately, the recent advances in deep learning made it ... [more ▼] View-invariant action recognition using a single RGB camera represents a very challenging topic due to the lack of 3D information in RGB images. Lately, the recent advances in deep learning made it possible to extract a 3D skeleton from a single RGB image. Taking advantage of this impressive progress, we propose a simple framework for fast and view-invariant action recognition using a single RGB camera. The proposed pipeline can be seen as the association of two key steps. The first step is the estimation of a 3D skeleton from a single RGB image using a CNN-based pose estimator such as VNect. The second one aims at computing view-invariant skeleton-based features based on the estimated 3D skeletons. Experiments are conducted on two well-known benchmarks, namely, IXMAS and Northwestern-UCLA datasets. The obtained results prove the validity of our concept, which suggests a new way to address the challenge of RGB-based view-invariant action recognition. [less ▲] Detailed reference viewed: 475 (23 UL)![]() Papadopoulos, Konstantinos ![]() ![]() ![]() in 18th International Conference on Computer Analysis of Images and Patterns SALERNO, 3-5 SEPTEMBER, 2019 (2019) In this paper, a novel approach for action detection from RGB sequences is proposed. This concept takes advantage of the recent development of CNNs to estimate 3D human poses from a monocular camera. To ... [more ▼] In this paper, a novel approach for action detection from RGB sequences is proposed. This concept takes advantage of the recent development of CNNs to estimate 3D human poses from a monocular camera. To show the validity of our method, we propose a 3D skeleton-based two-stage action detection approach. For localizing actions in unsegmented sequences, Relative Joint Position (RJP) and Histogram Of Displacements (HOD) are used as inputs to a k-nearest neighbor binary classifier in order to define action segments. Afterwards, to recognize the localized action proposals, a compact Long Short-Term Memory (LSTM) network with a de-noising expansion unit is employed. Compared to previous RGB-based methods, our approach offers robustness to radial motion, view-invariance and low computational complexity. Results on the Online Action Detection dataset show that our method outperforms earlier RGB-based approaches. [less ▲] Detailed reference viewed: 199 (11 UL)![]() Papadopoulos, Konstantinos ![]() ![]() ![]() in Sensors (2019) The Dense Trajectories concept is one of the most successful approaches in action recognition, suitable for scenarios involving a significant amount of motion. However, due to noise and background motion ... [more ▼] The Dense Trajectories concept is one of the most successful approaches in action recognition, suitable for scenarios involving a significant amount of motion. However, due to noise and background motion, many generated trajectories are irrelevant to the actual human activity and can potentially lead to performance degradation. In this paper, we propose Localized Trajectories as an improved version of Dense Trajectories where motion trajectories are clustered around human body joints provided by RGB-D cameras and then encoded by local Bag-of-Words. As a result, the Localized Trajectories concept provides an advanced discriminative representation of actions. Moreover, we generalize Localized Trajectories to 3D by using the depth modality. One of the main advantages of 3D Localized Trajectories is that they describe radial displacements that are perpendicular to the image plane. Extensive experiments and analysis were carried out on five different datasets. [less ▲] Detailed reference viewed: 342 (17 UL)![]() Baptista, Renato ![]() ![]() ![]() in Computer Methods and Programs in Biomedicine (2019) Background and Objective: With the increase in the number of stroke survivors, there is an urgent need for designing appropriate home-based rehabilitation tools to reduce health-care costs. The objective ... [more ▼] Background and Objective: With the increase in the number of stroke survivors, there is an urgent need for designing appropriate home-based rehabilitation tools to reduce health-care costs. The objective is to empower the rehabilitation of post-stroke patients at the comfort of their homes by supporting them while exercising without the physical presence of the therapist. Methods: A novel low-cost home-based training system is introduced. This system is designed as a composition of two linked applications: one for the therapist and another one for the patient. The therapist prescribes personalized exercises remotely, monitors the home-based training and re-adapts the exercises if required. On the other side, the patient loads the prescribed exercises, trains the prescribed exercise while being guided by color-based visual feedback and gets updates about the exercise performance. To achieve that, our system provides three main functionalities, namely: 1) Feedback proposals guiding a personalized exercise session, 2) Posture monitoring optimizing the effectiveness of the session, 3) Assessment of the quality of the motion. Results: The proposed system is evaluated on 10 healthy participants without any previous contact with the system. To analyze the impact of the feedback proposals, we carried out two different experimental sessions: without and with feedback proposals. The obtained results give preliminary assessments about the interest of using such feedback. Conclusions: Obtained results on 10 healthy participants are promising. This encourages to test the system in a realistic clinical context for the rehabilitation of stroke survivors. [less ▲] Detailed reference viewed: 179 (16 UL)![]() Baptista, Renato ![]() ![]() ![]() in 2018 Zooming Innovation in Consumer Electronics International Conference (ZINC), 30-31 May 2018 (2018, May 31) This paper presents an intuitive feedback tool able to implicitly guide motion with respect to a reference movement. Such a tool is important in multiple applications requiring assisting physical ... [more ▼] This paper presents an intuitive feedback tool able to implicitly guide motion with respect to a reference movement. Such a tool is important in multiple applications requiring assisting physical activities as in sports or rehabilitation. Our proposed approach is based on detecting key skeleton frames from a reference sequence of skeletons. The feedback is based on the 3D geometry analysis of the skeletons by taking into account the key-skeletons. Finally, the feedback is illustrated by a color-coded tool, which reflects the motion accuracy. [less ▲] Detailed reference viewed: 190 (7 UL)![]() Ghorbel, Enjie ![]() in Image and Vision Computing (2018), 77 Detailed reference viewed: 135 (5 UL)![]() Ghorbel, Enjie ![]() in Computer Vision and Image Understanding (2018), 175 Detailed reference viewed: 102 (4 UL)![]() ; ; Ghorbel, Enjie ![]() in IEEE Transactions on Cognitive and Developmental Systems (2018), 10(4), 894--902 Detailed reference viewed: 98 (3 UL) |
||