23 results on '"human silhouette"'
Search Results
2. SMDF: Spatial Mass Distribution Features and Deep Learning-Based Technique for Human Activity Recognition
- Author
-
Raychaudhuri, Amlan, Maity, Satyabrata, Chakrabarti, Amlan, and Bhattacharjee, Debotosh
- Published
- 2024
- Full Text
- View/download PDF
3. Human Silhouette and Skeleton Video Synthesis Through Wi-Fi Signals.
- Author
-
Avola, Danilo, Cascio, Marco, Cinque, Luigi, Fagioli, Alessio, and Foresti, Gian Luca
- Subjects
- *
WIRELESS Internet , *ELECTROMAGNETIC waves , *RADIO frequency , *VISIBLE spectra , *FEATURE extraction , *HUMAN skeleton - Abstract
The increasing availability of wireless access points (APs) is leading toward human sensing applications based on Wi-Fi signals as support or alternative tools to the widespread visual sensors, where the signals enable to address well-known vision-related problems such as illumination changes or occlusions. Indeed, using image synthesis techniques to translate radio frequencies to the visible spectrum can become essential to obtain otherwise unavailable visual data. This domain-to-domain translation is feasible because both objects and people affect electromagnetic waves, causing radio and optical frequencies variations. In the literature, models capable of inferring radio-to-visual features mappings have gained momentum in the last few years since frequency changes can be observed in the radio domain through the channel state information (CSI) of Wi-Fi APs, enabling signal-based feature extraction, e.g. amplitude. On this account, this paper presents a novel two-branch generative neural network that effectively maps radio data into visual features, following a teacher–student design that exploits a cross-modality supervision strategy. The latter conditions signal-based features in the visual domain to completely replace visual data. Once trained, the proposed method synthesizes human silhouette and skeleton videos using exclusively Wi-Fi signals. The approach is evaluated on publicly available data, where it obtains remarkable results for both silhouette and skeleton videos generation, demonstrating the effectiveness of the proposed cross-modality supervision strategy. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
4. A Novel Approach for Human Silhouette Extraction from Video Data
- Author
-
Raychaudhuri, Amlan, Maity, Satyabrata, Chakrabarti, Amlan, Bhattacharjee, Debotosh, Kacprzyk, Janusz, Series editor, Pal, Nikhil R., Advisory editor, Bello Perez, Rafael, Advisory editor, Corchado, Emilio S., Advisory editor, Hagras, Hani, Advisory editor, Kóczy, László T., Advisory editor, Kreinovich, Vladik, Advisory editor, Lin, Chin-Teng, Advisory editor, Lu, Jie, Advisory editor, Melin, Patricia, Advisory editor, Nedjah, Nadia, Advisory editor, Nguyen, Ngoc Thanh, Advisory editor, Wang, Jun, Advisory editor, Chaki, Rituparna, editor, Saeed, Khalid, editor, Cortesi, Agostino, editor, and Chaki, Nabendu, editor
- Published
- 2017
- Full Text
- View/download PDF
5. Human fall detection using slow feature analysis.
- Author
-
Fan, Kaibo, Wang, Ping, and Zhuang, Shuo
- Subjects
ACCIDENTAL falls ,ACCIDENTS ,HUMAN-computer interaction ,VIDEO surveillance ,MEDICAL care - Abstract
Falls are reported to be the leading causes of accidental deaths among elderly people. Automatic detection of falls from video sequences is an assistant technology for low-cost health care systems. In this paper, we present a novel slow feature analysis based framework for fall detection in a house care environment. Firstly, a foreground human body is extracted by a background subtraction technique. After morphological operations, the human silhouette is refined and covered by a fitted ellipse. Secondly, six shape features are quantified from the covered silhouette to represent different human postures. With the help of the learned slow feature functions, the shape feature sequences are transformed into slow feature sequences with discriminative information about human actions. To represent the fall incidents, the squared first order temporal derivatives of the slow features are accumulated into a classification vector. Lastly, falls are distinguished from other daily actions, such as walking, crouching, and sitting, by the trained directed acyclic graph support vector machine. Experiments on the multiple-camera fall dataset and the SDUFall dataset demonstrate that our method is comparable to other state-of-the-art methods, achieving 94.00% recognition rate on the former dataset and 96.57% on the latter one. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
6. Background Subtraction and Movement Detection Algorithms in Dynamic Environments – An Evaluation for Automated Threat Recognition Systems
- Author
-
Flizikowski, Adam, Maszewski, Mirosław, Wachowiak, Marcin, Taberski, Grzegorz, and S. Choras, Ryszard, editor
- Published
- 2014
- Full Text
- View/download PDF
7. Human Action Recognition Based on Tracking Features
- Author
-
Hernández, Javier, Montemayor, Antonio S., José Pantrigo, Juan, Sánchez, Ángel, Hutchison, David, Series editor, Kanade, Takeo, Series editor, Kittler, Josef, Series editor, Kleinberg, Jon M., Series editor, Mattern, Friedemann, Series editor, Mitchell, John C., Series editor, Naor, Moni, Series editor, Nierstrasz, Oscar, Series editor, Pandu Rangan, C., Series editor, Steffen, Bernhard, Series editor, Sudan, Madhu, Series editor, Terzopoulos, Demetri, Series editor, Tygar, Doug, Series editor, Vardi, Moshe Y., Series editor, Weikum, Gerhard, Series editor, Ferrández, José Manuel, editor, Álvarez Sánchez, José Ramón, editor, de la Paz, Félix, editor, and Toledo, F. Javier, editor
- Published
- 2011
- Full Text
- View/download PDF
8. Data Integration in Multimodal Home Care Surveillance and Communication System
- Author
-
Augustyniak, Piotr, Smoleń, Magdalena, Broniec, Anna, Chodak, Jacek, Kacprzyk, Janusz, editor, Piȩtka, Ewa, editor, and Kawa, Jacek, editor
- Published
- 2010
- Full Text
- View/download PDF
9. Fusion of Color/Infrared Video for Human Detection
- Author
-
Bhanu, Bir, Han, Ju, Bhanu, Bir, and Han, Ju
- Published
- 2010
- Full Text
- View/download PDF
10. Learning spatiotemporal representations for human fall detection in surveillance video.
- Author
-
Kong, Yongqiang, Huang, Jianhui, Huang, Shanshan, Wei, Zhengang, and Wang, Shengke
- Subjects
- *
CONVOLUTIONAL neural networks , *VIDEO surveillance , *AUTUMN - Abstract
Highlights • An effective background subtraction technique is proposed. • A novel view-independent CNNs classifier which is applied. • High-quality network inputs have low computational cost. • A simple voting classifier works fairly well in multi-camera system. Abstract In this paper, a computer vision based framework is proposed that detects falls from surveillance videos. Firstly, we employ background subtraction and rank pooling to model spatial and temporal representations in videos, respectively. We then introduce a novel three-stream Convolutional Neural Networks as an event classifier. Silhouettes and their motion history images serve as input to the first two streams, while dynamic images whose temporal duration is equal to motion history images, are used as input to the third stream. Finally, we apply voting on the results of event classification to perform multi-camera fall detection. The main novelty of our method against the conventional ones is that high-quality spatiotemporal representations in different levels are learned to take full advantage of the appearance and motion information. Extensive experiments have been conducted on two widely used fall datasets. The results have shown to demonstrate the effectiveness of the proposed method. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
11. Enhancing Real-Time Human Detection Based on Histograms of Oriented Gradients
- Author
-
Pedersoli, Marco, Gonzàlez, Jordi, Chakraborty, Bhaskar, Villanueva, Juan J., Kacprzyk, Janusz, editor, Kurzynski, Marek, editor, Puchala, Edward, editor, Wozniak, Michal, editor, and Zolnierek, Andrzej, editor
- Published
- 2007
- Full Text
- View/download PDF
12. Human Silhouette and Skeleton Video Synthesis Through Wi-Fi signals
- Author
-
Danilo Avola, Marco Cascio, Luigi Cinque, Alessio Fagioli, and Gian Luca Foresti
- Subjects
FOS: Computer and information sciences ,Computer Networks and Communications ,Radio Waves ,Computer Science - Artificial Intelligence ,Computer Vision and Pattern Recognition (cs.CV) ,Wi-Fi signal ,Computer Science - Computer Vision and Pattern Recognition ,Video synthesis ,General Medicine ,Human silhouette ,Artificial Intelligence (cs.AI) ,Skeleton ,Humans ,Wireless Technology - Abstract
The increasing availability of wireless access points (APs) is leading toward human sensing applications based on Wi-Fi signals as support or alternative tools to the widespread visual sensors, where the signals enable to address well-known vision-related problems such as illumination changes or occlusions. Indeed, using image synthesis techniques to translate radio frequencies to the visible spectrum can become essential to obtain otherwise unavailable visual data. This domain-to-domain translation is feasible because both objects and people affect electromagnetic waves, causing radio and optical frequencies variations. In the literature, models capable of inferring radio-to-visual features mappings have gained momentum in the last few years since frequency changes can be observed in the radio domain through the channel state information (CSI) of Wi-Fi APs, enabling signal-based feature extraction, e.g. amplitude. On this account, this paper presents a novel two-branch generative neural network that effectively maps radio data into visual features, following a teacher–student design that exploits a cross-modality supervision strategy. The latter conditions signal-based features in the visual domain to completely replace visual data. Once trained, the proposed method synthesizes human silhouette and skeleton videos using exclusively Wi-Fi signals. The approach is evaluated on publicly available data, where it obtains remarkable results for both silhouette and skeleton videos generation, demonstrating the effectiveness of the proposed cross-modality supervision strategy.
- Published
- 2022
13. Multi-view fall detection based on spatio-temporal interest points.
- Author
-
Su, Songzhi, Wu, Sin-Sian, Chen, Shu-Yuan, Duh, Der-Jyh, and Li, Shaozi
- Subjects
COMPUTER vision ,WEARABLE technology ,SPATIO-temporal variation ,CAMERA angles ,SUPPORT vector machines - Abstract
Many countries are experiencing a rapid increase in their elderly populations, increasing the demand for appropriate healthcare systems including fall-detection systems. In recent years, many fall-detection systems have been developed, although most require the use of wearable devices. Such systems function only when the subject is wearing the device. A vision-based system presents a more convenient option. However, visual features typically depend on camera view; a single, fixed camera may not properly identify falls occurring in various directions. Thus, this study presents a solution that involves using multiple cameras. The study offers two main contributions. First, in contrast to most vision-based systems that analyze silhouettes to detect falls, the present system proposes a novel feature for measuring the degree of impact shock that is easily detectable with a wearable device but more difficult with a computer vision system. In addition, the degree of impact shock is less sensitive to camera views and can be extracted more robustly than a silhouette. Second, the proposed method uses a majority-voting strategy based on multiple views to avoid performing the tedious camera calibration required by most multiple-camera approaches. Specifically, the proposed method is based on spatio-temporal interest points (STIPs). The number of local STIP clusters is designed to indicate the degree of impact shock and body vibration. Sequences of these features are concatenated into feature vectors that are then fed into a support vector machine to classify the fall event. A majority-voting strategy based on multiple views is then used for the final determination. The proposed method has been applied to a publicly available dataset to offer evidence that the proposed method outperforms existing methods based on the same data input. [ABSTRACT FROM AUTHOR]
- Published
- 2016
- Full Text
- View/download PDF
14. A Robust Visual Human Detection Approach With UKF-Based Motion Tracking for a Mobile Robot.
- Author
-
Gupta, Meenakshi, Behera, Laxmidhar, Subramanian, Venkatesh K., and Jamshidi, Mo M.
- Abstract
Robust tracking of a human in a video sequence is an essential prerequisite to an increasing number of applications, where a robot needs to interact with a human user or operates in a human-inhabited environment. This paper presents a robust approach that enables a mobile robot to detect and track a human using an onboard RGB-D sensor. Such robots could be used for security, surveillance, and assistive robotics applications. The proposed approach has real-time computation power through a unique combination of new ideas and well-established techniques. In the proposed method, background subtraction is combined with depth segmentation detector and template matching method to initialize the human tracking automatically. A novel concept of head and hand creation based on depth of interest is introduced in this paper to track the human silhouette in a dynamic environment, when the robot is moving. To make the algorithm robust, a series of detectors (e.g., height, size, and shape) is utilized to distinguish target human from other objects. Because of the relatively high computation time of the silhouette-matching-based method, a confidence level is defined, which allows using the matching-based method only where it is imperative. An unscented Kalman filter is used to predict the human location in the image frame to maintain the continuity of the robot motion. The efficacy of the approach is demonstrated through a real experiment on a mobile robot navigating in an indoor environment. [ABSTRACT FROM PUBLISHER]
- Published
- 2015
- Full Text
- View/download PDF
15. FASSD-Net Model for Person Semantic Segmentation
- Author
-
Gabriel Sanchez-Perez, Jesus Olivares-Mercado, Jose Portillo-Portillo, Luis Brandon Garcia-Ortiz, Aldo Hernandez-Suarez, Hector Perez-Meana, Karina Toscano-Medina, and Gibran Benitez-Garcia
- Subjects
Scheme (programming language) ,TK7800-8360 ,Computer Networks and Communications ,Computer science ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,02 engineering and technology ,Convolutional neural network ,cityscapes ,Task (project management) ,Silhouette ,0502 economics and business ,0202 electrical engineering, electronic engineering, information engineering ,Segmentation ,Electrical and Electronic Engineering ,computer.programming_language ,050210 logistics & transportation ,Class (computer programming) ,business.industry ,Deep learning ,05 social sciences ,deep learning ,Pattern recognition ,semantic segmentation ,Identification (information) ,Hardware and Architecture ,Control and Systems Engineering ,human silhouette ,Signal Processing ,020201 artificial intelligence & image processing ,Artificial intelligence ,Electronics ,business ,computer ,person class - Abstract
This paper proposes the use of the FASSD-Net model for semantic segmentation of human silhouettes, these silhouettes can later be used in various applications that require specific characteristics of human interaction observed in video sequences for the understanding of human activities or for human identification. These applications are classified as high-level task semantic understanding. Since semantic segmentation is presented as one solution for human silhouette extraction, it is concluded that convolutional neural networks (CNN) have a clear advantage over traditional methods for computer vision, based on their ability to learn the representations of appropriate characteristics for the task of segmentation. In this work, the FASSD-Net model is used as a novel proposal that promises real-time segmentation in high-resolution images exceeding 20 FPS. To evaluate the proposed scheme, we use the Cityscapes database, which consists of sundry scenarios that represent human interaction with its environment (these scenarios show the semantic segmentation of people, difficult to solve, that favors the evaluation of our proposal), To adapt the FASSD-Net model to human silhouette semantic segmentation, the indexes of the 19 classes traditionally proposed for Cityscapes were modified, leaving only two labels: One for the class of interest labeled as person and one for the background. The Cityscapes database includes the category “human” composed for “rider” and “person” classes, in which the rider class contains incomplete human silhouettes due to self-occlusions for the activity or transport used. For this reason, we only train the model using the person class rather than human category. The implementation of the FASSD-Net model with only two classes shows promising results in both a qualitative and quantitative manner for the segmentation of human silhouettes.
- Published
- 2021
16. Robust human silhouette extraction with Laplacian fitting.
- Author
-
Al-Maadeed, Somaya, Almotaeryi, Resheed, Jiang, Richard, and Bouridane, Ahmed
- Subjects
- *
ROBUST control , *FEATURE extraction , *LAPLACIAN matrices , *GESTURE , *IMAGE segmentation , *COMPUTATIONAL complexity - Abstract
Human silhouette extraction has been a primary step to estimate human poses or classify activities from videos. While the accuracy of human silhouettes has great impact on the follow-on human pose/gait estimation, it has been important to guarantee the highly-accurate extraction of human silhouettes. However, traditional methods such as motion segmentation can be fragile due to the complexity of real-world environment. In this paper, we propose an automated human silhouette extraction algorithm to attain this highly-demanded task. In our proposed scheme, the initial motion segmentation of foreground objects was roughly computed by Stauffer’s background subtraction using Gaussian mixtures, and then refined by the proposed Laplacian fitting scheme. In our method, the candidate regions of human objects are taken as the initial input, their Laplacian matrices are constructed, and Eigen mattes are then obtained by minimizing on Laplacian matrices. RANSAC algorithm is then applied to fit the Eigen mattes iteratively with inliers of the initially estimated motion blob. Finally, the foreground human silhouettes are obtained from the optimized matte fitting. Experimental results on a number of test videos validated that the proposed Laplacian fitting scheme enhances the accuracy in automated human silhouette extraction, exhibiting a potential use of our Laplacian fitting algorithm in many silhouette-based applications such as human pose estimation. [ABSTRACT FROM AUTHOR]
- Published
- 2014
- Full Text
- View/download PDF
17. Action recognition using graph embedding and the co-occurrence matrices descriptor.
- Author
-
Zheng, Feng, Shao, Ling, Song, Zhan, and Chen, Xi
- Subjects
- *
EMBEDDINGS (Mathematics) , *GRAPH theory , *MATRICES (Mathematics) , *COMPUTER vision , *ALGORITHMS , *NONLINEAR statistical models - Abstract
Recognizing actions from a monocular video is a very hot topic in computer vision recently. In this paper, we propose a new representation of actions, the co-occurrence matrices descriptor, on the intrinsic shape manifold learned by graph embedding. The co-occurrence matrices descriptor captures more temporal information than the bag of words (histogram) descriptor which only considers the spatial information, thus boosting the classification accuracy. In addition, we compare the performance of the co-occurrence matrices descriptor on different manifolds learned by various graph-embedding methods. Graph-embedding methods preserve as much of the significant structure of the high-dimensional data as possible in the low-dimensional map. The results show that nonlinear algorithms are more robust than linear ones. Furthermore, we conclude that the label information plays a critical role in learning more discriminating manifolds. [ABSTRACT FROM AUTHOR]
- Published
- 2011
- Full Text
- View/download PDF
18. A Study on Gait-Based Gender Classification.
- Author
-
Shiqi Yu, Tieniu Tan, Kaiqi Huang, Kui Jia, and Xinyu Wu
- Subjects
- *
GENDER , *GAIT in humans , *HUMAN body , *SILHOUETTES , *IMAGE processing - Abstract
Gender is an important cue in social activities. In this correspondence, we present a study and analysis of gender classification based on human gait. Psychological experiments were carried out. These experiments showed that humans can recognize gender based on gait information, and that contributions of different body components vary. The prior knowledge extracted from the psychological experiments can be combined with an automatic method to further improve classification accuracy. The proposed method which combines human knowledge achieves higher performance than some other methods, and is even more accurate than human observers. We also present a numerical analysis of the contributions of different human components, which shows that head and hair, back, chest and thigh are more discriminative than other components. We also did challenging cross-race experiments that used Asian gait data to classify the gender of Europeans, and vice versa. Encouraging results were obtained. All the above prove that gait-based gender classification is feasible in controlled environments. In real applications, it still suffers from many difficulties, such as view variation, clothing and shoes changes, or carrying objects. We analyze the difficulties and suggest some possible solutions. [ABSTRACT FROM AUTHOR]
- Published
- 2009
- Full Text
- View/download PDF
19. FASSD-Net Model for Person Semantic Segmentation.
- Author
-
Garcia-Ortiz, Luis Brandon, Portillo-Portillo, Jose, Hernandez-Suarez, Aldo, Olivares-Mercado, Jesus, Sanchez-Perez, Gabriel, Toscano-Medina, Karina, Perez-Meana, Hector, and Benitez-Garcia, Gibran
- Subjects
HUMAN activity recognition ,CONVOLUTIONAL neural networks ,COMPUTER vision ,DEEP learning ,HUMAN ecology - Abstract
This paper proposes the use of the FASSD-Net model for semantic segmentation of human silhouettes, these silhouettes can later be used in various applications that require specific characteristics of human interaction observed in video sequences for the understanding of human activities or for human identification. These applications are classified as high-level task semantic understanding. Since semantic segmentation is presented as one solution for human silhouette extraction, it is concluded that convolutional neural networks (CNN) have a clear advantage over traditional methods for computer vision, based on their ability to learn the representations of appropriate characteristics for the task of segmentation. In this work, the FASSD-Net model is used as a novel proposal that promises real-time segmentation in high-resolution images exceeding 20 FPS. To evaluate the proposed scheme, we use the Cityscapes database, which consists of sundry scenarios that represent human interaction with its environment (these scenarios show the semantic segmentation of people, difficult to solve, that favors the evaluation of our proposal), To adapt the FASSD-Net model to human silhouette semantic segmentation, the indexes of the 19 classes traditionally proposed for Cityscapes were modified, leaving only two labels: One for the class of interest labeled as person and one for the background. The Cityscapes database includes the category "human" composed for "rider" and "person" classes, in which the rider class contains incomplete human silhouettes due to self-occlusions for the activity or transport used. For this reason, we only train the model using the person class rather than human category. The implementation of the FASSD-Net model with only two classes shows promising results in both a qualitative and quantitative manner for the segmentation of human silhouettes. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
20. Human Action Recognition System Based on Silhouette
- Author
-
S. Maheswari and P. Arockia Jansi Rani
- Subjects
optical flow ,classification ,human silhouette ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Background subtraction - Abstract
Human action is recognized directly from the video sequences. The objective of this work is to recognize various human actions like run, jump, walk etc. Human action recognition requires some prior knowledge about actions namely, the motion estimation, foreground and background estimation. Region of interest (ROI) is extracted to identify the human in the frame. Then, optical flow technique is used to extract the motion vectors. Using the extracted features similarity measure based classification is done to recognize the action. From experimentations upon the Weizmann database, it is found that the proposed method offers a high accuracy., {"references":["A. Bobick and J. Davis. The recognition of human movement using temporal templates. IEEE Transactions on Pattern Analysis and Machine Intelligence, 23(3):257–267, 2002.","Antonios Oikonomopoulos, Ioannis Patras and Maja Pantic. \"Spatiotemporal Salient Points For visual Recognition of Human actions.\" IEEE Transactions on Image Processing Vol. 36, no. No. 3. (2006).","Antonios Oikonomopoulos, Maja Pantic, Ioannis Patras \"Sparse B-spline polynomial descriptors for human activity recognition\" in 2009.","Danielweinland, RemiRonfard and Edmond Boyer, \"A Survey of Vision- Based Methods for Action Representation, Segmentation and Recognition\", October 18, 2010.","Droogenbroeck, O. Barnich and M. Van. \"ViBe: A universal background subtraction algorithm for video sequences.\" IEEE Transactions on Image Processing, June 2011. 20(6):1709-1724.","Enric Meinhardt-Llopis, Javier Sanchez, Daniel Kondermann , \"Horn–Schunck Optical Flow with a Multi-Scale Strategy\", Image Processing On Line, 3 (2013), pp. 151–172.","I. Laptev, M. Marszalek, C. Schmid, and B. Rozenfeld. Learning realistic human actions from movies. In IEEE Conference on Computer Vision and Pattern Recognition, pages 1–8, 2008.","J. K. Aggarwal, Q. Cai, W. Liao, and B. Sabata, Articulated and elastic non-rigid motion: a review, in Workshop on Motion of Non-Rigid and Articulated Objects, Austin, TX, 1994, pp. 2–14.","J. Yamato, J. Ohya, and K. Ishii. Recognizing human action in time-sequential images using hidden markov model. In CVPR, 1992.\n[10]\tK. Schindler and L. van Gool. Action snippets: How many frames does human action recognition require? In CVPR, 2008.\n[11]\tKeigo Takahara, Takashi Toriu and Thi Thi Zin. \"Making Background Subtraction Robustto Various Illumination Changes.\" IJCSNS International Journal of Computer Science and Network Security, March 2011: VOL.11 No.3.\n[12]\tM. Blank, L. Gorelick, E. Shechtman, M. Irani, and R. Basri. \"Actions as space-time shapes.\" In ICCV, 2005.\n[13]\tM. Rodriguez, J. Ahmed, and M. Shah. Action MACH a spatio-temporal Maximum Average Correlation Height filter for action recognition. In IEEE Conference on Computer Vision and Pattern Recognition, pages 1–8, 2008.\n[14]\tMaja Pantic, Alexpentland and Thomas Huang, \"Human Computing and Machine Understanding of Human Behaviour\", ICMI'06, November 2-4, 2006.\n[15]\tN.A. Deepak and U.N.Sinha, \"Silhouette Based Human Motion Detection and Recognising their Actions from the captured Video Streams\", IntJ. Advanced Networking and Applications Volume: 02, Issue: 05, Pages: 817-823(2011).\n[16]\tR. Polana and R. Nelson. Low level recognition of human motion. In IEEE Workshop onNonrigid and Articulate Motion, 1994."]}
- Published
- 2016
- Full Text
- View/download PDF
21. Robust human silhouette extraction with Laplacian fitting
- Author
-
Richard Jiang, Resheed Almotaeryi, Ahmed Bouridane, and Somaya Al-Maadeed
- Subjects
Gaussian ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,RANSAC ,Silhouette ,Human silhouette ,symbols.namesake ,Artificial Intelligence ,Computer vision ,Segmentation ,Pose ,Pose estimation ,Mathematics ,ComputingMethodologies_COMPUTERGRAPHICS ,Background subtraction ,business.industry ,Pattern recognition ,Computer Science::Computer Vision and Pattern Recognition ,Signal Processing ,symbols ,Computer Vision and Pattern Recognition ,Artificial intelligence ,Laplacian matrix ,business ,Laplace operator ,Software - Abstract
Human silhouette extraction has been a primary step to estimate human poses or classify activities from videos. While the accuracy of human silhouettes has great impact on the follow-on human pose/gait estimation, it has been important to guarantee the highly-accurate extraction of human silhouettes. However, traditional methods such as motion segmentation can be fragile due to the complexity of real-world environment. In this paper, we propose an automated human silhouette extraction algorithm to attain this highly-demanded task. In our proposed scheme, the initial motion segmentation of foreground objects was roughly computed by Stauffer's background subtraction using Gaussian mixtures, and then refined by the proposed Laplacian fitting scheme. In our method, the candidate regions of human objects are taken as the initial input, their Laplacian matrices are constructed, and Eigen mattes are then obtained by minimizing on Laplacian matrices. RANSAC algorithm is then applied to fit the Eigen mattes iteratively with inliers of the initially estimated motion blob. Finally, the foreground human silhouettes are obtained from the optimized matte fitting. Experimental results on a number of test videos validated that the proposed Laplacian fitting scheme enhances the accuracy in automated human silhouette extraction, exhibiting a potential use of our Laplacian fitting algorithm in many silhouette-based applications such as human pose estimation. Scopus
- Published
- 2014
22. Contour segment analysis for human silhouette pre-segmentation
- Author
-
Migniot, Cyrille, Bertolino, Pascal, Chassery, Jean-Marc, GIPSA - Architecture, Géométrie, Perception, Images, Gestes (GIPSA-AGPIG), Département Images et Signal (GIPSA-DIS), Grenoble Images Parole Signal Automatique (GIPSA-lab), Université Stendhal - Grenoble 3-Université Pierre Mendès France - Grenoble 2 (UPMF)-Université Joseph Fourier - Grenoble 1 (UJF)-Institut polytechnique de Grenoble - Grenoble Institute of Technology (Grenoble INP )-Centre National de la Recherche Scientifique (CNRS)-Université Stendhal - Grenoble 3-Université Pierre Mendès France - Grenoble 2 (UPMF)-Université Joseph Fourier - Grenoble 1 (UJF)-Institut polytechnique de Grenoble - Grenoble Institute of Technology (Grenoble INP )-Centre National de la Recherche Scientifique (CNRS)-Grenoble Images Parole Signal Automatique (GIPSA-lab), and Université Stendhal - Grenoble 3-Université Pierre Mendès France - Grenoble 2 (UPMF)-Université Joseph Fourier - Grenoble 1 (UJF)-Institut polytechnique de Grenoble - Grenoble Institute of Technology (Grenoble INP )-Centre National de la Recherche Scientifique (CNRS)-Université Stendhal - Grenoble 3-Université Pierre Mendès France - Grenoble 2 (UPMF)-Université Joseph Fourier - Grenoble 1 (UJF)-Institut polytechnique de Grenoble - Grenoble Institute of Technology (Grenoble INP )-Centre National de la Recherche Scientifique (CNRS)
- Subjects
[INFO.INFO-TS]Computer Science [cs]/Signal and Image Processing ,human silhouette ,SVM ,human segmentation ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,silhouette ,[SPI.SIGNAL]Engineering Sciences [physics]/Signal and Image processing ,HOGs - Abstract
International audience; The union of Histograms of Oriented Gradients based descriptors and of a Support Vector Machine classifier is a classic and efficient method for human detection in the images. Conversely, as often in detection, accurate segmentation of these persons is not performed. Many applications however need it. this paper tackles the problem of giving rise to information that will guide the final segmentation step. It presents a method which uses the union mention above to relate to each contour segment a likelihood degree of being part of a human silhouette.
- Published
- 2010
23. Histogram of oriented rectangles: A new pose descriptor for human action recognition
- Author
-
Pinar Duygulu and Nazlı İkizler
- Subjects
Complex representation ,Dynamic time warping ,Computation time ,Computation ,Human Motion Understanding ,Temporal representations ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Compact representation ,Histogramming ,Temporal domain ,k-nearest neighbors algorithm ,Silhouette ,Human silhouette ,Nearest neighbor classification ,Gesture recognition ,Motion energy ,Statistical tests ,Models ,Histogram ,Parameter estimation ,Computer vision ,Robust recognition ,Pose Descriptor ,Mathematics ,Action sequences ,Accuracy rate ,business.industry ,Spatial domains ,Pattern recognition ,Descriptor ,Human-action recognition ,Rectangular patch ,Support vector machine ,Human form models ,Complex model ,Action Recognition ,Signal Processing ,Computer Vision and Pattern Recognition ,Artificial intelligence ,Data sets ,business ,Classifier (UML) ,Human pose - Abstract
Cataloged from PDF version of article. Most of the approaches to human action recognition tend to form complex models which require lots of parameter estimation and computation time. In this study, we show that, human actions can be simply represented by pose without dealing with the complex representation of dynamics. Based on this idea, we propose a novel pose descriptor which we name as Histogram-of-Oriented-Rectangles (HOR) for representing and recognizing human actions in videos. We represent each human pose in an action sequence by oriented rectangular patches extracted over the human silhouette. We then form spatial oriented histograms to represent the distribution of these rectangular patches. We make use of several matching strategies to carry the information from the spatial domain described by the HOR descriptor to temporal domain. These are (i) nearest neighbor classification, which recognizes the actions by matching the descriptors of each frame, (ii) global histogramming, which extends the idea of Motion Energy Image proposed by Bobick and Davis to rectangular patches, (iii) a classifier-based approach using Support Vector Machines, and (iv) adaptation of Dynamic Time Warping on the temporal representation of the HOR descriptor. For the cases when pose descriptor is not sufficiently strong alone, such as to differentiate actions "jogging" and "running", we also incorporate a simple velocity descriptor as a prior to the pose based classification step. We test our system with different configurations and experiment on two commonly used action datasets: the Weizmann dataset and the KTH dataset. Results show that our method is superior to other methods on Weizmann dataset with a perfect accuracy rate of 100%, and is comparable to the other methods on KTH dataset with a very high success rate close to 90%. These results prove that with a simple and compact representation, we can achieve robust recognition of human actions, compared to complex representations. (C) 2009 Elsevier B.V. All rights reserved.
- Published
- 2009
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.