149 results on '"foveation"'
Search Results
2. 27‐4: Design of a Statically Foveated Head‐Mounted Displays with a Novel Perceptual‐driven Approach.
- Author
-
Lyu, Pengyinjie and Hua, Hong
- Subjects
HIGH resolution imaging ,HEAD-mounted displays ,VISUAL acuity - Abstract
We introduce the design of statically foveated head‐mounted displays based on a new perceptual‐driven approach, aiming to offer a large field of view and high resolution with imperceptible image quality degradation eliminating the need for eye‐tracking or scanning mechanisms. The system incorporates a uniform‐resolution display panel and a spatially varying optical power eyepiece. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
3. Recovery of bacillary layer detachment associated with macula-off rhegmatogenous retinal detachment: Evidence of foveation mechanisms?
- Author
-
Andrea Govetto, Paolo Radice, Sara Lucchini, Daniela Bacherini, Stanislao Rizzo, and Alfonso Savastano
- Subjects
Retinal detachment ,Bacillary layer detachment ,Photoreceptors ,Foveation ,Fovea ,Optical coherence tomography ,Ophthalmology ,RE1-994 - Abstract
Purpose: To describe the mechanisms of postoperative foveal restoration in three patients with bacillary layer detachment (BALAD) associated with macula-off rhegmatogenous retinal detachment. Observations: BALAD associated with macula-off rhegmatogenous retinal detachment presented with two different morphologies: with an intact foveal roof (case 1) and lamellar, with an open foveal roof (cases 2 and 3). In case 1 visual acuity significantly improved and foveal morphology completely restored at postoperative month 6, with a marked increase in foveal thickness. Case 2 presented with a lamellar BALAD in a long-standing retinal detachment, and it was treated with scleral buckling with an unfavourable evolution into a full-thickness hole in the early postoperative period. It was then addressed with internal limiting membrane peeling and inverted flap, which resulted in the resolution of the lesion, but with limited postoperative visual and anatomical recovery. Case 3 lamellar BALAD was directly addressed with pars plana vitrectomy, ILM peeling and inverted flap, with a remarkable foveal anatomical restoration and visual acuity improvement over the follow-up period. Conclusions and importance: The process of foveation may play a key role in the healing process of BALAD associated with rhegmatogenous retinal detachment. Lamellar BALAD should be considered and treated as a FTMH associated with retinal detachment. The two BALAD subtypes may represent different clinical stages of the BALAD spectrum.
- Published
- 2023
- Full Text
- View/download PDF
4. Stakes of neuromorphic foveation: a promising future for embedded event cameras.
- Author
-
Gruel, Amélie, Hareb, Dalia, Grimaldi, Antoine, Martinet, Jean, Perrinet, Laurent, Linares-Barranco, Bernabé, and Serrano-Gotarredona, Teresa
- Subjects
- *
ARTIFICIAL neural networks , *CAMERAS , *COMPUTER vision , *NEUROSCIENCES , *ELECTRONIC data processing , *DIRECT action - Abstract
Foveation can be defined as the organic action of directing the gaze towards a visual region of interest to acquire relevant information selectively. With the recent advent of event cameras, we believe that taking advantage of this visual neuroscience mechanism would greatly improve the efficiency of event data processing. Indeed, applying foveation to event data would allow to comprehend the visual scene while significantly reducing the amount of raw data to handle. In this respect, we demonstrate the stakes of neuromorphic foveation theoretically and empirically across several computer vision tasks, namely semantic segmentation and classification. We show that foveated event data have a significantly better trade-off between quantity and quality of the information conveyed than high- or low-resolution event data. Furthermore, this compromise extends even over fragmented datasets. Our code is publicly available online at: https://github.com/amygruel/FoveationStakes_DVS. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
5. Visual Saliency Guided Foveated Video Compression
- Author
-
Shupei Zhang and Anup Basu
- Subjects
Foveation ,perceptual redundancy ,video compression ,visual saliency ,Electrical engineering. Electronics. Nuclear engineering ,TK1-9971 - Abstract
Video compression has become increasingly crucial as video resolution and bitrate have surged in recent years. However, most widely applied video compression methods do not fully exploit the characteristics of the Human Visual System (HVS) to reduce perceptual redundancy in videos. In this paper, we propose a novel video compression method that integrates visual saliency information with foveation to reduce perceptual redundancy. We present a new approach to subsample and restore the input image using saliency data, which allocates more space for salient regions and less for non-salient ones. We analyze the information entropy in video frames before and after applying our algorithm and demonstrate that the proposed method reduces redundancy. Through subjective and objective evaluations, we show that our method produces videos with superior perceptual visual quality. Moreover, our approach can be added to most existing video compression standards without altering their bitstream format.
- Published
- 2023
- Full Text
- View/download PDF
6. Impact Analysis of Foveation and Depth on the Perceptual Quality of Immersive Visual Content With Fixed Viewport
- Author
-
Chin-Han Shen and Hsu-Feng Hsiao
- Subjects
Depth map ,foveation ,perceptual quality analysis ,immersive visual contents ,Electrical engineering. Electronics. Nuclear engineering ,TK1-9971 - Abstract
With the rapid development of devices for virtual reality, massive amounts of multimedia data have been produced to enrich the immersive experience. To construct vivid virtual reality, high-resolution content is required to achieve a good user experience. With the limitations of hardware capability, efficient data compression is one of the keys to addressing the current dilemma. In our opinion, visual perception is a promising strategy for addressing this problem. In this study, the impact of human visual perception in immersive environment is studied. Two hypotheses accounting for their impact on the quality of immersive videos at a fixed viewport are proposed and validated by subjective evaluation and analysis. A series of videos are created with various blur distortions based on the stimulus triggered by depth perception and the rapid falloff of acuity in the peripheral vision. Stereoscopic videos with imposed blur distortions arranged with different severities are evaluated subjectively in an immersive viewing environment. We then follow statistical approaches to examine the collected subjective quality distributions to discuss the feasibility of video compression based on processed videos. To improve encoding efficiency, a heuristic mechanism for immersive visual content compression is proposed. We examine a simple compression framework to demonstrate that the proposed hypotheses are applicable for practical video compression.
- Published
- 2022
- Full Text
- View/download PDF
7. Subjective and Objective Quality Assessment of 2D and 3D Foveated Video Compression in Virtual Reality.
- Author
-
Jin, Yize, Chen, Meixu, Goodall, Todd, Patney, Anjul, and Bovik, Alan C.
- Subjects
- *
IMAGE compression , *VIRTUAL reality , *VIDEO compression , *VISUAL acuity , *DATABASES - Abstract
In Virtual Reality (VR), the requirements of much higher resolution and smooth viewing experiences under rapid and often real-time changes in viewing direction, leads to significant challenges in compression and communication. To reduce the stresses of very high bandwidth consumption, the concept of foveated video compression is being accorded renewed interest. By exploiting the space-variant property of retinal visual acuity, foveation has the potential to substantially reduce video resolution in the visual periphery, with hardly noticeable perceptual quality degradations. Accordingly, foveated image / video quality predictors are also becoming increasingly important, as a practical way to monitor and control future foveated compression algorithms. Towards advancing the development of foveated image / video quality assessment (FIQA / FVQA) algorithms, we have constructed 2D and (stereoscopic) 3D VR databases of foveated / compressed videos, and conducted a human study of perceptual quality on each database. Each database includes 10 reference videos and 180 foveated videos, which were processed by 3 levels of foveation on the reference videos. Foveation was applied by increasing compression with increased eccentricity. In the 2D study, each video was of resolution $7680\times 3840$ and was viewed and quality-rated by 36 subjects, while in the 3D study, each video was of resolution $5376\times 5376$ and rated by 34 subjects. Both studies were conducted on top of a foveated video player having low motion-to-photon latency (~50ms). We evaluated different objective image and video quality assessment algorithms, including both FIQA / FVQA algorithms and non-foveated algorithms, on our so called LIVE-Facebook Technologies Foveation-Compressed Virtual Reality (LIVE-FBT-FCVR) databases. We also present a statistical evaluation of the relative performances of these algorithms. The LIVE-FBT-FCVR databases have been made publicly available and can be accessed at https://live.ece.utexas.edu/research/LIVEFBTFCVR/index.html. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
8. A Log-Rectilinear Transformation for Foveated 360-degree Video Streaming.
- Author
-
Li, David, Du, Ruofei, Babu, Adharsh, Brumar, Camelia D., and Varshney, Amitabh
- Subjects
STREAMING video & television ,HEAD-mounted displays ,VIDEO codecs ,PARALLEL algorithms ,VIRTUAL reality ,COMPUTER graphics - Abstract
With the rapidly increasing resolutions of 360° cameras, head-mounted displays, and live-streaming services, streaming high-resolution panoramic videos over limited-bandwidth networks is becoming a critical challenge. Foveated video streaming can address this rising challenge in the context of eye-tracking-equipped virtual reality head-mounted displays. However, conventional log-polar foveated rendering suffers from a number of visual artifacts such as aliasing and flickering. In this paper, we introduce a new log-rectilinear transformation that incorporates summed-area table filtering and off-the-shelf video codecs to enable foveated streaming of 360° videos suitable for VR headsets with built-in eye-tracking. To validate our approach, we build a client-server system prototype for streaming 360° videos which leverages parallel algorithms over real-time video transcoding. We conduct quantitative experiments on an existing 360° video dataset and observe that the log-rectilinear transformation paired with summed-area table filtering heavily reduces flickering compared to log-polar subsampling while also yielding an additional 10% reduction in bandwidth usage. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
9. Asymmetric Foveated Just-Noticeable-Difference Model for Images With Visual Field Inhomogeneities.
- Author
-
Chen, Zhenzhong and Wu, Wei
- Subjects
- *
VISUAL acuity , *VISION , *EXPERIMENTAL design , *ANISOTROPY , *VERTICAL jump , *VISUAL fields - Abstract
Just noticeable difference (JND) models reveal visibility limitation of human visual system (HVS) and exploit perceptual redundancy in the visual signal. Conventional just noticeable difference (JND) model supposes the visual acuity is homogeneous with or without considerations on the isoeccentric locations. Recent research on vision science indicates that the effect of eccentricity on visual sensitivity is not homogeneous across the visual field (VF). Relevant studies have found visual acuity at the isoeccentric locations is better along the horizontal meridian than along the vertical meridian, i.e., horizontal-vertical anisotropy (HVA). Along the vertical meridian, better performance has been reported for the inferior meridian compared to the superior meridian, i.e., vertical-meridian asymmetry (VMA). These properties of HVS should be taken into account to yield a better JND estimation, for better exploiting the perceptual redundancy. In this paper, we use two hemi-ellipses to represent the iso-acuity contour, which can better express the HVA and VMA effects. Moreover, we design psychophysical experiments under different test conditions, where the results show the hemi-ellipse model performs better when compared to the circular model. Finally, we build an asymmetric foveated JND model using two hemi-ellipses for approximation. When comparing with traditional JND models, experimental results demonstrate the effectiveness of the proposed model. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
10. P‐23: A Subjective Method for Evaluating Foveated Image Quality in HMDs.
- Author
-
Thirumalai, Vijayaraghavan, Ribera, Javier, Xiang, Jie, Zhang, Junbin, Azimi, Maryam, Kamali, Jalil, and Nasiopoulos, Panos
- Subjects
HEAD-mounted displays ,VIRTUAL reality equipment ,IMAGE compression ,VISION disorders ,IMAGE ,TEST methods ,IMAGE quality analysis - Abstract
Analyzing the perceptual image quality rendered on a head‐mounted display (HMD) is of primary importance to provide a flawless visual quality to VR headset users. In this paper, we propose a new subjective testing method to evaluate the quality of foveated images in HMDs. The images we use are coded by progressively increasing the compression ratio from the fixation point to the periphery. Our method is based on a standardized flickering paradigm which temporally interleaves the original and reconstructed images to make the differences more prominent. Results indicate that observers can identify more visual impairments near the fixation point than in periphery. This new subjective testing approach is, therefore, effective in assessing barely visible image artifacts of foveated imaging in HMDs. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
11. Saccade Amplitude Statistics are Explained by Cortical Magnification
- Author
-
Raabe, Marcel C., Lopez, Francisco M., Yu, Zhengyang, Caplan, Spencer, Yu, Chen, Shi, Bertram Emil, Triesch, Jochen, Raabe, Marcel C., Lopez, Francisco M., Yu, Zhengyang, Caplan, Spencer, Yu, Chen, Shi, Bertram Emil, and Triesch, Jochen
- Abstract
How natural and artificial vision systems learn and develop depends on how they sample information from their environment. Humans actively do so through saccadic eye movements. The statistics of saccade amplitudes have been well-characterized in tightly controlled contexts such as viewing images on a computer screen. However, the degree to which such findings generalize to real-world contexts involving moving agents and objects is currently unknown. Here, we first analyze saccade amplitude statistics of both infants and adults during naturalistic free play. We find that these differ significantly from those previously reported for head-fixed picture viewing, with a relatively smaller/greater abundance of medium/large saccades. Next, we present a computational model that explains saccade amplitude statistics based on the foveated nature of vision and the associated space-variant magnification of different portions of the visual field in primary visual cortex. Finally, we demonstrate computationally efficient approximations to this space-variant sampling using a small number of discrete resolution levels. © 2023 IEEE.
- Published
- 2023
12. Mitigating Cybersickness in Virtual Reality Systems through Foveated Depth-of-Field Blur
- Author
-
Razeen Hussain, Manuela Chessa, and Fabio Solari
- Subjects
cycbersickness ,spatial blur ,depth-of-field ,foveation ,gaze-contingent ,eye-tracker ,Chemical technology ,TP1-1185 - Abstract
Cybersickness is one of the major roadblocks in the widespread adoption of mixed reality devices. Prolonged exposure to these devices, especially virtual reality devices, can cause users to feel discomfort and nausea, spoiling the immersive experience. Incorporating spatial blur in stereoscopic 3D stimuli has shown to reduce cybersickness. In this paper, we develop a technique to incorporate spatial blur in VR systems inspired by the human physiological system. The technique makes use of concepts from foveated imaging and depth-of-field. The developed technique can be applied to any eye tracker equipped VR system as a post-processing step to provide an artifact-free scene. We verify the usefulness of the proposed system by conducting a user study on cybersickness evaluation. We used a custom-built rollercoaster VR environment developed in Unity and an HTC Vive Pro Eye headset to interact with the user. A Simulator Sickness Questionnaire was used to measure the induced sickness while gaze and heart rate data were recorded for quantitative analysis. The experimental analysis highlighted the aptness of our foveated depth-of-field effect in reducing cybersickness in virtual environments by reducing the sickness scores by approximately 66%.
- Published
- 2021
- Full Text
- View/download PDF
13. An automated segmentation approach to calibrating infantile nystagmus waveforms.
- Author
-
Dunn, Matt J., Harris, Christopher M., Ennis, Fergal A., Margrain, Tom H., Woodhouse, J. Margaret, McIlreavy, Lee, and Erichsen, Jonathan T.
- Subjects
- *
EYE movements , *EYE tracking , *NYSTAGMUS , *CALIBRATION - Abstract
Infantile nystagmus (IN) describes a regular, repetitive movement of the eyes. A characteristic feature of each cycle of the IN eye movement waveform is a period in which the eyes are moving at minimal velocity. This so-called "foveation" period has long been considered the basis for the best vision in individuals with IN. In recent years, the technology for measuring eye movements has improved considerably, but there remains the challenge of calibrating the direction of gaze in tracking systems when the eyes are continuously moving. Identifying portions of the nystagmus waveform suitable for calibration typically involves time-consuming manual selection of the foveation periods from the eye trace. Without an accurate calibration, the exact parameters of the waveform cannot be determined. In this study, we present an automated method for segmenting IN waveforms with the purpose of determining the foveation positions to be used for calibration of an eye tracker. On average, the "point of regard" was found to be within 0.21° of that determined by hand-marking by an expert observer. This method enables rapid clinical quantification of waveforms and the possibility of gaze-contingent research paradigms being performed with this patient group. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
14. Application-Oriented Retinal Image Models for Computer Vision
- Author
-
Ewerton Silva, Ricardo da S. Torres, Allan Pinto, Lin Tzy Li, José Eduardo S. Vianna, Rodolfo Azevedo, and Siome Goldenstein
- Subjects
retinal image model ,space-variant computer vision ,foveation ,low-power ,energy consumption ,Chemical technology ,TP1-1185 - Abstract
Energy and storage restrictions are relevant variables that software applications should be concerned about when running in low-power environments. In particular, computer vision (CV) applications exemplify well that concern, since conventional uniform image sensors typically capture large amounts of data to be further handled by the appropriate CV algorithms. Moreover, much of the acquired data are often redundant and outside of the application’s interest, which leads to unnecessary processing and energy spending. In the literature, techniques for sensing and re-sampling images in non-uniform fashions have emerged to cope with these problems. In this study, we propose Application-Oriented Retinal Image Models that define a space-variant configuration of uniform images and contemplate requirements of energy consumption and storage footprints for CV applications. We hypothesize that our models might decrease energy consumption in CV tasks. Moreover, we show how to create the models and validate their use in a face detection/recognition application, evidencing the compromise between storage, energy, and accuracy.
- Published
- 2020
- Full Text
- View/download PDF
15. Foveation Pipeline for 360° Video-Based Telemedicine
- Author
-
Muhammad Firdaus Syawaludin, Myungho Lee, and Jae-In Hwang
- Subjects
HMD ,telemedicine ,foveation ,multi-resolution ,Chemical technology ,TP1-1185 - Abstract
Pan-tilt-zoom (PTZ) and omnidirectional cameras serve as a video-mediated communication interface for telemedicine. Most cases use either PTZ or omnidirectional cameras exclusively; even when used together, images from the two are shown separately on 2D displays. Conventional foveated imaging techniques may offer a solution for exploiting the benefits of both cameras, i.e., the high resolution of the PTZ camera and the wide field-of-view of the omnidirectional camera, but displaying the unified image on a 2D display would reduce the benefit of “omni-” directionality. In this paper, we introduce a foveated imaging pipeline designed to support virtual reality head-mounted displays (HMDs). The pipeline consists of two parallel processes: one for estimating parameters for the integration of the two images and another for rendering images in real time. A control mechanism for placing the foveal region (i.e., high-resolution area) in the scene and zooming is also proposed. Our evaluations showed that the proposed pipeline achieved, on average, 17 frames per second when rendering the foveated view on an HMD, and showed angular resolution improvement on the foveal region compared with the omnidirectional camera view. However, the improvement was less significant when the zoom level was 8× and more. We discuss possible improvement points and future research directions.
- Published
- 2020
- Full Text
- View/download PDF
16. Compressed Sensing Meets the Human Visual System
- Author
-
Ciocoiu, Iulian B., Hutchison, David, Series editor, Kanade, Takeo, Series editor, Kittler, Josef, Series editor, Kleinberg, Jon M., Series editor, Mattern, Friedemann, Series editor, Mitchell, John C., Series editor, Naor, Moni, Series editor, Nierstrasz, Oscar, Series editor, Pandu Rangan, C., Series editor, Steffen, Bernhard, Series editor, Sudan, Madhu, Series editor, Terzopoulos, Demetri, Series editor, Tygar, Doug, Series editor, Vardi, Moshe Y., Series editor, Weikum, Gerhard, Series editor, Lee, Geuk, editor, Howard, Daniel, editor, and Ślęzak, Dominik, editor
- Published
- 2011
- Full Text
- View/download PDF
17. Recovery of bacillary layer detachment associated with macula-off rhegmatogenous retinal detachment: Evidence of foveation mechanisms?
- Author
-
Govetto A, Radice P, Lucchini S, Bacherini D, Rizzo S, and Savastano A
- Abstract
Purpose: To describe the mechanisms of postoperative foveal restoration in three patients with bacillary layer detachment (BALAD) associated with macula-off rhegmatogenous retinal detachment., Observations: BALAD associated with macula-off rhegmatogenous retinal detachment presented with two different morphologies: with an intact foveal roof (case 1) and lamellar, with an open foveal roof (cases 2 and 3). In case 1 visual acuity significantly improved and foveal morphology completely restored at postoperative month 6, with a marked increase in foveal thickness. Case 2 presented with a lamellar BALAD in a long-standing retinal detachment, and it was treated with scleral buckling with an unfavourable evolution into a full-thickness hole in the early postoperative period. It was then addressed with internal limiting membrane peeling and inverted flap, which resulted in the resolution of the lesion, but with limited postoperative visual and anatomical recovery. Case 3 lamellar BALAD was directly addressed with pars plana vitrectomy, ILM peeling and inverted flap, with a remarkable foveal anatomical restoration and visual acuity improvement over the follow-up period., Conclusions and Importance: The process of foveation may play a key role in the healing process of BALAD associated with rhegmatogenous retinal detachment . Lamellar BALAD should be considered and treated as a FTMH associated with retinal detachment. The two BALAD subtypes may represent different clinical stages of the BALAD spectrum., Competing Interests: The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper., (© 2023 Published by Elsevier Inc.)
- Published
- 2023
- Full Text
- View/download PDF
18. Toddler-inspired embodied vision for learning object representations
- Author
-
Aubret, Arthur, Teulière, Céline, Triesch, Jochen, Teulière, Céline, Institut Pascal (IP), Centre National de la Recherche Scientifique (CNRS)-Université Clermont Auvergne (UCA)-Institut national polytechnique Clermont Auvergne (INP Clermont Auvergne), Université Clermont Auvergne (UCA)-Université Clermont Auvergne (UCA), and Frankfurt Institute for Advanced Studies (FIAS )
- Subjects
[INFO.INFO-CV] Computer Science [cs]/Computer Vision and Pattern Recognition [cs.CV] ,contrastive learning ,foveation ,[INFO.INFO-CV]Computer Science [cs]/Computer Vision and Pattern Recognition [cs.CV] ,embodied vision ,unsupervised learning ,depth of field ,object recognition - Abstract
International audience; Recent time-contrastive learning approaches manage to learn invariant object representations without supervision. This is achieved by mapping successive views of an object onto close-by internal representations. When considering this learning approach as a model of the development of human object recognition, it is important to consider what visual input a toddler would typically observe while interacting with objects. First, human vision is highly foveated, with high resolution only available in the central region of the field of view. Second, objects may be seen against a blurry background due to toddlers' limited depth of field. Third, during object manipulation a toddler mostly observes close objects filling a large part of the field of view due to their rather short arms. Here, we study how these effects impact the quality of visual representations learnt through time-contrastive learning. To this end, we let a visually embodied agent "play" with objects in different locations of a near photo-realistic flat. During each play session the agent views an object in multiple orientations before turning its body to view another object. The resulting sequence of views feeds a time-contrastive learning algorithm. Our results show that visual statistics mimicking those of a toddler improve object recognition accuracy in both familiar and novel environments. We argue that this effect is caused by the reduction of features extracted in the background, a neural network bias for large features in the image and a greater similarity between novel and familiar background regions. The results of our model suggest that several influences on toddler's visual input statistics support their unsupervised learning of object representations.
- Published
- 2022
19. Joint foveation-depth just-noticeable-difference model for virtual reality environment.
- Author
-
Liu, Di, Wang, Yingbin, and Chen, Zhenzhong
- Subjects
- *
DIGITAL image processing , *COMPUTER vision , *DIFFERENCE equations , *THRESHOLD (Perception) , *LUMINANCE (Video) , *DEPTH maps (Digital image processing) - Abstract
Highlights • Just noticeable difference (JND) increases when viewing depth increase in the VR environment. • Considering depth effect on visibility threshold can add more unperceived noises in images. • Foveation effect that JND increases with eccentricity is also observed in the VR environment. Abstract In this paper, we develop a joint foveation-depth just-noticeable-difference (FD-JND) model to quantify the perceptual redundancy of image in the VR display environment. The proposed FD-JND model is developed with considerations on the effects of both foveation and depth. More specifically, experiments for the VR environment on synthesized stimuli are conducted based on luminance masking and contrast masking and the FD-JND model is developed accordingly. Subjective quality discrimination experiments between the noise contaminated images and original ones validate favorableness of the proposed FD-JND model. [ABSTRACT FROM AUTHOR]
- Published
- 2018
- Full Text
- View/download PDF
20. Foveation-Based Wireless Soft Image Delivery.
- Author
-
Shen, Jian, Yu, Lei, Li, Li, and Li, Houqiang
- Abstract
In image compression and transmission, two basic data redundancies can be identified and exploited: statistical redundancy and perceptual redundancy. Statistical redundancy has been studied for a very long time and has been exploited in most image coding schemes. However, perceptual redundancy has not yet been completely exploited. Furthermore, perceptual redundancy is difficult to exploit by using digital coding techniques because power allocation is performed at the bit level for digital coding. However, as one of one-size-fits-all technique, soft transmission is suitable to exploit the perceptual redundancy because its power allocation is directly applied at the pixel level instead of the bit level. In this paper, we propose a novel image transmission scheme, named FoveaCast , for single-input single-output or multi-input multi-output broadcast systems that effectively utilizes the foveation characteristic of human vision and analog coding techniques to achieve higher visual perceptual quality. Hence, our scheme possesses not only high visual perceptual quality but also graceful quality adaptation. Experimental evaluations show that compared with three existing digital schemes and two existing analog schemes, SoftCast and ParCast, FoveaCast achieves better visual perceptual performance. Meanwhile, it achieves a graceful quality variation along with channel variation, behaving just like SoftCast and ParCast. [ABSTRACT FROM AUTHOR]
- Published
- 2018
- Full Text
- View/download PDF
21. The caudal fastigial nucleus and the steering of saccades toward a moving visual target.
- Author
-
Bourrelly, Clara, Quinet, Julie, and Goffart, Laurent
- Subjects
- *
CEREBELLUM diseases , *SACCADIC eye movements , *CELL nuclei , *NEURONS , *PREFRONTAL cortex - Abstract
The caudal fastigial nuclei (cFN) are the output nuclei by which the medio-posterior cerebellum influences the production of visual saccades. We investigated in two head-restrained monkeys their contribution to the generation of interceptive saccades toward a target moving centrifugally by analyzing the consequences of a unilateral inactivation (10 injection sessions). We describe here the effects on saccades made toward a centrifugal target that moved along the horizontal meridian with a constant (10, 20, or 40°/s), increasing (from 0 to 40°/s over 600 ms), or decreasing (from 40 to 0°/s over 600 ms) speed. After muscimol injection, the monkeys were unable to foveate the current location of the moving target. The horizontal amplitude of interceptive saccades was reduced during contralesional target motions and hypermetric during ipsilesional ones. For both contralesional and ipsilesional saccades, the magnitude of dysmetria increased with target speed. However, the use of accelerating and decelerating targets revealed that the dependence of dysmetria upon target velocity was not due to the current velocity but to the required amplitude of saccade. We discuss these results in the framework of two hypotheses, the so-called "dual drive" and "bilateral" hypotheses. [ABSTRACT FROM AUTHOR]
- Published
- 2018
- Full Text
- View/download PDF
22. Detection of Foveation Windows and Analysis of Foveation Sequences in Congenital Nystagmus
- Author
-
Pasquariello, Giulio, Cesarelli, Mario, Bifulco, Paolo, Fratini, Antonio, La Gatta, Antonio, Boccuzzi, Domenico, Magjarevic, R., editor, Nagel, J. H., editor, Vander Sloten, Jos, editor, Verdonck, Pascal, editor, Nyssen, Marc, editor, and Haueisen, Jens, editor
- Published
- 2009
- Full Text
- View/download PDF
23. Analysis of Foveation Sequences in Congenital Nystagmus
- Author
-
Pasquariello, Giulio, Bifulco, P., Cesarelli, M., Romano, M., Fratini, A., Magjarevic, R., editor, Nagel, J. H., editor, Katashev, Alexei, editor, Dekhtyar, Yuri, editor, and Spigulis, Janis, editor
- Published
- 2008
- Full Text
- View/download PDF
24. A JND Guided Foveation Video Coding
- Author
-
Zhang, Erli, Zhao, Debin, Zhang, Yongbing, Liu, Hongbin, Ma, Siwei, Wang, Ronggang, Hutchison, David, Series editor, Kanade, Takeo, Series editor, Kittler, Josef, Series editor, Kleinberg, Jon M., Series editor, Mattern, Friedemann, Series editor, Mitchell, John C., Series editor, Naor, Moni, Series editor, Nierstrasz, Oscar, Series editor, Pandu Rangan, C., Series editor, Steffen, Bernhard, Series editor, Sudan, Madhu, Series editor, Terzopoulos, Demetri, Series editor, Tygar, Doug, Series editor, Vardi, Moshe Y., Series editor, Weikum, Gerhard, Series editor, Huang, Yueh-Min Ray, editor, Xu, Changsheng, editor, Cheng, Kuo-Sheng, editor, Yang, Jar-Ferr Kevin, editor, Swamy, M. N. S., editor, Li, Shipeng, editor, and Ding, Jen-Wen, editor
- Published
- 2008
- Full Text
- View/download PDF
25. Authentication of Volume Data Using Wavelet-Based Foveation
- Author
-
Kankanhalli, M. S., Chang, E.-C., Guan, X., Huang, Z., Wu, Y., Hansmann, W., editor, Purgathofer, W., editor, Sillion, F., editor, Jorge, Joaquim, editor, Correia, Nuno, editor, Jones, Huw, editor, and Kamegai, Meera Blattner, editor
- Published
- 2002
- Full Text
- View/download PDF
26. Domain Specificity of Oculomotor Learning after Changes in Sensory Processing.
- Author
-
Tsank, Yuliy and Eckstein, Miguel P.
- Subjects
- *
OCULOMOTOR nerve , *MOTOR learning , *SENSORIMOTOR integration , *SACCADIC eye movements , *SCOTOMA - Abstract
Humans visually process the world with varying spatial resolution and can program their eye movements optimally to maximize information acquisition for a variety of everyday tasks. Diseases such as macular degeneration can change visual sensory processing, introducing central vision loss (a scotoma). However, humans can learn to direct a new preferred retinal location to regions of interest for simple visual tasks. Whether such learned compensatory saccades are optimal and generalize to more complex tasks, which require integrating information across a large area of the visual field, is not well understood. Here, we explore the possible effects of central vision loss on the optimal saccades during a face identification task, using a gaze-contingent simulated scotoma. We show that a new foveated ideal observer with a central scotoma correctly predicts that the human optimal point of fixation to identify faces shifts from just below the eyes to one that is at the tip of the nose and another at the top of the forehead. However, even after 5000 trials, humans of both sexes surprisingly do not change their initial fixations to adapt to the new optimal fixation points to faces. In contrast, saccades do change for tasks such as object following and to a lesser extent during search. Our findings argue against a central brain motor-compensatory mechanism that generalizes across tasks. They instead suggest task specificity in the learning of oculomotor plans in response to changes in front-end sensory processing and the possibility of separate domain-specific representations of learned oculomotor plans in the brain. [ABSTRACT FROM AUTHOR]
- Published
- 2017
- Full Text
- View/download PDF
27. The superior colliculus and the steering of saccades toward a moving visual target.
- Author
-
Goffart, Laurent, Cecala, Aaron L., and Gandhi, Neeraj J.
- Published
- 2017
- Full Text
- View/download PDF
28. Foveation: an alternative method to simultaneously preserve privacy and information in face images.
- Author
-
Alonso, Víctor E., Enríquez-Caldera, Rogerio, and Sucar, Luis Enrique
- Subjects
- *
HUMAN facial recognition software , *FACIAL expression , *EXPRESSIVE behavior , *PRIVACY , *COMPUTER software - Abstract
This paper presents a real-time foveation technique proposed as an alternative method for image obfuscation while simultaneously preserving privacy in face deidentification. Relevance of the proposed technique is discussed through a comparative study of the most common distortions methods in face images and an assessment on performance and effectiveness of privacy protection. All the different techniques presented here are evaluated when they go through a face recognition software. Evaluating the data utility preservation was carried out under gender and facial expression classification. Results on quantifying the tradeoff between privacy protection and image information preservation at different obfuscation levels are presented. Comparative results using the facial expression subset of the FERET database show that the technique achieves a good tradeoff between privacy and awareness with 30% of recognition rate and a classification accuracy as high as 88% obtained from the common figures of merit using the privacy-awareness map. [ABSTRACT FROM AUTHOR]
- Published
- 2017
- Full Text
- View/download PDF
29. A Biologically Inspired Framework for Visual Information Processing and an Application on Modeling Bottom-Up Visual Attention.
- Author
-
Aboudib, Ala, Gripon, Vincent, and Coppin, Gilles
- Abstract
Background: An emerging trend in visual information processing is toward incorporating some interesting properties of the ventral stream in order to account for some limitations of machine learning algorithms. Selective attention and cortical magnification are two such important phenomena that have been the subject of a large body of research in recent years. In this paper, we focus on designing a new model for visual acquisition that takes these important properties into account. Methods: We propose a new framework for visual information acquisition and representation that emulates the architecture of the primate visual system by integrating features such as retinal sampling and cortical magnification while avoiding spatial deformations and other side effects produced by models that tried to implement these two features. It also explicitly integrates the notion of visual angle, which is rarely taken into account by vision models. We argue that this framework can provide the infrastructure for implementing vision tasks such as object recognition and computational visual attention algorithms. Results: To demonstrate the utility of the proposed vision framework, we propose an algorithm for bottom-up saliency prediction implemented using the proposed architecture. We evaluate the performance of the proposed model on the MIT saliency benchmark and show that it attains state-of-the-art performance, while providing some advantages over other models. Conclusion: Here is a summary of the main contributions of this paper: (1) Introducing a new bio-inspired framework for visual information acquisition and representation that offers the following properties: (a) Providing a method for taking the distance between an image and the viewer into account. This is done by incorporating a visual angle parameter which is ignored by most visual acquisition models. (b) Reducing the amount of visual information acquired by introducing a new scheme for emulating retinal sampling and cortical magnification effects observed in the ventral stream. (2) Providing a concrete application of the proposed framework by using it as a substrate for building a new saliency-based visual attention model, which is shown to attain state-of-the-art performance on the MIT saliency benchmark. (3) Providing an online Git repository that implements the introduced framework that is meant to be developed as a scalable, collaborative project. [ABSTRACT FROM AUTHOR]
- Published
- 2016
- Full Text
- View/download PDF
30. Considerations and Framework for Foveated Imaging Systems †
- Author
-
Ram M. Narayanan, Timothy J. Kane, Terence F. Rice, and Michael J. Tauber
- Subjects
scale space ,receptive fields ,log polar ,foveation ,computer vision ,raptor eye ,wide angle ,local magnification ,distortion control ,Applied optics. Photonics ,TA1501-1820 - Abstract
Foveated sight as observed in some raptor eyes is a motivation for artificial imaging systems requiring both wide fields of view as well as specific embedded regions of higher resolution. These foveated optical imaging systems are applicable to many acquisition and tracking tasks and as such are often required to be relatively portable and operate in real-time. Two approaches to achieve foveation have been explored in the past: optical system design and back-end data processing. In this paper, these previous works are compiled and used to build a framework for analyzing and designing practical foveated imaging systems. While each approach (physical control of optical distortion within the lens design process, and post-processing image re-sampling) has its own pros and cons, it is concluded that a combination of both techniques will further spur the development of more versatile, flexible, and adaptable foveated imaging systems in the future.
- Published
- 2018
- Full Text
- View/download PDF
31. Stakes of foveation on event cameras
- Author
-
Amélie Gruel, Jean Martinet, Bernabé Linares-Barranco, Teresa Serrano-Gotarredona, Laboratoire d'Informatique, Signaux, et Systèmes de Sophia Antipolis (I3S), Université Nice Sophia Antipolis (... - 2019) (UNS), COMUE Université Côte d'Azur (2015-2019) (COMUE UCA)-COMUE Université Côte d'Azur (2015-2019) (COMUE UCA)-Centre National de la Recherche Scientifique (CNRS)-Université Côte d'Azur (UCA), Instituto de Microelectrónica de Sevilla (IMSE-CNM), Universidad de Sevilla-Centro Nacional de Microelectronica [Spain] (CNM)-Consejo Superior de Investigaciones Científicas [Madrid] (CSIC), Centre National de la Recherche Scientifique [CNRS], ANR-19-CHR3-0008,APROVIS3D,Traitement analogique de capteur visuels bio-inspirés pour la reconstruction 3D(2019), COMUE Université Côte d'Azur (2015 - 2019) (COMUE UCA)-COMUE Université Côte d'Azur (2015 - 2019) (COMUE UCA)-Centre National de la Recherche Scientifique (CNRS)-Université Côte d'Azur (UCA), Instituto de Microelectrónica de Sevilla (IMSE-CNM CSIC), and Centro Nacional de Microelectronica [Spain] (CNM)-Consejo Superior de Investigaciones Científicas [Madrid] (CSIC)
- Subjects
saliency ,Foveation ,segmentation ,[INFO.INFO-CV]Computer Science [cs]/Computer Vision and Pattern Recognition [cs.CV] ,[INFO]Computer Science [cs] ,saillance ,event cameras ,[INFO.INFO-AI]Computer Science [cs]/Artificial Intelligence [cs.AI] ,caméras événementielles ,fovéation - Abstract
International audience; Foveation is the organic action of directing the gaze towards a visual region of interest, to selectively acquire relevant information. In the recent advent of event cameras, we believe that developing such a mechanism would greatly improve the efficiency of event-data processing. Indeed, applying foveation to event-based data would allow to comprehend the visual scene while significantly reducing the amount of raw data to handle. We study the evolution of the accuracy of segmentation with respect to the amount of event data used, to demonstrate the stakes of foveation.; La fovéation consiste en la direction du regard vers une région visuelle d'intérêt pour acquérir sélectivement des informations pertinentes. Dans l'avènement récent des caméras événementielles, nous pensons que le développement d'un tel mécanisme améliorerait l'efficacité du traitement des données événementielles. En effet, appliquer la fovéation à celles-ci permettrait de comprendre la scène visuelle tout en réduisant considérablement la quantité de données brutes à traiter. Nous étudions l'évolution de la précision de la segmentation par rapport à la quantité de données utilisées, pour démontrer les enjeux de la fovéation.
- Published
- 2021
32. Stakes of foveation on event cameras
- Author
-
Gruel, Amélie, Martinet, Jean, Linares-Barranco, Bernabé, Serrano-Gotarredona, Teresa, Laboratoire d'Informatique, Signaux, et Systèmes de Sophia Antipolis (I3S), Université Nice Sophia Antipolis (... - 2019) (UNS), COMUE Université Côte d'Azur (2015-2019) (COMUE UCA)-COMUE Université Côte d'Azur (2015-2019) (COMUE UCA)-Centre National de la Recherche Scientifique (CNRS)-Université Côte d'Azur (UCA), Instituto de Microelectrónica de Sevilla (IMSE-CNM), Universidad de Sevilla-Centro Nacional de Microelectronica [Spain] (CNM)-Consejo Superior de Investigaciones Científicas [Madrid] (CSIC), Centre National de la Recherche Scientifique [CNRS], and ANR-19-CHR3-0008,APROVIS3D,Traitement analogique de capteur visuels bio-inspirés pour la reconstruction 3D(2019)
- Subjects
segmentation ,[INFO]Computer Science [cs] ,saillance ,caméras événementielles ,fovéation - Abstract
International audience; Foveation is the organic action of directing the gaze towards a visual region of interest, to selectively acquire relevant information. In the recent advent of event cameras, we believe that developing such a mechanism would greatly improve the efficiency of event-data processing. Indeed, applying foveation to event-based data would allow to comprehend the visual scene while significantly reducing the amount of raw data to handle. We study the evolution of the accuracy of segmentation with respect to the amount of event data used, to demonstrate the stakes of foveation.; La fovéation consiste en la direction du regard vers une région visuelle d'intérêt pour acquérir sélectivement des informations pertinentes. Dans l'avènement récent des caméras événementielles, nous pensons que le développement d'un tel mécanisme améliorerait l'efficacité du traitement des données événementielles. En effet, appliquer la fovéation à celles-ci permettrait de comprendre la scène visuelle tout en réduisant considérablement la quantité de données brutes à traiter.Nous étudions l'évolution de la précision de la segmentation par rapport à la quantité de données utilisées, pour démontrer les enjeux de la fovéation.
- Published
- 2021
33. Visual perception of noise in a simulated holographic display—A user study.
- Author
-
Georgiou, Andreas, Kollin, Joel, Hewitt, Charlie, Chakravarthula, Praneeth, and Guenter, Brian
- Subjects
- *
HOLOGRAPHIC displays , *HOLOGRAPHY , *HEAD-mounted displays , *IMAGE reconstruction , *EYE tracking , *NOISE , *VISUAL perception - Abstract
This work explores ways to bypass the fundamental image quality limitations of displays using Computer Generated Holograms (CGHs) and, specifically, the high-frequency noise associated with phase-only holograms. Although there is a wealth of literature on building experimental holographic systems, there are no user studies to assess the performance of a holographic system projecting a dynamic two-dimensional image. In this study, 18 participants blindly compared three groups of images displayed on a conventional monitor. The first group contained the original image, the second the simulated holographic reconstruction of the original image, and the third group had the foveated reconstruction of the original image, based on the pupil position. Holograms in the second group were computed using the Fienup algorithm and the third group using the Fienup with Perceptual Don't Care Areas (FiPDoC) algorithm, a novel algorithm that uses eye tracking to optimize image quality in CGHs. The aim of the study was to find out if the holographic display, assuming an ideal hardware, can be as good as a conventional display and whether eye tracking can help this goal. Most participants distinguished between the original image and the un-foveated simulated reconstruction. However, the participants could not differentiate between the original image and the foveated reconstruction. Thus, foveation may be essential in designing and building the first commercial holographic displays. • Holographically projected images cannot be noise free due to theoretical limitations. • Holographic images often "bury" noise in areas of high-frequency content. • Foveation can reduce perceived noise in images with uniform intensity or slow-changing features. • Holographic head-mounted displays can utilize on-board eye-tracking for foveation. • In a user study, most reconstructions were indistinguishable from the original images [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
34. An automated segmentation approach to calibrating infantile nystagmus waveforms
- Author
-
Fergal Ennis, J. Margaret Woodhouse, Lee McIlreavy, Chris Harris, Thomas Hengist Margrain, Matthew Dunn, and Jonathan T. Erichsen
- Subjects
genetic structures ,Computer science ,Visual Acuity ,Experimental and Cognitive Psychology ,Nystagmus ,Nystagmus, Pathologic ,Article ,050105 experimental psychology ,Automation ,03 medical and health sciences ,0302 clinical medicine ,Arts and Humanities (miscellaneous) ,Developmental and Educational Psychology ,medicine ,Calibration ,Humans ,Waveform ,0501 psychology and cognitive sciences ,Computer vision ,Eye Movement Measurements ,General Psychology ,business.industry ,Foveation ,05 social sciences ,Eye movement ,Tracking system ,Observer (special relativity) ,Gaze ,eye diseases ,Eye movements ,Eye tracking ,Psychology (miscellaneous) ,Artificial intelligence ,medicine.symptom ,business ,030217 neurology & neurosurgery - Abstract
Infantile nystagmus (IN) describes a regular, repetitive movement of the eyes. A characteristic feature of each cycle of the IN eye movement waveform is a period in which the eyes are moving at minimal velocity. This so-called “foveation” period has long been considered the basis for the best vision in individuals with IN. In recent years, the technology for measuring eye movements has improved considerably, but there remains the challenge of calibrating the direction of gaze in tracking systems when the eyes are continuously moving. Identifying portions of the nystagmus waveform suitable for calibration typically involves time-consuming manual selection of the foveation periods from the eye trace. Without an accurate calibration, the exact parameters of the waveform cannot be determined. In this study, we present an automated method for segmenting IN waveforms with the purpose of determining the foveation positions to be used for calibration of an eye tracker. On average, the “point of regard” was found to be within 0.21° of that determined by hand-marking by an expert observer. This method enables rapid clinical quantification of waveforms and the possibility of gaze-contingent research paradigms being performed with this patient group.
- Published
- 2019
35. Mitigating Cybersickness in Virtual Reality Systems through Foveated Depth-of-Field Blur
- Author
-
Manuela Chessa, Razeen Hussain, and Fabio Solari
- Subjects
Computer science ,Motion Sickness ,Headset ,Emotions ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Stereoscopy ,TP1-1185 ,02 engineering and technology ,Virtual reality ,Foveated imaging ,Biochemistry ,Article ,Analytical Chemistry ,law.invention ,Spatial blur ,03 medical and health sciences ,User-Computer Interface ,0302 clinical medicine ,law ,Human–computer interaction ,Eye-tracker ,Surveys and Questionnaires ,Cycbersickness ,Depth-of-field ,Foveation ,Gaze-contingent ,Shader ,Humans ,Virtual Reality ,0202 electrical engineering, electronic engineering, information engineering ,Electrical and Electronic Engineering ,Instrumentation ,Chemical technology ,020207 software engineering ,Gaze ,Atomic and Molecular Physics, and Optics ,Mixed reality ,Simulator sickness ,Eye tracking ,030217 neurology & neurosurgery - Abstract
Cybersickness is one of the major roadblocks in the widespread adoption of mixed reality devices. Prolonged exposure to these devices, especially virtual reality devices, can cause users to feel discomfort and nausea, spoiling the immersive experience. Incorporating spatial blur in stereoscopic 3D stimuli has shown to reduce cybersickness. In this paper, we develop a technique to incorporate spatial blur in VR systems inspired by the human physiological system. The technique makes use of concepts from foveated imaging and depth-of-field. The developed technique can be applied to any eye tracker equipped VR system as a post-processing step to provide an artifact-free scene. We verify the usefulness of the proposed system by conducting a user study on cybersickness evaluation. We used a custom-built rollercoaster VR environment developed in Unity and an HTC Vive Pro Eye headset to interact with the user. A Simulator Sickness Questionnaire was used to measure the induced sickness while gaze and heart rate data were recorded for quantitative analysis. The experimental analysis highlighted the aptness of our foveated depth-of-field effect in reducing cybersickness in virtual environments by reducing the sickness scores by approximately 66%.
- Published
- 2021
36. Foveated Compressed Sensing.
- Author
-
Ciocoiu, Iulian
- Subjects
- *
COMPRESSED sensing , *IMAGE reconstruction , *DISCRETE wavelet transforms , *DIGITAL image processing , *ORTHOGONAL matching pursuit - Abstract
Combining the principles behind Compressed Sensing theory with a wavelet-based implementation of a foveation operator inspired by the human visual system yields significant compression performances on both 1D and 2D signals. The solution provides spatially variable quality of the reconstructed information, enabling better approximation on specific regions of interest. Four distinct algorithms are compared in terms of reconstruction error and compression ratio on a set of ECG records and natural images. [ABSTRACT FROM AUTHOR]
- Published
- 2015
- Full Text
- View/download PDF
37. Spline interpolation to evaluate foveation parameters in Congenital Nystagmus recordings.
- Author
-
Cesarelli, Mario, Bifulco, Paolo, Romano, Maria, Fratini, Antonio, D'Addio, Giovanni, and Clemente, Fabrizio
- Abstract
Infantile Nystagmus Syndrome, or Congenital Nystagmus, is an ocular-motor disorder characterized by involuntary, conjugated and bilateral to and fro ocular oscillations. Good visual acuity in congenital nystagmus can be achieved during the foveation periods in which eye velocity slows down while the target image crosses the fovea. Visual acuity was found to be mainly dependent on the duration of the foveation periods. In this work a new approach is proposed for estimation of foveation parameters: a cubic spline interpolation of the nystagmus recording before localizing the start point of foveation window and to estimate its duration. The performances of the proposed algorithm were assessed in comparison with a previously developed algorithm, used here as gold standard. The obtained results suggest that the spline interpolation could be a useful tool to filter the eye movement recordings before applying an algorithm to estimate the foveation window parameters. [ABSTRACT FROM PUBLISHER]
- Published
- 2013
- Full Text
- View/download PDF
38. Modelling Foveated Depth-of-field Blur for Improving Depth Perception in Virtual Reality
- Author
-
Manuela Chessa, Fabio Solari, and Razeen Hussain
- Subjects
depth-of-field ,gaze contingency ,Computer science ,depth perception ,foveation ,image processing ,shader ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,02 engineering and technology ,Virtual reality ,050105 experimental psychology ,0202 electrical engineering, electronic engineering, information engineering ,0501 psychology and cognitive sciences ,Computer vision ,Depth of field ,Image resolution ,business.industry ,05 social sciences ,020207 software engineering ,Object (computer science) ,Visualization ,Human visual system model ,Artificial intelligence ,Depth perception ,business - Abstract
This paper presents a technique to incorporate spatial blur effects in virtual reality devices. The considered spatial blur is based on foveation and defocus blur: concepts inspired by the human visual system. The proposed technique can be applied to any head-mounted display as a post-processing step. Our foveated depth-of-field method removes intensity leakage artifacts in the transitory regions and works in real-time. We verify the usefulness of our technique by conducting a pilot study on depth perception in virtual environments. In the conducted user study, systems integrated with our blur effect provided a better estimation of object depth in the peripheral regions.
- Published
- 2020
39. Improved foveation- and saliency-based visual attention prediction under a quality assessment task.
- Author
-
Gide, Milind S. and Karam, Lina J.
- Abstract
Image quality assessment is one application out of many that can be aided by the use of computational saliency models. Existing visual saliency models have not been extensively tested under a quality assessment context. Also, these models are typically geared towards predicting saliency in non-distorted images. Recent work has also focussed on mimicking the human visual system in order to predict fixation points from saliency maps. One such technique (GAFFE) that uses foveation has been found to perform well for non-distorted images. This work extends the foveation framework by integrating it with saliency maps from well known saliency models. The performance of the foveated saliency models is evaluated based on a comparison with human ground-truth eye-tracking data. For comparison, the performance of the original non-foveated saliency predictions is also presented. It is shown that the integration of saliency models with a foveation based fixation finding framework significantly improves the prediction performance of existing saliency models over different distortion types. It is also found that the information maximization based saliency maps perform the best consistently over different distortion types and levels under this foveation based framework. [ABSTRACT FROM PUBLISHER]
- Published
- 2012
- Full Text
- View/download PDF
40. Subjective and objective video quality assessment of foveated videos in virtual reality
- Author
-
Jin, Yize
- Subjects
- Virtual reality, Foveation, Video quality assessment, Image processing, Video processing
- Abstract
Virtural Reality (VR) has experienced a substantial growth in popularity, due to recent advancements in consumer head-mounted displays (HMDs) and associated computing hardware technologies. In VR, the requirements of much higher resolution and smooth viewing experiences under rapid and often real-time changes in viewing direction, leads to significant challenges in compression and communication. To reduce the stresses of very high bandwidth consumption, the concept of foveated video compression is being accorded renewed interest. By exploiting the space-variant property of retinal visual acuity, foveation has the potential to substantially reduce video resolution in the visual periphery, with hardly noticeable perceptual quality degradations. Accordingly, foveated image / video quality predictors are also becoming increasingly important, as a practical way to monitor and control future foveated compression algorithms. Towards advancing the development of foveated compression / streaming algorithms, we first constructed 2D and (stereoscopic) 3D VR databases of foveated / compressed videos, called LIVE-Facebook Technologies Foveation-Compressed Virtual Reality (LIVE-FBT-FCVR) databases, and conducted a human study of perceptual quality on each database. Then we devised a no-reference (NR) foveated video quality assessment model, called FOVQA, which is based on new models of space-variant natural scene statistics (NSS) and natural video statistics (NVS). We find that FOVQA achieves state-of-the-art (SOTA) performance on the new 2D LIVE-FBT-FCVR database, as compared with other leading FVQA / VQA models. We also designed a full reference (FR) model called spatial-temporal foveated entropic differencing (STFED), which for the first time combines natural scene statistics with a contrast sensitivity function. The STFED model also achieved SOTA performance on the 2D LIVE-FBT-FCVR database in the category of FR models.
- Published
- 2022
41. Rapid shape detection signals in area V4.
- Author
-
Weiner, Katherine F. and Ghose, Geoffrey M.
- Subjects
BLINDFOLDS ,DEPTH perception ,VISION disorders ,EUKARYOTES ,CYTOLOGY - Abstract
Vision in foveate animals is an active process that requires rapid and constant decision-making. For example, when a new object appears in the visual field, we can quickly decide to inspect it by directing our eyes to the object's location. We studied the contribution of primate area V4 to these types of rapid foveation decisions. Animals performed a reaction time task that required them to report when any shape appeared within a peripherally-located noisy stimulus by making a saccade to the stimulus location. We found that about half of the randomly sampled V4 neurons not only rapidly and precisely represented the appearance of this shape, but they were also predictive of the animal's saccades. A neuron's ability to predict the animal's saccades was not related to the specificity with which the cell represented a single type of shape but rather to its ability to signal whether any shape was present. This relationship between sensory sensitivity and behavioral predictiveness was not due to global effects such as alertness, as it was equally likely to be observed for cells with increases and decreases in firing rate. Careful analysis of the timescales of reliability in these neurons implies that they reflect both feed forward and feedback shape detecting processes. In approximately 7% of our recorded sample, individual neurons were able to predict both the delay and precision of the animal's shape detection performance. This suggests that a subset of V4 neurons may have been directly and causally contributing to task performance and that area V4 likely plays a critical role in guiding rapid, form-based foveation decisions. [ABSTRACT FROM AUTHOR]
- Published
- 2014
- Full Text
- View/download PDF
42. 3D Visual Discomfort Prediction: Vergence , Foveation, and the Physiological Optics of Accommodation.
- Author
-
Park, Jincheol, Lee, Sanghoon, and Bovik, Alan Conrad
- Abstract
To achieve clear binocular vision, neural processes that accomplish accommodation and vergence are performed via two collaborative, cross-coupled processes: accommodation-vergence (AV) and vergence-accommodation (VA). However, when people watch stereo images on stereoscopic displays, normal neural functioning may be disturbed owing to anomalies of the cross-link gains. These anomalies are likely the main cause of visual discomfort experienced when viewing stereo images, and are called Accommodation-Vergence Mismatches (AVM). Moreover, the absence of any useful accommodation depth cues when viewing 3D content on a flat panel (planar) display induces anomalous demands on binocular fusion, resulting in possible additional visual discomfort. Most prior efforts in this direction have focused on predicting anomalies in the AV cross-link using measurements on a computed disparity map. We further these contributions by developing a model that accounts for both accommodation and vergence, resulting in a new visual discomfort prediction algorithm dubbed the 3D-AVM Predictor. The 3D-AVM model and algorithm make use of a new concept we call local 3D bandwidth (BW) which is defined in terms of the physiological optics of binocular vision and foveation. The 3D-AVM Predictor accounts for anomalous motor responses of both accommodation and vergence, yielding predictive power that is statistically superior to prior models that rely on a computed disparity distribution only. [ABSTRACT FROM PUBLISHER]
- Published
- 2014
- Full Text
- View/download PDF
43. A Retina-Based Perceptually Lossless Limit and a Gaussian Foveation Scheme With Loss Control.
- Author
-
Targino da Costa, Andre Luiz N. and Do, Minh N.
- Abstract
In this paper, we propose a new limit that promises theoretically achievable data reduction ratios up to approximately 9:1 with no perceptual loss in typical scenarios. Also, we introduce a novel Gaussian foveation scheme that provides experimentally achievable gains up to approximately 2 times the compression ratio of typical compression schemes with less perceptual loss than in typical transmissions. Both the proposed limit and foveation scheme shares the same background material: a model of image projection onto the retina; a model of cones distribution; and, subsequently, a proposed pointwise retina-based constraint called pixel efficiency. Quantitatively, the lattermost is globally processed to reveal the perceptual efficiency of a display. Analytical results indicate that in general the perceptual efficiency of displays are low for typical image sizes and viewing distances. Qualitatively, the pixel efficiency is used together with a lossy parameter to locally control the spatial resolution of a foveated image. Practical results show that proper use of the lossy parameter in the foveation filtering can increase the subjective quality of images. [ABSTRACT FROM PUBLISHER]
- Published
- 2014
- Full Text
- View/download PDF
44. Application-Oriented Retinal Image Models for Computer Vision
- Author
-
Rodolfo Azevedo, Ricardo da Silva Torres, Lin Tzy Li, Ewerton Silva, Siome Goldenstein, José Eduardo S. Vianna, and Allan Pinto
- Subjects
low-power ,Letter ,Computer science ,space-variant computer vision ,02 engineering and technology ,lcsh:Chemical technology ,Biochemistry ,Analytical Chemistry ,03 medical and health sciences ,Software ,energy consumption ,0202 electrical engineering, electronic engineering, information engineering ,Computer vision ,lcsh:TP1-1185 ,Electrical and Electronic Engineering ,Image sensor ,Face detection ,Instrumentation ,030304 developmental biology ,0303 health sciences ,business.industry ,retinal image model ,Energy consumption ,Atomic and Molecular Physics, and Optics ,Retinal image ,foveation ,020201 artificial intelligence & image processing ,Artificial intelligence ,business ,Energy (signal processing) - Abstract
Energy and storage restrictions are relevant variables that software applications should be concerned about when running in low-power environments. In particular, computer vision (CV) applications exemplify well that concern, since conventional uniform image sensors typically capture large amounts of data to be further handled by the appropriate CV algorithms. Moreover, much of the acquired data are often redundant and outside of the application’s interest, which leads to unnecessary processing and energy spending. In the literature, techniques for sensing and re-sampling images in non-uniform fashions have emerged to cope with these problems. In this study, we propose Application-Oriented Retinal Image Models that define a space-variant configuration of uniform images and contemplate requirements of energy consumption and storage footprints for CV applications. We hypothesize that our models might decrease energy consumption in CV tasks. Moreover, we show how to create the models and validate their use in a face detection/recognition application, evidencing the compromise between storage, energy, and accuracy. This is an open access article distributed under the Creative Commons Attribution License which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited
- Published
- 2020
45. WHAT THE EYE DID NOT SEE - A FUSION APPROACH TO IMAGE CODING.
- Author
-
ALSAM, ALI, RIVERTZ, HANS JAKOB, and SHARMA, PUNEET
- Subjects
- *
IMAGE compression , *RETINA , *IMAGE analysis , *PROBLEM solving , *MEDICAL artifacts , *PREDICTION models - Abstract
The concentration of the cones and ganglion cells is much higher in the fovea than the rest of the retina. This non-uniform sampling results in a retinal image that is sharp at the fixation point, where a person is looking, and blurred away from it. This difference between the sampling rates at the different spatial locations presents us with the question of whether we can employ this biological characteristic to achieve better image compression. This can be achieved by compressing an image less at the fixation point and more away from it. It is, however, known that the vision system employs more that one fixation to look at a single scene which presents us with the problem of combining images pertaining to the same scene but exhibiting different spatial contrasts. This article presents an algorithm to combine such a series of images by using image fusion in the gradient domain. The advantage of the algorithm is that unlike other algorithms that compress the image in the spatial domain our algorithm results in no artifacts. The algorithm is based on two steps, in the first we modify the gradients of an image based on a limited number of fixations and in the second we integrate the modified gradient. Results based on measured and predicted fixations verify our approach. [ABSTRACT FROM AUTHOR]
- Published
- 2013
- Full Text
- View/download PDF
46. Extent of foveal fixation with eye rotation in emmetropes and myopes.
- Author
-
Abdullah H, Verkicharla PK, and Ballae Ganeshrao S
- Subjects
- Fovea Centralis, Humans, Pilot Projects, Rotation, Young Adult, Emmetropia, Myopia
- Abstract
Purpose: This pilot study aimed to investigate the maximum extension of foveal fixation in the horizontal direction among young adults in both emmetropes and myopes., Methods: 35 participants (28 emmetropes and 7 myopes) were included. Participants with restricted extra-ocular mobility, end gaze nystagmus, and/or any other ocular pathology were excluded. Visual acuity (VA) was used as a surrogate measure of foveal fixation. VA was determined using a staircase procedure with 8 reversals. The average of the last 5 reversals was taken as the thresholds. VA acuity was measured at different gaze eccentricities along nasal and temporal visual field meridian. The eccentricity at which VA drops significantly was taken as the maximum extent of foveal fixation. A bilinear fit regression model was used to investigate the drop in the VA in both nasal and the temporal direction., Results: Emmetropes can foveate up to 35 ± 2° in nasal and 40 ± 3° in temporal direction and myopes can foveate up to 38° in both nasal and temporal directions. Paired student t-test showed a significant difference in foveal fixation between nasal and temporal direction for emmetropes (P<0.001) but not in myopes (P = 0.168). An unpaired student t-test showed a significant difference in foveal fixation for nasal direction between myopes and emmetropes (P = 0.01). However, no statistically significant difference was found in foveal fixation for temporal direction between myopes and emmetropes (P = 0.792)., Conclusion: The eye rotation does not necessarily match with the extent of foveal fixation at extreme eye rotation. Eyes can fixate only up to 35° nasally and 40° temporally maintaing their maximum visual acuity., Competing Interests: Conflicts of interest None The authors declare no competing interest., (Copyright © 2021 Spanish General Council of Optometry. Published by Elsevier España, S.L.U. All rights reserved.)
- Published
- 2022
- Full Text
- View/download PDF
47. Visual attention guided features selection with foveated images.
- Author
-
Beserra Gomes, Rafael, Motta de Carvalho, Bruno, and Marcos Garcia Gonçalves, Luiz
- Subjects
- *
VISUAL perception , *FEATURE selection , *IMAGE analysis , *AUTONOMOUS robots , *FEATURE extraction , *COMPUTER algorithms - Abstract
Abstract: Visual attention is a very important task in autonomous robotics, but, because of its complexity, the processing time required is significant. We propose an architecture for feature selection using foveated images that is guided by visual attention tasks and that reduces the processing time required to perform these tasks. Our system can be applied in bottom–up or top–down visual attention. The foveated model determines which scales are to be used on the feature extraction algorithm. The system is able to discard features that are not extremely necessary for the tasks, thus, reducing the processing time. If the fovea is correctly placed, then it is possible to reduce the processing time without compromising the quality of the tasks' outputs. The distance of the fovea from the object is also analyzed. If the visual system loses the tracking in top–down attention, basic strategies of fovea placement can be applied. Experiments have shown that it is possible to reduce up to 60% the processing time with this approach. To validate the method, we tested it with the feature algorithm known as speeded up robust features (SURF), one of the most efficient approaches for feature extraction. With the proposed architecture, we can accomplish real time requirements of robotics vision, mainly to be applied in autonomous robotics. [Copyright &y& Elsevier]
- Published
- 2013
- Full Text
- View/download PDF
48. GeoGCD: improved visual search via gaze-contingent display
- Author
-
Bektaş, Kenan, Çöltekin, Arzu, Krüger, Jens, Duchowski, Andrew T., Fabrikant, Sara Irina, Bektaş, Kenan, Çöltekin, Arzu, Krüger, Jens, Duchowski, Andrew T., and Fabrikant, Sara Irina
- Abstract
Gaze-Contingent Displays (GCDs) can improve visual search performance on large displays. GCDs, a Level Of Detail (LOD) management technique, discards redundant peripheral detail using various human visual perception models. Models of depth and contrast perception (e.g., depth-of-field and foveation) have often been studied to address the trade-off between the computational and perceptual benefits of GCDs. However, color perception models and combinations of multiple models have not received as much attention. In this paper, we present GeoGCD which uses individual contrast, color, and depth-perception models, and their combination to render scenes without perceptible latency. As proof-of-concept, we present a three-stage user evaluation built upon geographic image interpretation tasks. GeoGCD does not impair users' visual search performance or affect their display preferences. On the contrary, in some cases, it can significantly improve users' performance.
- Published
- 2019
49. Is it reliable to assess visual attention of drivers affected by Parkinson's disease from the backseat?—a simulator study.
- Author
-
Lee, Hoe C., Chee, Derserri Yanting, Selander, Helena, and Falkmer, Torbjorn
- Subjects
- *
AUTOMOBILE drivers' tests , *VISION testing , *EYE contact , *INTER-observer reliability , *PARKINSON'S disease , *DRIVERS' licenses , *HEALTH of automobile drivers , *OCCUPATIONAL therapy - Abstract
Background: Current methods of determining licence retainment or cancellation is through on-road driving tests. Previous research has shown that occupational therapists frequently assess drivers' visual attention while sitting in the back seat on the opposite side of the driver. Since the eyes of the driver are not always visible, assessment by eye contact becomes problematic. Such procedural drawbacks may challenge validity and reliability of the visual attention assessments. In terms of correctly classified attention, the aim of the study was to establish the accuracy and the inter-rater reliability of driving assessments of visual attention from the back seat. Furthermore, by establishing eye contact between the assessor and the driver through an additional mirror on the wind screen, the present study aimed to establish how much such an intervention would enhance the accuracy of the visual attention assessment. Methods: Two drivers with Parkinson's disease (PD) and six control drivers drove a fixed route in a driving simulator while wearing a head mounted eye tracker. The eye tracker data showed where the foveal visual attention actually was directed. These data were time stamped and compared with the simultaneous manual scoring of the visual attention of the drivers. In four of the drivers, one with Parkinson's disease, a mirror on the windscreen was set up to arrange for eye contact between the driver and the assessor. Interrater reliability was performed with one of the Parkinson drivers driving, but without the mirror. Results: Without mirror, the overall accuracy was 56% when assessing the three control drivers and with mirror 83%. However, for the PD driver without mirror the accuracy was 94%, whereas for the PD driver with a mirror the accuracy was 90%. With respect to the inter-rater reliability, a 73% agreement was found. Conclusion: If the final outcome of a driving assessment is dependent on the subcategory of a protocol assessing visual attention, we suggest the use of an additional mirror to establish eye contact between the assessor and the driver. The clinicians' observations on-road should not be a standalone assessment in driving assessments. Instead, eye trackers should be employed for further analyses and correlation in cases where there is doubt about a driver's attention. [ABSTRACT FROM AUTHOR]
- Published
- 2012
- Full Text
- View/download PDF
50. The mechanism of oscillopsia and its suppression.
- Author
-
Dell'Osso, L.F.
- Subjects
- *
VISION disorders , *NYSTAGMUS , *INFANT diseases , *MOTION perception (Vision) , *HYPOTHESIS , *DATA analysis , *VISUAL perception - Abstract
We studied the mechanisms of oscillopsia suppression in subjects with infantile nystagmus syndrome, fusion maldevelopment nystagmus syndrome, and acquired nystagmus (AN). Hypothetical possibilities for perceptual stability were the following: (1) epochs of clear and stable vision during foveation periods of nystagmus waveforms; (2) cancellation by efference copy of motor output; (3) a combination of the effects of both foveation-period stability and efference-copy cancellation; or (4) elevated motion-detection threshold and vision suppression. Observations, studies, and models of oscillopsia suppression allowed comparison of these possibilities. Data from individual subjects supported some of the putative hypotheses. However, only one hypothesis remained viable that could explain how all subjects maintained perceptual stability despite their different nystagmus types, waveforms, and variability. Robust suppression of oscillopsia was only possible using efference-copy feedback of the motor output containing these specific nystagmus signals to cancel that motion from the retinal error signals. In cases of AN, where oscillopsia could not be suppressed, the deficit was postulated to interfere with or lie outside of this efference-copy feedback loop. [ABSTRACT FROM AUTHOR]
- Published
- 2011
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.