1,453 results
Search Results
2. Foreword to the special section on best papers of the Eurographics 2022 Education Papers Program.
- Author
-
Paquette, Eric and Bourdin, Jean-Jacques
- Subjects
- *
COMPUTERS in education , *VIRTUAL reality , *COMPUTER graphics - Abstract
[Display omitted] • This special section includes papers on computer graphics education. • These papers are the best papers from the Eurographics 2022 Education Papers Program. • Specific topics include ray-tracing, the Vulkan API, and virtual reality. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
3. A photogrammetry-based verification of assumptions applied in the interpretation of paper architecture
- Author
-
Shih, Naai-Jung and Tsai, Yu-Tun
- Published
- 2002
- Full Text
- View/download PDF
4. GRSI Best Paper Award
- Published
- 2021
- Full Text
- View/download PDF
5. The LivePaper system: augmenting paper on an enhanced tabletop
- Author
-
Robinson, John A and Robertson, Charles
- Published
- 2001
- Full Text
- View/download PDF
6. Diffusion rendering of black ink paintings using new paper and ink models
- Author
-
Lee, Jintae
- Published
- 2001
- Full Text
- View/download PDF
7. <atl>A photogrammetry-based verification of assumptions applied in the interpretation of paper architecture
- Author
-
Shih, Naai-Jung and Tsai, Yu-Tun
- Subjects
- *
PHOTOGRAMMETRY , *COMPUTER-aided design , *COMPUTER-generated imagery - Abstract
This research investigated the space compositions of Chernikhov’s 101 Architectural Fantasies via computer-aided simulation to interpret the relationships between architectural components and spatial organization. An algorithmic approach and a perception approach were tested. Traditional analysis emphasized the simulation of corresponding objects by perspective deconstruction methods, which might not be able to show the exactly correct spatial relationship between objects. This research adopted photogrammetry to investigate the non-orthogonal spatial construction of 3D objects in 2D pictures. Research results showed that the algorithmic approach may derive different degrees of angles of parallel or intersected objects, and that observers tend to be misled by the effect of “orthogonal assumption” in terms of their own visual experiences. This finding revealed that Chernikhov had created unreasonable descriptions of space. This result was verified by the existence of false parallel and orthogonal relationships between drawn building parts. Three tests were conducted. Observers used a reverse verification process to analyze three-dimensional objects re-built in simulation. The verification mirrored a two-way construction relationship between 2D perspective and 3D models. [Copyright &y& Elsevier]
- Published
- 2002
- Full Text
- View/download PDF
8. GRSI Best Paper Award 2021.
- Subjects
- *
AWARDS - Published
- 2022
- Full Text
- View/download PDF
9. Freeform digital ink annotations in electronic documents: A systematic mapping study.
- Author
-
Sutherland, Craig J., Luxton-Reilly, Andrew, and Plimmer, Beryl
- Subjects
- *
ELECTRONIC paper , *ELECTRONIC records , *ANNOTATIONS , *RAPID prototyping , *INFORMATION display systems - Abstract
A variety of different approaches have been used to add digital ink annotations to text-based documents. While the majority of research in this field has focused on annotation support for static documents, a small number of studies have investigated support for documents in which the underlying content is changed. Although the approaches used to annotate static documents have been relatively successful, the annotation of dynamic text documents poses significant challenges which remain largely unsolved. However, it is difficult to clearly identify the successful techniques and the remaining challenges since there has not yet been a comprehensive review of digital ink annotation research. This paper reports the results of a systematic mapping study of existing work, and presents a taxonomy categorizing digital ink annotation research. [ABSTRACT FROM AUTHOR]
- Published
- 2016
- Full Text
- View/download PDF
10. COMPUTERS \amp GRAPHICS BEST PAPER AWARD
- Published
- 2006
- Full Text
- View/download PDF
11. Computers & Graphics best paper award (2004)
- Published
- 2005
- Full Text
- View/download PDF
12. Computer ; Graphics best paper award (2003)
- Published
- 2004
- Full Text
- View/download PDF
13. Computers & graphics best paper award 2002
- Published
- 2003
- Full Text
- View/download PDF
14. Best paper award 2001
- Published
- 2002
- Full Text
- View/download PDF
15. Call for Papers
- Published
- 2009
- Full Text
- View/download PDF
16. Call for Papers: SMI'10
- Published
- 2009
- Full Text
- View/download PDF
17. Call for papers
- Published
- 2006
- Full Text
- View/download PDF
18. Call for papers: Ankundigung EG2006
- Published
- 2005
- Full Text
- View/download PDF
19. Call for papers: Special Issue on "Mesh Analysis"
- Published
- 2005
- Full Text
- View/download PDF
20. Call for Papers: WSCG 2002
- Published
- 2001
- Full Text
- View/download PDF
21. Call for Papers
- Published
- 2001
- Full Text
- View/download PDF
22. COMPUTERS & GRAPHICS BEST PAPER AWARDS
- Published
- 2008
- Full Text
- View/download PDF
23. Call for papers: special issue of the Elsevier journal "Computers and Graphics", natural phenomena simulation
- Published
- 2005
- Full Text
- View/download PDF
24. First WSCG 2004 Call for Papers
- Published
- 2003
- Full Text
- View/download PDF
25. Computing and analyzing decision boundaries from shortest path maps.
- Author
-
Sharma, Ritesh and Kallmann, Marcelo
- Subjects
- *
CIVILIAN evacuation , *SCALAR field theory , *EMERGENCY management , *TOPOLOGICAL fields , *DATA visualization - Abstract
This paper proposes a methodology for computing, visualizing, and analyzing critical decision boundaries for the selection of shortest paths in a given environment. Decision boundaries are defined as the points in a map from which two or more different shortest paths exist towards a destination. This paper introduces the problem of visualizing their evolution, taking into account moving obstacles, moving goals, and as well multiple goals. The proposed visualizations enable analyzing which paths should be taken and at which departure times, such that a destination can be reached by the shortest possible path when taking into account a moving target or time-varying areas to be avoided. The proposed techniques are also applied to the analysis and improvement of exit placement in a given environment, in order to improve the evacuation flow in emergency situations. [Display omitted] • This research presents a unique method for detecting decision boundaries in a given environment, based on the analysis of the generator points of the Shortest Path Map (SPM) rather than employing traditional scalar field topological methods relying on cell neighborhood information which can be affected by the representation resolution. • The proposed approach introduces tools and techniques to visualize the evolution of decision boundaries when considering dynamically-changing obstacles and targets, and to design exit placement to equalize the escape flow distribution. • This novel approach supports decision-making applications related to navigation and environment modeling in emergency evacuation planning. • By analyzing and visualizing SPM decision boundaries, the lengths of globally-optimal Euclidean shortest paths are taken into account, instead of grid-based accumulated distances used in other approaches. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
26. Visualising geospatial time series datasets in realtime with the Digital Earth Viewer.
- Author
-
Buck, Valentin, Stäbler, Flemming, Mohrmann, Jochen, González, Everardo, and Greinert, Jens
- Subjects
- *
TIME series analysis , *UNDERWATER exploration , *INTERNET servers , *CONFERENCE papers , *VISUALIZATION , *GEOSPATIAL data - Abstract
A comprehensive study of the Earth System and its different environments requires understanding of multi-dimensional data acquired with a multitude of different sensors or produced by various models. Here we present a component-wise scalable web-based framework for simultaneous visualisation of multiple data sources. It helps contextualise mixed observation and simulation data in time and space. This work is an extended version of the conference paper (Buck et al., 2021). [Display omitted] • Open-source (EUPL) hybrid application for realtime visualisation of 4D geoscientific data. • Split into native server and web client to utilize the strengths of both platforms. • Desktop builds are released for Windows, Linux, and MacOS. • Viewer used on expedition cruises to plan underwater exploration missions • Presentation and data validation capabilities used by GLODAP [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
27. Computers & Graphics Most Downloaded Paper Award 2011
- Published
- 2013
- Full Text
- View/download PDF
28. Computers & Graphics Most Downloaded Paper Award 2010
- Published
- 2011
- Full Text
- View/download PDF
29. Computers & Graphics Best Paper Award 2009
- Published
- 2010
- Full Text
- View/download PDF
30. COMPUTERS & GRAPHICS BEST PAPER AWARD
- Published
- 2009
- Full Text
- View/download PDF
31. Call for papers: special issue on pervasive computing and ambient intelligence
- Published
- 2004
- Full Text
- View/download PDF
32. Call for papers for Volume 27 Number 3
- Published
- 2003
- Full Text
- View/download PDF
33. Image deraining based on dual-channel component decomposition.
- Author
-
Lin, Xiao, Xu, Duojiu, Tan, Peiwen, Ma, Lizhuang, and Wang, Zhi-Jie
- Subjects
- *
IMAGE reconstruction , *IMAGE processing , *VISIBILITY - Abstract
Image deraining aims to remove rain streaks from images and reduce information loss in outdoor images caused by rain. As a fundamental task in image processing, image deraining not only enhances the visibility of images but also provides necessary image restoration for advanced vision tasks. Existing image deraining models mostly train end-to-end models by minimizing the similarity between the output image of the model and the rain-free ground truth. Although these methods have achieved significant results, they often perform poorly in the face of dense and changing rain streak scenes. In this paper, we propose a novel method, called D ual-Channel C omponent D ecomposition Net work (DCD-Net). The basic idea of DCD-Net is to leverage the separability prior of rainy images, treats the rain-free background layer and the rain streak mask layer as two parallel component extraction tasks. To this end, it builds a dual-branch parallel networks that extract the rain-free background image and decouple the reconstruction information of the rain streak mask, respectively. It finally applies a composite multi-level contrastive supervision to the output of the above dual-branch parallel network, thereby achieving rain streak removal. Extensive experiments on various datasets demonstrate that the proposed model outperforms existing methods in deraining dense rain streak images. [Display omitted] • This paper proposes an image deraining method, called Dual-Channel Component Decomposition Network (DCD-Net). • DCD-Net treats the rain-free background layer and the rain streak mask layer as two parallel component extraction tasks. • DCD-Net obtains competitive performance in deraining complex and dense rain streak images. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
34. [formula omitted]GAN: Importance Weight and Wavelet feature guided Image-to-Image translation under limited data.
- Author
-
Yang, Qiuxia, Pu, Yuanyuan, Zhao, Zhengpeng, Xu, Dan, and Li, Siqi
- Subjects
- *
GENERATIVE adversarial networks , *MACHINE translating - Abstract
Image-to-Image (I2I) translation methods based on generative adversarial networks (GANs) require large amounts of training data, without which they will suffer from over-fitting and train divergence, and trained models are sub-optimal. In addition, it would be very difficult for the model to synthesize high-frequency signals, deteriorating the synthesis quality. To address these, this paper proposes W 2 GAN, which mainly introduces the ideas of the Importance Weight and Wavelet transformation to achieve the I2I translation trained on limited-data. Concretely, this paper first alleviates the over-fitting and train divergence by the adversarial loss with importance weight, which aims to improve the influence of the high-quality generated images during the training generator, thus enhancing the generator to deceive the discriminator. Then, the high-frequency features of the wavelet transformation are applied to the decoder, and wavelet-AdaIN normalization is proposed to prevent deficiency of high-frequency information, which adaptively integrates high-frequency statistical characteristics from generated features and real image high-frequency information. Qualitative and quantitative results on the AFHQ and CelebA-HQ datasets demonstrate the merits of the W 2 GAN. Noticeably, this paper achieves state-of-the-art FID and KID on AFHQ and CelebA-HQ datasets. [Display omitted] • In this paper, the GANs are trained under limited data. This paper overcomes the problem of the discriminator over-fitting during the training, which leads to the model divergence and the results degradation, stabilizing the training process and achieving higher quality results. From the qualitative and quantitative perspectives, this paper has achieved competitive results in current research. • This paper introduces the learnable importance weight to the adversarial loss, which aims to hope the high-quality images produce higher influence during the training generator. It relieves the problems of the training diverge and over-fitting. • This paper proposes a Wavelet-AdaIN Normalization to learn the high-frequency features, which adaptively integrates high-frequency statistical characteristics from generated features and real image high-frequency information. It encourages the generator to produce precise high-frequency signals with fine details. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
35. LBARNet: Lightweight bilateral asymmetric residual network for real-time semantic segmentation.
- Author
-
Hu, Xuegang and Zhou, Baoman
- Subjects
- *
DATA mining , *COMPUTER vision , *IMAGE segmentation , *VISUAL fields , *PIXELS , *LEARNING modules - Abstract
Real-time semantic segmentation, as a key technique for scene understanding, has been an important research topic in the field of computer vision in recent years. However, existing models are unable to achieve good segmentation accuracy on mobile devices due to their huge computational overhead, which makes it difficult to meet actual industrial requirements. To address the problems faced by current semantic segmentation tasks, this paper proposes a lightweight bilateral asymmetric residual network (LBARNet) for real-time semantic segmentation. First, we propose the bilateral asymmetric residual (BAR) module. This module learns multi-scale feature representations with strong semantic information at different stages of the semantic information extraction branch, thus improving pixel classification performance. Secondly, the spatial information extraction (SIE) module is constructed in the spatial detail extraction branch to capture multi-level local features of the shallow network to compensate for the lost geometric information in the downsampling stage. At the same time, we design the attention mechanism perception (AMP) module in the jump connection part to enhance the contextual representation. Finally, we design the dual branch feature fusion (DBF) module to exploit the correspondence between higher-order features and lower-order features to fuse spatial and semantic information appropriately. The experimental results show that LBARNet, without any pre-training and pre-processing and using only 0.6M parameters, achieves 70.8% mloU and 67.2% mloU on the Cityscapes dataset and Camvid dataset, respectively. LBARNet maintain a high segmentation accuracy while using a smaller number of parameters compared to most existing state-of-the-art models. [Display omitted] • This paper proposes a Bilateral Asymmetric Residual (BAR) module, a Spatial Information Extraction (SIE) module, an Attention Mechanism Perception (AMP) module and a Dual Branch Feature Fusion (DBF) module. • A Lightweight Bilateral Asymmetric Residual Network (LBARNet) for real-time image semantic segmentation is proposed in this article. • The experimental results show that LBARNet achieves 70.8% and 67.2% segmentation accuracy on two challenging datasets (Cityscapes and CamVid) using only 0.6M parameters. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
36. Deep learning of curvature features for shape completion.
- Author
-
Hernández-Bautista, Marina and Melero, Francisco Javier
- Subjects
- *
DEEP learning , *CURVATURE , *GEOMETRIC surfaces , *SURFACE reconstruction , *INPAINTING , *PARAMETERIZATION , *SURFACE geometry - Abstract
The paper presents a novel solution to the issue of incomplete regions in 3D meshes obtained through digitization. Traditional methods for estimating the surface of missing geometry and topology often yield unrealistic outcomes for intricate surfaces. To overcome this limitation, the paper proposes a neural network-based approach that generates points in areas where geometric information is lacking. The method employs 2D inpainting techniques on color images obtained from the original mesh parameterization and curvature values. The network used in this approach can reconstruct the curvature image, which then serves as a reference for generating a polygonal surface that closely resembles the predicted one. The paper's experiments show that the proposed method effectively fills complex holes in 3D surfaces with a high degree of naturalness and detail. This paper improves the previous work in terms of a more in-depth explanation of the different stages of the approach as well as an extended results section with exhaustive experiments. [Display omitted] • We perform 3D surface reconstructions using generative inpainting techniques. • 2D representation of a 3D surface geometry based on its curvature. • Application of a general purpose neural network for inpainting. • Our approach does not require dataset nor training time. • Results outperform state-of-the-art quality and naturalness of the reconstructions. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
37. A systematic review on open-set segmentation.
- Author
-
Nunes, Ian, Laranjeira, Camila, Oliveira, Hugo, and dos Santos, Jefersson A.
- Subjects
- *
COMPUTER vision , *REMOTE sensing , *RESEARCH personnel , *AUTONOMOUS vehicles , *VISUAL learning - Abstract
Open-set semantic segmentation remains yet a challenging task, not only due to the inherent challenges of pixel-wise classification but also the precise segmentation of categories not seen during training. The pursuit of that task is rapidly growing in the Computer Vision community, urging the need to organize the literature. In this paper, we extend our previous work by conducting a more comprehensive systematic mapping of the open-set segmentation literature between January 2001 and January 2023 and proposing a novel taxonomy. Our goal is to provide a broad understanding of current trends for the open-set semantic segmentation (OSS) task defined by existing approaches that may influence future methods. By characterizing methodologies in terms of open-set identification strategies, data inputs, and other relevant aspects, we present a structured view of how researchers are advancing in the field of open-set semantic segmentation. To the best of the authors' knowledge, this is the first systematic review of OSS methods. Moreover, we apply the proposed taxonomy to selected methods for open-set recognition, outlining important similarities and differences of such a closely related field. [Display omitted] • Systematic review of papers related to open-set semantic segmentation for the past 20 years. • The proposed taxonomy aims to organize the literature on open-set segmentation. • Seminal papers on open-set recognition are classified under the proposed taxonomy. • Applications like autonomous driving and remote sensing were found to commonly resort to the open set strategy. • Methods tackling open-world are becoming more commom; [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
38. LFPeers: Temporal similarity search and result exploration.
- Author
-
Sachdeva, Madhav, Burmeister, Jan, Kohlhammer, Jörn, and Bernard, Jürgen
- Subjects
- *
CONCEPT learning , *VISUAL analytics , *BEHAVIORAL assessment , *COVID-19 pandemic , *PEERS , *NATURAL gas prospecting - Abstract
In this paper, we introduce a general concept for the analysis of temporal and multivariate data and the system LFPeers that applies this concept to temporal similarity search and results exploration. The conceptual workflow divides the analysis in two phases: a search phase to find the most similar objects to a query object before a time point t 0 in the temporal data, and an exploration phase to analyze and contextualize this subset of objects after t 0. LFPeers enables users to search for peers through interactive similarity search and filtering, explore interesting behavior of this peer group, and learn from peers through the assessment of diverging behaviors. We present the conceptual workflow to learn from peers and the LFPeers system with novel interfaces for search and exploration in temporal and multivariate data. An earlier workshop publication for LFPeers included a usage scenario targeting epidemiologists and the public who want to learn from the Covid-19 pandemic and distinguish successful and ineffective measures. In this extended paper, we now show how our concept is generalized and applied by domain experts in two case studies, including a novel case on stocks data. Finally, we reflect on the new state of development and on the insights gained by the experts in the case studies on the search and exploration of temporal data to learn from peers. [Display omitted] • Conceptual framework for temporal and multimodal similarity search and exploration. • Visual Analytics system for cause–effect analysis on temporal data. • Interactive user-defined search and exploration of data objects by similarity. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
39. A framework for the efficient enhancement of non-uniform illumination underwater image using convolution neural network.
- Author
-
Zhang, Wenbo, Liu, Weidong, Li, Le, Jiao, Huifeng, Li, Yanli, Guo, Liwei, and Xu, Jingming
- Subjects
- *
CONVOLUTIONAL neural networks , *GENERATIVE adversarial networks , *LIGHT sources , *IMAGE enhancement (Imaging systems) , *COLOR in nature , *GAUSSIAN function - Abstract
In this paper, the non-uniform illumination enhancement problem of underwater images under the artificial light sources conditions is investigated based on Convolution Neural Network (CNN). First, we propose a trainable end-to-end enhancer called NUIENet, for enhancing the non-uniform illumination of underwater images. The proposed model consists of correction network and fusion layers. The correction network adopts the encoder–decoder structure with skip connections to enhance the features of different channels in the HSV domain, and then these enhanced features are fused by the fusion layers to obtain the desired high-quality images. Second, we built an underwater images dataset using Generative Adversarial Network (GAN) and Gaussian Function. Finally, both qualitative and quantitative experimental results show that the proposed method can produce better performance compared to other state-of-the-art enhancement methods on both real-word and synthetic underwater dataset. • This paper proposes a non-uniform illumination enhancer CNN-based which uses the encoder–decoder structure with skip connections to enhance the underwater images with NUI to the desired high-quality images. • To boost underwater imaging processing, we construct a dataset of the underwater image with NUI based on the GAN and Gaussian function which contains NUI images and their corresponding high-quality reference image. • Compared with other state-of-the-art NUIE methods, the proposed network achieves a nature color correction and superior or equivalent visibility improvement. [Display omitted] [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
40. Teaching the basics of computer graphics in virtual reality.
- Author
-
Heinemann, Birte, Görzen, Sergej, and Schroeder, Ulrik
- Subjects
- *
VIRTUAL reality , *COMPUTERS in education , *TECHNOLOGICAL innovations , *COMPUTER graphics , *SCHOOL environment , *TELEPORTATION - Abstract
New technology such as virtual reality can help computer graphics education, for example, by providing the opportunity to illustrate challenging 3D procedures. RePiX VR is a virtual reality tool for computer graphics education that focuses on teaching the core ideas of the rendering pipeline. This paper describes the development and two initial evaluations, which aimed to strengthen the usability, review requirements for different stakeholders, and build infrastructure for learning analytics and research. The integration of learning analytics raises the question of appropriate indicators to be approached through exploratory data analysis. In addition to learning analytics, the evaluation includes quantitative techniques to get insights about usability, and didactical feedback. This paper discusses advanced aspects of learning in VR and looks specifically at movement behavior. According to the evaluations, even learners without prior experience can utilize the VR tool to pick up the fundamentals of computer graphics. [Display omitted] • Evaluated educational VR environment for teaching Computer Graphics. • Teaching Computer Graphics in Virtual Reality is promising. • Learners have various movement and teleportation pattern and different interaction behavior. • A comparison of Desktop and VR users shows differences between groups, as well as a comparison of novices and experts. • The evaluation contains Multimodal Learning Analytics, Quantitative Feedback, and Usability aspects. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
41. DCU-NET: Self-supervised monocular depth estimation based on densely connected U-shaped convolutional neural networks.
- Author
-
Zheng, Qiumei, Yu, Tao, and Wang, Fenghua
- Subjects
- *
CONVOLUTIONAL neural networks , *MONOCULARS , *INFORMATION networks - Abstract
Depth estimation is crucial for scene understanding and downstream tasks, especially the self-supervised training methods showing great potential. The overall structure and local details of the scene are essential for improving the quality of depth estimation. The proposal of Monodepth2 has led to significant progress in self-supervised monocular depth estimation. However, Monodepth2 uses the most basic encoder–decoder architecture. The limited data flow information of the network leads to a large semantic gap between the encoder and the decoder, which reduces the accuracy of the network for fine-grained feature recognition. Monodepth2 adopts Resnet18 pre-trained on the Imagenet dataset as the encoder. This traditional convolutional pooling structure results in a loss of pixel information in the network at every scale. In order to solve this problem, this paper proposes an improved DepthNet. The network adopts Hrnet in semantic segmentation as the base encoder, which adopts an advanced multi-scale fusion method in the whole process, thus avoiding the loss of pixel information. An additional densely connected U-Net is employed at the decoder side to provide more information flow. Furthermore, the semantic gap between the encoder and decoder is reduced by adding different numbers of residual connections and channel attention on each layer. The network structure can be regarded as a collection of fully convolutional networks. Since the deep features of the network have a higher correlation with the vertical position, we add a spatial location attention module to the deep-level network to reduce this semantic gap. The approach performs significantly well on the KITTI dataset benchmark, with several performance criteria comparable to supervised monocular depth inference methods. • This work is a deep estimation network for scene reconstruction and scene understanding. This network redesigns the self-supervised monocular depth framework from an entirely new perspective. The network uses HrNet from the field of semantic segmentation as the base encoder, which employs a progressive multi-scale fusion approach throughout, thus avoiding the loss of pixel information. • An additional densely connected U-Net is used at the decoder side to provide further information flow. To reduce the semantic gap between codecs, we add a different number of residual connections and channel attention on each layer. The network is not trained with the help of other auxiliary networks, and the performance of the depth estimation is improved only by stimulating the network's potential. • This work achieves best-in-class accuracy in monocular depth estimation. When the model in this paper is used for 3D scene reconstruction, it can perform a complete recovery of the scene structure. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
42. BrightFormer: A transformer to brighten the image.
- Author
-
Wang, Yong, Li, Bo, and Yuan, Xinlin
- Subjects
- *
IMAGE intensifiers , *PIXELS - Abstract
Low-light image enhancement algorithms need recover overall information of images, including local details and global information. However, existing image enhancement methods mainly focus on local details or global information. Therefore, it is challenging to balance the two aspects at the same time. This paper proposes a local dual-branch network (BrightFormer) for image enhancement that combines convolutions and transformers as to solution. The salient features of this paper are: (1) convolution is adopted to refine high-frequency information so that local features are preserved and propagated throughout the network; (2) combining gated parameters with prior information on illumination (ill-map) in self-attention can not only improves the flexibility of feature expression but also extract global features more easily; (3) the obtained local details and global features are fused by spatial and channel attention in Feature equalization fusion unit (FEFU); (4) a Deep feedforward network (DFN) is utilized to encode the location information between adjacent pixels, and the GELU activation function is used to retain useful features and eliminate useless features with an attention-like mechanism. Experimental results show that BrightFormer achieves competitive performance on quantitative metrics and visual perception on the datasets such as LOL, MEF and LIME etc. [Display omitted] • Using cross-convolution can extract rich local features of the image. • Incorporates the gating mechanism and prior information on the illumination. • Different attention mechanisms are adopted for local and global features. • Use the GELU activation function as an attention mechanism. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
43. A novel isosurface segmentation method using common boundary tests.
- Author
-
Wang, Cuilan
- Subjects
- *
TEACHING aids , *VISUALIZATION , *GEOGRAPHIC boundaries - Abstract
Visualizing the isosurfaces that represent material boundaries is an important technique for understanding the features of interest in a scalar volumetric dataset. However, one isosurface may contain multiple types of boundaries, i.e., boundaries between different pairs of materials. In this paper, we present a novel isosurface segmentation method that aids in learning structural information of a dataset by separating different types of boundaries in one isosurface. This method uses common boundary tests to classify a point on the isosurface. The test determines whether a point on the isosurface that is at a boundary shared by both the isosurface and a reference isosurface. It uses a gradient-guided sampling approach and is based on material boundary properties. A new region growing algorithm is developed to improve the segmentation results. Our new method can also be used to segment an isosurface that passes through both material boundaries and the interior of a material. Two applications of the new method are also demonstrated in the paper. One is to render and segment section planes to enhance visualization. The other one is to obtain more accurate and meaningful isosurface statistics. • Segment an isosurface that contains multiple types of material boundaries. • Use region growing approach to improve the isosurface segmentation results. • Correctly segment the section planes to show the interior structure of the object. • Segment an isosurface that passes through both material boundaries and material interior. • Obtain more accurate isosurface statistics by averaging them over different portions of the segmented isosurface. [Display omitted] [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
44. Visualization of 3D forest fire spread based on the coupling of multiple weather factors.
- Author
-
Meng, Qingkuo, Huai, Yongjian, You, Jiawei, and Nie, Xiaoying
- Subjects
- *
FOREST fires , *WEATHER control , *WEATHER , *VISUALIZATION , *WIND speed , *RAINFALL - Abstract
Combustibles, topography, and weather factors are the three essential factors affecting forest fire behavior, and current forest fire spread models need to consider weather factors fully. This paper proposes a forest fire spread method based on environmental weather factors to present a visualized simulation of forest fire spread in the natural environment. Forest pyrolysis differs based on water content, so a single-tree pyrolysis model with temperature as its core has been constructed to describe the differences in forest pyrolysis during different seasons visually. In addition, based on the improved Huygens principle as the theoretical basis for forest fire spread, weather factors such as wind speed, wind direction, and precipitation controlled by weather are coupled with the forest fire spread process. And the forest fire spread in three-dimensional scenarios is simulated by considering environmental factors. The visualization of the forest fire extinguishing process caused by precipitation is realized. Finally, the interaction between rain and snow, terrain and trees is realized when precipitation affects the corresponding landscape and vegetation texture to enhance the realism of the constructed forest environment. In short, this paper proposes a forest fire spread method based on environmental weather factors, which intuitively expresses the influence of different weather factors on forest fire spread, thereby improving the immersive experience of the related senses and realizing realistic scene roaming. [Display omitted] • Based on the single tree pyrolysis model, differentiating forest burning. • Visualization of the influence of simulated weather factors on forest fire behavior. • Use texture mixing technology to construct 3D forest weather scene. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
45. LARFNet: Lightweight asymmetric refining fusion network for real-time semantic segmentation.
- Author
-
Hu, Xuegang and Gong, Juelin
- Subjects
- *
PROBLEM solving , *MULTICASTING (Computer networks) , *PYRAMIDS , *PIXELS - Abstract
In this paper, we propose a lightweight asymmetric refining fusion network (LARFNet) for real-time semantic segmentation to solve the problem that some existing models cannot achieve good segmentation accuracy with real-time inference speed in mobile devices due to the huge computational overhead. Specifically, LARFNet adopts an asymmetric encoder–decoder structure. The depth-wise separable asymmetric interaction module (DSAI module) is designed in the encoding process, which effectively extracted local and surrounding information under different receptive fields with optimized convolution in the condition of ensuring communication between channels. In the decoder, we design the bilateral pyramid pooling attention module (BPPA module) and the multi-stage refinement fusion module (MRF Module). The BPPA module is used to integrate the high-level output multi-scale context information. Based on spatial and channel attention mechanisms, the MRF module is proposed to refine the feature maps of different resolutions and guide the feature fusion. Experimental results show that LARFNet achieves 69.2% mIoU and 65.6% mIoU on Cityscapes and Camvid datasets at 127 FPS and 222 FPS respectively, only using a single NVIDIA GeForce GTX2080Ti GPU and 0.72M parameters without any pre-training or pre-processing. Compared with most of the existing state-of-the-art models, the proposed method realizes the efficient use of network parameters at a faster speed, reduces the number of network parameters, and still achieves the accuracy of good segmentation. [Display omitted] • In this paper, we propose a lightweight real-time semantic segmentation network which considers inference speed, number of model parameters and segmentation accuracy as a whole, named lightweight asymmetric refining fusion network (LARFNet). The network mainly consists of three modules. The DSAI module is used to extract different features. The BPPA module is used to provide pixel-level attention for features, and the MRF module is used to guide feature fusion after optimizing features. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
46. LSGRNet: Local Spatial Latent Geometric Relation Learning Network for 3D point cloud semantic segmentation
- Author
-
Luo, Liguo, Lu, Jian, Chen, Xiaogai, Zhang, Kaibing, and Zhou, Jian
- Published
- 2024
- Full Text
- View/download PDF
47. Immersive presentations of real-world medical equipment through interactive VR environment populated with the high-fidelity 3D model of mobile MRI unit
- Author
-
Tadeja, Sławomir Konrad, Bohné, Thomas, Godula, Kacper, Cybulski, Artur, and Woźniak, Magdalena Maria
- Published
- 2024
- Full Text
- View/download PDF
48. Frequency-aware network for low-light image enhancement.
- Author
-
Shang, Kai, Shao, Mingwen, Qiao, Yuanjian, and Liu, Huan
- Subjects
- *
IMAGE intensifiers , *IMAGE enhancement (Imaging systems) , *IMAGE reconstruction , *FREQUENCY-domain analysis - Abstract
Low-light images often suffer from severe visual degradation, affecting both human perception and high-level computer vision tasks. Most existing methods process images in the spatial domain, making it challenging to simultaneously improve brightness while suppressing noise. In this paper, we present a novel perspective to enhance images based on frequency domain characteristics. Specifically, we reveal that the low-frequency components are closely related to luminance and color, whereas the high-frequency components are not. Based on this observation, we propose the Frequency-aware Network (FaNet) for low-light image enhancement. By selectively adjusting low-frequency components, FaNet preserves more high-frequency details while achieving low-light image enhancement. Additionally, we employ a multi-scale framework and selective fusion for effective feature learning and image reconstruction. Experimental results demonstrate the superiority of the proposed method. [Display omitted] • We reveal that the luminance is closely related to low-frequency components. • We design a frequency-aware network to utilize frequency domain features. • A multi-scale framework and selective fusion is proposed for feature learning. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
49. Mixed reality human teleoperation with device-agnostic remote ultrasound: Communication and user interaction.
- Author
-
Black, David, Nogami, Mika, and Salcudean, Septimiu
- Subjects
- *
MIXED reality , *REMOTE control , *ULTRASONIC imaging , *TELECOMMUNICATION systems , *TELEROBOTICS , *HUMAN beings - Abstract
For many applications, remote guidance and telerobotics provide great advantages. For example, tele-ultrasound can bring much-needed expert healthcare to isolated communities. However, existing tele-guidance methods have serious limitations including either low precision for video conference-based systems, or high complexity and cost for telerobotics. A new concept called human teleoperation leverages mixed reality, haptics, and high-speed communication to provide tele-guidance that gives an expert nearly-direct remote control without requiring a robot. This paper provides an overview of the human teleoperation concept and its application to tele-ultrasound. The concept and its impact are discussed. A new approach to remote streaming and control of point-of-care ultrasound systems independent of their manufacturer is described, as is a high-speed communication system for the HoloLens 2 that is compatible with ResearchMode API sensor stream access. Details of these systems are shown in supplementary video demonstrations. Novel interaction methods enabled by HoloLens 2-based pose tracking are also introduced and tests of the communication and user interaction are presented. The results show continued improvement of the system compared to previous work in instrumentation, HCI, and communication. The system thus has good potential for tele-ultrasound, as well as possible other applications of human teleoperation including remote maintenance, inspection, and training. The remote ultrasound streaming and control application is made available open source. [Display omitted] • System improvements to human teleoperation demonstrate its feasibility. • Other devices such as the Nreal Light can be used for implementation. • New device-agnostic remote ultrasound streaming and control demonstrated. • HoloLens pose tracking enables human teleoperation with limited compute resources. • HoloLens-based communication system provides effective sensor data streaming. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
50. Voting-based patch sequence autoregression network for adaptive point cloud completion.
- Author
-
Wu, Hang and Miao, Yubin
- Subjects
- *
POINT cloud , *PETRI nets , *NETWORK performance - Abstract
Point cloud completion aims to estimate the whole shapes of objects from their partial scans, and one of the main obstacles that prevents current methods from being applied in real-world scenarios is the variety of structural losses in real-scanned objects, which can hardly be fully included and reflected by the training samples. In this paper, we introduce Patch Sequence Autoregression Network (PSA-Net), a learning-based method that can be trained without the partial point clouds in dataset and is inherently adaptable to input scans with different levels of shape incompleteness: It makes restoring the unseen parts of objects be equivalent to predicting the missing tokens in local patch embedding sequences, and such prediction can start from any initial states. Specifically, we first introduce a Sequential Patch AutoEncoder that reconstructs complete point clouds from quantized patch feature sequences. Second, we establish a Mixed Patch Autoregression pipeline that can flexibly infer the whole sequence from any number of known tokens at any positions. Third, we propose a Voting-Based Mapping module that makes input points softly vote for their possible related tokens in sequences based on their local areas, which transforms partial point clouds to masked sequences in test. Quantitative and qualitative evaluations on two synthetic and four real-world datasets illustrate the competitive performances of our network when comparing with existing approaches. [Display omitted] • A Sequential Patch AutoEncoder for shape generation from quantized feature sequence. • A Mixed Patch Autoregression pipeline for token prediction from any initial states. • A Voting-based Mapping module for transformation from partial shapes to sequences. • Competitive performances on two synthetic and four real-world datasets. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.