image stream
Recently Published Documents


TOTAL DOCUMENTS

28
(FIVE YEARS 9)

H-INDEX

5
(FIVE YEARS 0)

2021 ◽  
Vol 154 ◽  
pp. 107592
Author(s):  
Yifan Dong ◽  
Tangbin Xia ◽  
Dong Wang ◽  
Xiaolei Fang ◽  
Lifeng Xi
Keyword(s):  

PLoS ONE ◽  
2021 ◽  
Vol 16 (2) ◽  
pp. e0246336
Author(s):  
Håkan Wieslander ◽  
Carolina Wählby ◽  
Ida-Maria Sintorn

Microscopy imaging experiments generate vast amounts of data, and there is a high demand for smart acquisition and analysis methods. This is especially true for transmission electron microscopy (TEM) where terabytes of data are produced if imaging a full sample at high resolution, and analysis can take several hours. One way to tackle this issue is to collect a continuous stream of low resolution images whilst moving the sample under the microscope, and thereafter use this data to find the parts of the sample deemed most valuable for high-resolution imaging. However, such image streams are degraded by both motion blur and noise. Building on deep learning based approaches developed for deblurring videos of natural scenes we explore the opportunities and limitations of deblurring and denoising images captured from a fast image stream collected by a TEM microscope. We start from existing neural network architectures and make adjustments of convolution blocks and loss functions to better fit TEM data. We present deblurring results on two real datasets of images of kidney tissue and a calibration grid. Both datasets consist of low quality images from a fast image stream captured by moving the sample under the microscope, and the corresponding high quality images of the same region, captured after stopping the movement at each position to let all motion settle. We also explore the generalizability and overfitting on real and synthetically generated data. The quality of the restored images, evaluated both quantitatively and visually, show that using deep learning for image restoration of TEM live image streams has great potential but also comes with some limitations.


2020 ◽  
Vol 34 (05) ◽  
pp. 9185-9192
Author(s):  
Ruize Wang ◽  
Zhongyu Wei ◽  
Piji Li ◽  
Qi Zhang ◽  
Xuanjing Huang

Visual storytelling aims at generating a story from an image stream. Most existing methods tend to represent images directly with the extracted high-level features, which is not intuitive and difficult to interpret. We argue that translating each image into a graph-based semantic representation, i.e., scene graph, which explicitly encodes the objects and relationships detected within image, would benefit representing and describing images. To this end, we propose a novel graph-based architecture for visual storytelling by modeling the two-level relationships on scene graphs. In particular, on the within-image level, we employ a Graph Convolution Network (GCN) to enrich local fine-grained region representations of objects on scene graphs. To further model the interaction among images, on the cross-images level, a Temporal Convolution Network (TCN) is utilized to refine the region representations along the temporal dimension. Then the relation-aware representations are fed into the Gated Recurrent Unit (GRU) with attention mechanism for story generation. Experiments are conducted on the public visual storytelling dataset. Automatic and human evaluation results indicate that our method achieves state-of-the-art.


Author(s):  
Ruize Wang ◽  
Zhongyu Wei ◽  
Ying Cheng ◽  
Piji Li ◽  
Haijun Shan ◽  
...  

Author(s):  
Jay Watson

The early years of the “talkies,” which correspond with Faulkner’s surge into a fully realized literary modernism, brought technical problems that the cinema was slow to work out, especially the challenge of synchronizing the film soundtrack with its image stream to achieve verisimilitude. This technical crisis pointed to new creative opportunities for artists imaginative enough to seize the possibilities and extend montage effects across the visual and auditory realms. As sound film struggled through its growing pains, Faulkner experimented with new stylistic techniques of punctuation that introduced new discontinuities between speech and speaker, voice and subject, sound and source, into literary narration and onto the printed page, making his own unique contribution to his era’s aesthetic repertoire. This transmedial embrace of asynchrony went hand in hand with a new appreciation for the affective and thematic potential of silence, another aesthetic development that leaves its mark on Faulkner’s contemporaneous fictions.


Author(s):  
Pengcheng Yang ◽  
Fuli Luo ◽  
Peng Chen ◽  
Lei Li ◽  
Zhiyi Yin ◽  
...  

The visual storytelling (VST) task aims at generating a reasonable and coherent paragraph-level story with the image stream as input. Different from caption that is a direct and literal description of image content, the story in the VST task tends to contain plenty of imaginary concepts that do not appear in the image. This requires the AI agent to reason and associate with the imaginary concepts based on implicit commonsense knowledge to generate a reasonable story describing the image stream. Therefore, in this work, we present a commonsense-driven generative model, which aims to introduce crucial commonsense from the external knowledge base for visual storytelling. Our approach first extracts a set of candidate knowledge graphs from the knowledge base. Then, an elaborately designed vision-aware directional encoding schema is adopted to effectively integrate the most informative commonsense. Besides, we strive to maximize the semantic similarity within the output during decoding to enhance the coherence of the generated text. Results show that our approach can outperform the state-of-the-art systems by a large margin, which achieves a 29\% relative improvement of CIDEr score. With additional commonsense and semantic-relevance based objective, the generated stories are more diverse and coherent.


Sign in / Sign up

Export Citation Format

Share Document