Towards learning line descriptors from patches: a new paradigm and large-scale dataset

Author(s):  
Hongmin Liu ◽  
Yujie Liu ◽  
Miaomiao Fu ◽  
Yuhui Wei ◽  
Zhanqiang Huo ◽  
...  
2019 ◽  
Vol 11 (1) ◽  
pp. 1-18 ◽  
Author(s):  
V. Vijayakumar ◽  
Subramaniyaswamy Vairavasundaram ◽  
R. Logesh ◽  
A. Sivapathi

With the massive growth of the internet, a new paradigm of recommender systems (RS's) is introduced in various real time applications. In the research for better RS's, especially in the travel domain, the evolution of location-based social networks have helped RS's to understand the changing interests of users. In this article, the authors present a new travel RS employed on the mobile device to generate personalized travel planning comprising of multiple Point of Interests (POIs). The recommended personalized list of travel locations will be predicted by generating a heat map of already visited POIs and the highly relevant POIs will be selected for recommendation as destinations. To enhance the recommendation quality, this article exploits the temporal features for increased user visits. A personalized travel plan is recommended to the user based on the user selected POIs and the proposed travel RS is experimentally evaluated with the real-time large-scale dataset. The obtained results of the developed RS are found to be proficient by means of improved diversity and accuracy of generated recommendations.


Author(s):  
Yu Wu ◽  
Furu Wei ◽  
Shaohan Huang ◽  
Yunli Wang ◽  
Zhoujun Li ◽  
...  

Open domain response generation has achieved remarkable progress in recent years, but sometimes yields short and uninformative responses. We propose a new paradigm, prototypethen-edit for response generation, that first retrieves a prototype response from a pre-defined index and then edits the prototype response according to the differences between the prototype context and current context. Our motivation is that the retrieved prototype provides a good start-point for generation because it is grammatical and informative, and the post-editing process further improves the relevance and coherence of the prototype. In practice, we design a contextaware editing model that is built upon an encoder-decoder framework augmented with an editing vector. We first generate an edit vector by considering lexical differences between a prototype context and current context. After that, the edit vector and the prototype response representation are fed to a decoder to generate a new response. Experiment results on a large scale dataset demonstrate that our new paradigm significantly increases the relevance, diversity and originality of generation results, compared to traditional generative models. Furthermore, our model outperforms retrieval-based methods in terms of relevance and originality.


2014 ◽  
pp. 97-104 ◽  
Author(s):  
Electo Eduardo Silv Lora ◽  
Mateus Henrique Rocha ◽  
José Carlos Escobar Palacio ◽  
Osvaldo José Venturini ◽  
Maria Luiza Grillo Renó ◽  
...  

The aim of this paper is to discuss the major technological changes related to the implementation of large-scale cogeneration and biofuel production in the sugar and alcohol industry. The reduction of the process steam consumption, implementation of new alternatives in driving mills, the widespread practice of high steam parameters use in cogeneration facilities, the insertion of new technologies for biofuels production (hydrolysis and gasification), the energy conversion of sugarcane trash and vinasse, animal feed production, process integration and implementation of the biorefinery concept are considered. Another new paradigm consists in the wide spreading of sustainability studies of products and processes using the Life Cycle Assessment (LCA) and the implementation of sustainability indexes. Every approach to this issue has as an objective to increase the economic efficiency and the possibilities of the sugarcane as a main source of two basic raw materials: fibres and sugar. The paper briefly presents the concepts, indicators, state-of-the-art and perspectives of each of the referred issues.


Author(s):  
Jin Zhou ◽  
Qing Zhang ◽  
Jian-Hao Fan ◽  
Wei Sun ◽  
Wei-Shi Zheng

AbstractRecent image aesthetic assessment methods have achieved remarkable progress due to the emergence of deep convolutional neural networks (CNNs). However, these methods focus primarily on predicting generally perceived preference of an image, making them usually have limited practicability, since each user may have completely different preferences for the same image. To address this problem, this paper presents a novel approach for predicting personalized image aesthetics that fit an individual user’s personal taste. We achieve this in a coarse to fine manner, by joint regression and learning from pairwise rankings. Specifically, we first collect a small subset of personal images from a user and invite him/her to rank the preference of some randomly sampled image pairs. We then search for the K-nearest neighbors of the personal images within a large-scale dataset labeled with average human aesthetic scores, and use these images as well as the associated scores to train a generic aesthetic assessment model by CNN-based regression. Next, we fine-tune the generic model to accommodate the personal preference by training over the rankings with a pairwise hinge loss. Experiments demonstrate that our method can effectively learn personalized image aesthetic preferences, clearly outperforming state-of-the-art methods. Moreover, we show that the learned personalized image aesthetic benefits a wide variety of applications.


2021 ◽  
Vol 7 (3) ◽  
pp. 50
Author(s):  
Anselmo Ferreira ◽  
Ehsan Nowroozi ◽  
Mauro Barni

The possibility of carrying out a meaningful forensic analysis on printed and scanned images plays a major role in many applications. First of all, printed documents are often associated with criminal activities, such as terrorist plans, child pornography, and even fake packages. Additionally, printing and scanning can be used to hide the traces of image manipulation or the synthetic nature of images, since the artifacts commonly found in manipulated and synthetic images are gone after the images are printed and scanned. A problem hindering research in this area is the lack of large scale reference datasets to be used for algorithm development and benchmarking. Motivated by this issue, we present a new dataset composed of a large number of synthetic and natural printed face images. To highlight the difficulties associated with the analysis of the images of the dataset, we carried out an extensive set of experiments comparing several printer attribution methods. We also verified that state-of-the-art methods to distinguish natural and synthetic face images fail when applied to print and scanned images. We envision that the availability of the new dataset and the preliminary experiments we carried out will motivate and facilitate further research in this area.


Author(s):  
Anil S. Baslamisli ◽  
Partha Das ◽  
Hoang-An Le ◽  
Sezer Karaoglu ◽  
Theo Gevers

AbstractIn general, intrinsic image decomposition algorithms interpret shading as one unified component including all photometric effects. As shading transitions are generally smoother than reflectance (albedo) changes, these methods may fail in distinguishing strong photometric effects from reflectance variations. Therefore, in this paper, we propose to decompose the shading component into direct (illumination) and indirect shading (ambient light and shadows) subcomponents. The aim is to distinguish strong photometric effects from reflectance variations. An end-to-end deep convolutional neural network (ShadingNet) is proposed that operates in a fine-to-coarse manner with a specialized fusion and refinement unit exploiting the fine-grained shading model. It is designed to learn specific reflectance cues separated from specific photometric effects to analyze the disentanglement capability. A large-scale dataset of scene-level synthetic images of outdoor natural environments is provided with fine-grained intrinsic image ground-truths. Large scale experiments show that our approach using fine-grained shading decompositions outperforms state-of-the-art algorithms utilizing unified shading on NED, MPI Sintel, GTA V, IIW, MIT Intrinsic Images, 3DRMS and SRD datasets.


2021 ◽  
Vol 13 (5) ◽  
pp. 905
Author(s):  
Chuyi Wu ◽  
Feng Zhang ◽  
Junshi Xia ◽  
Yichen Xu ◽  
Guoqing Li ◽  
...  

The building damage status is vital to plan rescue and reconstruction after a disaster and is also hard to detect and judge its level. Most existing studies focus on binary classification, and the attention of the model is distracted. In this study, we proposed a Siamese neural network that can localize and classify damaged buildings at one time. The main parts of this network are a variety of attention U-Nets using different backbones. The attention mechanism enables the network to pay more attention to the effective features and channels, so as to reduce the impact of useless features. We train them using the xBD dataset, which is a large-scale dataset for the advancement of building damage assessment, and compare their result balanced F (F1) scores. The score demonstrates that the performance of SEresNeXt with an attention mechanism gives the best performance, with the F1 score reaching 0.787. To improve the accuracy, we fused the results and got the best overall F1 score of 0.792. To verify the transferability and robustness of the model, we selected the dataset on the Maxar Open Data Program of two recent disasters to investigate the performance. By visual comparison, the results show that our model is robust and transferable.


Sign in / Sign up

Export Citation Format

Share Document