Deep Neural Framework with Visual Attention and Global Context for Predicting Image Aesthetics

نویسندگان

چکیده

Computational inference of aesthetics has recently become a hot topic due to its usefulness in widely applications such as evaluating image quality, retouching and retrieving image. Owing the subjectivity this problem, there is no general framework predict aesthetics. In paper, we propose deep neural with visual attention module, self-generated global features hybrid loss address problem. Specifically, can be any state-of-the-art convolution classification network compatible attention. Further, feature compensates for context information during training stage, guides learn similarity between predicted aesthetic scores ground-truths through fusing soft-max-entropy Earth Mover’s Distance(EMD). With above-mentioned improvements, proposed capable effectively predicting an efficient way. our experiments, release real-world dataset that contains 1,800 2K photos labeled by several experienced photographers, then provide thorough ablation study design choices better understand superiority brought each part framework, comparisons methods on fraction metrics. The experimental results two datasets demonstrate both accuracy efficiency achieve favorably performance.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Brain-Inspired Deep Networks for Image Aesthetics Assessment

Image aesthetics assessment has been challenging due to its subjective nature. Inspired by the scientific advances in the human visual perception and neuroaesthetics, we design Brain-Inspired Deep Networks (BDN) for this task. BDN first learns attributes through the parallel supervised pathways, on a variety of selected feature dimensions. A high-level synthesis network is trained to associate ...

متن کامل

Incorporating Global Visual Features into Attention-based Neural Machine Translation

We introduce multi-modal, attentionbased Neural Machine Translation (NMT) models which incorporate visual features into different parts of both the encoder and the decoder. Global image features are extracted using a pre-trained convolutional neural network and are incorporated (i) as words in the source sentence, (ii) to initialise the encoder hidden state, and (iii) as additional data to init...

متن کامل

Cystoscopy Image Classication Using Deep Convolutional Neural Networks

In the past three decades, the use of smart methods in medical diagnostic systems has attractedthe attention of many researchers. However, no smart activity has been provided in the eld ofmedical image processing for diagnosis of bladder cancer through cystoscopy images despite the highprevalence in the world. In this paper, two well-known convolutional neural networks (CNNs) ...

متن کامل

Deep Convolutional Neural Networks with Layer-Wise Context Expansion and Attention

In this paper, we propose a deep convolutional neural network (CNN) with layer-wise context expansion and location-based attention, for large vocabulary speech recognition. In our model each higher layer uses information from broader contexts, along both the time and frequency dimensions, than its immediate lower layer. We show that both the layer-wise context expansion and the location-based a...

متن کامل

Anomaly-based Web Attack Detection: The Application of Deep Neural Network Seq2Seq With Attention Mechanism

Today, the use of the Internet and Internet sites has been an integrated part of the people’s lives, and most activities and important data are in the Internet websites. Thus, attempts to intrude into these websites have grown exponentially. Intrusion detection systems (IDS) of web attacks are an approach to protect users. But, these systems are suffering from such drawbacks as low accuracy in ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: IEEE Access

سال: 2021

ISSN: ['2169-3536']

DOI: https://doi.org/10.1109/access.2020.3015060