Bidirectional Focused Semantic Alignment Attention Network for Cross-Modal Retrieval

Author(s):  
Shuli Cheng ◽  
Liejun Wang ◽  
Anyu Du ◽  
Yongming Li
2020 ◽  
Vol 34 (07) ◽  
pp. 11189-11196 ◽  
Author(s):  
Ya Jing ◽  
Chenyang Si ◽  
Junbo Wang ◽  
Wei Wang ◽  
Liang Wang ◽  
...  

Text-based person search aims to retrieve the corresponding person images in an image database by virtue of a describing sentence about the person, which poses great potential for various applications such as video surveillance. Extracting visual contents corresponding to the human description is the key to this cross-modal matching problem. Moreover, correlated images and descriptions involve different granularities of semantic relevance, which is usually ignored in previous methods. To exploit the multilevel corresponding visual contents, we propose a pose-guided multi-granularity attention network (PMA). Firstly, we propose a coarse alignment network (CA) to select the related image regions to the global description by a similarity-based attention. To further capture the phrase-related visual body part, a fine-grained alignment network (FA) is proposed, which employs pose information to learn latent semantic alignment between visual body part and textual noun phrase. To verify the effectiveness of our model, we perform extensive experiments on the CUHK Person Description Dataset (CUHK-PEDES) which is currently the only available dataset for text-based person search. Experimental results show that our approach outperforms the state-of-the-art methods by 15 % in terms of the top-1 metric.


Author(s):  
Holger Gevensleben ◽  
Gunther H. Moll ◽  
Hartmut Heinrich

Im Rahmen einer multizentrischen, randomisierten, kontrollierten Studie evaluierten wir die klinische Wirksamkeit eines Neurofeedback-Trainings (NF) bei Kindern mit einer Aufmerksamkeitsdefizit-/Hyperaktivitätsstörung (ADHS) und untersuchten die einem erfolgreichen Training zugrunde liegenden neurophysiologischen Wirkmechanismen. Als Vergleichstraining diente ein computergestütztes Aufmerksamkeitstraining, das dem Setting des Neurofeedback-Trainings in den wesentlichen Anforderungen und Rahmenbedingungen angeglichen war. Auf Verhaltensebene (Eltern- und Lehrerbeurteilung) zeigte sich das NF-Training nach Trainingsende dem Kontrolltraining sowohl hinsichtlich der ADHS-Kernsymptomatik als auch in assoziierten Bereichen überlegen. Für das Hauptzielkriterium (Verbesserung im FBB-HKS Gesamtwert) ergab sich eine mittlere Effektstärke (von 0.6). Sechs Monate nach Trainingsende (follow-up) konnte das gleiche Ergebnismuster gefunden werden. Die Ergebnisse legen somit den Schluss nahe, dass NF einen klinisch wirksamen Therapiebaustein zur Behandlung von Kindern mit ADHS darstellt. Auf neurophysiologischer Ebene (EEG; ereignisbezogene Potentiale, EPs) konnten für die beiden Neurofeedback-Protokolle Theta/Beta-Training und Training langsamer kortikaler Potentiale spezifische Effekte aufgezeigt werden. So war für das Theta/Beta-Training beispielsweise die Abnahme der Theta-Aktivität mit einer Reduzierung der ADHS-Symptomatik assoziiert. Für das SCP-Training wurde u. a. im Attention Network Test eine Erhöhung der kontingenten negativen Variation beobachtet, die die mobilisierten Ressourcen bei Vorbereitungsprozessen widerspiegelt. EEG- und EP-basierte Prädiktorvariablen konnten ermittelt werden. Der vorliegende Artikel bietet einen Gesamtüberblick über die in verschiedenen Publikationen unserer Arbeitsgruppe beschriebenen Ergebnisse der Studie und zeigt zukünftige Fragestellungen auf.


2018 ◽  
Vol 32 (5) ◽  
pp. 541-553 ◽  
Author(s):  
Nadine M. Richard ◽  
Charlene O'Connor ◽  
Ayan Dey ◽  
Ian H. Robertson ◽  
Brian Levine

2019 ◽  
Vol 69 (10) ◽  
pp. 423
Author(s):  
Manuel Vázquez Marrufo ◽  
Macarena García-Valdecasas Colell ◽  
Alejandro Galvao Carmona ◽  
Esteban Sarrias Arrabal ◽  
Javier Tirapu Ustárroz

Author(s):  
Chengzhu Yu ◽  
Heng Lu ◽  
Na Hu ◽  
Meng Yu ◽  
Chao Weng ◽  
...  

Author(s):  
Jincan Deng ◽  
Liang Li ◽  
Beichen Zhang ◽  
Shuhui Wang ◽  
Zhengjun Zha ◽  
...  

2021 ◽  
Vol 226 (4) ◽  
pp. 989-1006
Author(s):  
Ilenia Salsano ◽  
Valerio Santangelo ◽  
Emiliano Macaluso

AbstractPrevious studies demonstrated that long-term memory related to object-position in natural scenes guides visuo-spatial attention during subsequent search. Memory-guided attention has been associated with the activation of memory regions (the medial-temporal cortex) and with the fronto-parietal attention network. Notably, these circuits represent external locations with different frames of reference: egocentric (i.e., eyes/head-centered) in the dorsal attention network vs. allocentric (i.e., world/scene-centered) in the medial temporal cortex. Here we used behavioral measures and fMRI to assess the contribution of egocentric and allocentric spatial information during memory-guided attention. At encoding, participants were presented with real-world scenes and asked to search for and memorize the location of a high-contrast target superimposed in half of the scenes. At retrieval, participants viewed again the same scenes, now all including a low-contrast target. In scenes that included the target at encoding, the target was presented at the same scene-location. Critically, scenes were now shown either from the same or different viewpoint compared with encoding. This resulted in a memory-by-view design (target seen/unseen x same/different view), which allowed us teasing apart the role of allocentric vs. egocentric signals during memory-guided attention. Retrieval-related results showed greater search-accuracy for seen than unseen targets, both in the same and different views, indicating that memory contributes to visual search notwithstanding perspective changes. This view-change independent effect was associated with the activation of the left lateral intra-parietal sulcus. Our results demonstrate that this parietal region mediates memory-guided attention by taking into account allocentric/scene-centered information about the objects' position in the external world.


Sign in / Sign up

Export Citation Format

Share Document