scholarly journals The (ir)reproducibility of published analyses: A case study of 57 JML articles published between 2019 and 2021

2021 ◽  
Author(s):  
Anna Laurinavichyute ◽  
Shravan Vasishth

In 2019 the Journal of Memory and Language instituted an open data and code policy; this policy requires that, as a rule, code and data be released at the latest upon publication. Does this policy lead to reproducible results? We looked at whether 57 papers published between 2019 and 2021 were reproducible, in the sense that the published summary statistics should be possible to regenerate given the data, and given the code, when code was provided. We found that for 10 out of the 57 papers, data sets were inaccessible; 29 of the remaining 47 papers provided code, of which 16 were reproducible. Of the 18 papers that did not provide code, one was reproducible. Overall, the reproducibility rate was about 30%. This estimate is similar to the ones reported for psychology, economics, and other areas, but it is probably possible to do better. We provide some suggestions on how reproducibility can be improvedin future work.

2019 ◽  
Author(s):  
Benedict C Jones ◽  
Lisa Marie DeBruine ◽  
Urszula M marcinkowska

Secondary data analyses (analyses of open data from published studies) can play a critical role in hypothesis generation and in maximizing the contribution of collected data to the accumulation of scientific knowledge. However, assessing the evidentiary value of results from secondary data analyses is often challenging because analytical decisions can be biased by knowledge of the results of (and analytical choices made in) the original study and by unacknowledged exploratory analyses of open data sets (Scott & Kline, 2019; Weston, Ritchie, Rohrer, & Przybylski, 2018). Using the secondary data analyses reported by Gangestad et al. (this issue) as a case study, we outline several approaches that, if implemented, would allow readers to assess the evidentiary value of results from secondary data analyses with greater confidence.


Information ◽  
2021 ◽  
Vol 12 (7) ◽  
pp. 258
Author(s):  
Paolo Fosci ◽  
Giuseppe Psaila

How to exploit the incredible variety of JSON data sets currently available on the Internet, for example, on Open Data portals? The traditional approach would require getting them from the portals, then storing them into some JSON document store and integrating them within the document store. However, once data are integrated, the lack of a query language that provides flexible querying capabilities could prevent analysts from successfully completing their analysis. In this paper, we show how the J-CO Framework, a novel framework that we developed at the University of Bergamo (Italy) to manage large collections of JSON documents, is a unique and innovative tool that provides analysts with querying capabilities based on fuzzy sets over JSON data sets. Its query language, called J-CO-QL, is continuously evolving to increase potential applications; the most recent extensions give analysts the capability to retrieve data sets directly from web portals as well as constructs to apply fuzzy set theory to JSON documents and to provide analysts with the capability to perform imprecise queries on documents by means of flexible soft conditions. This paper presents a practical case study in which real data sets are retrieved, integrated and analyzed to effectively show the unique and innovative capabilities of the J-CO Framework.


2013 ◽  
Vol 9 (4) ◽  
pp. 19-43 ◽  
Author(s):  
Bo Hu ◽  
Nuno Carvalho ◽  
Takahide Matsutsuka

In light of the challenges of effectively managing Big Data, the authors are witnessing a gradual shift towards the increasingly popular Linked Open Data (LOD) paradigm. LOD aims to impose a machine-readable semantic layer over structured as well as unstructured data and hence automate some data analysis tasks that are not designed for computers. The convergence of Big Data and LOD is, however, not straightforward: the semantic layer of LOD and the Big Data large scale storage do not get along easily. Meanwhile, the sheer data size envisioned by Big Data denies certain computationally expensive semantic technologies, rendering the latter much less efficient than their performance on relatively small data sets. In this paper, the authors propose a mechanism allowing LOD to take advantage of existing large-scale data stores while sustaining its “semantic” nature. The authors demonstrate how RDF-based semantic models can be distributed across multiple storage servers and the authors examine how a fundamental semantic operation can be tuned to meet the requirements on distributed and parallel data processing. The authors' future work will focus on stress test of the platform in the magnitude of tens of billions of triples, as well as comparative studies in usability and performance against similar offerings.


2017 ◽  
Vol 118 (7/8) ◽  
pp. 420-432 ◽  
Author(s):  
Stuti Saxena

Purpose While “transparency-by-design” serves as the antecedent of any Open Government Data (OGD) initiative (Janssen et al., 2017), its logical objective would be the extent to which data “usage” is facilitated. This paper aims to underscore the significance, drivers and barriers to ensure “usage” of data sets conceding that re-use of data sets is one of the key objectives of any OGD initiative. Design/methodology/approach With a documentary analysis approach, the OGD initiative of Sri Lanka is investigated for the present purpose. Furthermore, the theoretical model of citizen engagement in OGD suggested by Sieber and Johnson (2015) is being referred to appreciate the extent to which the usage of data sets is facilitated via the OGD platform. Findings There are drivers as well as barriers as far as facilitating the usage of the data sets in the Sri Lankan OGD initiative is concerned. For instance, some of the drivers are the provision of suggesting data sets or the possibility of referring to data sets which are historical in nature. However, there are countless barriers to usage than the drivers. For instance, there is absence of metadata in the data sets; the data sets are not updated regularly; there are historical data; the formats of the data sets are limited in nature and are not user-friendly; there is no facility of conducting data visualization or analytics, and there is no collaborative approach towards building further the OGD initiative. Research limitations/implications As only one case study is probed in the paper, further research is warranted to undertake a comparative approach by taking two or more case studies into consideration. Practical implications This study holds relevance for Sri Lankan Government and other stakeholders (policy makers, citizens, developers and the like) so far as furthering of user engagement in OGD initiative is concerned. Social implications Facilitating more usage by citizens would increase their engagement, and they might derive value out of the data sets. At the same time, the government’s objective of ensuring increased usage of the data sets would be better realized. Originality/value “Transparency-by-design” approach had its focus on the publishing phase of OGD, and this paper seeks to provide its logical conclusion by emphasizing upon “usage by stakeholders” because by opening data sets, the government has the target to ensure that these open data sets are being used and re-used. Therefore, it is the outcome which is being discussed with the support of a case study set in the background of Sri Lanka’s Open Data initiative. Besides, this is the first study which probes the OGD initiatives of Sri Lanka – therein lies the major contribution of the study.


Author(s):  
Edgar Meij ◽  
Marc Bron ◽  
Laura Hollink ◽  
Bouke Huurnink ◽  
Maarten de Rijke
Keyword(s):  

Author(s):  
Harrison Togia ◽  
Oceana P. Francis ◽  
Karl Kim ◽  
Guohui Zhang

Hazards to roadways and travelers can be drastically different because hazards are largely dependent on the regional environment and climate. This paper describes the development of a qualitative method for assessing infrastructure importance and hazard exposure for rural highway segments in Hawai‘i under different conditions. Multiple indicators of roadway importance are considered, including traffic volume, population served, accessibility, connectivity, reliability, land use, and roadway connection to critical infrastructures, such as hospitals and police stations. The method of evaluating roadway hazards and importance can be tailored to fit different regional hazard scenarios. It assimilates data from diverse sources to estimate risks of disruption. A case study for Highway HI83 in Hawai‘i, which is exposed to multiple hazards, is conducted. Weakening of the road by coastal erosion, inundation from sea level rise, and rockfall hazards require adaptation solutions. By analyzing the risk of disruption to highway segments, adaptation approaches can be prioritized. Using readily available geographic information system data sets for the exposure and impacts of potential hazards, this method could be adapted not only for emergency management but also for planning, design, and engineering of resilient highways.


Sensors ◽  
2021 ◽  
Vol 21 (15) ◽  
pp. 5204
Author(s):  
Anastasija Nikiforova

Nowadays, governments launch open government data (OGD) portals that provide data that can be accessed and used by everyone for their own needs. Although the potential economic value of open (government) data is assessed in millions and billions, not all open data are reused. Moreover, the open (government) data initiative as well as users’ intent for open (government) data are changing continuously and today, in line with IoT and smart city trends, real-time data and sensor-generated data have higher interest for users. These “smarter” open (government) data are also considered to be one of the crucial drivers for the sustainable economy, and might have an impact on information and communication technology (ICT) innovation and become a creativity bridge in developing a new ecosystem in Industry 4.0 and Society 5.0. The paper inspects OGD portals of 60 countries in order to understand the correspondence of their content to the Society 5.0 expectations. The paper provides a report on how much countries provide these data, focusing on some open (government) data success facilitating factors for both the portal in general and data sets of interest in particular. The presence of “smarter” data, their level of accessibility, availability, currency and timeliness, as well as support for users, are analyzed. The list of most competitive countries by data category are provided. This makes it possible to understand which OGD portals react to users’ needs, Industry 4.0 and Society 5.0 request the opening and updating of data for their further potential reuse, which is essential in the digital data-driven world.


2021 ◽  
Vol 10 (4) ◽  
pp. 251
Author(s):  
Christina Ludwig ◽  
Robert Hecht ◽  
Sven Lautenbach ◽  
Martin Schorcht ◽  
Alexander Zipf

Public urban green spaces are important for the urban quality of life. Still, comprehensive open data sets on urban green spaces are not available for most cities. As open and globally available data sets, the potential of Sentinel-2 satellite imagery and OpenStreetMap (OSM) data for urban green space mapping is high but limited due to their respective uncertainties. Sentinel-2 imagery cannot distinguish public from private green spaces and its spatial resolution of 10 m fails to capture fine-grained urban structures, while in OSM green spaces are not mapped consistently and with the same level of completeness everywhere. To address these limitations, we propose to fuse these data sets under explicit consideration of their uncertainties. The Sentinel-2 derived Normalized Difference Vegetation Index was fused with OSM data using the Dempster–Shafer theory to enhance the detection of small vegetated areas. The distinction between public and private green spaces was achieved using a Bayesian hierarchical model and OSM data. The analysis was performed based on land use parcels derived from OSM data and tested for the city of Dresden, Germany. The overall accuracy of the final map of public urban green spaces was 95% and was mainly influenced by the uncertainty of the public accessibility model.


Sign in / Sign up

Export Citation Format

Share Document