scholarly journals INFRASTRUCTURE FOR LARGE SCALE HARVESTING OF WEBOMETRIC INDICATORS

2020 ◽  
pp. 86-97
Author(s):  
D. V. Kosyakov

The main purpose of scientific rankings is not only comparison, but also stimulating the development of universities and research institutes, improving the effectiveness of their educational and research processes. However, existing webometric ratings are not suitable for this task because of the lack of the possibility of their analysis and verification. The solution is to change the approach to webometric measurements, to perform them based on the principles of regular collection and openness of source data. The article discusses the indicators used in the tasks of academic webometrics, their reliability and stability, substantiates the need for regular collection of the values of these indicators to improve the quality of data and analyze their dynamics. Based on the experience of implementing a project to research the Russian academic web space (http://www. webometrix.ru), problems are analyzed that arise when collecting the values of webometric indicators of sites of research organizations and institutions of higher education on a global scale with a frequency of at least 1 time per month.The author describes a distributed system for collecting values of webometric indicators and evaluates the necessary degree of parallelization of the process. The developed approach is universal for the tasks of collecting large amounts of information by the method of extracting data from website pages, and the collection of webometric data is also relevant for the tasks of analyzing other thematic segments of web space, such as library websites.

2018 ◽  
Vol 39 (2) ◽  
pp. 231-240
Author(s):  
Danping Ren ◽  
Kun Guo ◽  
Qiuyan Yao ◽  
Jijun Zhao

AbstractThe impairment-aware routing and wavelength assignment algorithm with probe flow (P-IA-RWA) can make an accurate estimation for the transmission quality of the link when the connection request comes. But it also causes some problems. The probe flow data introduced in the P-IA-RWA algorithm can result in the competition for wavelength resources. In order to reduce the competition and the blocking probability of the network, a new P-IA-RWA algorithm with segmentation monitoring-control mechanism (SMC-P-IA-RWA) is proposed. The algorithm would reduce the holding time of network resources for the probe flow. It segments the candidate path suitably for the data transmitting. And the transmission quality of the probe flow sent by the source node will be monitored in the endpoint of each segment. The transmission quality of data can also be monitored, so as to make the appropriate treatment to avoid the unnecessary probe flow. The simulation results show that the proposed SMC-P-IA-RWA algorithm can effectively reduce the blocking probability. It brings a better solution to the competition for resources between the probe flow and the main data to be transferred. And it is more suitable for scheduling control in the large-scale network.


2017 ◽  
Author(s):  
Wil Roebroeks ◽  
Sabine Gaudzinski-Windheuser ◽  
Michael Baales ◽  
Ralf-Dietrich Kahlke

AbstractThe database regarding the earliest occupation of Europe has increased significantly in quantity and quality of data points over the last two decades, mainly through the addition of new sites as a result of long-term systematic excavations and large-scale prospections of Early and early Middle Pleistocene exposures. The site distribution pattern suggests an ephemeral presence of hominins in the south of Europe from around one million years ago, with occasional short northward expansions along the western coastal areas when temperate conditions permitted. From around 600,000-700,000 years ago Acheulean artefacts appear in Europe and somewhat later hominin presence seems to pick up, with more sites and now some also present in colder climatic settings. It is again only later, around 350,000 years ago, that the first sites show up in more continental, central parts of Europe, east of the Rhine. A series of recent papers on the Early Pleistocene palaeontological site of Untermassfeld (Germany) makes claims that are of great interest for studies of earliest Europe and are at odds with the described pattern: the papers suggest that Untermassfeld has yielded stone tools and humanly modified faunal remains, evidence for a one million years old hominin presence in European continental mid-latitudes, and additional evidence that hominins were well-established in Europe already around that time period. Here we evaluate these claims and demonstrate that these studies are severely flawed in terms of data on provenance of the materials studied and in the interpretation of faunal remains and lithics as testifying to a hominin presence at the site. In actual fact any reference to the Untermassfeld site as an archaeological one is unwarranted. Furthermore, it is not the only European Early Pleistocene site where inferred evidence for hominin presence is problematic. The strength of the spatiotemporal patterns of hominin presence and absence depend on the quality of the data points we work with, and data base maintenance, including critical evaluation of new sites, is crucial to advance our knowledge of the expansions and contractions of hominin ranges during the Pleistocene.


2021 ◽  
Author(s):  
Sven Hilbert ◽  
Stefan Coors ◽  
Elisabeth Barbara Kraus ◽  
Bernd Bischl ◽  
Mario Frei ◽  
...  

Classical statistical methods are limited in the analysis of highdimensional datasets. Machine learning (ML) provides a powerful framework for prediction by using complex relationships, often encountered in modern data with a large number of variables, cases and potentially non-linear effects. ML has turned into one of the most influential analytical approaches of this millennium and has recently become popular in the behavioral and social sciences. The impact of ML methods on research and practical applications in the educational sciences is still limited, but continuously grows as larger and more complex datasets become available through massive open online courses (MOOCs) and large scale investigations.The educational sciences are at a crucial pivot point, because of the anticipated impact ML methods hold for the field. Here, we review the opportunities and challenges of ML for the educational sciences, show how a look at related disciplines can help learning from their experiences, and argue for a philosophical shift in model evaluation. We demonstrate how the overall quality of data analysis in educational research can benefit from these methods and show how ML can play a decisive role in the validation of empirical models. In this review, we (1) provide an overview of the types of data suitable for ML, (2) give practical advice for the application of ML methods, and (3) show how ML-based tools and applications can be used to enhance the quality of education. Additionally we provide practical R code with exemplary analyses, available at https: //osf.io/ntre9/?view only=d29ae7cf59d34e8293f4c6bbde3e4ab2.


F1000Research ◽  
2015 ◽  
Vol 4 ◽  
pp. 1075 ◽  
Author(s):  
Camilla L.C. Ip ◽  
Matthew Loose ◽  
John R. Tyson ◽  
Mariateresa de Cesare ◽  
Bonnie L. Brown ◽  
...  

The advent of a miniaturized DNA sequencing device with a high-throughput contextual sequencing capability embodies the next generation of large scale sequencing tools. The MinION™ Access Programme (MAP) was initiated by Oxford Nanopore Technologies™ in April 2014, giving public access to their USB-attached miniature sequencing device. The MinION Analysis and Reference Consortium (MARC) was formed by a subset of MAP participants, with the aim of evaluating and providing standard protocols and reference data to the community. Envisaged as a multi-phased project, this study provides the global community with the Phase 1 data from MARC, where the reproducibility of the performance of the MinION was evaluated at multiple sites. Five laboratories on two continents generated data using a control strain of Escherichia coli K-12, preparing and sequencing samples according to a revised ONT protocol. Here, we provide the details of the protocol used, along with a preliminary analysis of the characteristics of typical runs including the consistency, rate, volume and quality of data produced. Further analysis of the Phase 1 data presented here, and additional experiments in Phase 2 of E. coli from MARC are already underway to identify ways to improve and enhance MinION performance.


Author(s):  
Suresh Neethirajan

The world's growing population is highly dependent on animal agriculture. Animal products provide nutrient-packed meals that help to sustain individuals of all ages in communities across the globe. As the human demand for animal proteins grows, the agricultural industry must continue to advance its efficiency and quality of production. One of the most commonly farmed livestock is poultry and their significance is felt on a global scale. Current poultry farming practices result in the premature death and rejection of billions of chickens on an annual basis before they are processed for meat. This loss of life is concerning regarding animal welfare, agricultural efficiency, and economic impacts. The best way to prevent these losses is through the individualistic and/or group level assessment of animal on a continuous basis. On large-scale farms, such attention to detail was generally considered to be inaccurate and inefficient, but with the integration of Artificial Intelligence (AI) assisted technology individualized and per-herd assessments of livestock are possible and accurate. Various studies have shown cameras linked with specialized systems of AI can properly analyze flocks for health concerns, thus improving the survival rate and product quality of farmed poultry. Building on the recent advancements, this review explores the aspects of AI in the detection, counting and tracking of the poultry in commercial and research-based applications.


Author(s):  
Kai Geschuhn ◽  
Andreas Vogler

Abstract This paper makes the strong, fact-based case for a large-scale transformation of the current corpus of scientific subscription journals to an open access business model. The existing journals, with their well-tested functionalities, should be retained and developed to meet the demands of 21st-century research, while the underlying payment streams undergo a major restructuring. There is sufficient momentum for this decisive push toward open access publishing. The diverse existing initiatives must be coordinated so as to converge on this clear goal. The international nature of research implies that this transformation will be achieved on a truly global scale only through a consensus of the world’s most eminent research organizations. All the indications are that the money already invested in the research publishing system is sufficient to enable a transformation that will be sustainable for the future. There needs to be a shared understanding that the money currently locked in the journal subscription system must be withdrawn and re-purposed for open access publishing services. The current library acquisition budgets are the ultimate reservoir for enabling the transformation without financial or other risks. The goal is to preserve the established service levels provided by publishers that are still requested by researchers, while redefining and reorganizing the necessary payment streams. By disrupting the underlying business model, the viability of journal publishing can be preserved and put on a solid footing for the scholarly developments of the future.


2020 ◽  
Author(s):  
Alban de Lavenne ◽  
Alena Bartosova ◽  
Johan Strömqvist ◽  
Berit Arheimer

<p>Water quality modelling is very often constrained beforehand by the performance of discharge modelling. When the model is set up at global scale, this discharge performance usually greatly varies in space and can therefore be a limiting factor in many regions around the globe. Besides discharge performances, the quality of the observations themselves can also be highly heterogeneous in space and among datasets. Modeller then has to find a compromise between being restrictive on the quality of his dataset or having a good spatio-temporal representativeness of the various hydrological conditions at global scale. This often relies on subjective thresholds.</p> <p>This work proposed a more objective calibration strategy that aims to consider both aspects explicitly: observation quality and model performance on discharge. It leads to the construction of two scores that are assigned to each water quality station quantifying their reliability for model calibration and evaluation. The average of those two scores is then used as a weight in the objective function to emphasise the training on the most reliable stations.</p> <p>The strategy is implemented for sediment modelling using the WW-HYPE model (Arheimer et al., 2019) at global scale. The score on discharge simulation performance is based on the regionalisation of the Nash-Sutcliffe Efficiency that is spatially interpolated at every sediment monitoring stations using inverse Ghosh distance weighting (de Lavenne et al., 2016). The score on the observation quality is based on the location of the station with respect to the catchment outlet and on the amount of data. A multi-objective calibration is performed to optimise parameters on two global databases, one on long terms sediment loads (730 stations) and one on sediment concentrations time series (1440 stations). The sensitivity of the model to this calibration strategy is analysed according to model performances and model outputs, such as sediment loads at global scale, in order to discuss the importance of considering this heterogeneity of the reliability of monitoring stations.</p> <p>References</p> <p>Arheimer, B., Pimentel, R., Isberg, K., Crochemore, L., Andersson, J. C. M., Hasan, A., and Pineda, L. (2019), Global catchment modelling using World-Wide HYPE (WWH), open data and stepwise parameter estimation, Hydrol. Earth Syst. Sci. Discuss., doi:10.5194/hess-2019-111.</p> <p>de Lavenne, A., J. O. Skøien, C. Cudennec, F. Curie, and F. Moatar (2016), Transferring measured discharge time series: Large-scale comparison of Top-kriging to geomorphology-based inverse modeling, Water Resour. Res., 52, 5555–5576, doi:10.1002/2016WR018716.</p>


2016 ◽  
Vol 16 (12) ◽  
pp. 5010-5020 ◽  
Author(s):  
Mihaela I. Chidean ◽  
Eduardo Morgado ◽  
Margarita Sanroman-Junquera ◽  
Julio Ramiro-Bargueno ◽  
Javier Ramos ◽  
...  

Author(s):  
B. L. Armbruster ◽  
B. Kraus ◽  
M. Pan

One goal in electron microscopy of biological specimens is to improve the quality of data to equal the resolution capabilities of modem transmission electron microscopes. Radiation damage and beam- induced movement caused by charging of the sample, low image contrast at high resolution, and sensitivity to external vibration and drift in side entry specimen holders limit the effective resolution one can achieve. Several methods have been developed to address these limitations: cryomethods are widely employed to preserve and stabilize specimens against some of the adverse effects of the vacuum and electron beam irradiation, spot-scan imaging reduces charging and associated beam-induced movement, and energy-filtered imaging removes the “fog” caused by inelastic scattering of electrons which is particularly pronounced in thick specimens.Although most cryoholders can easily achieve a 3.4Å resolution specification, information perpendicular to the goniometer axis may be degraded due to vibration. Absolute drift after mechanical and thermal equilibration as well as drift after movement of a holder may cause loss of resolution in any direction.


Author(s):  
A. Babirad

Cerebrovascular diseases are a problem of the world today, and according to the forecast, the problem of the near future arises. The main risk factors for the development of ischemic disorders of the cerebral circulation include oblique and aging, arterial hypertension, smoking, diabetes mellitus and heart disease. An effective strategy for the prevention of cerebrovascular events is based on the implementation of large-scale risk control measures, including the use of antiagregant and anticoagulant therapy, invasive interventions such as atheromectomy, angioplasty and stenting. In this connection, the efforts of neurologists, cardiologists, angiosurgery, endocrinologists and other specialists are the basis for achieving an acceptable clinical outcome. A review of the SF-36 method for assessing the quality of life in patients with the effects of transient ischemic stroke is presented. The assessment of quality of life is recognized in world medical practice and research, an indicator that is also used to assess the quality of the health system and in general sociological research.


Sign in / Sign up

Export Citation Format

Share Document