complex data sets
Recently Published Documents


TOTAL DOCUMENTS

132
(FIVE YEARS 46)

H-INDEX

14
(FIVE YEARS 2)

2022 ◽  
pp. 67-76
Author(s):  
Dineshkumar Bhagwandas Vaghela

The term big data has come due to rapid generation of data in various organizations. In big data, the big is the buzzword. Here the data are so large and complex that the traditional database applications are not able to process (i.e., they are inadequate to deal with such volume of data). Usually the big data are described by 5Vs (volume, velocity, variety, variability, veracity). The big data can be structured, semi-structured, or unstructured. Big data analytics is the process to uncover hidden patterns, unknown correlations, predict the future values from large and complex data sets. In this chapter, the following topics will be covered more in detail. History of big data and business analytics, big data analytics technologies and tools, and big data analytics uses and challenges.


2021 ◽  
Vol 16 (1) ◽  
pp. 5-15
Author(s):  
Alexander Refsum Jensenius

Music researchers work with increasingly large and complex data sets. There are few established data handling practices in the field and several conceptual, technological, and practical challenges. Furthermore, many music researchers are not equipped for (or interested in) the craft of data storage, curation, and archiving. This paper discusses some of the particular challenges that empirical music researchers face when working towards Open Research practices: handling (1) (multi)media files, (2) privacy, and (3) copyright issues. These are exemplified through MusicLab, an event series focused on fostering openness in music research. It is argued that the "best practice" suggested by the FAIR principles is too demanding in many cases, but "good enough practice" may be within reach for many. A four-layer data handling "recipe" is suggested as concrete advice for achieving "good enough practice" in empirical music research.


Author(s):  
Matthias König ◽  
Jan Grzegorzewski ◽  
Martin Golebiewski ◽  
Henning Hermjakob ◽  
Mike Hucka ◽  
...  

Science continues to become more interdisciplinary and to involve increasingly complex data sets. Many projects in the biomedical and health-related sciences follow or aim to follow the principles of FAIR data sharing, which has been demonstrated to foster collaboration, to lead to better research outcomes, and to help ensure reproducibility of results. Data generated in the course of biomedical and health research present specific challenges for FAIR sharing in the sense that they are heterogeneous and highly sensitive to context and the needs of protection and privacy. Data sharing must respect these features without impeding timely dissemination of results, so that they can contribute to time-critical advances in medical therapy and treatment. Modeling and simulation of biomedical processes have become established tools, and a global community has been developing algorithms, methodologies, and standards for applying biomedical simulation models in clinical research. However, it can be difficult for clinician scientists to follow the specific rules and recommendations for FAIR data sharing within this domain. We seek to clarify the standard workflow for sharing experimental and clinical data with the simulation modeling community. By following these recommendations, data sharing will be improved, collaborations will become more effective, and the FAIR publication and subsequent reuse of data will become possible at the level of quality necessary to support biomedical and health-related sciences.


2021 ◽  
Vol 921 (2) ◽  
pp. 177
Author(s):  
Regina Sarmiento ◽  
Marc Huertas-Company ◽  
Johan H. Knapen ◽  
Sebastián F. Sánchez ◽  
Helena Domínguez Sánchez ◽  
...  

Abstract As available data sets grow in size and complexity, advanced visualization tools enabling their exploration and analysis become more important. In modern astronomy, integral field spectroscopic galaxy surveys are a clear example of increasing high dimensionality and complex data sets, which challenges the traditional methods used to extract the physical information they contain. We present the use of a novel self-supervised machine-learning method to visualize the multidimensional information on stellar population and kinematics in the MaNGA survey in a 2D plane. Our framework is insensitive to nonphysical properties such as the size of the integral field unit and is therefore able to order galaxies according to their resolved physical properties. Using the extracted representations, we study how galaxies distribute based on their resolved and global physical properties. We show that even when exclusively using information about the internal structure, galaxies naturally cluster into two well-known categories, rotating main-sequence disks and massive slow rotators, from a purely data-driven perspective, hence confirming distinct assembly channels. Low-mass rotation-dominated quenched galaxies appear as a third cluster only if information about the integrated physical properties is preserved, suggesting a mixture of assembly processes for these galaxies without any particular signature in their internal kinematics that distinguishes them from the two main groups. The framework for data exploration is publicly released with this publication, ready to be used with the MaNGA or other integral field data sets.


Author(s):  
James Harnly ◽  
Matthew J. Picklo ◽  
Kenneth F. Kalscheur ◽  
Andrew Magnuson ◽  
Michael R. Bukowski ◽  
...  

Molecules ◽  
2021 ◽  
Vol 26 (19) ◽  
pp. 5989
Author(s):  
Giuseppina Ioele ◽  
Fedora Grande ◽  
Michele De Luca ◽  
Maria Antonietta Occhiuzzi ◽  
Antonio Garofalo ◽  
...  

The present paper provides an updated overview of the methodologies applied in photodegradation studies of non-steroidal anti-inflammatory drugs. Photostability tests, performed according to international standards, have clearly demonstrated the photolability of many drugs belonging to this class, observed during the preparation of commercial forms, administration or when dispersed in the environment. The photodegradation profile of these drugs is usually monitored by spectrophotometric or chromatographic techniques and in many studies the analytical data are processed by chemometric procedures. The application of multivariate analysis in the resolution of often-complex data sets makes it possible to estimate the pure spectra of the species involved in the degradation process and their concentration profiles. Given the wide use of these drugs, several pharmaceutical formulations have been investigated to improve their photostability in solution or gel, as well as the pharmacokinetic profile. The use of lipid nanocarriers as liposomes, niosomes or solid lipid nanoparticles has demonstrated to both minimize photodegradation and improve the controlled release of the entrapped drugs.


2021 ◽  
Vol 11 (18) ◽  
pp. 8772
Author(s):  
Laura Raya ◽  
Sara A. Boga ◽  
Marcos Garcia-Lorenzo ◽  
Sofia Bayona

Technological advances enable the capture and management of complex data sets that need to be correctly understood. Visualisation techniques can help in complex data analysis and exploration, but sometimes the visual channel is not enough, or it is not always available. Some authors propose using the haptic channel to reinforce or substitute the visual sense, but the limited human haptic short-term memory still poses a challenge. We present the haptic tuning fork, a reference signal displayed before the haptic information for increasing the discriminability of haptic icons. With this reference, the user does not depend only on short-term memory. We have decided to evaluate the usefulness of the haptic tuning fork in impedance kinesthetic devices as these are the most common. Furthermore, since the renderable signal ranges are device-dependent, we introduce a methodology to select a discriminable set of signals called the haptic scale. Both the haptic tuning fork and the haptic scale proved their usefulness in the performed experiments regarding haptic stimuli varying in frequency.


Leonardo ◽  
2021 ◽  
pp. 1-7
Author(s):  
JoAnn Kuchera-Morin

Abstract This paper discusses the creation and development of a large distributed immersive multimedia computation system and environment based on the discipline of orchestral music composition, concert hall design, and performance. Just as the orchestra evolved through mechanical engineering to become a large distributed multi-user instrument whose information can be transmitted either by a client-server model as in orchestra-conductor, or a client-to-client model, as in an instrumental ensemble, large-scale distributed multi-media computational platforms can be modeled in the same way, facilitating the users as performers of the system. Multiple researchers can mine large, complex data sets to uncover important relationships in their spatio-temporal structures.


Author(s):  
Matthias König ◽  
Jan Grzegorzewski ◽  
Martin Golebiewski ◽  
Henning Hermjakob ◽  
Mike Hucka ◽  
...  

Science continues to become more interdisciplinary and to involve increasingly complex data sets. Many projects in the biomedical and health related sciences adhere to the principles of FAIR data sharing, or aim to follow them. Data sharing has been proven to foster collaboration, to lead to better research outcomes, and to help ensure reproducibility of results. Data generated in biomedical and health research are specific in the sense that they are heterogeneous, often big, and highly sensitive in terms of data protection needs and contextuality. Data sharing has to respect these features, but at the same time advances in medical therapy and treatment are time-critical. Modeling and simulation of biomedical processes have become an established tool, and a global community has been developing algorithms, methodologies, and standards for applying biomedical simulation models in clinical research. However, it can be difficult for clinician scientists to follow the specific rules and recommendations for FAIR data sharing within the domain. With this paper, we aim to clarify the standard workflow for sharing experimental and clinical data with the simulation modeling community. By following these recommendations, data sharing will be improved, collaborations will become more effective, and the FAIR publication and subsequent reuse of data will become possible at the level of quality necessary in biomedical and health related sciences.


Sign in / Sign up

Export Citation Format

Share Document