DataDesk: an interactive package for data exploration, display, model building, and data analysis

2012 ◽  
Vol 4 (4) ◽  
pp. 407-414 ◽  
Author(s):  
Paul F. Velleman
Author(s):  
Patrick J. Ogao ◽  
Connie A. Blok

Measurements from dynamic environmental phenomena have resulted in the acquisition and generation of an enormous amount of data. This upsurge in data availability can be attributed to the interdisciplinary nature of environmental problem solving and the wide range of acquisition technology involved. In essence, users are dealing with data that is complex in nature, multidimensional and probably of a temporal nature. Also, the frequency by which this data is acquired far exceeds the rate at which it is being explored, a factor that has accelerated the search for innovative approaches and tools in spatial data analysis. These attempts have seen both analytical and visual techniques being used as aids in presentation and scientific data exploration. Examples are seen in techniques as in: data mining, data exploration and visualization.


1987 ◽  
Vol 19 (9) ◽  
pp. 1153-1160 ◽  
Author(s):  
M J Baxter

Methods of assessing model fit for models of spatial flows frequently do not take account of spatial structure. A nonparametric test, based on the signs of the residuals from a fit, is presented for detecting patterns in the residuals. It can be thought of as a general test of misspecification that can allow for spatial structure effects. It is argued that the test is most usefully applied with a model-building approach to data analysis.


Author(s):  
Filomena Maggino

In the sphere of social research, one of the principal methodological approaches is that of the survey. In the performance of surveys, statistics play a fundamental role. This publication is intended to highlight this role, with special reference to data analysis, placing the emphasis on the objectives of such analysis, the logic, the procedures necessary to achieve them, and the interpretation of results. The material has been divided into two volumes, which while being independent are also logically connected. The first volume is organised in two sections: the first is designed to provide a synthetic overview of the statistical survey in terms of methods and instruments, while the second describes the principal instruments of statistical analysis which enable a description of the data. The first volume comes complete with a CD with hypertext. The second volume presents the statistical approaches which enable the achievement of specific objectives, seen in terms of data exploration and verification of the results obtained.


2020 ◽  
Author(s):  
◽  
Siyang Zhang

Safety is the top concern in transportation, especially in work zones, as work zones deviate from regular driving environment and driver behavior is very different. In order to protect workers and create a safer work zone environment, new technologies are proposed by agencies and deployed to work zones, however, some are without scientific study before deployment. Therefore, quantitative studies need to be conducted to show the effectiveness of technologies. Driving simulator is a safe and cost-effective way to test effectiveness of new designs and compare different configurations. Field study is another scientific way of testing, as it provides absolute validity, while simulator study provides relative validity. The synergy of field and simulator studies construct a precise experiment as field study calibrates simulator design and validates simulator results. Two main projects, Evaluation of Automated Flagger Assistance Devices (AFADs), and Evaluation of Green Lights on Truck-Mounted Attenuator (TMA), are discussed in this dissertation to illustrate the investigation of smart work zone technologies using mixed simulator and field studies, along with one simulator project investigating interaction between human driven car and autonomous truck platoon in work zones. Both field and simulator studies indicated that AFADs improved stationary work zone safety by enhancing visibility, isolating workers from immediate traffic, and conveying clear guidance message to traffic. The results of green light on TMAs implied an inverse relationship between visibility/awareness of work zone and arrow board recognition/easy on eyes, but did not show if any of the light configurations is superior. Results anticipated for autonomous truck platoon in work zones are drivers behave more uniformly after being educated about the meaning of signage displayed on the back of truck, and performance measured with signage would be more preferable than those without signage. Applications of statistics are extension of studies, including experimental design, survey design, and data analysis. Data obtained from AFAD and Green Light projects were utilized to illustrate the methodologies of data analysis and model building, which incorporated simulator data, biofeedback and survey response to interpret the relationship among driver perspective and mental status, and driving behavior. From the studies conducted, it could be concluded that mixed simulator and field study is a good fit for smart work zone technologies investigation. Simulators provide a safe environment, flexibility and cost-effectiveness, while field studies calibrate and validate simulator setup and its results. The collaboration of two forms of study generates legitimate and convincing results for investigations. Applying statistical methodologies into transportation simulator and field studies is a good way to make experiment and survey design more rational, and the statistical methods are applicable for further data analysis.


2018 ◽  
Author(s):  
Joy Merwin Monteiro ◽  
Jeremy McGibbon ◽  
Rodrigo Caballero

Abstract. sympl (System for Modelling Planets) and climt (Climate Modelling and diagnostics Toolkit) represent an attempt to rethink climate modelling frameworks from the ground up. The aim is to use expressive data structures available in the scientific Python ecosystem along with best practices in software design to build models that are self-documenting, highly inter-operable and that provide fine grained control over model components and behaviour. We believe that such an approach towards building models is essential to allow scientists to easily and reliably combine model components to represent the climate system at a desired level of complexity, and to enable users to fully understand what the model is doing. sympl is a framework which formulates the model in terms of a "state" which gets evolved forward in time by TimeStepper and Implicit components, and which can be modified by Diagnostic components. TimeStepper components in turn rely on Prognostic components to compute tendencies. Components contain all the information about the kinds of inputs they expect and outputs that they provide. Components can be used interchangeably, even when they rely on different units or array configurations. sympl provides basic functions and objects which could be used by any type of Earth system model. climt is an Earth system modelling toolkit that contains scientific components built over the sympl base objects. Components can be written in any language accessible from Python, and Fortran/C libraries are accessed via Cython. climt aims to provide different user APIs which trade-off simplicity of use against flexibility of model building, thus appealing to a wide audience. Model building, configuration and execution is through a Python script (or Jupyter Notebook), enabling researchers to build an end-to-end Python based pipeline along with popular Python based data analysis tools. Because of the modularity of the individual components, using online data analysis, visualisation or assimilation algorithms and tools with sympl/climt components is extremely simple.


Sign in / Sign up

Export Citation Format

Share Document