accuracy measure
Recently Published Documents


TOTAL DOCUMENTS

140
(FIVE YEARS 73)

H-INDEX

12
(FIVE YEARS 3)

2022 ◽  
Vol 31 (1) ◽  
pp. 1-37
Author(s):  
Chao Liu ◽  
Xin Xia ◽  
David Lo ◽  
Zhiwe Liu ◽  
Ahmed E. Hassan ◽  
...  

To accelerate software development, developers frequently search and reuse existing code snippets from a large-scale codebase, e.g., GitHub. Over the years, researchers proposed many information retrieval (IR)-based models for code search, but they fail to connect the semantic gap between query and code. An early successful deep learning (DL)-based model DeepCS solved this issue by learning the relationship between pairs of code methods and corresponding natural language descriptions. Two major advantages of DeepCS are the capability of understanding irrelevant/noisy keywords and capturing sequential relationships between words in query and code. In this article, we proposed an IR-based model CodeMatcher that inherits the advantages of DeepCS (i.e., the capability of understanding the sequential semantics in important query words), while it can leverage the indexing technique in the IR-based model to accelerate the search response time substantially. CodeMatcher first collects metadata for query words to identify irrelevant/noisy ones, then iteratively performs fuzzy search with important query words on the codebase that is indexed by the Elasticsearch tool and finally reranks a set of returned candidate code according to how the tokens in the candidate code snippet sequentially matched the important words in a query. We verified its effectiveness on a large-scale codebase with ~41K repositories. Experimental results showed that CodeMatcher achieves an MRR (a widely used accuracy measure for code search) of 0.60, outperforming DeepCS, CodeHow, and UNIF by 82%, 62%, and 46%, respectively. Our proposed model is over 1.2K times faster than DeepCS. Moreover, CodeMatcher outperforms two existing online search engines (GitHub and Google search) by 46% and 33%, respectively, in terms of MRR. We also observed that: fusing the advantages of IR-based and DL-based models is promising; improving the quality of method naming helps code search, since method name plays an important role in connecting query and code.


2021 ◽  
Vol 2021 ◽  
pp. 1-16
Author(s):  
Mostafa K. El-Bably ◽  
Muhammad I. Ali ◽  
El-Sayed A. Abo-Tabl

There are many approaches to deal with vagueness and ambiguity including soft sets and rough sets. Feng et al. initiated the concept of possible hybridization of soft sets and rough sets. They introduced the concept of soft rough sets, in which parameterized subsets of a universe set serve as the building blocks for lower and upper approximations of a subset. Topological notions play a vital role in rough sets and soft rough sets. So, the basic objectives of the current work are as follows: first, we find answers to some very important questions, such as how to determine the probability that a subset of the universe is definable. Some more similar questions are answered in rough sets and their extensions. Secondly, we enhance soft rough sets from topological perspective and introduce topological soft rough sets. We explore some of their properties to improve existing techniques. A comparison has been made with some existing studies to show that accuracy measure of proposed technique shows an improvement. Proposed technique has been employed in decision-making problem for diagnosing heart failure. For this two algorithms have been given.


2021 ◽  
Vol 19 (4) ◽  
pp. e0210-e0210
Author(s):  
Tamara C. Maltauro ◽  

Aim of study: To evaluate the influence of the parameters of the geostatistical model and the initial sample configuration used in the optimization process; and to propose and evaluate the resizing of a sample configuration, reducing its sample size, for simulated data and for the study of the spatial variability of soil chemical attributes under a non-stationary with drift process from a commercial soybean cultivation area. Area of study: Cascavel, Brazil Material and methods: For both, the simulated data and the soil chemical attributes, the Genetic Algorithm was used for sample resizing, maximizing the overall accuracy measure. Main results: The results obtained from the simulated data showed that the practical range did not influence in a relevant way the optimization process. Moreover, the local variations, such as variance or sampling errors (nugget effect), had a direct relationship with the reduction of the sample size, mainly for the smaller nugget effect. For the soil chemical attributes, the Genetic Algorithm was efficient in resizing the sampling configuration, since it generated sampling configurations with 30 to 35 points, corresponding to 29.41% to 34.31% of the initial configuration, respectively. In addition, comparing the optimized and initial configurations, similarities were obtained regarding spatial dependence structure and characterization of spatial variability of soil chemical attributes in the study area. Research highlights: The optimization process showed that it is possible to reduce the sample size, allowing for lesser financial investments with data collection and laboratory analysis of soil samples in future experiments.


Author(s):  
Grant M. Walker ◽  
Alexandra Basilakos ◽  
Julius Fridriksson ◽  
Gregory Hickok

Purpose: Meaningful changes in picture naming responses may be obscured when measuring accuracy instead of quality. A statistic that incorporates information about the severity and nature of impairments may be more sensitive to the effects of treatment. Method: We analyzed data from repeated administrations of a naming test to 72 participants with stroke aphasia in a clinical trial for anomia therapy. Participants were divided into two groups for analysis to demonstrate replicability. We assessed reliability among response type scores from five raters. We then derived four summary statistics of naming ability and their changes over time for each participant: (a) the standard accuracy measure, (b) an accuracy measure adjusted for item difficulty, (c) an accuracy measure adjusted for item difficulty for specific response types, and (d) a distance measure adjusted for item difficulty for specific response types. While accuracy measures address the likelihood of a correct response, the distance measure reflects that different response types range in their similarity to the target. Model fit was assessed. The frequency of significant improvements and the average magnitude of improvements for each summary statistic were compared between treatment groups and a control group. Effect sizes for each model-based statistic were compared with the effect size for the standard accuracy measure. Results: Interrater and intrarater reliability were near perfect, on average, though compromised somewhat by phonological-level errors. The effects of treatment were more evident, in terms of both frequency and magnitude, when using the distance measure versus the other accuracy statistics. Conclusions: Consideration of item difficulty and response types revealed additional effects of treatment on naming scores beyond those observed for the standard accuracy measure. The results support theories that assume naming ability is decomposable into subabilities rather than being monolithic, suggesting new opportunities for measuring treatment outcomes. Supplemental Material https://doi.org/10.23641/asha.17019515


2021 ◽  
Vol 13 (21) ◽  
pp. 4424
Author(s):  
Mariusz Specht

In navigation, the Twice the Distance Root Mean Square (2DRMS) is commonly used as a position accuracy measure. Its determination, based on statistical methods, assumes that the position errors are normally distributed and are often not reflected in actual measurements. As a result of the widespread adoption of this measure, the positioning accuracy of navigation systems is overestimated by 10–15%. In this paper, a new method is presented for determining the navigation system positioning accuracy based on a reliability model where the system’s operation and failure statistics are referred to as life and failure times. Based on real measurements, the method proposed in this article will be compared with the classical method (based on the 2DRMS measure). Real (empirical) measurements made by the principal modern navigation positioning systems were used in the analyses: Global Positioning System (GPS) (168’286 fixes), Differential Global Positioning System (DGPS) (900’000 fixes) and European Geostationary Navigation Overlay Service (EGNOS) (900’000 fixes). Research performed on real data, many of which can be considered representative, have shown that the reliability method provides a better (compared to the 2DRMS measure) estimate of navigation system positioning accuracy. Thanks to its application, it is possible to determine the position error distribution of the navigation system more precisely when compared to the classical method, as well as to indicate those applications that can be used by this system, ensuring the safety of the navigation process.


Author(s):  
M.K. El-Bably ◽  
T.M. Al-shami ◽  
A.S. Nawar ◽  
A. Mhemdi

The main aims of this paper are to show that some results presented in [1] are erroneous. To this end, we provide some counterexamples to demonstrate our claim, and give the correct form of the incorrect results in [1]. Also, some improvements for the definition of accuracy measure is proposed. Furthermore, we show that the relationships given in the three figures need not be true in general, and determine the conditions under which they are correct. Finally, a medical application in the decision-making of the diagnosis of dengue fever is examined.


Symmetry ◽  
2021 ◽  
Vol 13 (11) ◽  
pp. 2010
Author(s):  
Muhammad Asim Bilal ◽  
Muhammad Shabir ◽  
Ahmad N. Al-Kenani

Yager recently introduced the q-rung orthopair fuzzy set to accommodate uncertainty in decision-making problems. A binary relation over dual universes has a vital role in mathematics and information sciences. During this work, we defined upper approximations and lower approximations of q-rung orthopair fuzzy sets using crisp binary relations with regard to the aftersets and foresets. We used an accuracy measure of a q-rung orthopair fuzzy set to search out the accuracy of a q-rung orthopair fuzzy set, and we defined two types of q-rung orthopair fuzzy topologies induced by reflexive relations. The novel concept of a rough q-rung orthopair fuzzy set over dual universes is more flexible when debating the symmetry between two or more objects that are better than the prevailing notion of a rough Pythagorean fuzzy set, as well as rough intuitionistic fuzzy sets. Furthermore, using the score function of q-rung orthopair fuzzy sets, a practical approach was introduced to research the symmetry of the optimal decision and, therefore, the ranking of feasible alternatives. Multiple criteria decision making (MCDM) methods for q-rung orthopair fuzzy sets cannot solve problems when an individual is faced with the symmetry of a two-sided matching MCDM problem. This new approach solves the matter more accurately. The devised approach is new within the literature. In this method, the main focus is on ranking and selecting the alternative from a collection of feasible alternatives, reckoning for the symmetry of the two-sided matching of alternatives, and providing a solution based on the ranking of alternatives for an issue containing conflicting criteria, to assist the decision-maker in a final decision.


2021 ◽  
Vol 2021 ◽  
pp. 1-13
Author(s):  
Mona Hosny

Ideal is a fundamental concept in topological spaces and plays an important role in the study of topological problems. This motivated us to use two ideals to generate different topologies to take the advantage of the two ideals at the same time. Two ideals represent two opinions instead of one opinion which is very useful for using the insights of two groups of experts to study the problem and elicit decisions based on their common vision. Topology is a rich source for constructs that is helpful to enrich the original model of approximations spaces. Rough set theory has inbuilt topological concepts. Hence, the main purpose of this paper is to point out that the concept of rough sets has a purely topological aspects nature. To do so, new approximations spaces are introduced and defined based on the topologies generated by two ideals. The results in this paper show that the topological concepts can be a powerful method to study rough set models. The basic properties of these approximations are studied and compared to the previous ones and shown to be more general. The importance of the current paper is not only introducing a new kind of rough set based on bi-ideals, increasing the accuracy measure, and reducing the boundary region of the sets which is the main aim of rough set but also introducing a chemical application to explain the concepts.


2021 ◽  
Vol 13 (17) ◽  
pp. 3488
Author(s):  
Keren Goldberg ◽  
Ittai Herrmann ◽  
Uri Hochberg ◽  
Offer Rozenstein

The overarching aim of this research was to develop a method for deriving crop maps from a time series of Sentinel-2 images between 2017 and 2018 to address global challenges in agriculture and food security. This study is the first step towards improving crop mapping based on phenological features retrieved from an object-based time series on a national scale. Five main crops in Israel were classified: wheat, barley, cotton, carrot, and chickpea. To optimize the object-based classification process, different characteristics and inputs of the mean shift segmentation algorithm were tested, including vegetation indices, three-band combinations, and high/low emphasis on the spatial and spectral characteristics. Four known vegetation indices (VIs)-based time series were tested. Additionally, we compared two widely used machine learning methods for crop classification, support vector machine (SVM) and random forest (RF), in addition to a newer classifier, extreme gradient boosting (XGBoost). Lastly, we examined two accuracy measures—overall accuracy (OA) and area under the curve (AUC)—in order to optimally estimate the accuracy in the case of imbalanced class representation. Mean shift best performed when emphasizing both the spectral and spatial characteristics while using the green, red, and near-infrared (NIR) bands as input. Both accuracy measures showed that RF and XGBoost classified different types of crops with significantly greater success than achieved by SVM. Nevertheless, AUC was better able to represent the significant differences between the classification algorithms than OA was. None of the VIs showed a significantly higher contribution to the classification. However, normalized difference infrared index (NDII) with XGBoost classifier showed the highest AUC results (88%). This study demonstrates that the short-wave infrared (SWIR) band with XGBoost improves crop type classification results. Furthermore, the study emphasizes the importance of addressing imbalanced classification datasets by using a proper accuracy measure. Since object-based classification and phenological features derived from a VI-based time series are widely used to produce crop maps, the current study is also relevant for operational agricultural management and informatics at large scales.


Sign in / Sign up

Export Citation Format

Share Document