نتایج جستجو برای: evaluation metrics

تعداد نتایج: 878773  

2015
Radek Pelánek

Researchers use many different metrics for evaluation of performance of student models. The aim of this paper is to provide an overview of commonly used metrics, to discuss properties, advantages, and disadvantages of different metrics, to summarize current practice in educational data mining, and to provide guidance for evaluation of student models. In the discussion we mention the relation of...

Journal: :CoRR 2018
Mucahid Kutlu Vivek Khetan Matthew Lease

Because researchers typically do not have the time or space to present more than a few evaluation metrics in any published study, it can be difficult to assess relative effectiveness of prior methods for unreported metrics when baselining a new method or conducting a systematic meta-review. While sharing of study data would help alleviate this, recent attempts to encourage consistent sharing ha...

طباطبائی, راضیه, فیضی درخشی, محمدرضا, معصومی, سعید,

Nowadays summarizers and machine translators have attracted much attention to themselves, and many activities on making such tools have been done around the world. For Farsi like the other languages there have been efforts in this field. So evaluating such tools has a great importance. Human evaluations of machine summarization are extensive but expensive. Human evaluations can take months to f...

2007
Ricardo Vilalta

This paper provides a characterization of bias for evaluation metrics in classiica-tion (e.g., Information Gain, Gini, 2 , etc.). Our characterization provides a uniform representation for all traditional evaluation metrics. Such representation leads naturally to a measure for the distance between the bias of two evaluation metrics. We give a practical value to our measure by observing if the d...

2000
Ricardo Vilalta Daniel Oblinger

This paper provides a characterization of bias for evaluation metrics in classiication (e.g., Information Gain, Gini, 2 , etc.). Our characterization provides a uniform representation for all traditional evaluation metrics. Such representation leads naturally to a measure for the distance between the bias of two evaluation metrics. We give a practical value to our measure by observing if the di...

Journal: :Computational Intelligence 2003
Ricardo Vilalta Daniel Oblinger

This paper provides a characterization of bias for evaluation metrics in classification (e.g., Information Gain, Gini, χ, etc.). Our characterization provides a uniform representation for all traditional evaluation metrics. Such representation leads naturally to a measure for the distance between the bias of two evaluation metrics. We give a practical value to our measure by observing the dista...

2016
Josiah Wang Robert J. Gaizauskas

The task of automatically generating sentential descriptions of image content has become increasingly popular in recent years, resulting in the development of large-scale image description datasets and the proposal of various metrics for evaluating image description generation systems. However, not much work has been done to analyse and understand both datasets and the metrics. In this paper, w...

2006
Jie Yu Jaume Amores Nicu Sebe Qi Tian

In this work, we present a general guideline to establish the relation between a distribution model and its corresponding similarity estimation. A rich set of distance metrics, such as Harmonic distance and Geometric distance, is derived according to Maximum Likelihood theory. These metrics can provide a more accurate model than the conventional Euclidean distance and Manhattan distance. Becaus...

2014
Francisco Guzmán Shafiq R. Joty Lluís Màrquez i Villodre Preslav Nakov

We present experiments in using discourse structure for improving machine translation evaluation. We first design two discourse-aware similarity measures, which use all-subtree kernels to compare discourse parse trees in accordance with the Rhetorical Structure Theory. Then, we show that these measures can help improve a number of existing machine translation evaluation metrics both at the segm...

Journal: :CoRR 2013
Robin Aly Maria Eskevich Roeland Ordelman Gareth J. F. Jones

This report describes metrics for the evaluation of the effectiveness of segment-based retrieval based on existing binary information retrieval metrics. This metrics are described in the context of a task for the hyperlinking of video segments. This evaluation approach re-uses existing evaluation measures from the standard Cranfield evaluation paradigm. Our adaptation approach can in principle ...

نمودار تعداد نتایج جستجو در هر سال

با کلیک روی نمودار نتایج را به سال انتشار فیلتر کنید