Skip to content

Releases: obss/jury

v2.1.2

14 Nov 12:08
2cf2de8
Compare
Choose a tag to compare

What's Changed

Full Changelog: 2.1.1...2.1.2

v2.1.1

10 Nov 12:53
c6db3e8
Compare
Choose a tag to compare

What's Changed

Full Changelog: 2.1.0...2.1.1

v2.1.0

25 Oct 15:10
8801eb1
Compare
Choose a tag to compare

What's New 🚀

Tasks 📝

We added task based new metric system which allows us to evaluate different type of inputs rather than old system which could only evaluate from strings (generated text) for only language generation tasks. Hence, jury now is able to support broader set of metrics works with different types of input.

With this, on jury.Jury API, the consistency of set of tasks given is under control. Jury will raise an error if any pair of metrics are not consistent with each other in terms of task (evaluation input).

AutoMetric ✨

  • AutoMetric is introduced as a main factory class for automatically loading metrics, as a side note load_metric is still available for backward compatibility and is preferred (it uses AutoMetric under the hood).
  • Tasks are now distinguished within metrics. For example, precision can be used for language-generation or sequence-classification task, where one evaluates from string (generated text) while other one evaluates from integers (class labels).
  • On configuration file, metrics can be now stated with HuggingFace's datasets' metrics initializiation parameters. The keyword arguments for metrics that are used on computation are now separated in "compute_kwargs" key.

Full Changelog: 2.0.0...2.1.0

v2.0.0

11 Oct 17:47
bca6ca4
Compare
Choose a tag to compare

Jury 2.0.0 is out 🎉🥳

New Metric System

  • datasets package Metric implementation is adopted (and extended) to provide high performance 💯 and more unified interface 🤗.
  • Custom metric implementation changed accordingly (it now requires 3 abstract methods to be implemented).
  • Jury class is now callable (implements call() method to be used thoroughly) though evaluate() method is still available for backward compatibility.
  • In the usage of evaluate of Jury, predictions and references parameters are restricted to be passed as keyword arguments to prevent confusion/wrong computations (like datasets' metrics).
  • MetricCollator is removed, the methods for metrics are attached directly to Jury class. Now, metric addition and removal can be performed from a Jury instance directly.
  • Jury now supports reading metrics from string, list and dictionaries. It is more generic to input type of metrics given along with parameters.

New metrics

  • Accuracy, F1, Precision, Recall are added to Jury metrics.
  • All metrics on datasets package are still available on jury through the use of jury.load_metric()

Development

  • Test cases are improved with fixtures, and test structure is enchanced.
  • Expected outputs are now required for tests as a json with proper name.

v1.1.2

15 Sep 13:26
Compare
Choose a tag to compare
  • SQuAD bug fixed for evaluating with multiple references.
  • Test design & cases revised with fixtures (improvement).

v1.1.1

15 Aug 10:54
Compare
Choose a tag to compare
  • Malfunctioning multiple prediction calculation caused by multiple reference input for BLEU and SacreBLEU is fixed.
  • CLI Implementation is completed. 🎉

v1.0.1

13 Aug 17:22
Compare
Choose a tag to compare
  • Fix for nltk version (Colab is fixed as well).

v1.0.0

09 Aug 16:29
Compare
Choose a tag to compare

Release Notes

  • New metric structure is completed.
    • Custom metric support is improved and no longer required to extend datasets.Metric, rather uses jury.metrics.Metric.
    • Metric usage is unified with compute, preprocess and postprocess functions, which the only required implementation for custom metric is compute.
    • Both string and Metric objects can be passed to Jury(metrics=metrics) now in a mixed fashion.
    • load_metric function was rearranged to capture end score results and several metrics added accordingly (e.g. load_metric("squad_f1") will load squad metric which returns F1-score).
  • Example notebook has added to example.
    • MT and QA tasks were illustrated.
    • Custom metric creation added as example.

Acknowledgments

@fcakyon @cemilcengiz @devrimcavusoglu

v0.0.6

28 Jul 21:21
f06bba7
Compare
Choose a tag to compare
Release v0.0.6 (#28)

Co-authored-by: cemilcengiz <cemil.cengiz94@gmail.com>

v0.0.5

27 Jul 10:38
6e85666
Compare
Choose a tag to compare
Jury moved. black added. (#17)

* requirements.txt updated (tqdm version loosen).

* requirements.txt updated. Removed packages that `datasets` depends on.

* Jury moved under __init__.py. black package added to requirements-dev.txt.

* v0.0.5

* requirements.txt updated. Removed packages that `datasets` depends on.

* Jury moved under __init__.py. black package added to requirements-dev.txt.

* v0.0.5