Skip to content

1.10.1

Compare
Choose a tag to compare
@eladven eladven released this 01 Jul 08:04
· 357 commits to main since this release
59b0a62

Main Changes

  • Continued with major improvements to the documentation including a new code examples section with standalone python code that shows how to perform evaluation, add new datasets, compare formats, use LLM as judges , and more. Cards for datasets from huggingface have detailed descriptions. New documentation of RAG tasks and metrics.
  • load_dataset can now load cards defined in a python file (and not only in the catalog). See example.
  • The evaluation results returned from evaluate now include two fields predictions and processed_predictions. See example.
  • The fields can have defaults, so if they are not specified in the card, they get a default value. For example, multi-class classification has text as the default text_type. See example.

Non backward compatible changes

You need to recreate the any cards/metrics you added by running prepare//.py file. You can create all cards simply by running python utils/prepare_all_artifacts.py . This will avoid the type error.

The AddFields operator was renamed Set and CopyFields operator was renamed Copy. Note previous code should continue to work, but we renamed all existing code in the unitxt and fm-eval repos.

New Features

Bug Fixes

Documentation

New Assets

Testing and CI/CD

New Contributors

Full Changelog: 1.10.0...1.10.1