You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
-[Evaluation](documentation/evaluation.md) – metrics and evaluation utilities.
21
26
22
27
## Installation and Development
23
28
24
29
Please see [CONTRIBUTING.md](CONTRIBUTING.md)
25
30
26
-
27
31
## Usage
28
32
Once you have [setup TGI container](#setting-up-the-tgi-container-with-hugging-face-models), you can proceed to score and the documents and trainer and classifier
29
33
@@ -32,12 +36,28 @@ Once you have [setup TGI container](#setting-up-the-tgi-container-with-hugging-f
Outputs: `${source_filename}.jsonl` with predicted scores in `annotated_data/`.
59
+
60
+
### 5. Measure Interrater Reliability
41
61
If you have a dataset with scores annotated by multiple annotators, you can compute metrics to measure the interrater reliability with the command interrater_reliability. If you want to compare the scores in a single file (e.g. the human annotated ground truth data), run:
This service relies on **TGI containers** (Text Generation Inference), which can be downloaded from [Hugging Face](https://huggingface.co). Follow the steps below to download and run the TGI container.
0 commit comments