Evaluation code for various unsupervised automated metrics for Natural Language Generation.
-
Updated
Aug 20, 2024 - Python
Evaluation code for various unsupervised automated metrics for Natural Language Generation.
ROUGE automatic summarization evaluation toolkit. Support for ROUGE-[N, L, S, SU], stemming and stopwords in different languages, unicode text evaluation, CSV output.
Python implementation of ROUGE
Python ROUGE Score Implementation for Chinese Language Task (official rouge score)
A serverless set of functions for evaluating whether incoming messages to an LLM system seem to contain instances of prompt injection; uses cascading cosine similarity and ROUGLE-L calculation against known good and bad prompts
ROUGE L metric implementation using tensorflow ops
Evaluation tools for image captioning. Including BLEU, ROUGE-L, CIDEr, METEOR, SPICE scores.
Evaluation and agreement scripts for the DISCOSUMO project. Each evaluation script takes both manual annotations as automatic summarization output. The formatting of these files is highly project-specific. However, the evaluation functions for precision, recall, ROUGE, Jaccard, Cohen's kappa and Fleiss' kappa may be applicable to other domains too.
This is a Pytorch implementation of a summarization model that is fine-tuned on the top of Google-T5 pre-trained model.
Text summarization of Hotel reviews using Huggingface Transformers T5 and Bart
A news headline generator finetuned on T5-base.
Add a description, image, and links to the rouge-l topic page so that developers can more easily learn about it.
To associate your repository with the rouge-l topic, visit your repo's landing page and select "manage topics."