Update README.md
Browse files
README.md
CHANGED
|
@@ -3,34 +3,47 @@ library_name: transformers
|
|
| 3 |
base_model: google/pegasus-xsum
|
| 4 |
tags:
|
| 5 |
- generated_from_trainer
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 6 |
model-index:
|
| 7 |
- name: a-text-summarizer
|
| 8 |
results: []
|
| 9 |
---
|
| 10 |
|
| 11 |
-
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
| 12 |
-
should probably proofread and complete it, then remove this comment. -->
|
| 13 |
-
|
| 14 |
# a-text-summarizer
|
| 15 |
|
| 16 |
-
This model is a fine-tuned version of
|
|
|
|
| 17 |
It achieves the following results on the evaluation set:
|
| 18 |
- Loss: 2.3989
|
| 19 |
|
| 20 |
## Model description
|
| 21 |
|
| 22 |
-
|
| 23 |
|
| 24 |
## Intended uses & limitations
|
| 25 |
|
| 26 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 27 |
|
| 28 |
## Training and evaluation data
|
| 29 |
|
| 30 |
-
|
|
|
|
|
|
|
|
|
|
| 31 |
|
| 32 |
## Training procedure
|
| 33 |
|
|
|
|
|
|
|
| 34 |
### Training hyperparameters
|
| 35 |
|
| 36 |
The following hyperparameters were used during training:
|
|
|
|
| 3 |
base_model: google/pegasus-xsum
|
| 4 |
tags:
|
| 5 |
- generated_from_trainer
|
| 6 |
+
- summarization
|
| 7 |
+
- transformers
|
| 8 |
+
- fine-tuned
|
| 9 |
+
- google-pegasus-xsum
|
| 10 |
+
- ccdv/govreport-summarization
|
| 11 |
+
|
| 12 |
model-index:
|
| 13 |
- name: a-text-summarizer
|
| 14 |
results: []
|
| 15 |
---
|
| 16 |
|
|
|
|
|
|
|
|
|
|
| 17 |
# a-text-summarizer
|
| 18 |
|
| 19 |
+
This model is a fine-tuned version of the google/pegasus-xsum model (https://huggingface.co/google/pegasus-xsum).
|
| 20 |
+
It has been trained to generate summaries for governmental reports based on the GovReport summarization dataset (https://huggingface.co/datasets/ccdv/govreport-summarization).
|
| 21 |
It achieves the following results on the evaluation set:
|
| 22 |
- Loss: 2.3989
|
| 23 |
|
| 24 |
## Model description
|
| 25 |
|
| 26 |
+
This is a summarization model fine-tuned on the ccdv/govreport-summarization dataset.
|
| 27 |
|
| 28 |
## Intended uses & limitations
|
| 29 |
|
| 30 |
+
This model is intended for generating concise summaries of governmental reports or similar long-form documents in an official or formal American English register.
|
| 31 |
+
|
| 32 |
+
The model's performance is limited by the data it was trained on (GovReport summarization dataset). It may not generalize well to other domains or types of text.
|
| 33 |
+
Summarization models can sometimes hallucinate information or produce summaries that are not entirely accurate.
|
| 34 |
+
Potential biases present in the training data may be reflected in the generated summaries. Further analysis is needed to identify and mitigate potential biases.
|
| 35 |
|
| 36 |
## Training and evaluation data
|
| 37 |
|
| 38 |
+
The model was fine-tuned on a subset of the ccdv/govreport-summarization dataset.
|
| 39 |
+
Specifically, a subset of 5000 training examples and 500 validation examples were used for fine-tuning.
|
| 40 |
+
|
| 41 |
+
The GovReport dataset contains governmental reports and their corresponding summaries.
|
| 42 |
|
| 43 |
## Training procedure
|
| 44 |
|
| 45 |
+
The model was fine-tuned using the Hugging Face transformers library and Trainer API.
|
| 46 |
+
|
| 47 |
### Training hyperparameters
|
| 48 |
|
| 49 |
The following hyperparameters were used during training:
|