| language: en | |
| tags: | |
| - text-simplification | |
| - bart | |
| license: apache-2.0 | |
| # Text Simplification Model (H100 Trained) | |
| ## Training Results | |
| - **Training Loss**: 0.2796 | |
| - **Training Time**: 22:39 (3 epochs) | |
| - **Dataset**: GEM/wiki_auto_asset_turk (483,801 samples) | |
| - **GPU**: NVIDIA H100 80GB | |
| - **Batch Size**: 64 | |
| ## Usage | |
| ```python | |
| from transformers import BartTokenizer, BartForConditionalGeneration | |
| model = BartForConditionalGeneration.from_pretrained("Lorobert/text-simplification-runpod") | |
| tokenizer = BartTokenizer.from_pretrained("Lorobert/text-simplification-runpod") | |
| text = "Complex sentence here." | |
| inputs = tokenizer(text, return_tensors="pt", max_length=128, truncation=True) | |
| outputs = model.generate(**inputs, max_length=128, num_beams=4) | |
| simplified = tokenizer.decode(outputs[0], skip_special_tokens=True) | |
| print(simplified) | |
| ``` | |