Skip to content

This repository contains a notebook for fine-tuning the bart-large-xsum model using the BBC News Summary dataset from Kaggle. This project's goal is to enhance the summarization capabilities of the BART model, enabling it to generate concise and coherent summaries from news articles. If you're looking to fine-tune BART, this Repo is a good start.

Notifications You must be signed in to change notification settings

Seif-Yasser-Ahmed/Bart-Fine-Tuning-Text-Summarizer

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

11 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Bart Fine Tuning Text Summarizer

This model is a fine-tuned version of facebook/bart-large-xsum on the BBC News Summary dataset.

Downloads

If you'd like to use this model in a pipeline, you can load it easily with Hugging Face's transformers library:

from transformers import pipeline
summarizer = pipeline("summarization", model="Seif-Yasser/bart-large-xsum-finetuned-xsum")

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 16
  • eval_batch_size: 16
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 1
  • mixed_precision_training: Native AMP

Training results

Training Loss Epoch Step Validation Loss Rouge1 Rouge2 Rougel Rougelsum Gen Len
No log 1.0 84 0.8199 42.0809 35.1683 34.0117 35.9952 59.7371

Framework versions

  • Transformers 4.44.0
  • Pytorch 2.4.0
  • Datasets 2.21.0
  • Tokenizers 0.19.1

About

This repository contains a notebook for fine-tuning the bart-large-xsum model using the BBC News Summary dataset from Kaggle. This project's goal is to enhance the summarization capabilities of the BART model, enabling it to generate concise and coherent summaries from news articles. If you're looking to fine-tune BART, this Repo is a good start.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published