Skip to content

Transfer Learning for Text Classification using Transformer-based Models

Notifications You must be signed in to change notification settings

rachida-saroui/transfer-learning-text-classification

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 

Repository files navigation

Transfer Learning for Text Classification with Transformer-based Models

This repository contains the code and experiments for the ALTEGRAD 2024 project by Rachida Saroui, focusing on using transfer learning to improve text classification tasks. This project utilizes pre-trained transformer models and custom architectures such as AttentionBiGRU, HAN (Hierarchical Attention Networks), and TimeDistributed layers to build powerful classifiers for document classification tasks.

Key Features:

  • Transfer Learning: Leveraging pre-trained transformer models for efficient training on a smaller dataset.
  • Custom Architectures: Includes AttentionBiGRU and Hierarchical Attention Networks (HAN) for better document classification performance.
  • Dynamic Dataset Handling: A custom PyTorch dataset loader to efficiently manage and preprocess documents and labels.

Requirements

To run this project, you'll need the following dependencies:

  • Python 3.8+
  • PyTorch 1.10+
  • Transformers 4.3.0+
  • NumPy
  • Pandas
  • Matplotlib (optional, for plotting)
  • scikit-learn (for evaluation metrics)

About

Transfer Learning for Text Classification using Transformer-based Models

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published