Skip to content

This GitHub repo fine-tunes DistilGPT-2 to generate stories from prompts. It includes model selection, hyperparameter setup, and data preprocessing for efficient and creative story generation.

Notifications You must be signed in to change notification settings

KamleshKumar427/DistilGpt2_fineTunning

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 
 
 
 
 

Repository files navigation

DistilGpt2_fineTunning

This GitHub repository hosts a comprehensive project that **fine-tunes the DistilGPT-2 **model for generating coherent and creative stories from prompts. The repository includes model selection code using Hugging Face's Transformers library, specifying the "distilgpt2" model, and configuring essential hyperparameters for training.

The training process is well-documented in the given report, with key settings like the number of training epochs, learning rate, batch sizes, and more. Additionally, data preprocessing steps, such as text cleaning and bracket removal, are incorporated, ensuring high-quality input data for the model.

This repository provides a complete solution for training and inferencing a story generation model.

**NOTE: Finetunned models can be found here: You can download and use: **

https://drive.google.com/drive/folders/109xaME-XbXxAACZmXkI-X4fKUhHYdJlu?usp=sharing

Two tunned models:

  1. "final_model" (trained with 30K examples for 10 epochs) (Best)
  2. "1_epoch_final_model" (train with 273K examples for 1 epoch only)

About

This GitHub repo fine-tunes DistilGPT-2 to generate stories from prompts. It includes model selection, hyperparameter setup, and data preprocessing for efficient and creative story generation.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published