Edit model card

GPT-2 Fine-Tuned Model

This is a fine-tuned version of the GPT-2 model designed for text generation tasks. The model has been fine-tuned to improve its performance on generating coherent and contextually relevant text.

Model Details

  • Model Name: GPT-2 Fine-Tuned
  • Base Model: gpt2
  • Architecture: GPT2LMHeadModel
  • Tokenization: Supported
    • pad_token_id: 50256
    • bos_token_id: 50256
    • eos_token_id: 50256

Supported Tasks

This model supports the following task:

  • Text Generation

Configuration

Model Configuration (config.json)

  • Hidden Size: 768
  • Number of Layers: 12
  • Number of Attention Heads: 12
  • Vocab Size: 50257
  • Token Type IDs: Not used

Generation Configuration (generation_config.json)

  • Sampling Temperature: 0.7
  • Top-p (nucleus sampling): 0.9
  • Pad Token ID: 50256
  • Bos Token ID: 50256
  • Eos Token ID: 50256

Usage

To use this model for text generation via the Hugging Face API, use the following Python code snippet:

import requests

api_url = "https://api-inference.huggingface.co/models/rahul77/gpt-2-finetune"
headers = {
    "Authorization": "Bearer YOUR_API_TOKEN",  # Replace with your Hugging Face API token
    "Content-Type": "application/json"
}

data = {
    "inputs": "What is a large language model?",
    "parameters": {
        "max_length": 50
    }
}

response = requests.post(api_url, headers=headers, json=data)

if response.status_code == 200:
    print(response.json())
else:
    print(f"Error: {response.status_code}")
    print(response.json())
Downloads last month
58
Inference Examples
Inference API (serverless) is not available, repository is disabled.

Model tree for rahul77/gpt-2-finetune

Finetuned
this model

Dataset used to train rahul77/gpt-2-finetune