| | --- |
| | language: en |
| | tags: |
| | - exbert |
| |
|
| | license: gpl |
| | --- |
| | |
| |
|
| | # TCP 2023 for NTU students |
| |
|
| | Fine tuning pre-trained language models for text generation. |
| |
|
| | Pretrained model on Chinese language using a GPT2 for Large Language Head Model objective(GPT2LMHeadModel). |
| |
|
| | ## Model description |
| |
|
| | TCP 2023 is a transformers model that has undergone fine-tuning using the GPT-2 architecture. |
| | It was initially pretrained on an extensive corpus of Chinese data in a self-supervised manner. |
| | This implies that the pretraining process involved using raw text data without any human annotations, allowing the model to make use of a wide range of publicly available data. |
| | The model leveraged an automatic process to derive inputs and corresponding labels from these texts. |
| | To be more specific, the pretraining aimed at predicting the subsequent word in sentences. |
| | it was trained to guess the next word in sentences. |
| |
|
| | ### How to use |
| |
|
| | You can use this model directly with a pipeline for text generation. Since the generation relies on some randomness, we |
| | set a seed for reproducibility: |
| |
|
| | ```python |
| | >>> from transformers import GPT2LMHeadModel, AutoTokenizer, pipeline |
| | |
| | >>> model_name = "DavidLanz/tcp2023" |
| | |
| | >>> model = GPT2LMHeadModel.from_pretrained(model_name) |
| | >>> tokenizer = AutoTokenizer.from_pretrained(model_name) |
| | |
| | >>> text_generator = pipeline('text-generation', model=model, tokenizer=tokenizer) |
| | >>> generated_text = text_generator(input_text, max_length=max_len, num_return_sequences=1) |
| | >>> print(generated_text[0]['generated_text']) |
| | |