The Development of a Sepedi Text Generation Model Using Transformers
| dc.contributor.author | Ramalepe, Simon P | |
| dc.contributor.author | Modipa, Thipe I | |
| dc.contributor.author | Davel, Marelie H | |
| dc.date.accessioned | 2023-07-31T07:42:56Z | |
| dc.date.available | 2023-07-31T07:42:56Z | |
| dc.date.issued | 2022 | |
| dc.description.abstract | Text generation is one of the important sub-tasks of natural language generation (NLG), and aims to produce humanly readable text given some input text. Deep learning approaches based on neural networks have been proposed to solve text generation tasks. Although these models can generate text, they do not necessarily capture long-term dependencies accurately, making it difficult to coherently generate longer sentences. Transformer-based models have shown significant improvement in text generation. However, these models are computationally expensive and data hungry. In this study, we develop a Sepedi text generation model using a Transformer based approach and explore its performance. The developed model has one Transformer block with causal masking on the attention layers and two separate embedding layers. To train the model, we use the National Centre for Human Language Technology (NCHLT) Sepedi text corpus. Our experimental setup varied the model embedding size, batch size and the sequence length. The final model was able to reconstruct unseen test data with 75% accuracy: the highest accuracy achieved to date, using a Sepedi corpus. | en_US |
| dc.description.sponsorship | Southern Africa Telecommunication Networks and Applications Conference (SATNAC) 2022 | en_US |
| dc.identifier.citation | Ramalepe, SM et.al.2022.The Development of a Sepedi Text Generation Model Using Transformers | en_US |
| dc.identifier.uri | http://hdl.handle.net/10394/41890 | |
| dc.language.iso | en | en_US |
| dc.publisher | Southern Africa Telecommunication Networks and Applications Conference (SATNAC) 2022 | en_US |
| dc.subject | Transformers | en_US |
| dc.subject | Generative pre-trained Trans former | en_US |
| dc.subject | Natural Language Generation | en_US |
| dc.subject | Text generation | en_US |
| dc.title | The Development of a Sepedi Text Generation Model Using Transformers | en_US |
| dc.type | Article | en_US |
Files
Original bundle
1 - 1 of 1
Loading...
- Name:
- Ramalepe-2022-development-sepedi-text.pdf
- Size:
- 690.43 KB
- Format:
- Adobe Portable Document Format
- Description:
License bundle
1 - 1 of 1
Loading...
- Name:
- license.txt
- Size:
- 1.61 KB
- Format:
- Item-specific license agreed upon to submission
- Description:
