This project explores new T5 model by Hugging Face (https://huggingface.co/docs/transformers/model_doc/t5).
You can download ConceptNet assertions (/~https://github.com/commonsense/conceptnet5/wiki/Downloads) and save in "data" folder.
Pre-processing steps:
- Filter out all non-english assertions
- Randomly select 5 triplets from each Subject
- Randomly mask 2 words in each sentence
The pre-training is done following the example of T5 (https://huggingface.co/docs/transformers/model_doc/t5#training) on Masked Language Modeling (MLM) using previously pre-processed ConceptNet triplets.
In the end T5 model is fine-tuned on TellMeWhy dataset (https://stonybrooknlp.github.io/tellmewhy/) for Q&A task.