Skip to content
This repository has been archived by the owner on Jul 26, 2022. It is now read-only.

Course project on prompt engineering for automated verifiability checking of online user comments.

Notifications You must be signed in to change notification settings

RamonMeffert/ltp-verifiability

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

62 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

A Comparison of Prompt Engineering Methods for Checking Verifiability of Online User Comments

Md Zobaer Hossain, Linfei Zhang, Robert van Timmeren and Ramon Meffert, June 2022

This repository contains the source code for the experiments, data processing and data analysis conducted as part of our course project for the 2021–2022 edition of the Language Technology Project course at the University of Groningen.

Data

All files related to the datasets are located in the datasets folder. We have taken the original dataset files and transformed them into the HuggingFace dataset format. All dataset folders contain the original dataset files, an analysis notebook and a demo file showing how you use the dataset.

Experiments

All code for experiments is located in the experiments folder. Information on how to reproduce the experiments is available in the readme in that folder.

Results

The results for all methods can be found in the results folder. Information about the results is available in the readme in that folder.


References

Black, S., G. Leo, P. Wang, C. Leahy, and S. Biderman (2021, March). GPT-Neo: Large scale autoregressive language modelling with mesh-tensorflow. https://doi.org/105281/zenodo.5297715.

Devlin, J., M.-W. Chang, K. Lee, and K. Toutanova (2019, June). BERT: Pre- training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), Minneapolis, Minnesota, pp. 4171–4186. Association for Computational Linguistics.

Gao, T., A. Fisch, and D. Chen (2021, August). Making pre-trained language models better few-shot learners. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), Online, pp. 3816–3830. Association for Computational Linguistics.

Liu, Y., M. Ott, N. Goyal, J. Du, M. Joshi, D. Chen, O. Levy, M. Lewis, L. Zettlemoyer, and V. Stoyanov (2019). RoBERTa: A robustly optimized BERT pretraining approach. CoRR abs/1907.11692.

Park, J., & Cardie, C. (2014). Identifying Appropriate Support for Propositions in Online User Comments. Proceedings of the First Workshop on Argumentation Mining, 29–38. https://doi.org/10/gg29gq

Schick, T. and H. Schütze (2021). Exploiting Cloze-Questions for Few-Shot Text Classification and Natural Language Inference. In Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, Online, pp. 255–269. Association for Computational Linguistics.

About

Course project on prompt engineering for automated verifiability checking of online user comments.

Topics

Resources

Stars

Watchers

Forks