Back
Projects

Machine-Generated Text Attribution
Trained BERT Sequence classifier on a novel dataset created using 5 generative models (GPT-2-small, GPT-2XL, Phi-2, Falcon-7B, Mistral-7B-it) by prompting with text sourced from Wikipedia and GSM8K Math dataset, achieving F1-score of 0.65. and studied the impact of input length, prompt domain, and LLM parameter size on attribution accuracy.
