My goal is to fine-tune a pre-trained LLM on a dataset about Manchester United's (MU's) 2021/22 season (they had a poor season). I want to be able to prompt the fine-tuned model with questions such as "How can MU improve?", or "What are MU's biggest weaknesses?". The ideal responses would be insightful/logical and +100 words
You can try to see how far you can get with LLMs and prompting (e.g., use Alpaca-LoRA or libraries like LangChain and FastChat).
However, if you want to persist with an approach similar to your current one, given the limited data you have, I would highly recommend considering a zero-shot approach. This means you must fine-tune your T5 model on a large Q&A dataset that is unrelated to your problem domain, and then test it on your current annotated data. If you are satisfied with the model's performance, you can stop there.
You can refer to my paper To tune or not to tune? Zero-shot models for legal case entailment, where I deal with a very similar problem. The conclusion of the paper is that if you don't have enough data for fine-tuning, it is sometimes better to simply forgo the target domain and fine-tune your models on a well-established dataset, even if it may be on a completely different subject.
As for how you should structure your test data, I can't provide a specific answer because it's highly dependent on what is happening in your code. It's difficult to prescribe what kind of preprocessing should be done in a high-level discussion like this.
If you love us? You can donate to us via Paypal or buy me a coffee so we can maintain and grow! Thank you!
Donate Us With