Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fine-tuning codellama with own dataset for code review #257

Open
DAP5555 opened this issue Dec 15, 2024 · 0 comments
Open

Fine-tuning codellama with own dataset for code review #257

DAP5555 opened this issue Dec 15, 2024 · 0 comments

Comments

@DAP5555
Copy link

DAP5555 commented Dec 15, 2024

Hi :)
I am planning to fine tune Code LLaMA 7B using my own dataset to evaluate its effectiveness for code review tasks. I came across a paper where the authors attempted a similar approach using data from CodeReviewer (https://arxiv.org/abs/2203.09095) .

My plan is to gather data from the pull requests in our repositories and create JSON files in the following format:
{
"oldf": "... contents of another old file ...",
"patch": "@@ -25,13 +25,16 @@ ...", // Diffs from the repo
"msg": "we call cities + towns ...",
"id": 12959,
"y": 1
}
Will this data format work for fine tuning Code LLaMA, or will I need to adjust the dataset to a specific format for it to be compatible?
Thanks in advance

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant