Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Questions regarding the pre-trained models #3

Open
SargamMenghani opened this issue Sep 13, 2019 · 1 comment
Open

Questions regarding the pre-trained models #3

SargamMenghani opened this issue Sep 13, 2019 · 1 comment

Comments

@SargamMenghani
Copy link

SargamMenghani commented Sep 13, 2019

Hello, we had a few questions regarding the model weights that have been provided:

  1. We tried to evaluate the Lakh400kPretrainOnly model (using the reproduce_paper_eval.sh script) and got the following results (because we ran this using reproduce_paper_eval.sh, the validation and test sets that we are using are from the nesmdb dataset):
valid loss  1.86 | valid ppl  6.448 | test loss  1.71 | test ppl  5.541 

Can you please confirm that these match with what you were seeing?

  1. We also tried to generate a few samples using the Lakh400kPretrainOnly model. Out of the 25 samples that we generated, 15 samples only had the "WT" and "NO" channel notes and no "P1"/"P2"/"TR" channel notes. Do you have an intuition as to why this might be happening?

  2. We are also having trouble in generating chiptunes from the LakhNES model. The chiptunes don't sound correct to us and we think that we might be missing some steps in the generation process or performing a step incorrectly. Here are the steps that we followed for generating the chiptunes (if you can point to us anything that we are missing/doing incorrectly in this process, it will help us a lot.) :
    a) we downloaded the LakhNES model from the link provided in the repo
    b) we ran python generate.py model/pretrained/LakhNES/ --out_dir ./generated/LakhNES --num 25 (this ran without any errors)
    c) we used tx1_to_midi function (in tx1_midi.py script) to convert the generated tx1 files to midi (we then used the timidity software to listen to these midi files)

If you are able to provide the 775k LakhNES data (by LakhNES data we mean the LakhMIDI examples mapped to NES channels) that you used for pre-training, it will help us a lot. We have generated our own 775k LakhNES examples using your scripts but because there is some randomness involved in mapping the instruments, we are not entirely sure if what we have matches with what you had used.

Thank you in advance!

@ruohoruotsi
Copy link

Hi there,
using the Lakh400kPretrainOnly, I also got similar results to you (using the ./reproduce....sh script)

====================================================================================================
| valid loss  1.86 | valid ppl     6.438 | test loss  1.72 | test ppl     5.601 
====================================================================================================

I'm about to use it to generate some some samples, so I don't have responses to 2. or 3. yet.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants