Skip to content

Basic question on fine tuning...my results are not even close... #84

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
mikedorin opened this issue Mar 25, 2025 · 0 comments
Open

Basic question on fine tuning...my results are not even close... #84

mikedorin opened this issue Mar 25, 2025 · 0 comments

Comments

@mikedorin
Copy link

I have a small dataset. I have about 100 words in the train directory and 31 words in the validate directory.
I run the steps for a fine-tune with no issues.
It runs for 90 epochs then stops, the output is below.

I run:
python -m allosaurus.run --lang bura1267 --model=mad999 --device_id=1 -i train/index_7_botella.wav
I sort of expect to see something in the ballpark of : ah ak al ad aq ar

instead, I see:
ag ab

I have updated bura1267 with my own list of phones, by the way.

Do I need more data? Do I need to prep my wave files better? Any suggestions?
Thank you,
Mike

training results:
saving model
epoch[batch]: 88[0000] | train loss 3.06233 train per 0.78426
epoch88 | validate per : 0.94444
epoch[batch]: 89[0000] | train loss 3.04346 train per 0.80758
epoch89 | validate per : 0.93056
epoch[batch]: 90[0000] | train loss 3.02284 train per 0.79883
epoch90 | validate per : 0.93750
no improvements for several epochs, early stopping now

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant