Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fine-tuning or Retrain #8

Open
mahsayedsalem opened this issue Jan 25, 2019 · 2 comments
Open

Fine-tuning or Retrain #8

mahsayedsalem opened this issue Jan 25, 2019 · 2 comments

Comments

@mahsayedsalem
Copy link

Thank you for the great work!

I wonder if I needed to fine tune the model or retrain it, how to pre-process my data? Specially the network's output, what does the "28" vector represents?

@Barqawiz
Copy link
Owner

Barqawiz commented Jan 25, 2019

To preprocess the data for training or prediction use:
def prepare_input(self, input_sent):

If you want information about each input max length, check following code part (each model version tried different length):

if version == 1: self.max_sentence = 495 elif version == 2: self.max_sentence = 315 elif version == 3: self.max_sentence = 315

For more information about outputs, print (output_int_to_vocab) array:
output_int_to_vocab = helper.load_binary('output_int_to_vocab',dictionary_folder)

Thanks for the interest in the library will try to include a notebook about fine-tuning for better contribution.

@AbdallahNasir
Copy link

Thanks @Barqawiz

prepare_input will convert the string to integers, I guess what we are looking for is a method to load the final text, with diacritics, input, and create two files, one plain, and the result, which contains the list of diacritics, the same thing that the keras model outputs after using logits_to_text.

For sure you had used some preparation scripts. Having those is helpful if you are willing to share. If it is okay, please :)

It may be too much to ask for the training code I know :D

Best wishes,

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants