Opts need a major cleanup #60
Labels
bug
Something isn't working
documentation
Improvements or additions to documentation
enhancement
New feature or request
Going through the existing catalogue of options listed in our docs, a number of them seem to not be plugged in.
The list below is most likely not exhaustive.
Data/Tasks
skip_empty_level
: would be a useful feature but is actually currently not supported. I would suggest replacing with a transform if we ever end up needing it.Vocab
share_vocab
is likely not functional (in tests, in build_vocab.py that we no longer use, and in one transform)vocab_paths
,vocab_size_multiple
, and{src,tgt}_min_word_frequency}
are only listed in opts.pyPruning
The flags here are not plugged to anything, Seem to be somewhat duplicate with Transform/Filter (first)
Embeddings
All flags here have been nuked. There seems to be a fixme in modules/embeddings.py. I'd suggest commenting them out with a fixme and creating an issue, if still relevant
Transform/Filter
We have multiple groups with the same name, which isn't great. They're likely functional but need to be renamed according to the transform they actually correspond to
Model- Embeddings
I am not sure why this is a separate group from Embedding.
position_encoding
True by defaultshare_embeddings
doesn't seem to be usedupdate_vocab
doesn't seem to be usedModel-Embedding Features
We don't support features. This should go.
Model- Task
We don't support CLM / no-decoder models. This should go.
Model- Encoder-Decoder
model_type
only offers text option; seems duplicate with General-data_type
. Do we want to keep it in place given Supporting other data types (e.g. video) #53 ? Do we want to remove it for now?encoder_type
I don't know that we support mean encoders?decoder_type
only has one option. Might be relevant for External dependencies for layer architectures #56layers
is deprecated. can we remove it?{enc,dec}_layers
: should mention stacks or modules in their help textbridge
,n_node
,n_steps
: misleading and almost certainly only in optsbridge_extra_node
,bidir_edges
,state_dim
,n_edge_types
,src_ggnn_size
: only in optsModel- Attention
global_attention
,global_attention_function
: only in optsmax_relative_positions
andaan_useffn
seems like they are used, but i have doubts? at least I'm not familiar with those.Model - Alignement
There's a typo in the opts group name (took me a while, it's the expected French orthography). Otherwise seems supported, would be worth a check later on as I'm not familiar with these.
Generator
i recall the copy_attn being half functional in onmt, would need a deeper check
General
save_all_gpus
: do we want to support that?data_type
: cf. above wrtmodel_type
in Encoder-Decodergpuid
deprecated and should be removedgpu_backend
has only one option. Do we want to support anything else than NCCL?Initialization
pre_word_vecs_{enc,dec}
only in opts, same as with flags in Embeddings aboveOptimization- Type
This contain both flags regarding batching (batch size, batch type, etc) and optimizers hparams, as well as miscellanies like early stopping, etc. Not sure why it's separate from the Optimization- Rate group.
truncated_decoder
is apparently not compatible withgpu_ranks
/ accum, so it should probably be removedLogging
report_stats_from_parameters=
has a trailing equal signDynamic data
Should be merged with batching options
The text was updated successfully, but these errors were encountered: