Configuring sequifier preprocessing, training and inference



Sequifier enables training and infering many-to-many sequence autoregression models using the decoder-only transformer architecture. It also enables the preprocessing of the data into from a parsimonious representation into the one consumed by the training step, while model inference outputs the data with the same representation as fed into the preprocessing step. Each of these steps is configurable in many ways, and this document lays out what the configuration parameters are, what purpose they serve, and any relevant considerations in setting them.


Preprocessing



Training


General training config parameters

Model Spec


Training Spec


Data driven config parameters

These parameters are typically read from the data driven config file created through the preprocessing step, with the path to that config file passed in the training config. For transparency and in case the preprocessing step isn't used, they are listed here.



Inference

There are additional parameters to the inference step that are typically read from the data driven config, as listed above. If the data driven config path isn't passed, these need to be set manually for inference as well