Handles everything you need to assemble a mini-batch of inputs and targets, as well as decode the dictionary produced
HF_CausalLMBeforeBatchTransform(
hf_arch,
hf_tokenizer,
max_length = NULL,
padding = TRUE,
truncation = TRUE,
is_split_into_words = FALSE,
n_tok_inps = 1,
ignore_token_id = -100,
...
)
architecture
tokenizer
maximum length
padding or not
truncation or not
to split into words
number tok inputs
ignore token id
additional arguments
None
as a byproduct of the tokenization process in the `encodes` method.