checkpoint = "google/mt5-small" tokenizer = MT5Tokenizer.from_pretrained(checkpoint, legacy=False) model = MT5ForConditionalGeneration.from_pretarined(checkpoint)
- This model is katakana-oriented.
- CharacTER score = 0.1033
Training
- Trainer: Seq2seqTrainer
- Epochs: 200
- Optimizer: AdamW
- Learning Rate: 2e-5
- Weight decay: 0.01
- Warm-ups: 0.05*total_steps
- Scheduler: cosine
- Downloads last month
- 21
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support