Skip to content

Numbers normalisation #1065

Closed
Closed
@zidsi

Description

@zidsi

Checks

  • This template is only for research question, not usage problems, feature requests or bug reports.
  • I have thoroughly reviewed the project documentation and read the related paper(s).
  • I have searched for existing issues, including closed ones, no similar questions.
  • I am using English to submit this issue to facilitate community communication.

Question details

FYI - I've trained model with no number to string (words) normalisation text preprocessing for bravity. It picked pronaunciation well as well as context conditioned "word form" - which for same "number form" might differ in target language.

I wonder if anyone did some ablation study on this aspect? Which text preprocessing is a "must", which "nice to have" and which could be safely skipped?

Would mixed training samples (word and number representation for same speech target) confuse the model or help?

Metadata

Metadata

Assignees

No one assigned

    Labels

    questionFurther information is requested

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions