Token Masking


Révision datée du 10 mai 2025 à 13:32 par Arianne (discussion | contributions) (Page créée avec « == en construction == == Définition == XXXXXXX Voir aussi '''grand modèle de langues''', '''BERT''', '''segmentation''' == Français == ''' XXXXXXXX''' == Anglais == ''' Token Masking''' ''Masking sensitive tokens from users’ data helps reduce the privacy risks and prevent any personal information being leaked or extracted from adversaries. Such token masking task shall be performed without human-in-the-loop since practitioners are not allowe... »)
(diff) ← Version précédente | Voir la version actuelle (diff) | Version suivante → (diff)

en construction

Définition

XXXXXXX

Voir aussi grand modèle de langues, BERT, segmentation

Français

XXXXXXXX

Anglais

Token Masking

Masking sensitive tokens from users’ data helps reduce the privacy risks and prevent any personal information being leaked or extracted from adversaries. Such token masking task shall be performed without human-in-the-loop since practitioners are not allowed to have the access to annotate or label private data of users.

Token masking provides privacy protections, however, the resulting masked corpus might not be suitable to be directly used for training NLP models for downstream tasks.

Sources

Source: Arxiv

Source: TowardsAI

Contributeurs: Arianne