roberta#
- modeling
- tokenizer
RobertaTokenizer
RobertaChineseTokenizer
RobertaChineseTokenizer.vocab_size
RobertaChineseTokenizer.get_vocab()
RobertaChineseTokenizer.convert_tokens_to_string()
RobertaChineseTokenizer.num_special_tokens_to_add()
RobertaChineseTokenizer.build_inputs_with_special_tokens()
RobertaChineseTokenizer.build_offset_mapping_with_special_tokens()
RobertaChineseTokenizer.create_token_type_ids_from_sequences()
RobertaChineseTokenizer.get_special_tokens_mask()
RobertaBPETokenizer
RobertaBPETokenizer.get_vocab()
RobertaBPETokenizer.build_inputs_with_special_tokens()
RobertaBPETokenizer.get_offset_mapping()
RobertaBPETokenizer.build_offset_mapping_with_special_tokens()
RobertaBPETokenizer.get_special_tokens_mask()
RobertaBPETokenizer.create_token_type_ids_from_sequences()
RobertaBPETokenizer.convert_tokens_to_string()
RobertaBPETokenizer.num_special_tokens_to_add()
RobertaBPETokenizer.prepare_for_tokenization()