modeling¶
-
class
BertModel
(vocab_size, hidden_size=768, num_hidden_layers=12, num_attention_heads=12, intermediate_size=3072, hidden_act='gelu', hidden_dropout_prob=0.1, attention_probs_dropout_prob=0.1, max_position_embeddings=512, type_vocab_size=16, initializer_range=0.02, pad_token_id=0)[source]¶ Bases:
paddlenlp.transformers.bert.modeling.BertPretrainedModel
-
class
BertPretrainedModel
(name_scope=None, dtype='float32')[source]¶ Bases:
paddlenlp.transformers.model_utils.PretrainedModel
An abstract class for pretrained BERT models. It provides BERT related
model_config_file
,resource_files_names
,pretrained_resource_files_map
,pretrained_init_configuration
,base_model_prefix
for downloading and loading pretrained models. SeePretrainedModel
for more details.-
base_model_class
¶
-
-
class
BertForPretraining
(bert)[source]¶ Bases:
paddlenlp.transformers.bert.modeling.BertPretrainedModel
-
class
BertPretrainingCriterion
(vocab_size)[source]¶ Bases:
paddle.fluid.dygraph.layers.Layer
-
forward
(prediction_scores, seq_relationship_score, masked_lm_labels, next_sentence_labels, masked_lm_scale)[source]¶ Defines the computation performed at every call. Should be overridden by all subclasses.
- Parameters
*inputs (tuple) – unpacked tuple arguments
**kwargs (dict) – unpacked dict arguments
-
-
class
BertPretrainingHeads
(hidden_size, vocab_size, activation, embedding_weights=None)[source]¶ Bases:
paddle.fluid.dygraph.layers.Layer
-
class
BertForSequenceClassification
(bert, num_classes=2, dropout=None)[source]¶ Bases:
paddlenlp.transformers.bert.modeling.BertPretrainedModel
Model for sentence (pair) classification task with BERT. :param bert: An instance of BertModel. :type bert: BertModel :param num_classes: The number of classes. Default 2 :type num_classes: int, optional :param dropout: The dropout probability for output of BERT.
If None, use the same value as
hidden_dropout_prob
ofBertModel
instancebert
. Default None
-
class
BertForTokenClassification
(bert, num_classes=2, dropout=None)[source]¶ Bases:
paddlenlp.transformers.bert.modeling.BertPretrainedModel
-
class
BertForQuestionAnswering
(bert, dropout=None)[source]¶ Bases:
paddlenlp.transformers.bert.modeling.BertPretrainedModel