paddlenlp.taskflow.models.dependency_parsing_model 源代码

# coding:utf-8
# Copyright (c) 2021  PaddlePaddle Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License"
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
#     http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

import paddle
import paddle.nn as nn
from paddlenlp.transformers import AutoModel


[文档]class BiAffineParser(nn.Layer): """DDParser""" def __init__(self, encoding_model, n_rels, n_words, pad_index, bos_index, eos_index, n_mlp_arc=500, n_mlp_rel=100): super(BiAffineParser, self).__init__() self.pad_index = pad_index self.bos_index = bos_index self.eos_index = eos_index if encoding_model == "lstm-pe": self.embed = LSTMByWPEncoder(n_words, pad_index) else: # encoding_model is "ernie-3.0-medium-zh", "ernie-1.0" or other models: pretrained_model = AutoModel.from_pretrained(encoding_model) self.embed = ErnieEncoder(pad_index, pretrained_model) # MLP layer self.mlp_arc_h = MLP(n_in=self.embed.mlp_input_size, n_out=n_mlp_arc) self.mlp_arc_d = MLP(n_in=self.embed.mlp_input_size, n_out=n_mlp_arc) self.mlp_rel_h = MLP(n_in=self.embed.mlp_input_size, n_out=n_mlp_rel) self.mlp_rel_d = MLP(n_in=self.embed.mlp_input_size, n_out=n_mlp_rel) # Biaffine layer self.arc_attn = BiAffine(n_in=n_mlp_arc, bias_x=True, bias_y=False) self.rel_attn = BiAffine(n_in=n_mlp_rel, n_out=n_rels, bias_x=True, bias_y=True)
[文档] def forward(self, words, wp): words, x = self.embed(words, wp) mask = paddle.logical_and(words != self.pad_index, words != self.eos_index) arc_h = self.mlp_arc_h(x) arc_d = self.mlp_arc_d(x) rel_h = self.mlp_rel_h(x) rel_d = self.mlp_rel_d(x) # Get arc and rel scores from the bilinear attention # Shape: (batch_size, seq_len, seq_len) s_arc = self.arc_attn(arc_d, arc_h) # Shape: (batch_size, seq_len, seq_len, n_rels) s_rel = paddle.transpose(self.rel_attn(rel_d, rel_h), perm=[0, 2, 3, 1]) # Set the scores that exceed the length of each sentence to -1e5 s_arc_mask = paddle.unsqueeze(mask, 1) s_arc = s_arc * s_arc_mask + paddle.scale( paddle.cast(s_arc_mask, "int32"), scale=1e5, bias=-1, bias_after_scale=False ) mask = paddle.cast( paddle.logical_and( paddle.logical_and(words != self.pad_index, words != self.bos_index), words != self.eos_index, ), "int32", ) arc_preds = paddle.argmax(s_arc, axis=-1) rel_preds = paddle.argmax(s_rel, axis=-1) return arc_preds, rel_preds, s_arc, mask
[文档]class MLP(nn.Layer): """MLP""" def __init__(self, n_in, n_out): super(MLP, self).__init__() self.linear = nn.Linear( n_in, n_out, weight_attr=nn.initializer.XavierNormal(), ) self.leaky_relu = nn.LeakyReLU(negative_slope=0.1)
[文档] def forward(self, x): # Shape: (batch_size, output_size) x = self.linear(x) x = self.leaky_relu(x) return x
[文档]class BiAffine(nn.Layer): """BiAffine""" def __init__(self, n_in, n_out=1, bias_x=True, bias_y=True): super(BiAffine, self).__init__() self.n_in = n_in self.n_out = n_out self.bias_x = bias_x self.bias_y = bias_y self.weight = self.create_parameter(shape=[n_out, n_in + bias_x, n_in + bias_y], dtype="float32")
[文档] def forward(self, x, y): if self.bias_x: x = paddle.concat([x, paddle.ones_like(x[:, :, :1])], axis=-1) if self.bias_y: y = paddle.concat([y, paddle.ones_like(x[:, :, :1])], axis=-1) # Shape x: (batch_size, num_tokens, input_size + bias_x) b = x.shape[0] o = self.weight.shape[0] # Shape x: (batch_size, output_size, num_tokens, input_size + bias_x) x = paddle.expand(paddle.unsqueeze(x, axis=1), shape=(x.shape[0], o, x.shape[1], x.shape[2])) # Shape y: (batch_size, output_size, num_tokens, input_size + bias_y) y = paddle.expand(paddle.unsqueeze(y, axis=1), shape=(y.shape[0], o, y.shape[1], y.shape[2])) # Shape weight: (batch_size, output_size, input_size + bias_x, input_size + bias_y) weight = paddle.expand( paddle.unsqueeze(self.weight, axis=0), shape=(b, self.weight.shape[0], self.weight.shape[1], self.weight.shape[2]), ) # Shape: (batch_size, output_size, num_tokens, num_tokens) s = paddle.matmul(paddle.matmul(x, weight), paddle.transpose(y, perm=[0, 1, 3, 2])) # Remove dim 1 if n_out == 1 if s.shape[1] == 1: s = paddle.squeeze(s, axis=1) return s
[文档]class ErnieEncoder(nn.Layer): def __init__(self, pad_index, pretrained_model): super(ErnieEncoder, self).__init__() self.pad_index = pad_index self.ptm = pretrained_model self.mlp_input_size = self.ptm.config["hidden_size"]
[文档] def forward(self, words, wp): x, _ = self.ptm(words) x = paddle.reshape( index_sample(x, wp), shape=[wp.shape[0], wp.shape[1], x.shape[2]], ) words = index_sample(words, wp) return words, x
[文档]class LSTMByWPEncoder(nn.Layer): def __init__(self, n_words, pad_index, lstm_by_wp_embed_size=200, n_embed=300, n_lstm_hidden=300, n_lstm_layers=3): super(LSTMByWPEncoder, self).__init__() self.pad_index = pad_index self.word_embed = nn.Embedding(n_words, lstm_by_wp_embed_size) self.lstm = nn.LSTM( input_size=lstm_by_wp_embed_size, hidden_size=n_lstm_hidden, num_layers=n_lstm_layers, direction="bidirectional", ) self.mlp_input_size = n_lstm_hidden * 2
[文档] def forward(self, words, wp): word_embed = self.word_embed(words) mask = words != self.pad_index seq_lens = paddle.sum(paddle.cast(mask, "int32"), axis=-1) x, _ = self.lstm(word_embed, sequence_length=seq_lens) x = paddle.reshape( index_sample(x, wp), shape=[wp.shape[0], wp.shape[1], x.shape[2]], ) words = paddle.index_sample(words, wp) return words, x
[文档]def index_sample(x, index): """Select input value according to index Arags: input: input matrix index: index matrix Returns: output >>> input [ [1, 2, 3], [4, 5, 6] ] >>> index [ [1, 2], [0, 1] ] >>> index_sample(input, index) [ [2, 3], [4, 5] ] """ x_s = x.shape dim = len(index.shape) - 1 assert x_s[:dim] == index.shape[:dim] if len(x_s) == 3 and dim == 1: r_x = paddle.reshape(x, shape=[-1, x_s[1], x_s[-1]]) else: r_x = paddle.reshape(x, shape=[-1, x_s[-1]]) index = paddle.reshape(index, shape=[len(r_x), -1, 1]) # Generate arange index, shape like index arr_index = paddle.arange(start=0, end=len(index), dtype=index.dtype) arr_index = paddle.unsqueeze(arr_index, axis=[1, 2]) arr_index = paddle.expand(arr_index, index.shape) # Generate new index new_index = paddle.concat((arr_index, index), -1) new_index = paddle.reshape(new_index, (-1, 2)) # Get output out = paddle.gather_nd(r_x, new_index) if len(x_s) == 3 and dim == 2: out = paddle.reshape(out, shape=[x_s[0], x_s[1], -1]) else: out = paddle.reshape(out, shape=[x_s[0], -1]) return out