ELECTRA Model Summary
The following table summarizes the currently supported ELECTRA models and their corresponding pretrained weights in PaddleNLP. For model details, please refer to the corresponding links.
Note: Both models follow ELECTRA-style architecture. The discriminator is trained for sequence classification tasks, while the generator is used for token-level predictions in the ELECTRA framework.
| Chinese | Generator, 12-layer, 64-hidden, |
| | 1-heads, 3M parameters. |
| | Trained on Chinese legal corpus. |
| | |
| | Please refer to: |
| | hfl/chinese-legal-electra-small-generator
.. _hfl/chinese-electra-180g-base-discriminator: https://huggingface.co/hfl/chinese-electra-180g-base-discriminator
.. _hfl/chinese-electra-180g-small-ex-discriminator: https://huggingface.co/hfl/chinese-electra-180g-small-ex-discriminator
.. _hfl/chinese-legal-electra-small-generator: https://huggingface.co/hfl/chinese-legal-electra-small-generator