Skip to content

Commit ed68f99

Browse files
committed
update README
1 parent d0a5ea9 commit ed68f99

File tree

8 files changed

+24
-9
lines changed

8 files changed

+24
-9
lines changed
File renamed without changes.
File renamed without changes.
File renamed without changes.
File renamed without changes.
File renamed without changes.
File renamed without changes.
File renamed without changes.

model_zoo/ernie-3.0/README.md

Lines changed: 24 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -1,14 +1,20 @@
1-
# ERNIE 3.0
1+
# ERNIE 3.0 轻量级模型
22

3-
## ERNIE 3.0 CLUE Benchmark
3+
## 模型介绍
44

5-
ERNIE 3.0 开源 Base(12L768H) 和 Medium(6L768H) 两个模型,它们在 CLUE 的各**验证集**上有如下效果:
5+
<p align="center">
6+
<img width="644" alt="image" src="https://user-images.githubusercontent.com/1371212/168254282-e6901f9a-fd9a-4bbd-a5b6-e847b7f90bb4.png">
7+
</p>
8+
9+
## 模型效果
10+
11+
ERNIE 3.0 开源 Base(12L768H) 和 Medium(6L768H) 两个模型,它们在 CLUE **验证集**上指标如下:
612

713
<table style="width:100%;" cellpadding="2" cellspacing="0" border="1" bordercolor="#000000">
814
<tbody>
915
<tr>
1016
<td style="text-align:center">
11-
<span style="font-size:18px;">Config</span>
17+
<span style="font-size:18px;">Arch</span>
1218
</td>
1319
<td style="text-align:center">
1420
<span style="font-size:18px;">Model</span>
@@ -470,10 +476,11 @@ python infer.py --task_name tnews --model_path best_models/TNEWS/compress/0.75/h
470476

471477
压缩后的性能:
472478

473-
GPU
479+
GPU:
480+
TBD
474481

475-
476-
CPU
482+
CPU:
483+
TBD
477484

478485
**压缩 API 使用TIPS:**
479486

@@ -485,8 +492,16 @@ CPU
485492

486493
### 推理部署
487494

488-
needs yl
495+
TBD
489496

490497
#### Paddle Serving 部署
491498

492-
needs lq
499+
TBD
500+
501+
## Reference
502+
503+
* Sun Y, Wang S, Feng S, et al. ERNIE 3.0: Large-scale Knowledge Enhanced Pre-training for Language Understanding and Generation[J]. arXiv preprint arXiv:2107.02137, 2021.
504+
505+
* Su W, Chen X, Feng S, et al. ERNIE-Tiny: A Progressive Distillation Framework for Pretrained Transformer Compression[J]. arXiv preprint arXiv:2106.02241, 2021.
506+
507+
* Wang S, Sun Y, Xiang Y, et al. ERNIE 3.0 Titan: Exploring Larger-scale Knowledge Enhanced Pre-training for Language Understanding and Generation[J]. arXiv preprint arXiv:2112.12731, 2021.

0 commit comments

Comments
 (0)