Skip to content
This repository was archived by the owner on Jul 21, 2025. It is now read-only.

Commit 234968b

Browse files
authored
add links of other readme in the master readme (#70)
* add links of other readme in the master readme * modify links of training/inference readme * modify links of training/inference readme
1 parent 64912ff commit 234968b

File tree

1 file changed

+12
-6
lines changed

1 file changed

+12
-6
lines changed

README.md

Lines changed: 12 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -11,8 +11,8 @@ LightSeq is a high performance training and inference library for sequence proce
1111
in CUDA.
1212
It enables highly efficient computation of modern NLP models such as **BERT**, **GPT**,
1313
**Transformer**, etc.
14-
It is therefore best useful for Machine Translation, *Text Generation*, *Dialog*, *Language
15-
Modelling*, *Sentiment analysis*, and other related tasks with sequence data.
14+
It is therefore best useful for *Machine Translation*, *Text Generation*, *Dialog*, *Language
15+
Modelling*, *Sentiment Analysis*, and other related tasks with sequence data.
1616

1717
The library is built on top of CUDA official
1818
library([cuBLAS](https://docs.nvidia.com/cuda/cublas/index.html),
@@ -23,11 +23,14 @@ addition to model components, the inference library also provide easy-to deploy
2323
Server](https://docs.nvidia.com/deeplearning/sdk/inference-server-archived/tensorrt_inference_server_120/tensorrt-inference-server-guide/docs/quickstart.html).
2424
With LightSeq, one can easily develop modified Transformer architecture with little additional code.
2525

26+
## Features
27+
### [Training](./lightseq/training)
2628
The following is a support matrix of LightSeq **training** library compared with
2729
[DeepSpeed](https://github.com/microsoft/DeepSpeed).
2830

2931
![features](./docs/training/images/features.png)
3032

33+
### [Inference](./lightseq/inference)
3134
The following is a support matrix of LightSeq **inference** library compared with
3235
[TurboTransformers](https://github.com/Tencent/TurboTransformers) and
3336
[FasterTransformer](https://github.com/NVIDIA/DeepLearningExamples/tree/master/FasterTransformer).
@@ -37,7 +40,7 @@ The following is a support matrix of LightSeq **inference** library compared wit
3740

3841
## Performance
3942

40-
### Training
43+
### [Training](./lightseq/training)
4144
Here we present the experimental results on WMT14 English to German translation task based on Transformer-big models. We train Transformer models of different sizes on eight NVIDIA Tesla V100/NVIDIA Ampere A100 GPUs with data parallel and fp16 mixed precision.
4245
[Fairseq](https://github.com/pytorch/fairseq) with [Apex](https://github.com/NVIDIA/apex) is choosed as our baseline.
4346

@@ -47,7 +50,7 @@ We compute speedup on different batch size using the WPS (real words per second)
4750

4851
More results is available [here](./docs/training/performance.md)
4952

50-
### Inference
53+
### [Inference](./lightseq/inference)
5154
Here we present the experimental results on neural machine translation based on Transformer-base models using beam search methods.
5255
We choose Tensorflow and
5356
[FasterTransformer](https://github.com/NVIDIA/DeepLearningExamples/tree/master/FasterTransformer) as a comparison.
@@ -79,6 +82,8 @@ sh examples/training/fairseq/ls_fairseq_wmt14en2de.sh
7982

8083
To compare lightseq with fairseq, delete the arguments with `ls_`prefix to using the original fairseq implementation
8184

85+
More usage is available [here](./lightseq/training/README.md).
86+
8287
### Fast inference from HuggingFace bart
8388

8489
We provide an end2end bart-base example to see how fast Lightseq is compared to HuggingFace. First you should install these requirements.
@@ -97,14 +102,15 @@ python ls_bart.py
97102

98103
LightSeq installation from pypi only supports python 3.6 to 3.8 on Linux for now. Consider compiling from source if you have other environments.
99104

105+
More usage is available [here](./lightseq/inference/README.md).
100106

101107
## Cite Us
102108

103109
If you use LightSeq in your research, please cite the following paper.
104110

105111
```
106112
@InProceedings{wang2021lightseq,
107-
title = "{L}ight{S}eq: A High Performance Inference Library for Transformers",
113+
title = "{L}ight{S}eq: A High Performance Inference Library for Transformers",
108114
author = "Wang, Xiaohui and Xiong, Ying and Wei, Yang and Wang, Mingxuan and Li, Lei",
109115
booktitle = "Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies: Industry Papers (NAACL-HLT)",
110116
month = jun,
@@ -117,4 +123,4 @@ If you use LightSeq in your research, please cite the following paper.
117123
## Contact
118124

119125
Any questions or suggestions, please feel free to contact us at
120-
wangxiaohui.neo@bytedance.com, xiongying.taka@bytedance.com, weiyang.god@bytedance.com, wangmingxuan.89@bytedance.com, lileilab@bytedance.com
126+
wangxiaohui.neo@bytedance.com, xiongying.taka@bytedance.com, qian.xian@bytedance.com, weiyang.god@bytedance.com, wangmingxuan.89@bytedance.com, lileilab@bytedance.com

0 commit comments

Comments
 (0)