Skip to content

Commit 434dc2a

Browse files
authored
Update README.md
1 parent 2929ec4 commit 434dc2a

File tree

1 file changed

+21
-9
lines changed

1 file changed

+21
-9
lines changed

README.md

Lines changed: 21 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -48,23 +48,28 @@ TL;DR: A realistic and stylized talking head video generation method from a sing
4848
- [ ] Generating 3D face from Audio.
4949
- [x] Generating 4D free-view talking examples from audio and a single image.
5050
- [ ] Gradio/Colab Demo.
51+
- [ ] training code of each componments.
52+
- [ ] Audio-driven Anime Avatar.
5153
- [ ] integrade with stable-diffusion-web-ui. (stay tunning!)
5254

5355
https://user-images.githubusercontent.com/4397546/222513483-89161f58-83d0-40e4-8e41-96c32b47bd4e.mp4
5456

5557

56-
- [ ] training code of each componments.
57-
58+
## 🔮 Inference Demo!
5859

59-
## 🔮 Test!
6060

6161
#### Requirements
62+
<details><summary>CLICK ME</summary>
6263

63-
* Python
64+
* Python 3.8
6465
* PyTorch
6566
* ffmpeg
6667

67-
#### Conda Installation
68+
</details>
69+
70+
#### Dependence Installation
71+
72+
<details><summary>CLICK ME</summary>
6873

6974
```
7075
git clone https://github.com/Winfredy/SadTalker.git
@@ -81,7 +86,10 @@ conda install dlib
8186
pip install -r requirements.txt
8287
```
8388

84-
#### Models
89+
</details>
90+
91+
#### Trained Models
92+
<details><summary>CLICK ME</summary>
8593

8694
Please download our [pre-trained model](https://drive.google.com/drive/folders/1Wd88VDoLhVzYsQ30_qDVluQr_Xm46yHT?usp=sharing) and put it in ./checkpoints.
8795

@@ -97,10 +105,14 @@ Please download our [pre-trained model](https://drive.google.com/drive/folders/1
97105
|checkpoints/BFM | 3DMM library file.
98106
|checkpoints/hub | Face detection models used in [face alignment](https://github.com/1adrianb/face-alignment).
99107

108+
</details>
109+
100110
#### Generating 2D face from a single Image
101111

102-
```
103-
python inference.py --driven_audio <audio.wav> --source_image <video.mp4 or picture.png> --result_dir <a file to store results>
112+
```bash
113+
python inference.py --driven_audio <audio.wav> \
114+
--source_image <video.mp4 or picture.png> \
115+
--result_dir <a file to store results>
104116
```
105117

106118
#### Generating 3D face from Audio
@@ -110,7 +122,7 @@ To do ...
110122
#### Generating 4D free-view talking examples from audio and a single image
111123

112124
We use `camera_yaw`, `camera_pitch`, `camera_roll` to control camera pose. For example, `--camera_yaw -20 30 10` means the camera yaw degree changes from -20 to 30 and then changes from 30 to 10.
113-
```
125+
```bash
114126
python inference.py --driven_audio <audio.wav> \
115127
--source_image <video.mp4 or picture.png> \
116128
--result_dir <a file to store results> \

0 commit comments

Comments
 (0)