You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+5-5Lines changed: 5 additions & 5 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,14 +1,14 @@
1
1
# Vector-Quantized Contrastive Predictive Coding
2
2
3
-
To learn discrete representations of speech for the [ZeroSpeech challenges](https://zerospeech.com/), we propose vector-quantized contrastive predictive coding.
4
-
An encoder maps input speech into a discrete sequence of codes.
5
-
Next, an autoregressive model summarises the latent representation (up until time t) into a context vector.
6
-
Using this context, the model learns to discriminate future frames from negative examples sampled randomly from other utterances.
7
-
Finally, an RNN based vocoder is trained to generate audio from the discretized representation.
3
+
Train and evaluate the VQ-VAE model for our submission to the [ZeroSpeech 2020 challenge](https://zerospeech.com/).
4
+
Voice conversion samples can be found [here](https://bshall.github.io/VectorQuantizedCPC/).
5
+
Pretrained weights for the 2019 English and Indonesian datasets can be found [here](https://github.com/bshall/VectorQuantizedCPC/releases/tag/v0.1).
6
+
Leader-board for the ZeroSpeech 2020 challenge can be found [here](https://zerospeech.com/2020/results.html).
8
7
9
8
<palign="center">
10
9
<img width="784" height="340" alt="VQ-CPC model summary"
0 commit comments