From e10186229123db4bbec91dac18afd59f375b3acf Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Carl=20Thom=C3=A9?= Date: Fri, 29 Sep 2023 00:01:26 +0200 Subject: [PATCH] Spelling --- audiocraft/models/lm.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/audiocraft/models/lm.py b/audiocraft/models/lm.py index 8cefd2c5..473d14bc 100644 --- a/audiocraft/models/lm.py +++ b/audiocraft/models/lm.py @@ -393,7 +393,7 @@ def generate(self, check: bool = False, callback: tp.Optional[tp.Callable[[int, int], None]] = None) -> torch.Tensor: """Generate tokens sampling from the model given a prompt or unconditionally. Generation can - be perform in a greedy fashion or using sampling with top K and top P strategies. + be performed in a greedy fashion or using sampling with top K and top P strategies. Args: prompt (torch.Tensor, optional): Prompt tokens of shape [B, K, T].