From 12d04ce8ed6186ff176ab7266d5b336107a48ca8 Mon Sep 17 00:00:00 2001 From: Fangjun Kuang Date: Thu, 26 Sep 2024 16:51:43 +0800 Subject: [PATCH] Fix running MeloTTS models on GPU. (#1379) We need to use opset 18 to export the model to onnx. --- scripts/melo-tts/export-onnx.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/scripts/melo-tts/export-onnx.py b/scripts/melo-tts/export-onnx.py index dfbcae953..6db8335a7 100755 --- a/scripts/melo-tts/export-onnx.py +++ b/scripts/melo-tts/export-onnx.py @@ -229,7 +229,7 @@ def main(): torch_model = ModelWrapper(model) - opset_version = 13 + opset_version = 18 x = torch.randint(low=0, high=10, size=(60,), dtype=torch.int64) print(x.shape) x_lengths = torch.tensor([x.size(0)], dtype=torch.int64)