-
Notifications
You must be signed in to change notification settings - Fork 116
/
run.py
140 lines (120 loc) · 4.28 KB
/
run.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
import dataclasses
import json
import logging
from argparse import ArgumentParser
from pathlib import Path
from voicevox_core import AccelerationMode, AudioQuery, wav_from_s16le
from voicevox_core.blocking import Onnxruntime, OpenJtalk, Synthesizer, VoiceModelFile
@dataclasses.dataclass
class Args:
mode: AccelerationMode
vvm: Path
onnxruntime: str
dict_dir: Path
text: str
out: Path
style_id: int
streaming: bool
@staticmethod
def parse_args() -> "Args":
argparser = ArgumentParser()
argparser.add_argument(
"--mode",
default="AUTO",
type=AccelerationMode,
help='モード ("AUTO", "CPU", "GPU")',
)
argparser.add_argument(
"vvm",
type=Path,
help="vvmファイルへのパス",
)
argparser.add_argument(
"--onnxruntime",
default=Onnxruntime.LIB_VERSIONED_FILENAME,
help="ONNX Runtimeのライブラリのfilename",
)
argparser.add_argument(
"--dict-dir",
default="./open_jtalk_dic_utf_8-1.11",
type=Path,
help="Open JTalkの辞書ディレクトリ",
)
argparser.add_argument(
"--text",
default="この音声は、ボイスボックスを使用して、出力されています。",
help="読み上げさせたい文章",
)
argparser.add_argument(
"--out",
default="./output.wav",
type=Path,
help="出力wavファイルのパス",
)
argparser.add_argument(
"--style-id",
default=0,
type=int,
help="話者IDを指定",
)
argparser.add_argument(
"--streaming",
action="store_true",
help="ストリーミング生成",
)
args = argparser.parse_args()
return Args(
args.mode,
args.vvm,
args.onnxruntime,
args.dict_dir,
args.text,
args.out,
args.style_id,
args.streaming,
)
def main() -> None:
logging.basicConfig(format="[%(levelname)s] %(name)s: %(message)s")
logger = logging.getLogger(__name__)
logger.setLevel("DEBUG")
logging.getLogger("voicevox_core_python_api").setLevel("DEBUG")
logging.getLogger("voicevox_core").setLevel("DEBUG")
args = Args.parse_args()
logger.info("%s", f"Loading ONNX Runtime ({args.onnxruntime=})")
onnxruntime = Onnxruntime.load_once(filename=args.onnxruntime)
logger.debug("%s", f"{onnxruntime.supported_devices()=}")
logger.info("%s", f"Initializing ({args.mode=}, {args.dict_dir=})")
synthesizer = Synthesizer(
onnxruntime, OpenJtalk(args.dict_dir), acceleration_mode=args.mode
)
logger.debug("%s", f"{synthesizer.metas=}")
logger.debug("%s", f"{synthesizer.is_gpu_mode=}")
logger.info("%s", f"Loading `{args.vvm}`")
with VoiceModelFile.open(args.vvm) as model:
synthesizer.load_voice_model(model)
logger.info("%s", f"Creating an AudioQuery from {args.text!r}")
audio_query = synthesizer.create_audio_query(args.text, args.style_id)
logger.info("%s", f"Synthesizing with {display_as_json(audio_query)}")
if args.streaming:
logger.info("%s", "In streaming mode")
chunk_sec = 1.0
audio_feature = synthesizer.precompute_render(audio_query, args.style_id)
chunk_frames = int(audio_feature.frame_rate * chunk_sec)
pcm = b""
for i in range(0, audio_feature.frame_length, chunk_frames):
logger.info("%s", f"{i/audio_feature.frame_length:.2%}")
pcm += synthesizer.render(
audio_feature, i, min(i + chunk_frames, audio_feature.frame_length)
)
logger.info("%s", f"100%")
wav = wav_from_s16le(
pcm, audio_query.output_sampling_rate, audio_query.output_stereo
)
else:
wav = synthesizer.synthesis(audio_query, args.style_id)
args.out.write_bytes(wav)
logger.info("%s", f"Wrote `{args.out}`")
def display_as_json(audio_query: AudioQuery) -> str:
return json.dumps(dataclasses.asdict(audio_query), ensure_ascii=False)
if __name__ == "__main__":
main()