From 4e2b22260b0e5c4aa9ad98c9a4f6e5885dede836 Mon Sep 17 00:00:00 2001 From: nl <3210346136@qq.com> Date: Wed, 18 Sep 2019 14:29:55 +0800 Subject: [PATCH] =?UTF-8?q?=E9=80=82=E5=BA=94=E6=80=A7=E4=BF=AE=E6=94=B9?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- SpeechModel24.py | 1 - SpeechModel25.py | 1 - SpeechModel251.py | 1 - SpeechModel251_p.py | 1 - SpeechModel252.py | 5 ++--- SpeechModel26.py | 1 - SpeechModel261.py | 3 +-- SpeechModel261_p.py | 3 +-- 8 files changed, 4 insertions(+), 12 deletions(-) diff --git a/SpeechModel24.py b/SpeechModel24.py index f9e125b..c9883ca 100644 --- a/SpeechModel24.py +++ b/SpeechModel24.py @@ -304,7 +304,6 @@ class ModelSpeech(): # 语音模型类 def RecognizeSpeech(self, wavsignal, fs): ''' 最终做语音识别用的函数,识别一个wav序列的语音 - 不过这里现在还有bug ''' #data = self.data diff --git a/SpeechModel25.py b/SpeechModel25.py index fe6f999..f121479 100644 --- a/SpeechModel25.py +++ b/SpeechModel25.py @@ -313,7 +313,6 @@ class ModelSpeech(): # 语音模型类 def RecognizeSpeech(self, wavsignal, fs): ''' 最终做语音识别用的函数,识别一个wav序列的语音 - 不过这里现在还有bug ''' #data = self.data diff --git a/SpeechModel251.py b/SpeechModel251.py index 0f2831e..a304e62 100644 --- a/SpeechModel251.py +++ b/SpeechModel251.py @@ -339,7 +339,6 @@ class ModelSpeech(): # 语音模型类 def RecognizeSpeech(self, wavsignal, fs): ''' 最终做语音识别用的函数,识别一个wav序列的语音 - 不过这里现在还有bug ''' #data = self.data diff --git a/SpeechModel251_p.py b/SpeechModel251_p.py index d79f67d..20b7394 100644 --- a/SpeechModel251_p.py +++ b/SpeechModel251_p.py @@ -327,7 +327,6 @@ class ModelSpeech(): # 语音模型类 def RecognizeSpeech(self, wavsignal, fs): ''' 最终做语音识别用的函数,识别一个wav序列的语音 - 不过这里现在还有bug ''' #data = self.data diff --git a/SpeechModel252.py b/SpeechModel252.py index b914af1..8b904ad 100644 --- a/SpeechModel252.py +++ b/SpeechModel252.py @@ -31,9 +31,9 @@ class ModelSpeech(): # 语音模型类 def __init__(self, datapath): ''' 初始化 - 默认输出的拼音的表示大小是1422,即1421个拼音+1个空白块 + 默认输出的拼音的表示大小是1424,即1423个拼音+1个空白块 ''' - MS_OUTPUT_SIZE = 1422 + MS_OUTPUT_SIZE = 1424 self.MS_OUTPUT_SIZE = MS_OUTPUT_SIZE # 神经网络最终输出的每一个字符向量维度的大小 #self.BATCH_SIZE = BATCH_SIZE # 一次训练的batch self.label_max_string_length = 64 @@ -331,7 +331,6 @@ class ModelSpeech(): # 语音模型类 def RecognizeSpeech(self, wavsignal, fs): ''' 最终做语音识别用的函数,识别一个wav序列的语音 - 不过这里现在还有bug ''' #data = self.data diff --git a/SpeechModel26.py b/SpeechModel26.py index 95097c3..9860243 100644 --- a/SpeechModel26.py +++ b/SpeechModel26.py @@ -331,7 +331,6 @@ class ModelSpeech(): # 语音模型类 def RecognizeSpeech(self, wavsignal, fs): ''' 最终做语音识别用的函数,识别一个wav序列的语音 - 不过这里现在还有bug ''' #data = self.data diff --git a/SpeechModel261.py b/SpeechModel261.py index 2ed70c9..eb0e716 100644 --- a/SpeechModel261.py +++ b/SpeechModel261.py @@ -35,7 +35,7 @@ class ModelSpeech(): # 语音模型类 初始化 默认输出的拼音的表示大小是1422,即1421个拼音+1个空白块 ''' - MS_OUTPUT_SIZE = 1422 + MS_OUTPUT_SIZE = 1424 self.MS_OUTPUT_SIZE = MS_OUTPUT_SIZE # 神经网络最终输出的每一个字符向量维度的大小 #self.BATCH_SIZE = BATCH_SIZE # 一次训练的batch self.label_max_string_length = 64 @@ -353,7 +353,6 @@ class ModelSpeech(): # 语音模型类 def RecognizeSpeech(self, wavsignal, fs): ''' 最终做语音识别用的函数,识别一个wav序列的语音 - 不过这里现在还有bug ''' #data = self.data diff --git a/SpeechModel261_p.py b/SpeechModel261_p.py index bbd8708..300f825 100644 --- a/SpeechModel261_p.py +++ b/SpeechModel261_p.py @@ -35,7 +35,7 @@ class ModelSpeech(): # 语音模型类 初始化 默认输出的拼音的表示大小是1422,即1421个拼音+1个空白块 ''' - MS_OUTPUT_SIZE = 1422 + MS_OUTPUT_SIZE = 1424 self.MS_OUTPUT_SIZE = MS_OUTPUT_SIZE # 神经网络最终输出的每一个字符向量维度的大小 #self.BATCH_SIZE = BATCH_SIZE # 一次训练的batch self.label_max_string_length = 64 @@ -344,7 +344,6 @@ class ModelSpeech(): # 语音模型类 def RecognizeSpeech(self, wavsignal, fs): ''' 最终做语音识别用的函数,识别一个wav序列的语音 - 不过这里现在还有bug ''' #data = self.data