Skip to content

Commit 5c5fe43

Browse files
committed
push
1 parent 4e74427 commit 5c5fe43

File tree

2 files changed

+44
-26
lines changed

2 files changed

+44
-26
lines changed

train.py

+29-23
Original file line numberDiff line numberDiff line change
@@ -4,6 +4,12 @@
44
import numpy as np
55
import sys
66

7+
from dataclasses import dataclass, field
8+
9+
@dataclass
10+
class CustomTrainingArguments(TrainingArguments):
11+
save_steps: int = field(default=100)
12+
713
# model_path = sys.argv[1]
814

915
def setup_seed(seed):
@@ -49,33 +55,33 @@ def read_data(dataset_name, wav_split):
4955

5056

5157
# for debug
52-
read_data('Shanghai_Dialect_Dict', 1)
53-
read_data('Shanghai_Dialect_Dict', 2)
54-
random.shuffle(train_data)
55-
eval_ratio = 0.05
56-
index = int(len(train_data) * eval_ratio)
57-
eval_data = train_data[:10]
58-
train_data = train_data[10:20]
59-
batch_size = 1
60-
eval_steps = 100
61-
fp16 = False
62-
63-
# for train
64-
# read_data('Shanghai_Dialect_Conversational_Speech_Corpus', 1)
65-
# read_data('Shanghai_Dialect_Conversational_Speech_Corpus', 2)
66-
# read_data('Shanghai_Dialect_Scripted_Speech_Corpus_Daily_Use_Sentence', 1)
67-
# read_data('Shanghai_Dialect_Scripted_Speech_Corpus_Daily_Use_Sentence', 2)
6858
# read_data('Shanghai_Dialect_Dict', 1)
6959
# read_data('Shanghai_Dialect_Dict', 2)
70-
# read_data('Shanghai_Dialect_Zhongguoyuyan', 1)
71-
60+
# random.shuffle(train_data)
7261
# eval_ratio = 0.05
7362
# index = int(len(train_data) * eval_ratio)
74-
# eval_data = train_data[:index]
75-
# train_data = train_data[index:]
76-
# batch_size = 32
63+
# eval_data = train_data[:10]
64+
# train_data = train_data[10:20]
65+
# batch_size = 1
7766
# eval_steps = 100
78-
# fp16 = True
67+
# fp16 = False
68+
69+
# for train
70+
read_data('Shanghai_Dialect_Conversational_Speech_Corpus', 1)
71+
read_data('Shanghai_Dialect_Conversational_Speech_Corpus', 2)
72+
read_data('Shanghai_Dialect_Scripted_Speech_Corpus_Daily_Use_Sentence', 1)
73+
read_data('Shanghai_Dialect_Scripted_Speech_Corpus_Daily_Use_Sentence', 2)
74+
read_data('Shanghai_Dialect_Dict', 1)
75+
read_data('Shanghai_Dialect_Dict', 2)
76+
read_data('Shanghai_Dialect_Zhongguoyuyan', 1)
77+
78+
eval_ratio = 0.05
79+
index = int(len(train_data) * eval_ratio)
80+
eval_data = train_data[:index]
81+
train_data = train_data[index:]
82+
batch_size = 32
83+
eval_steps = 100
84+
fp16 = True
7985

8086

8187
print('eval_data_len:', len(eval_data))
@@ -84,7 +90,7 @@ def read_data(dataset_name, wav_split):
8490
# gradient_checkpointing=True,
8591
# gradient_accumulation_steps=2,
8692

87-
training_args = TrainingArguments(
93+
training_args = CustomTrainingArguments(
8894
save_steps=eval_steps,
8995
group_by_length=True,
9096
num_train_epochs=200,

transcribe.py

+15-3
Original file line numberDiff line numberDiff line change
@@ -1,12 +1,24 @@
11
import torch
22
from huggingsound import SpeechRecognitionModel
3-
3+
import os
44
device = "cuda" if torch.cuda.is_available() else "cpu"
55
batch_size = 1
66
# model = SpeechRecognitionModel("wbbbbb/wav2vec2-large-chinese-zh-cn", device=device)
7-
model = SpeechRecognitionModel("checkpoint-wav2vec2-large-xlsr-53-chinese-zh-cn-2022-09-02-10_08_26", device=device)
7+
model = SpeechRecognitionModel("checkpoint-wav2vec2-large-xlsr-53-chinese-zh-cn-2023-08-31-09:02:50", device=device)
88
# audio_paths = ["Shanghai_Dialect_Dict/Split_WAV/1.wav", "Shanghai_Dialect_Dict/Split_WAV/2.wav"]
9-
audio_paths = ["Shanghai_Dialect_Dict/Split_WAV/1.wav"]
9+
audio_paths = ["Shanghai_Dialect_Dict/Split_WAV1/1.wav"]
10+
11+
# audio_paths = []
12+
# for x in os.listdir('/data/xumh/asr/zhuanrengongzhuananjian/zhuananjian'):
13+
# audio_paths.append('/data/xumh/asr/zhuanrengongzhuananjian/zhuananjian/' + x)
14+
# print(audio_paths)
15+
16+
audio_paths = []
17+
for x in os.listdir('/data/xumh/asr/zhuanrengongzhuananjian/zhuanrengong'):
18+
audio_paths.append('/data/xumh/asr/zhuanrengongzhuananjian/zhuanrengong/' + x)
19+
print(audio_paths)
20+
21+
1022

1123
transcriptions = model.transcribe(audio_paths, batch_size=batch_size)
1224

0 commit comments

Comments
 (0)