mirror of
https://github.com/k2-fsa/icefall.git
synced 2025-09-18 21:44:18 +00:00
remove unused arguments
This commit is contained in:
parent
90bc61e970
commit
0a1080e7b8
@ -285,7 +285,6 @@ def fast_beam_search(
|
|||||||
def decode_one_chunk(
|
def decode_one_chunk(
|
||||||
params: AttributeDict,
|
params: AttributeDict,
|
||||||
model: nn.Module,
|
model: nn.Module,
|
||||||
sp: spm.SentencePieceProcessor,
|
|
||||||
decode_streams: List[DecodeStream],
|
decode_streams: List[DecodeStream],
|
||||||
) -> List[int]:
|
) -> List[int]:
|
||||||
"""Decode one chunk frames of features for each decode_streams and
|
"""Decode one chunk frames of features for each decode_streams and
|
||||||
@ -296,8 +295,6 @@ def decode_one_chunk(
|
|||||||
It's the return value of :func:`get_params`.
|
It's the return value of :func:`get_params`.
|
||||||
model:
|
model:
|
||||||
The neural model.
|
The neural model.
|
||||||
sp:
|
|
||||||
The BPE model.
|
|
||||||
decode_streams:
|
decode_streams:
|
||||||
A List of DecodeStream, each belonging to a utterance.
|
A List of DecodeStream, each belonging to a utterance.
|
||||||
Returns:
|
Returns:
|
||||||
@ -469,7 +466,7 @@ def decode_dataset(
|
|||||||
|
|
||||||
while len(decode_streams) >= params.num_decode_streams:
|
while len(decode_streams) >= params.num_decode_streams:
|
||||||
finished_streams = decode_one_chunk(
|
finished_streams = decode_one_chunk(
|
||||||
params, model, sp, decode_streams
|
params=params, model=model, decode_streams=decode_streams
|
||||||
)
|
)
|
||||||
for i in sorted(finished_streams, reverse=True):
|
for i in sorted(finished_streams, reverse=True):
|
||||||
hyp = decode_streams[i].hyp
|
hyp = decode_streams[i].hyp
|
||||||
@ -488,7 +485,9 @@ def decode_dataset(
|
|||||||
|
|
||||||
# decode final chunks of last sequences
|
# decode final chunks of last sequences
|
||||||
while len(decode_streams):
|
while len(decode_streams):
|
||||||
finished_streams = decode_one_chunk(params, model, sp, decode_streams)
|
finished_streams = decode_one_chunk(
|
||||||
|
params=params, model=model, decode_streams=decode_streams
|
||||||
|
)
|
||||||
for i in sorted(finished_streams, reverse=True):
|
for i in sorted(finished_streams, reverse=True):
|
||||||
hyp = decode_streams[i].hyp
|
hyp = decode_streams[i].hyp
|
||||||
if params.decoding_method == "greedy_search":
|
if params.decoding_method == "greedy_search":
|
||||||
|
@ -291,7 +291,6 @@ def fast_beam_search(
|
|||||||
def decode_one_chunk(
|
def decode_one_chunk(
|
||||||
params: AttributeDict,
|
params: AttributeDict,
|
||||||
model: nn.Module,
|
model: nn.Module,
|
||||||
sp: spm.SentencePieceProcessor,
|
|
||||||
decode_streams: List[DecodeStream],
|
decode_streams: List[DecodeStream],
|
||||||
) -> List[int]:
|
) -> List[int]:
|
||||||
"""Decode one chunk frames of features for each decode_streams and
|
"""Decode one chunk frames of features for each decode_streams and
|
||||||
@ -302,8 +301,6 @@ def decode_one_chunk(
|
|||||||
It's the return value of :func:`get_params`.
|
It's the return value of :func:`get_params`.
|
||||||
model:
|
model:
|
||||||
The neural model.
|
The neural model.
|
||||||
sp:
|
|
||||||
The BPE model.
|
|
||||||
decode_streams:
|
decode_streams:
|
||||||
A List of DecodeStream, each belonging to a utterance.
|
A List of DecodeStream, each belonging to a utterance.
|
||||||
Returns:
|
Returns:
|
||||||
@ -477,7 +474,7 @@ def decode_dataset(
|
|||||||
|
|
||||||
while len(decode_streams) >= params.num_decode_streams:
|
while len(decode_streams) >= params.num_decode_streams:
|
||||||
finished_streams = decode_one_chunk(
|
finished_streams = decode_one_chunk(
|
||||||
params, model, sp, decode_streams
|
params=params, model=model, decode_streams=decode_streams
|
||||||
)
|
)
|
||||||
for i in sorted(finished_streams, reverse=True):
|
for i in sorted(finished_streams, reverse=True):
|
||||||
hyp = decode_streams[i].hyp
|
hyp = decode_streams[i].hyp
|
||||||
@ -496,7 +493,9 @@ def decode_dataset(
|
|||||||
|
|
||||||
# decode final chunks of last sequences
|
# decode final chunks of last sequences
|
||||||
while len(decode_streams):
|
while len(decode_streams):
|
||||||
finished_streams = decode_one_chunk(params, model, sp, decode_streams)
|
finished_streams = decode_one_chunk(
|
||||||
|
params=params, model=model, decode_streams=decode_streams
|
||||||
|
)
|
||||||
for i in sorted(finished_streams, reverse=True):
|
for i in sorted(finished_streams, reverse=True):
|
||||||
hyp = decode_streams[i].hyp
|
hyp = decode_streams[i].hyp
|
||||||
if params.decoding_method == "greedy_search":
|
if params.decoding_method == "greedy_search":
|
||||||
|
@ -292,7 +292,6 @@ def fast_beam_search(
|
|||||||
def decode_one_chunk(
|
def decode_one_chunk(
|
||||||
params: AttributeDict,
|
params: AttributeDict,
|
||||||
model: nn.Module,
|
model: nn.Module,
|
||||||
sp: spm.SentencePieceProcessor,
|
|
||||||
decode_streams: List[DecodeStream],
|
decode_streams: List[DecodeStream],
|
||||||
) -> List[int]:
|
) -> List[int]:
|
||||||
"""Decode one chunk frames of features for each decode_streams and
|
"""Decode one chunk frames of features for each decode_streams and
|
||||||
@ -303,8 +302,6 @@ def decode_one_chunk(
|
|||||||
It's the return value of :func:`get_params`.
|
It's the return value of :func:`get_params`.
|
||||||
model:
|
model:
|
||||||
The neural model.
|
The neural model.
|
||||||
sp:
|
|
||||||
The BPE model.
|
|
||||||
decode_streams:
|
decode_streams:
|
||||||
A List of DecodeStream, each belonging to a utterance.
|
A List of DecodeStream, each belonging to a utterance.
|
||||||
Returns:
|
Returns:
|
||||||
@ -478,7 +475,7 @@ def decode_dataset(
|
|||||||
|
|
||||||
while len(decode_streams) >= params.num_decode_streams:
|
while len(decode_streams) >= params.num_decode_streams:
|
||||||
finished_streams = decode_one_chunk(
|
finished_streams = decode_one_chunk(
|
||||||
params, model, sp, decode_streams
|
params=params, model=model, decode_streams=decode_streams
|
||||||
)
|
)
|
||||||
for i in sorted(finished_streams, reverse=True):
|
for i in sorted(finished_streams, reverse=True):
|
||||||
hyp = decode_streams[i].hyp
|
hyp = decode_streams[i].hyp
|
||||||
@ -497,7 +494,9 @@ def decode_dataset(
|
|||||||
|
|
||||||
# decode final chunks of last sequences
|
# decode final chunks of last sequences
|
||||||
while len(decode_streams):
|
while len(decode_streams):
|
||||||
finished_streams = decode_one_chunk(params, model, sp, decode_streams)
|
finished_streams = decode_one_chunk(
|
||||||
|
params=params, model=model, decode_streams=decode_streams
|
||||||
|
)
|
||||||
for i in sorted(finished_streams, reverse=True):
|
for i in sorted(finished_streams, reverse=True):
|
||||||
hyp = decode_streams[i].hyp
|
hyp = decode_streams[i].hyp
|
||||||
if params.decoding_method == "greedy_search":
|
if params.decoding_method == "greedy_search":
|
||||||
|
@ -304,7 +304,6 @@ def fast_beam_search(
|
|||||||
def decode_one_chunk(
|
def decode_one_chunk(
|
||||||
params: AttributeDict,
|
params: AttributeDict,
|
||||||
model: nn.Module,
|
model: nn.Module,
|
||||||
sp: spm.SentencePieceProcessor,
|
|
||||||
decode_streams: List[DecodeStream],
|
decode_streams: List[DecodeStream],
|
||||||
) -> List[int]:
|
) -> List[int]:
|
||||||
"""Decode one chunk frames of features for each decode_streams and
|
"""Decode one chunk frames of features for each decode_streams and
|
||||||
@ -315,8 +314,6 @@ def decode_one_chunk(
|
|||||||
It's the return value of :func:`get_params`.
|
It's the return value of :func:`get_params`.
|
||||||
model:
|
model:
|
||||||
The neural model.
|
The neural model.
|
||||||
sp:
|
|
||||||
The BPE model.
|
|
||||||
decode_streams:
|
decode_streams:
|
||||||
A List of DecodeStream, each belonging to a utterance.
|
A List of DecodeStream, each belonging to a utterance.
|
||||||
Returns:
|
Returns:
|
||||||
@ -490,7 +487,7 @@ def decode_dataset(
|
|||||||
|
|
||||||
while len(decode_streams) >= params.num_decode_streams:
|
while len(decode_streams) >= params.num_decode_streams:
|
||||||
finished_streams = decode_one_chunk(
|
finished_streams = decode_one_chunk(
|
||||||
params, model, sp, decode_streams
|
params=params, model=model, decode_streams=decode_streams
|
||||||
)
|
)
|
||||||
for i in sorted(finished_streams, reverse=True):
|
for i in sorted(finished_streams, reverse=True):
|
||||||
hyp = decode_streams[i].hyp
|
hyp = decode_streams[i].hyp
|
||||||
@ -509,7 +506,9 @@ def decode_dataset(
|
|||||||
|
|
||||||
# decode final chunks of last sequences
|
# decode final chunks of last sequences
|
||||||
while len(decode_streams):
|
while len(decode_streams):
|
||||||
finished_streams = decode_one_chunk(params, model, sp, decode_streams)
|
finished_streams = decode_one_chunk(
|
||||||
|
params=params, model=model, decode_streams=decode_streams
|
||||||
|
)
|
||||||
for i in sorted(finished_streams, reverse=True):
|
for i in sorted(finished_streams, reverse=True):
|
||||||
hyp = decode_streams[i].hyp
|
hyp = decode_streams[i].hyp
|
||||||
if params.decoding_method == "greedy_search":
|
if params.decoding_method == "greedy_search":
|
||||||
|
Loading…
x
Reference in New Issue
Block a user