프롬프트 지원
whisper에서도 prompt를 사용하여 좀더 다양한 효과를 줄수 있어보인다.
https://github.com/openai/whisper/blob/main/whisper/transcribe.py
def transcribe(
model: "Whisper",
audio: Union[str, np.ndarray, torch.Tensor],
*,
verbose: Optional[bool] = None,
temperature: Union[float, Tuple[float, ...]] = (0.0, 0.2, 0.4, 0.6, 0.8, 1.0),
compression_ratio_threshold: Optional[float] = 2.4,
logprob_threshold: Optional[float] = -1.0,
no_speech_threshold: Optional[float] = 0.6,
condition_on_previous_text: bool = True,
initial_prompt: Optional[str] = None,
word_timestamps: bool = False,
prepend_punctuations: str = "\"'“¿([{-",
append_punctuations: str = "\"'.。,,!!??::”)]}、",
**decode_options,
):
"""
initial_prompt: Optional[str]
Optional text to provide as a prompt for the first window. This can be used to provide, or
"prompt-engineer" a context for transcription, e.g. custom vocabularies or proper nouns
to make it more likely to predict those word correctly.
"""prompt를 지원
tag : v0.8 을 보자.
query string으로 받아서 옵션에 넣어주자.
테스트해보자.
프롬프트를 넣엇을때 성과가 어떤지는 테스트가 좀 안된다. //todo

Last updated
Was this helpful?