It's disingenuous because literally none of the code transcribes or translates audio.
This is NOT an app that transcribes, or translates, audio.
This is a front end to another companies service.
In its defense, it is a useful front end, because getting whisper running locally was a pain in the butt because of py-torch's specific python requirements (not too old, not too new... juuuuust right).
This app also looks like it does very useful things with what whisper outputs.
But it is 100% disingenous because it does none of the things it markets itself as doing. I was expecting it to run whisper locally, not call out to a paid service.
Download Whisper and the models, run it in a Docker container as a server, and it's Open Source.
Honestly, try see it as a favor that it's using OpenAI's endpoint, since some of us won't think it's feasible to have a GPU-loaded server running 24/7 just for some occasional transcriptions.
This is NOT an app that transcribes, or translates, audio.
This is a front end to another companies service.
In its defense, it is a useful front end, because getting whisper running locally was a pain in the butt because of py-torch's specific python requirements (not too old, not too new... juuuuust right).
This app also looks like it does very useful things with what whisper outputs.
But it is 100% disingenous because it does none of the things it markets itself as doing. I was expecting it to run whisper locally, not call out to a paid service.