Currently, it appears that audio is handled as a binary attachment, and is then transcribed. ref
|
Future<void> _onTranslateStt(XFile file) async { |
For multimodal models such as Gemini, audio as an input is natively supported.
The expectation is that instead of an audio attachment that is transcribed, the audio should be used as the input to the model directly rather than the transcription.