The more I become accustomed to voice input for my notes, the more I wonder about the optimal integrated workflow with Logseq. Has anyone developed a speech-to-text process alongside Logseq that goes beyond simple copy-pasting?
Not exactly developed but I sometimes use the Recorder App on my Pixel and just share the transcript to Logseq. It’s not perfect but does the job for me.
Yep. That’s what many of us do. In my case I use Tana Capture for the voice memos, and once I have the curated output, if suitable, paste it using Markdown in Logseq.
But the “copy-pasting” strategy seems a bit old-fashioned… tbh.
MacOS has native speech recognition, and MacWhisper for example works well offline across macOS apps too.
But, what is the Logseq role in that situation: copy-pasting directly from the transcript? some text parsing in between?
It directly takes the output of the transcription, you are dictating into Logseq itself. MacWhisper lets you setup a whole range of different AI models to process the transcript in between, but I haven’t done that yet.
You could use the logseq api to add notes, i played a bit with it, it is not perfect, but would be a way to use it. Main problem was to control, when to break a text into peaces.