Another day, another AI app idea. As I have been discussing earlier I do a daily walk in the woods. From time to time I walk around with my headphones on and start to talk about my day into the voice recorder on my iPhone. I enjoy structuring my thoughts this way, as I get to really think through ideas I had that day. However that is where most of the times it ends. Very sporadically I end up transcribing and summarizing my audio file with ChatGPT, but most of the time I don’t want OpenAI to know my raw thoughts. The recordings are messy and often contradictory. There is no follow up and the recording just becomes stale over time as I don’t do anything with it.

That is something I wanted to fix, but in a way that is very low friction. What if I could have a small MacOS app that runs in my menu bar that I could drop the audio file in. And then starts to automatically transcribe and summarize my ramblings on my local machine. That would be super helpful turning the chaos in my brain into something coherent, but what if that is automatically added to my Obsidian vault as a daily note for the day the audio file was recorded. Even better!

So I set out to start creating a prompt for OpenAI Codex. I find that I easily default to OpenAI tools as my plus account accommodates my needs perfectly fine and reduces my cognitive load. This lowers the barrier to actually start building and makes me go super fast. I wish to start experimenting with Claude Code as well, but so far I don’t have had a very good use case as Codex works really well IMO.

Attached Document

Execution Plan: Created by Codex based on my answers

content/experiments/experiment-11-local-transscription-summary-mac-toolbar-app/plan.md · showing first 15 lines

Then the back and forth between Codex and me started again. This is becoming something I start to get used to, it comes with a version that is about 80% correct and then it needs some iterations on some missed dependencies and crashes. Things like that still happen, but after a few tries it gets it fixed and working correctly.

After about 90 minutes I had the final version running with the result I wanted. I dragged the audio file from Voice Memo app into the drop zone. It started downloading the local models needed, transcripts my audio file and summarized it. All removing repetition and giving me insights I was unaware of before reading the summary. Then as the finishing touch it was automatically being added as a daily note in my note taking app.

The main question I had was would my ramblings and the way I speak change now that I know that the result will become structured and coherent?

Before my thoughts during the walks were fleeting and most of them disappeared, but now everything becomes indexed and searchable. So I ask myself is forgetting thoughts a feature of thinking?

transcription-menubar-drop-zone.png

Key insight

When I structure my thinking it uncovers patterns that hide in my ramblings. By automating this structuring via the summarizing with AI, am I working toward selection? If that is the case I still forget things as they are not selected. So what happens when selection is done automatically and externalized? Does it allow for contradictions?