Remix.run Logo
mparas 8 hours ago

I'm a solo dev from India who moved to Berlin a few years ago. I built Sensonym because I was bored with every language app feeling the same, and I wanted to find out what happens when you make vocabulary learning physical.

The idea is that tying vocabulary to physical actions makes it stick better. So I tried mapping every word to a physical action using phone sensors. Some examples:

- To learn the word for "drink", you tilt your phone toward your mouth like a glass

- To learn the word for "blow", you blow into the microphone

- To learn the word for "listen", you bring the phone to your ear

- To learn the word for "eat", you plug in your charger

- To learn the word for "remember", you take a screenshot

The app contains two modes: story mode, where sensor interactions and vocabulary are woven into the narrative, and training mode for quick single-word drills.

Sensonym supports 10 languages and is live in Germany (iOS + Android), expanding soon. If you're outside Germany, you can sign up on the website to get notified when it launches in your region or contact me at hello@sensonym.com to get a beta test invite.

I would love to hear your honest feedback. What do you think of the general approach (sensor interactions and stories)? Do the sensor-word mappings feel intuitive or forced? Any interaction ideas I'm missing?