100 points · 1 submission
with AWS Kiro
We built a voice-first AI teaching platform that turns any topic into a live, interactive lesson instead of a dead chatbot conversation. The problem it solves is that most learning tools are still text-heavy, passive, and not great for people who learn by speaking, seeing, doing, and asking follow-up questions in real time. Our app gives learners a live tutor they can talk to naturally, respond to with voice, text, drawings, image markup, and interactive canvas activities, then saves the session as a polished study article they can revisit later. It uses ElevenLabs at the core of the experience: ElevenLabs powers the natural tutor voice with low-latency text-to-speech, real-time speech-to-text for learner responses, and the voice-first turn-taking flow that makes the tutor feel conversational instead of robotic. It uses this week’s sponsor, Vercel, to deploy the full Next.js app, run the API routes that orchestrate tutoring sessions, and provide web analytics so we can track visits and unique users after launch. Around that, we also use AI planning, image search/generation, and persistent lesson history to make the product feel like a real multimodal teacher, not just a voice wrapper around an LLM. Shorter version if needed: We built a voice-first multimodal AI tutor that teaches any topic through natural conversation, interactive canvas tasks, visuals, and saved lesson articles. It solves the problem of passive text-only learning by making lessons feel live, adaptive, and reusable. ElevenLabs powers the real-time speech-to-text, natural tutor voice, and voice-driven lesson flow, while Vercel powers deployment, backend routes, and analytics for tracking usage after release.
Submitted 23 Apr 2026