1,200 points · 6 submissions
with v0
AmazonAI is a fully voice-controlled e-commerce platform built for the ElevenLabs x v0 Hackathon that revolutionizes online shopping through conversational AI. Users can browse products, apply filters, add items to cart, and complete checkout entirely through natural voice commands eliminating the need for clicking, typing, or navigating complex menus. This hands-free approach makes shopping accessible for everyone, whether they're cooking, driving, or have accessibility needs. Core Features Voice-Powered Product Discovery: Users can say "Show me dresses under $50" or "Only red ones with free shipping" and the AI agent instantly applies filters and displays matching products. The agent understands natural language preferences for price, color, brand, category, and shipping options in real-time. Context-Aware Shopping: When viewing a product detail page, users simply say "Add this to cart" and the agent recognizes the current product without requiring a product name. The system automatically navigates to the cart and confirms the action. Conversational Checkout Flow: The agent guides users step-by-step through checkout. Users provide delivery details conversationally ("My name is John Smith", "Email is [john@gmail.com](mailto:john@gmail.com)", "Address is 45 Park Street, Mumbai") and the form fills automatically without manual input. Smart Navigation: Voice commands like "View my cart", "Go back to shopping", "Proceed to checkout", and "Place my order" allow seamless navigation throughout the shopping journey. Real-Time Feedback: A debug panel tracks all voice tool calls, errors, and agent responses in real-time, helping users understand what the agent is processing. User Journey User connects the voice agent on the Shop page. User speaks natural language commands like "Show me keyboards under 100 dollars with free delivery". Agent applies filters and displays results. User browses products and says "Tell me about the wireless keyboard". Agent navigates to product details and reads specifications. User says "Add this to my cart". Agent adds item and navigates to cart page. User reviews order summary and says "Proceed to checkout". Agent navigates to checkout form and asks "What's your full name?". User provides details conversationally, agent fills form fields automatically. User says "Pay by cash on delivery" and agent selects payment method. User says "Place my order". Agent confirms and displays order confirmation. Key Benefits Accessibility: Hands-free shopping for people with mobility challenges, visual impairments, or situational limitations. Speed: Voice commands are faster than traditional clicking and typing for browsing and checkout. Convenience: Users can shop while multitasking, driving, cooking, or with hands full. Natural Interaction: Conversational AI understands colloquial language and context, not rigid command structures. Error Resilience: Smart fallbacks for ambiguous voice input, multiple product matches, and network issues. AmazonAI demonstrates how conversational AI can transform e-commerce into an inclusive, accessible, and frictionless experience. it proves that voice-powered shopping isn't just a nice-to-have feature but a fundamental reimagining of how people interact with online stores.
Submitted 7 May 2026
with AWS Kiro
Most candidates don't fail tech interviews because they lack knowledge they fail because the pressure of someone sitting across from them makes everything they know disappear. Confidence breaks, answers get jumbled, and one bad interview kills the motivation to even try the next one. The real problem isn't preparation it's that there's no safe space to practice under real pressure without being judged. NeuralPrep fixes this by creating a personalized, voice-based interview environment where candidates can practice as many times as they need without fear. Upload your resume, and the system parses it, chunks it into semantic sections, embeds it using Gemini, and indexes it into Qdrant then an ElevenLabs Conversational AI agent grills you on your actual experience over a live WebRTC voice session. It's not generic flashcard prep. The AI references your specific projects, challenges your claims, and adapts its questioning based on your responses. When it detects you're stuck silence, filler words, hesitation it shifts from interviewer to coach, helps you recover, and builds your confidence back up in the same session. After every session, you get instant AI-scored feedback on communication, technical depth, and structure, plus an AI mentor that reviews your history and gives targeted coaching. ElevenLabs made the voice layer possible Conversational AI with dynamic variables means the agent knows your name, your resume, your weak spots, and adjusts in real time. Kiro made the development process structured 7 specs drove the entire build through requirements, design, and tasks, while agent hooks automated commits and health checks, steering kept conventions tight, and the ElevenLabs Power plus Postman MCP handled integrations without leaving the IDE.
with turbopuffer
SonicReal is an AI engine that turns your photos or text into immersive, 3D soundscapes. Imagine looking at a picture of a rainy forest and instantly hearing the rain hitting leaves, distant birds, and soft, moody music. Using Gemini to "see" the scene and turbopuffer to search millions of real-world sound patterns, the app directs ElevenLabs to build a layered, cinematic audio experience. It doesn't just play a sound; it orchestrates a reality. Use Case: It’s perfect for game developers and content creators who need instant, high-quality audio that perfectly matches their visuals without hours of manual editing.
with Replit
Dating today is exhausting. More often than not, we end up choosing people based on a quick glance a photo, a vibe rather than their values. And honestly, that makes sense. It’s hard to truly understand someone without spending time talking to them. But what if we could change that? What if, instead of endlessly swiping, you only interacted with people who genuinely share your values, your energy, and your outlook on life? That’s exactly what we’re building. Our approach starts differently. Instead of leading with photos, we begin by understanding *you*. Through a series of thoughtful, judgment-free questions, we get to know who you are as a person. What are you looking for in a partner? What kind of life do you envision? What are your expectations around career, responsibilities, and compatibility? What are your deal-breakers—your “icks”? These are the things that actually define meaningful connections. Based on your answers, we match you with people who align with your values and mindset not just your surface-level preferences. But here’s where it gets really interesting. We use advanced conversational AI combined with voice cloning to create a personalized version of you an AI that talks like you, sounds like you, and represents your personality. So instead of awkward first chats or forced small talk, your AI interacts on your behalf. When you connect with someone, you’re not just reading text you’re experiencing a conversation. You hear their voice, their tone, their personality. And they hear yours. It feels natural, human, and far more meaningful than traditional messaging. At this stage, neither of you sees each other’s full profile. There’s no pressure, no judgment—just authentic interaction based on who you are. If the conversation clicks if the energy feels right you can choose to like each other. Once there’s mutual interest, both of you unlock the next step: real profiles, real conversations, and the opportunity to take things forward. And if the other person prefers, they can even interact with your AI first before deciding to match making the entire process more comfortable and less intimidating. We’re building all of this using powerful tools like ElevenLabs for voice cloning and conversational AI, combined with an integrated development environment that lets us design, build, and launch everything seamlessly in one place. The result? A dating experience that prioritizes compatibility over appearance, depth over small talk, and genuine connection over endless swiping. Because finding the right person shouldn’t feel like a chore. It should feel like meeting someone who truly gets you.
with Cloudflare
Deverse is a persistent 3D playground where developers collaborate with autonomous AI agents. Instead of chat windows, we’ve built a spatial world where you walk up to AI engineers like Aria (Frontend) and Kai (Backend) and talk to them naturally using your voice. It provides platform for developers around the corner so they can collaborate with their fellow developers and can build together. Also team working remotely can access private arena so they can have their privacy. We have set of autonomous agent that are master of their respective field and they can really help with u so u can interact with them in both arena's and good part u can give them your own preference and they will act like that . We use ElevenLabs to give our agents a human soul. By integrating the ElevenLabs Text-to-Speech Streaming API with the eleven lab's model, our agents respond with ultra-realistic voices in real-time. This creates a "Voice-to-Voice" loop that feels like a natural conversation with a real senior engineer. Deverse is 100% "Edge-Native" to ensure zero-latency 3D interaction: Durable Objects: Syncs the 3D world state and AI characters across users globally. Vectorize: Serves as the AI's long-term memory for Spatial RAG, remembering your project details across sessions. Workers AI: Powers Whisper (Voice-to-Text) and Llama 3 (Agent Reasoning) right at the edge. D1 & R2: Handles developer profiles and 3D world infrastructure with high performance.
with Firecrawl
Serenity Voice is a continuous AI companion that replaces your expensive theraphy sessions, with an always-present, deeply personalized companion. The problem: Every day, people struggle with stress, anxiety, and self-reflection but meaningful support is either expensive (therapy) or shallow (generic wellness apps). Users end up journaling alone, repeating their thoughts without receiving context-aware guidance that actually evolves with them. Serenity eliminates this entirely. Speak or write naturally and a pipeline of intelligent systems builds a living understanding of you: ElevenLabs transcribes your voice in real-time, capturing tone and emotional nuance Your journal entries and conversations are stored in a persistent “Memory Bank,” building long-term context about your thoughts, patterns, and experiences A reflection agent analyzes emotional trends, recurring themes, and behavioral signals over time An AI reasoning layer synthesizes your past present state delivering deeply personalized, context-aware guidance ElevenLabs generates natural, human-like voice responses for a calming, immersive experience The loop continues across sessions continuously evolving with you, not resetting every time How it uses ElevenLabs: Voice-to-text transcription (STT) captures real-time thoughts with emotional fidelity, while text-to-speech (TTS) delivers warm, natural responses using streaming APIs enabling fluid, human-like conversations that feel safe and reflective. How it uses Firecrawl: Each interaction can trigger Firecrawl to retrieve relevant, real-world wellness knowledge grounding the AI’s responses in psychological insights, coping strategies, and evidence-based practices. This ensures Serenity’s guidance is not just empathetic, but also informed and actionable. Built as a continuous system: Serenity is designed as a long-term companion not a chatbot. It grows with you, remembers you, and adapts to you delivering personalized care anytime, anywhere.
Submitted 23 Apr 2026
Submitted 16 Apr 2026
Submitted 4 Apr 2026
Submitted 2 Apr 2026
Submitted 25 Mar 2026