Skip to main content

    95 Ideas

    Native WhatsApp Forms (In-App Forms)New

    Currently, when running WhatsApp ad campaigns, users must send a keyword or message first before a ManyChat flow can start. This extra step creates friction in the lead onboarding process and can negatively affect conversion rates — especially in paid campaigns.Proposed Solution:Introduce native WhatsApp Forms (in-app forms) that open directly inside WhatsApp when a user clicks a link or ad.These forms would allow users to fill in fields such as name, email, preferences, or other custom data without needing to send a message first.Why This Matters:I build automated WhatsApp funnels via ManyChat for coaches and consultants in the DACH region, and this feature would be a real gamechanger for WhatsApp automation and advertising.Platforms like Spoki already offer this kind of native WhatsApp form experience — enabling smoother lead capture and higher conversion rates.With native WhatsApp forms, advertisers could: Simplify the ad-to-chat flow dramatically Collect structured data directly in WhatsApp Improve the first-contact experience for new leads Optimize campaign performance and reduce drop-offs Questions: Is native in-WhatsApp form support currently on the ManyChat roadmap for 2025? If not yet, is there any workaround (e.g., via structured messages, templates, or custom API logic) to simulate this behavior? Impact:This feature would make WhatsApp ads much easier to optimize, improve lead quality, and reduce friction in funnel automation — especially for businesses running WhatsApp-based lead generation at scale.

    Rodrigo Many
    Rodrigo ManyUp-and-Comer

    WhatsApp and Instagram still do not support audio on automation platforms (Make.com or N8N)New

    English:Currently, sending audio files is not supported on WhatsApp and Instagram channels when used through integration platforms such as Make.com and N8N. This limitation significantly compromises the experience, as audio is one of the most popular forms of communication among users—especially on WhatsApp here in Brazil.Therefore, we request that this feature be enabled, allowing audio messages to be sent natively. Make itself already supports this feature, but since ManyChat doesn't support the audio format, all automation is rendered impossible.Enabling this feature would make communication much more natural, practical, and aligned with the already established usage patterns on these platforms. It's worth noting that this is a basic feature, already present in competing solutions, which reinforces its importance and urgency.We thank you in advance for your attention and hope that this improvement will be prioritized. Portuguese BR: Atualmente, o envio de arquivos de áudio não é suportado nos canais do WhatsApp e Instagram quando utilizados via plataformas de integração como Make.com e N8N. Essa limitação compromete bastante a experiência, já que o áudio é uma das formas de comunicação mais utilizadas pelos usuários — especialmente no WhatsApp aqui no Brasil.Solicitamos, portanto, a habilitação dessa funcionalidade, permitindo que mensagens de áudio sejam enviadas de forma nativa. Hoje, o próprio Make já oferece suporte a esse recurso, mas como o ManyChat não libera o formato de áudio, toda a automação acaba inviabilizada.A liberação dessa função tornaria a comunicação muito mais natural, prática e alinhada ao padrão de uso já consolidado nessas plataformas. Vale destacar que trata-se de um recurso básico, já presente em soluções concorrentes, o que reforça sua importância e urgência.Desde já, agradecemos pela atenção e esperamos que essa melhoria seja priorizada. Font: https://manychat.github.io/dynamic_block_docs/channels/

    fvayas
    fvayasExplorer

    Enable multimodal AI in ManyChat understand voice notes and imagesNew

    Summary:I propose that ManyChat’s integrated AI can understand audio (transcribe + extract intent) and images (OCR / recognize context). Today users prefer sending voice notes and images; allowing ManyChat to process them natively will reduce friction and enable new support and sales flows.Problem: Many users send voice messages and images because it’s more convenient; bots currently ask them to type or repeat information. This causes delays and extra work for agents. Proposal: Enable processing of voice messages to produce a transcription and the user’s main intent. Enable basic image recognition (read text in photos like tickets/receipts and detect image type: product / receipt / ID). Allow flows to combine voice + image + text to make decisions (for example: detect a complaint and automatically create a ticket). Use cases: Support: customer sends a product photo and says by voice “it arrived broken” → bot identifies order and creates a refund proposal or ticket. Commerce: user sends a photo of a product and asks the price by voice → bot replies with options and a purchase button. Pre-human assistance: automatic summary of audio + photo so the agent sees the essentials before replying. Accessibility: people who have trouble writing use voice and images to complete forms. Closing / Request:Please consider prioritizing multimodal capabilities (processing audio and images) in ManyChat’s AI. I can provide concrete flow examples if the product team wants them.

    fvayas
    fvayasExplorer

    Enable multimodal AI in ManyChat — understand voice notes and imagesNew

    Summary:I propose that ManyChat’s integrated AI can understand audio (transcribe + extract intent) and images (OCR / recognize context). Today users prefer sending voice notes and images; allowing ManyChat to process them natively will reduce friction and enable new support and sales flows.Problem: Many users send voice messages and images because it’s more convenient; bots currently ask them to type or repeat information. This causes delays and extra work for agents. Proposal: Enable processing of voice messages to produce a transcription and the user’s main intent. Enable basic image recognition (read text in photos like tickets/receipts and detect image type: product / receipt / ID). Allow flows to combine voice + image + text to make decisions (for example: detect a complaint and automatically create a ticket). Use cases: Support: customer sends a product photo and says by voice “it arrived broken” → bot identifies order and creates a refund proposal or ticket. Commerce: user sends a photo of a product and asks the price by voice → bot replies with options and a purchase button. Pre-human assistance: automatic summary of audio + photo so the agent sees the essentials before replying. Accessibility: people who have trouble writing use voice and images to complete forms. Closing / Request:Please consider prioritizing multimodal capabilities (processing audio and images) in ManyChat’s AI. I can provide concrete flow examples if the product team wants them.