Have something to say?

Tell us how we could make the product more useful to you.

Private Feedback Notifications: Dynamic "Reply-To" Header + Automated Status Sync + Message Content Logging

Problem Statement: Standard private feedback notifications currently use a static ESP system email (repliesESP@mySaaS.com) in the Reply-To header. This prevents support teams from replying directly to customers, forcing agents to manually copy-paste the customer's email address. Proposed Solution: Implement a dynamic Reply-To header that uses the respondent's email address. To ensure this does not interfere with existing configurations for other Agencies/WL-SaaS, this feature could be disabled by default. UI Configuration & Path: The setting could be located under the form customization settings: Path: Feedback Forms > Content & Flow > Private Feedback Feature Toggle: "Advanced Direct Customer Reply by Email" Description: "When enabled, replying to a private feedback email notification will go directly to the customer's email address instead of the system default." Technical Requirements: Default State: Disabled (Off). Logic: If Enabled, the SMTP Reply-To header must be populated with the value from the {Email} field submitted in the feedback form. If Disabled, it should fall back to the current system default. Phase 2: Automated "Replied" Status Sync To avoid manual status updates inside EMR, can we implement an Email Proxy/Relay? Logic: Instead of the customer's direct email, the Reply-To would be a unique system-generated alias (e.g.,{feedback_id}@reply.revw.me). Action: When the server receives an email at this alias, the system should: 1. Parse the feedback_id and change the status of that entry to "Replied". 2. Forward the email content to the actual customer's email address. This would create a seamless loop where the software tracks external email interactions automatically. Phase 3: Message Content Logging: Action: When the system receives the proxy email, it should also parse the body of the reply. Goal: Store the agent's response text within the EMR database so the conversation history is visible in the dashboard. Requirement: Implement logic to strip signatures and original notification text to log only the new content. If needed, we can start with Phase 1 and 2 to solve the immediate workflow pain, leaving the content parsing for a later sprint.

Leandro Teixeira 1 day ago

πŸ’‘ Feature Request

Monthly AI Insights report displays weekly data instead of monthly data

The AI Insights (analysis) and the AI Insights reports are currently linked in an incompatible way. When a monthly AI Insights report is created, the data included in that report still always refers to a weekly period. In other words, the report is sent monthly, but the information shown only covers one week rather than the full month. This makes the report frequency largely irrelevant, except when a weekly frequency is selected, since that is the only case in which it matches the analysis, which is weekly by default. Given this dependency between the AI analysis and the data displayed in the report, I suggest resolving the issue in one of the following ways: Keep the current link, but allow the AI analysis execution frequency to be configured as weekly, biweekly, or monthly. Break this link, allowing the report to generate an AI analysis based on its own frequency. In other words, the report period should match the period of the data shown in the report. Personally, I believe the second option makes more sense. β€”β€”β€”β€”β€”β€”β€”β€”β€”β€”β€”β€”β€”β€”β€”β€”β€”β€”β€”β€”β€”β€”β€”β€”β€”β€”β€”-- Formal Bug Report: Steps to Reproduce Go to the AI Insights Reports configuration. Create a new report with the frequency set to Monthly. Wait for the report to be generated and sent. Open the generated report and review the data period covered. Expected Result A report configured with a Monthly frequency should include data covering the full monthly period. Actual Result A report configured with a Monthly frequency is sent monthly, but the data included in the report only covers a weekly period instead of the full month. Additional Notes The AI analysis appears to be tied to a weekly default execution period, and the report inherits that same weekly dataset regardless of the selected report frequency. As a result, the selected report frequency becomes effectively irrelevant unless the report is also configured as Weekly. Suggested Solutions Allow the AI analysis frequency to be configured as Weekly, Biweekly, or Monthly. Decouple the report from the current analysis frequency, so that the report period matches the data period shown in the report. Preferred Solution Decoupling the report from the analysis frequency makes the most sense, so that each report generates data according to its own selected period.

Leandro Teixeira 1 day ago

πŸ› Bug Reports

Add scheduled batch sending to review request campaigns

Type: Feature request Module: Campaign editor β€” Planification (scheduling) Some users want to send all their review requests at a chosen moment instead of in a constant trickle. Two reasons come up: Deliverability β€” avoid off-hours and saturated time windows. Conversion β€” reach customers when they are actually available to read and click (typically evenings or weekends for the beauty/wellness vertical). Concrete example: a salon owner wants every review request collected during the week to be sent on Sunday at 7pm, when their clients are home and relaxed. Proposed solution Add a sending mode selector inside the existing Planification module of the campaign editor: Instant send β€” current behaviour (delay in hours after the trigger event). Scheduled batch send β€” new mode. Accumulates eligible requests and releases them at the configured time slot. Scope (v1) Inside scheduled batch mode: Recurrence preset dropdown β€” Once / Daily / Weekly on [day] / Monthly / Custom. Pattern inspired by Google Calendar's recurrence picker. Custom recurrence panel β€” pick days of the week, set the exact time, choose timezone. Live preview line β€” "All pending requests will be sent Sunday at 7:00 PM. Next batch: Sunday, May 10." Existing "stop sequence on review attempt" logic must keep working β€” pending batched requests are dropped if the contact already left a review. tHt UX reference See attached mockup. Layout follows the current Planification module; recurrence UI mirrors Google Calendar's Personnaliser panel for familiarity. User value Better open / click / review rates through smarter timing. Marketing-grade control over send moments. Brings EMR closer to standard email/SMS campaign tools.

Jean-Gabriel 3 days ago

πŸ’‘ Feature Request

In Progress

Email, Status, and Activity Sync to CRM via Zapier

Right now, there is no way to extract the email address, β€œstatus,” or most recent activity from EMR using review exports, Zapier, or the API. Why This Is Important When using a CRM to trigger a review request via EMR, it can’t get data on the request’s status or the reviewer’s last action. This makes it difficult to track progress or segment contacts by their status. If the CRM could access this data, it would have complete details on the review requests and be able to follow them from start to finish. Possible Applications for This Sync segment contacts by review rating segment contacts by the date of the last review activity trigger automations to upsell and/or cross sell to satisfied customer send a β€œreward” for customers who have completed the review process Data to Be Shared from EMR to CRM via Zapier Email address (to identify the contact within the CRM) Status (to indicate the contact’s response to a feedback form, such as a rating between 1 and 5 stars) Last activity (to show the final result of the request, like being redirected to leave a review) Date of last activity (to record when the contact last engaged with a review request) All these details can currently be viewed here within EMR (see screenshot): /request-reviews/contacts?statusFilter=completed

Dominic Leger 4 days ago

πŸ’‘ Feature Request

In Progress

Request to Make AI Hub an Independent Feature in the Platform

Description Currently, AI Hub is integrated as a complementary feature within the API & Webhooks module. This means that when API access is sold, clients automatically gain access to AI Hub without the ability to control or monetize it separately. Since other AI-related features like AI Insights and Search AI are offered as independent features, I request that AI Hub also be separated from the API module and offered as a standalone feature. Benefits of this Separation Allows agencies and the platform to monetize AI Hub as an autonomous product with its own plans and pricing. Provides greater control over AI Hub activation and access for clients based on their plans or needs, without it being tied to API access purchase. Enables a more flexible and segmented commercial strategy, similar to that applied to AI Insights and Search AI. Current Impact The direct dependency of AI Hub on the API module limits monetization potential and bundles both features together, which may not be ideal for all clients. Separating AI Hub would allow it to be offered as an upsell or premium add-on, improving profitability and sales options.

DJ Fonseca 5 days ago

πŸ’‘ Feature Request

GBP keyword brand name logic

The new feature now assumes that the location name = business name for branded keywords. Personally I don’t find many cases where you’d set the location name in the dashboard to the actual branded business name (I could also be wrong). Suggestion: use the Organization name across all locations as the default branded keyword, or pull the Place name for each location directly from Google for a more accurate default value (I assume the name is already saved in the database when the Google connection is set up?). It would also be helpful to have a one time (persistent until clicked) visual prompt on the dashboard to check the branded keywords. Great addition to the dashboard though, thanks a lot!

Severi 7 days ago

1

πŸ’‘ Feature Request

Feature Request – AI Fake Review Detector & Google Flag Assistant

This is just a suggestion from a platform user, I hope it’s useful for the roadmap. A quick feature idea that I think would be a real differentiator for EMR: An AI layer that automatically scans incoming reviews, detects potential Google policy violations (fake reviews, competitor attacks, spam, hate speech), scores each review with a Flaggability Score (0-100%), and generates a ready-to-submit flag report citing the exact Google policy being violated. Why it matters: β€’ Most businesses never flag violations because they don’t know how β€’ A structured, policy-cited report gets significantly better results from Google than manual flags β€’ Fake and competitor reviews are a growing problem in many markets β€’ Adds clear, defensible value that helps agencies justify their monthly retainers Thanks for everything you’re building.​​​​​​​​​​​​​​​​

Eng.Sultan Alharbi 7 days ago

πŸ’‘ Feature Request

Webhook event for opt-in submissions + API access to consent records

Currently the platform stores opt-in consent data behind the scenes, but there's no way for agencies to access it programmatically. For US agencies operating under TCPA rules, having a real-time and exportable consent trail is becoming a hard requirement. Two requests: Add a new webhook event (e.g. "optin-created") that fires when a customer submits the campaign opt-in page. The payload should include: contact name, phone/email, timestamp, IP address, and the consent text they agreed to. This would let agencies automatically store consent records on their side without needing to raise a support ticket each time. As a simpler first step, include consent record fields (opted-in timestamp, IP, consent language) in the API response when listing campaign contacts. This would be a strong differentiator for EMR β€” most competing platforms don't offer webhook-level consent tracking, and US agencies are increasingly looking for this as TCPA enforcement tightens. Having it built in would give EMR a compliance edge that competitors like Birdeye and Podium don't currently offer.

Yoan Pumar 8 days ago

πŸ’‘ Feature Request

Custom Package Visibility

Hi, I would like to submit feedback regarding Custom Package visibility. Right now, whenever a Sales Agent attempts to add a new account to his/her dashboard, they can see all of the available custom plans - even those that are designated as β€˜hidden’. It would be helpful if we could designate certain plans to ONLY sales agents, and other plans to the general public. Secondly, it would also be useful if we could toggle off any client-facing visibility to the billing section. When our agency is fully responsible for managing customer reviews and direct billing, it becomes a challenge for/when they see the available plans (with pricing). Hiding the custom plans only solves half the problem, because now those customers who have signed up directly can no longer see the other plans to upgrade to. https://www.loom.com/share/5dcb9289baea4aac85642747214efe86

Review Pulse 360 9 days ago

πŸ“₯ Feedback

Further Search AI bugs/issues

➑️ Sometimes it happens that instead of answering the question the AI will reply with something like this: Would you like me to find and list the one clearly best-rated garage door installer in Tampere, or would you prefer a short top-3 list (company + rating + source and contact information)? I'll tell you the search criteria (for example, Google/Fonecta/Urakkamaailma ratings) before searching, if you'd like. This is probably just a prompt issue, maybe embedded in the prompt should be instructions to answer directly and to not ask followup questions? ➑️ What causes this outcome? Should this not be the raw output even for non-detections? Happens mostly with Gemini and ChatGPT, but also with Perplexity even though much more rare. ➑️ Still plenty of prompts with mixed English and profile language. Or are the prompt templates in English and we see the translations just in the UI?

Severi 10 days ago

1

πŸ› Bug Reports