Pulse Survey: A Small Business Guide
How small businesses run pulse surveys without enterprise software. Question library, cadence, response rates, action on feedback, and common mistakes.
Pulse Survey
A practical guide for small businesses running pulse surveys without enterprise software
The first time I ran a pulse survey at one of my early companies, I made every classic mistake. I sent a 35-question survey to my 18-person team and called it lightweight. I made it weekly because someone told me weekly was best practice. I never communicated what I was going to do with the results, and I never communicated what I actually did with the results, because I never actually did anything with the results. By the third cycle, response rate had dropped from 80% on the first survey to under 30% on the third. The team had correctly read the signal: the survey was not connected to action, and engaging with it was a waste of their time. Two of my best engineers told me later they had stopped responding because they realized I was not going to do anything with what they wrote. The total cost was about 12 hours of my own time across three months, plus the engagement loss from running a program that signaled to the team that their feedback did not matter.
Most articles about pulse surveys are written by enterprise survey software vendors who have an incentive to make the practice sound more complex and platform-dependent than it is. Reading them as a small business operator running a 12-person or 30-person team is misleading. The dynamics at small business scale are different in ways that matter. Most enterprise pulse survey advice fails when ported down without adjustment, and the version that works at 5-100 person companies is shorter, simpler, less platform-dependent, and far more focused on the action loop than on data sophistication. The right pulse survey practice produces meaningful signal and visible action; the wrong one produces survey fatigue and lost trust.
This guide covers what a pulse survey actually is at small business scale, how it differs from annual engagement surveys and other feedback formats, why pulse surveys matter for small teams specifically, the four main types (engagement, onboarding, wellbeing, eNPS) and when to use each, how to pick the right cadence for your team, a question library of 25+ tested questions organized by category, the question design principles that distinguish good surveys from bad ones, how to actually get high response rates, anonymity and confidentiality practices, the action loop that makes surveys worth running, how to avoid survey fatigue, the DIY-versus-platform decision, common mistakes that destroy programs, and how pulse surveys fit into broader engagement practice. I built FirstHR for small businesses operating at exactly this scale, and the perspective here is shaped by what works in the field across teams from 10 to 100 employees.
What a Pulse Survey Actually Is
Three things a pulse survey is not, despite frequent confusion. First, it is not the same as an annual engagement survey. Annual engagement surveys are 40-100 questions, run once per year, designed for comprehensive baseline measurement. Pulse surveys are 5-10 questions, run monthly or quarterly, designed for continuous signal. The two are complementary rather than competitive; mature programs often run both. Second, it is not the same as a one-on-one conversation. The 1-on-1 covers individual context, performance, development; pulse surveys cover team-level patterns that aggregate across individuals. Confusing them produces 1-on-1s that feel like checkboxes and pulse surveys that feel intrusive. Third, it is not the same as a customer Net Promoter Score adapted for employees. While eNPS is one type of pulse question, a complete pulse program covers multiple dimensions (engagement, manager relationship, team, growth, wellbeing) that a single NPS question cannot capture.
The simplest working definition I use: a pulse survey is a short feedback instrument designed to maintain a continuous signal about how the team is doing, used as part of a feedback loop where leadership commits to specific actions based on results. The phrase "feedback loop" is doing real work in that definition. A pulse survey without an action loop is data collection theater; the team learns within 2-3 cycles that the survey is performative, response rates collapse, and the program produces negative engagement effect. The loop is what makes the practice worth running.
Pulse Survey vs Annual Engagement vs Other Feedback Formats
Pulse surveys sit alongside several other feedback formats, each with different purposes and tradeoffs. Understanding which format fits which need is essential to building a coherent feedback program rather than running everything inefficiently.
| Format | Length | Cadence | Best for | Limitations |
|---|---|---|---|---|
| Annual engagement survey | 40-100 questions | Once per year | Comprehensive baseline, strategic planning input | Slow feedback loop; data is stale by the time it is analyzed |
| Pulse survey | 5-10 questions | Monthly or quarterly | Continuous signal, early problem detection | Limited depth per cycle; requires sustained action loop |
| 1-on-1 meetings | 30-60 minutes | Weekly or bi-weekly | Individual context, performance, development | Manager-employee dyad only; no team-level pattern visibility |
| Stay interviews | 30-60 minutes | Quarterly | Retention risk identification at individual level | Time-intensive; requires specific manager skill |
| Exit interviews | 30-60 minutes | At separation | Understanding why people leave | Lagging indicator; the person is already leaving |
| 360-degree feedback | 20-40 questions per rater | Annual or bi-annual | Multi-source view of individual performance | Heavy administrative cost; not for tracking team-level engagement |
| Onboarding pulse | 5-7 questions | Days 7/30/60/90 | New hire experience and early intervention | Specific to onboarding window; not for ongoing team monitoring |
The pattern across these formats: they are complementary, not substitutable. A team running only annual engagement surveys gets a slow feedback loop and stale data. A team running only pulse surveys lacks the comprehensive baseline that informs strategy. A team running only 1-on-1s lacks team-level pattern visibility. The integrated practice that works combines pulse surveys for continuous signal, 1-on-1s for individual context, annual engagement surveys (when team size justifies them) for baseline depth, and stay interviews for retention risk. The employee feedback guide covers the day-to-day feedback skill that complements pulse survey programs, and the one-on-one meeting guide covers the recurring conversation cadence.
Why Pulse Surveys Matter for Small Teams
The case for pulse surveys at enterprise scale is well-documented in business literature. The case at small business scale is actually stronger, but it is rarely written about because most pulse survey content is produced by vendors selling to large companies. Three dynamics make pulse surveys particularly leveraged at 10-100 person companies.
First, each engagement signal matters more. On a 1,000-person team, one disengaged person represents 0.1% of the workforce; their disengagement is unlikely to spread visibly before HR catches it. On a 12-person team, one disengaged person represents 8% of the workforce, their attitude affects every meeting they attend, and disengagement spreads quickly through small teams. The early signal that a pulse survey provides is correspondingly more valuable; catching disengagement at month 3 rather than discovering it at the resignation in month 9 is the difference between fixable and irreversible.
Second, small businesses cannot absorb the cost of preventable resignations. Work Institute research on voluntary turnover consistently identifies engagement-driven factors (manager relationship, growth opportunity, recognition) as major contributors to resignations. The cost of replacing a knowledge worker is typically estimated at 50-200% of annual salary, and at small business scale that math becomes existential. A single preventable resignation on a 12-person team often costs more than years of investment in pulse survey practice.
Third, small business teams operate without the management infrastructure that catches engagement problems naturally. Enterprise companies have HR business partners, manager training programs, formal feedback mechanisms, and engagement professionals whose job is to monitor team health. Small businesses have the founder, who is also handling sales, product, and operations. The pulse survey is one of the few mechanisms that provides systematic engagement signal without requiring dedicated headcount. Gallup research consistently shows that the manager-employee relationship accounts for a substantial portion of variance in engagement; pulse surveys make that relationship visible at scale.
The Four Types of Pulse Surveys
Most teams confusingly try to capture everything in a single pulse survey type, producing surveys that do not work well for any specific purpose. The clearer approach is to distinguish among the four main types of pulse surveys, run each at appropriate cadence, and design questions that match the specific purpose. The four types covered below address different signals and require different tactics.
Three principles for choosing among the types. First, start with engagement pulse if you are starting a program from scratch. Engagement covers the broadest range of important signals and produces the highest leverage from a single survey type. Second, add onboarding pulse as soon as your hiring volume justifies it (typically 5+ hires per year). Onboarding pulse catches problems early when they are still fixable; without it, onboarding gaps surface only at exit interviews when it is too late. Third, treat eNPS as a leading indicator alongside other formats, not as a complete program. The single-question simplicity is appealing but produces shallow signal; combine eNPS with at least one other type to get usable depth.
The wellbeing pulse deserves special treatment because the topic is sensitive and the question design is harder than for other types. SHRM guidance on connecting organizational pulse to business outcomes reinforces that wellbeing-focused pulse surveys benefit from particularly careful question construction, robust anonymity protections, and clear action commitments. The employee wellness guide covers the broader context that wellbeing pulse surveys should support rather than substitute for.
The Right Cadence for Your Team
Cadence is the second most consequential design decision after type selection. Too-frequent cadence produces survey fatigue and falling response rates; too-infrequent cadence produces signal that is too stale to support timely action. The right cadence depends on team size, change rate, and action capacity.
The single most common cadence mistake is running surveys frequently to feel rigorous without the capacity to act on results between cycles. A team running monthly surveys but only acting on results quarterly is essentially running quarterly surveys with three months of fatigue overhead. The discipline that matters: match cadence to action capacity, not to data appetite. If you can only commit to action twice per year, run pulse surveys twice per year. If you can act monthly, run monthly. The cadence that fits your action capacity produces dramatically better outcomes than the cadence that fits your aspirations.
For small teams of 10-30 people, the practical recommendation is quarterly engagement pulse plus onboarding pulse for new hires, scaling up to monthly engagement pulse only when team size or change rate justify the additional administrative cost. For teams of 30-100, monthly engagement pulse plus onboarding pulse plus event-driven wellbeing pulse during high-pressure periods covers most of the engagement signal needs. SHRM research on pulse survey frequency reinforces that the optimal cadence varies by team and that experimenting with different cadences often produces better results than committing rigidly to one frequency.
25+ Pulse Survey Questions by Category
The question library below covers the five main dimensions of engagement pulse surveys: meaningful work, manager relationship, team experience, growth, and wellbeing. Pick 5-10 questions across categories rather than going deep on one. Use 5-point Likert scales (Strongly disagree to Strongly agree) for most questions plus one or two open-ended questions for context.
| Category | Question | Type |
|---|---|---|
| Meaningful work | I find my work meaningful and engaging. | Likert 1-5 |
| Meaningful work | I have what I need to do my job well. | Likert 1-5 |
| Meaningful work | I understand how my work contributes to the company's goals. | Likert 1-5 |
| Meaningful work | My role plays to my strengths most of the time. | Likert 1-5 |
| Meaningful work | What is one thing about your work that energizes you right now? | Open-ended |
| Manager relationship | My manager gives me useful feedback about my work. | Likert 1-5 |
| Manager relationship | I trust my manager to advocate for my development. | Likert 1-5 |
| Manager relationship | My manager makes time for our 1-on-1s consistently. | Likert 1-5 |
| Manager relationship | I can raise concerns with my manager without fear of negative consequences. | Likert 1-5 |
| Team experience | Our team works well together to get things done. | Likert 1-5 |
| Team experience | I feel like I belong on this team. | Likert 1-5 |
| Team experience | We hold each other accountable in healthy ways. | Likert 1-5 |
| Team experience | Conflicts in our team get addressed rather than ignored. | Likert 1-5 |
| Growth | I see meaningful opportunities for growth in this role. | Likert 1-5 |
| Growth | I am learning the skills I want to develop in my career. | Likert 1-5 |
| Growth | I have had at least one growth-focused conversation in the past quarter. | Likert 1-5 |
| Wellbeing | My workload is sustainable for me right now. | Likert 1-5 |
| Wellbeing | I am able to disconnect from work when I need to. | Likert 1-5 |
| Wellbeing | I am not currently experiencing burnout signals. | Likert 1-5 |
| Wellbeing | What would help you do your best work that we are not currently doing? | Open-ended |
| eNPS | How likely are you to recommend this company as a place to work? (0-10) | NPS scale |
| eNPS | What is the main reason for your score? | Open-ended |
| Onboarding (Day 7) | I have what I need to do my job in this first week. | Likert 1-5 |
| Onboarding (Day 7) | I understand what is expected of me in my first 30 days. | Likert 1-5 |
| Onboarding (Day 30) | I am clear on my role and how I contribute to the team. | Likert 1-5 |
| Onboarding (Day 30) | What is one thing that would make your onboarding better? | Open-ended |
| Onboarding (Day 90) | I am confident I made the right decision joining this company. | Likert 1-5 |
| Onboarding (Day 90) | What surprised you most about your first 90 days? | Open-ended |
Three principles for using the question library. First, repeat core questions across cycles for trend analysis. The same 5-7 questions every month allow you to see whether engagement is rising or falling on specific dimensions. Second, rotate optional questions to keep the survey fresh. After core questions, add 1-2 rotating questions covering specific topics relevant to recent events or current concerns. Third, always include at least one open-ended question. The Likert numbers are easier to chart but the open-ended responses contain most of the actual signal; teams that aggregate open-ended responses into themes consistently produce more value from surveys than teams that only track score deltas.
Question Design Principles
Question design matters more than most founders realize. Biased questions produce biased answers, leading questions train the team to give the answers they think are wanted, double-barreled questions produce uninterpretable results, and overly abstract questions produce inconsistent interpretation across respondents. Five principles distinguish good pulse survey questions from bad ones.
First, avoid leading language. "How satisfied are you with our excellent benefits?" is leading; "How satisfied are you with the benefits we offer?" is neutral. The leading version produces answers calibrated to the language; the neutral version produces actual signal.
Second, avoid double-barreled questions. "My manager gives me useful feedback and supports my development" combines two ideas; if a respondent agrees with one and disagrees with the other, their score is uninterpretable. Split into two questions even though it adds length.
Third, use specific behavior-anchored language. "I trust my manager" is abstract and produces inconsistent interpretation. "I can raise concerns with my manager without fear of negative consequences" describes specific behavior and produces consistent interpretation.
Fourth, match the scale to the question. Likert scales (Strongly disagree to Strongly agree) work for most attitudinal questions. NPS scales (0-10) work for recommendation questions specifically. Frequency scales (Never to Always) work for behavior questions. Mismatched scales produce confused respondents and degraded data quality.
Fifth, test new questions before launching them in production. Send a draft to 3-5 trusted team members and ask them what each question means and how they would answer. The interpretation you get back often differs from your intended interpretation; iterate before running the question with the full team.
How to Actually Get High Response Rates
Response rate is the foundation of a pulse survey program; below 50%, the data quality degrades to the point where action becomes risky because the responses may not represent the team. Below 30%, the program is structurally broken. Three structural factors drive response rates more than any tactic.
First, demonstrated action on previous surveys. The strongest predictor of future response rate is whether past responses produced visible change. Teams that experience their feedback producing action engage with future surveys at much higher rates; teams that experience their feedback disappearing into a void disengage. The action loop is the response rate driver.
Second, perceived anonymity. Team members must trust that responses cannot be traced to them, particularly when responses involve criticism. The structural anonymity practices that matter: do not collect identifying information unless necessary, suppress segment results below 8-10 respondents, store raw data with appropriate access controls, and explicitly communicate the anonymity practices in survey communications.
Third, visible leadership participation. If the founder or CEO publicly discusses results, commits to specific actions, and references those actions in subsequent communications, the signal that the survey matters becomes credible. Surveys that arrive without leadership engagement signal that the leadership does not actually care about the results.
Tactical adjustments help at the margin. Send the survey from a real person rather than a no-reply address. Send brief reminders to non-respondents two days before closing. Keep surveys to 3-5 minutes. Avoid running surveys during peak deadline weeks or holiday periods. SHRM survey templates and guidance reinforce these tactical adjustments. But none of these tactical adjustments compensate for the structural factor of whether the team believes responses produce action.
Anonymity and Confidentiality
Anonymity dramatically increases response honesty, particularly for questions about manager relationship, leadership decisions, or workplace concerns that team members fear could produce retaliation. Without trusted anonymity, pulse surveys produce calibrated responses that match what team members think is wanted rather than what they actually believe; the resulting data has limited useful signal.
Five anonymity practices matter most at small business scale. First, do not collect identifying information unless necessary. Most pulse surveys do not need name, email, or detailed demographic information; collecting it weakens anonymity without improving signal quality. Second, suppress segment results below 8-10 respondents. Slicing results by small teams of 3-5 people often de-anonymizes individual responses, particularly when combined with demographic intersection; the structural fix is to not report results below the suppression threshold.
Third, store raw data with appropriate access controls. Raw response data should be accessible only to the people running the survey program; managers should not have access to raw data from their direct reports' responses. Fourth, explicitly communicate the anonymity practices. Tell the team in survey communications how anonymity is protected, who has access to data, and what aggregations will be reported. The explicit communication is what makes the protection credible.
Fifth, never use survey responses for individual performance management. The temptation to use anonymous responses to identify and address specific individuals is strong; resisting it is what preserves the anonymity that makes the program valuable. Once team members suspect that responses are being traced back to specific individuals, response honesty collapses across the entire program, not just for that incident.
Acting on Results: The Most Skipped Step
The five-step action loop below is the most consequential part of a pulse survey program. Most failing programs collapse not because of bad question design or poor cadence but because the action loop is missing. The team responds, leadership reviews, nothing visible happens, response rates drop. The discipline of consistently executing the action loop is what makes pulse surveys worth running.
Three principles for the action loop. First, speed signals seriousness. The seven-day window between survey closing and team communication is non-negotiable; longer delays signal that the responses are not being treated urgently. Second, specific actions beat broad commitments. "We will work on improving manager relationships" is too vague to be evaluable; "We are launching a manager training program starting next month" is specific and evaluable. The team can tell the difference. Third, communicate non-action transparently. Some surfaced issues will not be addressed; saying so explicitly preserves trust, while staying silent on them produces the impression that the survey was theater.
The most common action loop failure at small business scale is the founder reading the results, having a series of internal meetings about them, producing thoughtful internal analysis, committing to a thoughtful response strategy, and then never communicating any of it to the team. The internal work was real; the team's experience was that nothing happened. From the team's perspective, the indistinguishable-from-doing-nothing internal work is functionally equivalent to actually doing nothing. Gallup research on engagement drivers consistently identifies the perception that leadership listens and acts as one of the strongest engagement factors; the action loop is what converts that perception from theory into team experience.
Survey Fatigue and How to Avoid It
Survey fatigue is the gradual decline in response quality and rate as team members become exhausted with feedback collection. The classic signs: response rates dropping cycle over cycle, open-ended responses becoming shorter and less thoughtful, increasing patterns of straight-line responses (all 3s or all 5s) that signal disengaged completion, growing complaints in 1-on-1s about survey overhead.
Five practices reduce survey fatigue most effectively. First, keep surveys short. 5-10 questions, 3-5 minutes maximum. Length is the single biggest fatigue driver; cutting from 20 questions to 10 typically increases response rate by 15-25 points. Second, match cadence to action capacity. Do not run monthly surveys if you can only act on results quarterly; the unmatched cadence produces fatigue without compensating benefit. Third, close the loop quickly. The action loop both produces engagement (which reduces fatigue) and signals that responses matter (which increases willingness to engage with future surveys).
Fourth, vary question content across cycles while keeping core questions stable. The core questions you need for trend analysis should remain consistent; supplementary questions can rotate to address topical concerns. The variation prevents the survey from feeling stale while preserving the trend data. Fifth, demonstrate that responses produce action. The strongest fatigue antidote is the visible execution of changes resulting from previous feedback; teams that experience their input producing change tolerate higher cadences without fatigue.
Survey fatigue is not really about survey volume; it is about the perception that responses do not matter. Teams running monthly 7-question surveys with strong action loops report less fatigue than teams running quarterly 25-question surveys with weak action loops. The structural variable is the action loop, not the cadence or length.
DIY vs Platform Decisions
The choice between running pulse surveys through free tools and investing in dedicated pulse survey software is one of the most common questions at small business scale. The decision depends on team size, program maturity, and analysis overhead.
Below 30 employees, free survey tools or built-in form features in existing platforms typically work fine. The administrative overhead of manual approaches is low at small scale (analyzing 15 responses takes about an hour), and the cost of a dedicated platform usually exceeds the time saved. The key requirement is whether the free approach supports the structural anonymity practices needed; some free tools collect identifying information by default that compromises anonymity, and those should be avoided regardless of price.
Around 30-50 employees, the analysis overhead of manual approaches starts to exceed the time savings, and dedicated platforms become worth considering. The capabilities that matter most: automated reminders, segment suppression for anonymity, trend reporting across cycles, integration with existing HRIS, and templates for common pulse survey types. Most platforms in this category cost $5-15 per employee per month; for a 40-person team, that is $200-600 per month, which usually justifies itself in time savings if the program is running structured ongoing surveys.
Above 50 employees, dedicated platform support is generally worthwhile, and the conversation shifts from whether to use a platform to which platform fits the specific program needs. The features that distinguish platforms: depth of question library and templates, sophistication of analysis (theme extraction from open-ended responses, predictive modeling for retention risk), integration depth with HRIS and other tools, and quality of action-tracking workflows.
The most important decision at any team size is not which tool you use; it is whether you have built the action loop that makes any tool worth using. A team running structured action loops with free tools dramatically outperforms a team running sophisticated platforms without action loops.
Common Mistakes That Make Pulse Survey Programs Fail
The same patterns show up in almost every failing pulse survey program I have observed at small business scale. Each is preventable. Naming them is half the work; the other half is structuring the program to avoid them from the start.
The mistake that catches founders most often is running surveys without closing the loop. The instinct is rational: collecting data feels like making progress, and the action loop feels like additional work after the work is supposedly done. The math runs the other way. Survey data without action produces negative engagement effect because team members correctly read the signal that their input does not matter; the response rate collapse within 2-3 cycles is the team's accurate assessment of the program. The fix is mechanical: do not launch a pulse survey program until you have committed to the seven-day action loop, and do not run subsequent cycles if the previous cycle's actions have not been executed and communicated.
The second most damaging mistake is asking too many questions. The instinct is to capture as much data as possible while you have the team's attention; the result is fatigue that undermines the entire program. The structural rule: 5-10 questions is the right window for ongoing pulse surveys. If you have 20 questions you want to ask, run a separate annual engagement survey at lower cadence rather than bloating the pulse survey. The discipline of staying short is what makes the practice sustainable.
How Pulse Surveys Fit Into Broader Engagement Practice
Pulse surveys are one component of an engagement practice that includes several complementary elements. Treating pulse surveys as the standalone solution for engagement consistently fails; treating them as one layer in a coherent practice consistently works.
Three layers matter most at small business scale. First, structural foundations: clear roles, weekly 1-on-1s, sustainable workload, real feedback. Without these, no amount of pulse survey data will produce engagement improvement because the underlying conditions of work are eroding the foundation faster than data collection can address it. The employee feedback guide covers the day-to-day feedback skill, and the one-on-one meeting guide covers the recurring conversation cadence.
Second, recognition practice: specific, behavior-anchored, frequent positive feedback that calibrates the team to interpret leadership attention as support rather than threat. Recognition is what makes pulse survey results feel like investment in people who are valued rather than performative attention to people who are otherwise ignored. The employee recognition guide covers the daily practice that complements pulse survey programs.
Third, retention practice: deliberate attention to the factors that drive voluntary turnover at small business scale. Pulse surveys produce signal; retention practice produces action on that signal. Without retention practice, surveys generate data that is not converted into outcomes. SHRM's toolkit on managing employee performance reinforces that the integrated practice across these layers produces stronger outcomes than any single component alone. The employee retention strategies guide covers the broader retention practice context.
The Long-Term View on Pulse Surveys
The teams I have watched build durable pulse survey practice over years share three traits. First, they treat the action loop as the most important part of the program: closing the loop within seven days, picking specific actions, executing visibly, referencing previous actions in subsequent cycles. Second, they keep the practice simple: 5-10 questions, monthly or quarterly cadence, free tools or modest platform investment, focused on signal rather than sophistication. Third, they integrate pulse surveys with the broader engagement practice rather than treating them as standalone: pulse surveys complement 1-on-1s, recognition, retention work, and structural foundations rather than substituting for any of them. The compounding effect over years is significant; teams running consistent pulse surveys with strong action loops produce dramatically better engagement outcomes than teams of similar size that either skip the practice or run it poorly.
The teams I have watched struggle share a different set of traits. They run elaborate annual surveys while ignoring the continuous signal that pulse surveys provide. They launch pulse surveys without committing to the action loop. They design surveys with too many questions and watch response rates collapse. They suppress legitimate concerns by collecting demographic information that compromises anonymity. They invest in sophisticated platforms while skipping the basic discipline of communicating results to the team. None of these patterns are stupid; all of them are common; all of them are correctable, but the correction requires accepting that pulse surveys are fundamentally about the action loop rather than the data collection.
The honest message I would give my earlier self at the third-cycle-collapse stage: the pulse survey practice that compounds over years is quieter and less sophisticated than dramatic data programs. Keep it short. Run it monthly or quarterly. Close the loop within seven days. Pick 1-3 specific actions and execute visibly. Reference past actions in new cycles. Protect anonymity rigorously. Integrate with broader engagement practice. The practice is not novel; the discipline of doing it consistently is what separates teams that build genuine engagement signal from teams that produce data theater while underlying engagement erodes.
How FirstHR Fits
FirstHR covers the foundation underneath sustainable pulse survey practice at small business scale: structured onboarding workflows that establish the role clarity and expectations pulse surveys monitor, employee profiles with role context, document management for team norms and policies, integrated HRIS that gives the foundation a single home rather than scattered across tools. We are actively building feedback collection capabilities into the platform as part of expanding from onboarding-first into broader people operations support; pulse survey question libraries, scheduled feedback collection, and action tracking workflows are part of the roadmap. The platform is currently expanding into 1:1 management as part of the broader people foundation we serve. Pricing stays flat: $98/month for up to 10 employees, $198/month for up to 50, regardless of features used.
Frequently Asked Questions
What is a pulse survey?
A pulse survey is a short, frequent employee feedback survey designed to capture real-time signals about engagement, sentiment, and team experience. Pulse surveys typically contain 5-10 questions and take 3-5 minutes to complete. Unlike annual engagement surveys (40-100 questions, run once per year), pulse surveys are designed for ongoing use at monthly or quarterly cadence. The defining features are short length, focused scope, and frequent administration. The point is not to collect comprehensive data once a year but to maintain a continuous signal about how the team is doing, allowing leadership to spot problems early and address them before they compound into resignations or culture damage.
How often should you run a pulse survey?
Match the cadence to team size and stability. For most small teams of 20-100 people, monthly is the right window: frequent enough to capture meaningful signal change, infrequent enough to avoid survey fatigue. For smaller teams of 10-20, quarterly often works better because each survey can be longer and the signal-to-noise ratio is naturally lower at small sample sizes. Weekly cadence rarely justifies the fatigue cost at small business scale and should be reserved for short bursts during specific high-change periods (post-leadership change, crisis response). The biggest single cadence mistake is running surveys frequently to feel rigorous without the capacity to act on results between cycles; cadence should match your action capacity, not your data appetite.
What is the difference between a pulse survey and an annual engagement survey?
Three structural differences. First, length: pulse surveys are 5-10 questions, annual surveys are 40-100. Second, frequency: pulse surveys run monthly or quarterly, annual surveys run once per year. Third, purpose: pulse surveys maintain a continuous signal and enable rapid response to emerging issues; annual surveys produce comprehensive baseline data and inform yearly strategy. The two are complementary rather than competitive. Mature engagement programs typically run an annual deep-dive survey for breadth plus pulse surveys for depth-of-time. At small business scale, most teams should start with pulse surveys (lower setup cost, faster feedback loop) and add annual surveys only when team size justifies the additional process.
What questions should you ask in a pulse survey?
Five categories cover most pulse survey needs. Engagement (do you find your work meaningful, do you have what you need to do your job well). Manager relationship (does your manager give you useful feedback, do you trust your manager). Team experience (does your team work well together, do you feel like you belong here). Growth (do you see opportunities for development, are you learning). Wellbeing (is your workload sustainable, are you experiencing burnout signals). Pick 5-10 questions from across these categories rather than going deep on one. Use 5-point Likert scales for most questions, plus one or two open-ended questions for context. Repeat the core questions across cycles for trend analysis; rotate optional questions to keep the survey fresh.
How do you avoid survey fatigue?
Five practices reduce survey fatigue most effectively. First, keep surveys short: 5-10 questions, 3-5 minutes maximum. Second, match cadence to action capacity: do not run monthly surveys if you can only act on results quarterly. Third, close the loop quickly: send results summary and committed actions within a week of survey closing. Fourth, vary question content across cycles while keeping core questions stable for trend analysis. Fifth, demonstrate that responses produce action: when team members see specific changes resulting from feedback, they engage with future surveys at much higher rates. Survey fatigue is not really about survey volume; it is about the perception that responses do not matter. Teams that experience their feedback producing action rarely report fatigue even at relatively high cadences.
How do you get high pulse survey response rates?
Three structural factors drive response rates more than any tactic. First, demonstrated action on previous surveys: the strongest predictor of future response is whether past responses produced visible change. Second, perceived anonymity: team members must trust that responses cannot be traced to them, particularly when responses involve criticism. Third, leadership participation visible: if the founder or CEO publicly discusses results and commitments, the signal that the survey matters becomes credible. Tactical adjustments help at the margin. Send the survey from a real person rather than a no-reply address. Send brief reminders to non-respondents two days before closing. Keep surveys to 3-5 minutes. Avoid running surveys during peak deadline weeks or holiday periods. But none of these tactical adjustments compensate for the structural factor of whether the team believes responses produce action.
Should pulse surveys be anonymous?
Yes, with one caveat. Anonymity dramatically increases response honesty, particularly for questions about manager relationship, leadership decisions, or workplace concerns that team members fear could produce retaliation. The structural anonymity practices that matter: do not collect identifying information unless necessary, suppress segment results below 8-10 respondents to prevent de-anonymization through demographic intersection, store raw data with appropriate access controls, and explicitly communicate the anonymity practices in survey communications. The caveat: some specific situations benefit from non-anonymous feedback (1-on-1 contexts, individual development conversations, performance feedback). These are not pulse surveys; they are different feedback formats with different purposes. For pulse surveys specifically, anonymity is the default and should only be removed for specific defensible reasons.
What is a good pulse survey response rate?
At small business scale, 70-85% response rate is achievable for established programs running consistent action loops. New programs typically start at 50-70% and improve over the first 3-4 cycles as the team learns that responses produce action. Below 50% response rate signals significant problems with the program: either the surveys are too long, the cadence is too frequent, the team does not perceive that responses produce action, or the anonymity is not trusted. The response rate itself is a signal worth monitoring; a sudden drop in response rate often indicates that something has changed in the team's relationship to the survey program, not just to the survey content.
Do you need software to run pulse surveys at a small business?
Not strictly. Many small teams run effective pulse survey programs through free survey tools or built-in form features in their existing platforms. The investment in dedicated pulse survey software typically becomes worthwhile around 30-50 employees when the analysis overhead of manual approaches starts to exceed the cost of a platform. Below that threshold, the time savings from a dedicated platform often do not justify the additional cost and complexity. The decision is less about company size than about whether the team is running a structured ongoing program versus occasional one-time surveys: structured programs benefit from platform support; occasional surveys often run fine on free tools. The most important decision is not which tool you use; it is whether you have built the action loop that makes any tool worth using.
How do you act on pulse survey results?
Five-step action loop that works at small business scale. First, close the loop within 7 days of survey closing: send a brief summary to the entire team covering response rate, top three themes, and committed actions. Second, pick 1-3 specific actions rather than 10 vague intentions; depth of action matters more than breadth. Third, communicate what you are not addressing and why: silence on surfaced issues damages trust more than clear non-action would. Fourth, make changes visible in the work itself rather than only in policy documents. Fifth, reference the action in the next survey cycle to demonstrate continuity. The pattern that fails consistently is collecting data, holding leadership meetings about it, producing thoughtful analysis, and never communicating any of it to the team. The pattern that works treats the survey as one part of a feedback loop where the team experiences their input producing visible change.