Design

Voice Best Practices

Updated: Mar 13, 2026
Every technology comes with its own set of limitations and challenges that must be addressed in order to ensure optimal performance and usability.

Speech recognition / Text-to-Speech (TTS)

Background noise and room acoustics

Advantages

  • Training on data with background noise and varying room acoustics produces more robust speech recognition systems better equipped for real-world scenarios.
  • Incorporating environmental audio into training helps the system distinguish speech from non-speech sounds, improving accuracy.

Challenges

  • Performance degrades in noisy environments (for example, traffic, crowds, machinery). To mitigate, adjust microphone sensitivity based on the environment's noise level to optimize audio capture.
  • Echoes or reverberation reduce clarity. Provide users with visual or auditory cues when high noise is detected, prompting them to move to a quieter location or adjust their microphone.
  • Compressed audio degrades recognition performance.

Variability in speaker accents, volume, and pace

Advantages

  • Training on diverse accents, volumes, and pacing improves recognition accuracy even in noisy or varied conditions.
  • Systems can handle a wider variety of accents, reducing accent-related errors.
  • Leads to more natural speech recognition by learning realistic speech patterns.

Challenges

  • Struggles with regional accents, slang, or non-native speakers. Include diverse accents and dialects in training data, or train separate models for specific accents to capture regional nuances.
  • Difficulty distinguishing homophones (for example, "their" vs. "there"). Incorporate linguistic features such as part-of-speech tagging and dependency parsing to improve contextual disambiguation.
  • Mixing languages mid-sentence confuses monolingual models. Train on multilingual datasets and design models that explicitly handle code-switching with separate decoders for each language.
  • Age, gender, vocal clarity, or speech impairments affect performance. Use speaker embeddings to capture speaker-specific characteristics and include diverse speakers in training data.
  • Performance drops with emotional, excited, or stressed speech. Include emotionally expressive speech in training data and integrate emotion recognition modules.

Limited vocabulary or domain knowledge

Advantages

  • A focused vocabulary allows more efficient processing with fewer computational resources.
  • Better handling of specialized terminology in domains such as medicine or law.
  • Reduced system complexity makes development and maintenance easier.

Challenges

  • Fails to infer intent from conversational history or situational context. Implement contextual understanding models that analyze history and situational cues to better infer intent.
  • Misinterprets figurative language, humor, or sarcasm (for example, "That's just great" said sarcastically). Incorporate NLP techniques for detecting figurative language and tone.
  • Struggles with ambiguous phrases (for example, "I need a light" — lamp or cigarette?). Use disambiguation techniques such as follow-up questions or multiple response options.
  • Cannot distinguish homophones without context. Use deep learning models with linguistic knowledge for improved accuracy.
  • Limited ability to handle slang and colloquialisms. Expand training data with informal language from social media and online forums.
  • Difficulty handling mid-sentence language switching. Implement language identification and switching mechanisms.

Technical limitations

Advantages

  • Technical constraints encourage focus on the most important speech features, improving robustness.
  • Optimizations driven by constraints can increase processing efficiency.
  • Constraints push improvements in noise handling.

Challenges

  • Real-time recognition involves trade-offs between speed and accuracy. Implement caching, optimize algorithms, and use parallel processing to reduce latency.
  • Continuous speech is more challenging than recognizing discrete commands. Use contextual information such as conversation history, and incorporate pause detection and sentence segmentation.
  • Cannot effectively handle multiple people speaking simultaneously. Use speaker identification and separation techniques such as independent component analysis or deep learning-based methods.
  • Systems trained for general use perform poorly in specialized domains. Train domain-specific language models and incorporate domain knowledge graphs and ontologies.

Sociocultural challenges

Advantages

  • Addressing sociocultural factors improves cultural sensitivity and performance in diverse contexts.
  • Provides a more personalized and engaging user experience across cultural backgrounds.
  • Increases accessibility for users with disabilities or language barriers.
  • Builds trust and credibility with users from diverse backgrounds.

Challenges

  • Better performance for dominant languages with more training data. Incorporate underrepresented languages and use transfer learning and multilingual models to improve coverage.
  • Difficulty understanding culturally-specific expressions. Use cultural knowledge graphs and contextual information for disambiguation, and provide feedback mechanisms for users to correct misunderstandings.
  • Systems often perform better for well-represented demographic groups. Use data augmentation and adversarial training to reduce bias, and incorporate fairness metrics into evaluation pipelines.
  • Users who need speech recognition most (those with disabilities) often experience lower accuracy. Implement specialized accessibility features, provide alternative input methods, and partner with disability organizations for feedback.

Generative speech / Text-to-Speech (TTS)

Naturalness and expressiveness

Advantages

  • Creates more engaging and immersive interactions that feel like human conversations.
  • Natural-sounding output increases user trust and credibility.
  • Expressive content creates an emotional connection, leading to more memorable experiences.

Challenges

  • TTS systems often struggle to convey emotions, empathy, and personality. Implement advanced prosody modeling trained on emotionally rich datasets, incorporate emotion embeddings, and add user feedback loops for fine-tuning.
  • Limited expressive range in pitch, tone, and rhythm. Use voice cloning and style transfer techniques, augment training data with diverse speaking styles, and train on multi-speaker datasets.

Audio quality and realism

Advantages

  • Realistic audio creates an emotional connection for a more memorable experience.
  • High-quality audio enables more realistic simulations useful for training and testing.

Challenges

  • Audio artifacts like glitches, pops, or hisses detract from the experience. Apply post-processing filters and use high-quality vocoders for smoother output.
  • Generated speech may lack human nuances such as breathing sounds and mouth noises. Develop advanced acoustic models and incorporate realistic soundscapes for more natural output.

Language and accent support

Advantages

  • Multiple language support makes the system accessible to a wider range of users.
  • Supporting different accents increases cultural sensitivity.
  • Multilingual support provides a competitive advantage through more personalized services.

Challenges

  • Many TTS systems only support a limited number of languages. Leverage transfer learning and cross-lingual training, and encourage community contributions to expand language support.
  • May not accurately capture nuances of different accents and dialects. Implement accent adaptation techniques and train on dialect-specific datasets.

Contextual understanding and coherence

Advantages

  • Contextual understanding enables more accurate and relevant speech generation.
  • Coherent output provides more useful information to users.
  • Better handling of ambiguity through contextual awareness.

Challenges

  • TTS systems may misinterpret or mispronounce words without understanding context. Use contextual embeddings and semantic analysis to improve accuracy.
  • Generated speech may have awkward pauses, incorrect intonation, or unnatural structure. Develop end-to-end TTS models and implement dynamic prosody adjustment for natural flow.

Data requirements and training

Advantages

  • Well-trained models generate more coherent, contextually relevant content.
  • Well-structured data requirements streamline training, reducing time and resources.
  • Diverse, representative datasets help reduce bias in output.

Challenges

  • Training high-quality TTS models requires large, expensive datasets. Use data augmentation and synthetic data generation to supplement real-world data.
  • Training is complex, requiring expertise in machine learning, signal processing, and linguistics. Leverage transfer learning from pre-trained models and implement automated hyperparameter tuning.

Large Language Models (LLMs)

Despite their impressive capabilities, large language models (LLMs) have several limitations that affect their performance and applicability.

Data quality and bias

Advantages

  • High-quality data is essential for accurate and reliable language models.
  • Identifying and mitigating bias reduces the risk of discriminatory outcomes.
  • Reduces risk of misinformation and disinformation.

Challenges

  • LLMs can inherit biases from training data, leading to unfair outcomes. Carefully curate diverse, representative training data and implement bias detection methods.
  • May perform poorly outside their training domain. Use domain adaptation techniques and data augmentation to improve generalization.

Lack of common sense and world knowledge

Advantages

  • A broader approach to world knowledge can encourage more novel and creative responses.
  • Greater flexibility and adaptability to new situations and contexts.
  • Less prone to perpetuating biases from narrow world views.

Challenges

  • Limited understanding of the physical world; struggles with tasks requiring common sense. Train on tasks requiring common sense and incorporate knowledge graphs for structured world knowledge.
  • Difficulty with abstract reasoning. Add reasoning modules and evaluate using human evaluators to assess common sense capabilities.

Vulnerability to adversarial attacks

Advantages

  • Studying vulnerabilities leads to more effective defenses.
  • Identifying weaknesses improves overall system security.

Challenges

  • Vulnerable to adversarial examples specifically designed to mislead the model. Use adversarial training and input validation to improve robustness.
  • May not be robust to data poisoning or model stealing attacks. Apply model regularization and ensemble methods to strengthen defenses.

Explainability and transparency

Advantages

  • Provides insights into decision-making, building user trust.
  • Increases accountability for model actions and decisions.
  • Helps identify and mitigate bias, leading to fairer outcomes.

Challenges

  • Difficult to interpret why a model made a particular prediction. Use attention mechanisms, feature importance scores, and interpretability techniques such as LIME or SHAP.
  • May not provide transparent explanations for decisions. Generate transparency reports for stakeholders covering performance, limitations, and biases.

Evaluation metrics and benchmarking

Advantages

  • Identifies areas for improvement, leading to better performance.
  • Enables comparison of different models for specific tasks.
  • Quickly highlights model strengths and weaknesses, increasing development efficiency.

Challenges

  • Often evaluated using limited metrics (for example, perplexity and accuracy) that miss the full range of capabilities. Use multi-metric evaluation and include human evaluators.
  • Difficult to benchmark across different architectures. Develop comprehensive benchmarking suites with varied tasks and datasets, and continuously monitor performance.

Ethical considerations

Advantages

  • Ensures respect for human rights such as privacy and freedom of expression.
  • Promotes fairness and non-discrimination in model design.
  • Protects personal data from harmful or exploitative use.

Challenges

  • Can perpetuate biases and unfairness if not designed carefully. Develop fairness metrics, implement bias detection, and generate transparency reports.
  • Can displace human workers in automatable industries. Develop mitigation strategies such as retraining programs and skills-development initiatives, and implement human oversight mechanisms.

Natural Language Processing (NLP)

Ambiguity and uncertainty

Advantages

  • Handling ambiguity improves model robustness with uncertain input.
  • Acknowledging language ambiguity leads to more realistic modeling.
  • Improves generalization to new and unseen data.

Challenges

  • Words can have multiple meanings, making disambiguation challenging. Use contextual information, part-of-speech tagging, and named entity recognition to reduce ambiguity.
  • Struggles with sarcasm, idioms, and figurative language. Leverage conversation history and external knowledge for better contextual understanding.

Limited domain knowledge

Advantages

  • Focused domain training yields higher accuracy for recognized intents and entities.
  • Reduces ambiguity in user requests.

Challenges

  • May not understand domain-specific terminology. Use domain-specific training data and knowledge graphs to represent domain knowledge.
  • Lacks common sense and real-world experience compared to humans. Use transfer learning to leverage pre-trained models and fine-tune on domain-specific data.

Noise and errors

Advantages

  • Accounting for noise improves accuracy and reduces misinterpretation.
  • Noise-tolerant systems are more robust to input variability.

Challenges

  • Struggles with typos, grammatical errors, and other text noise. Use preprocessing techniques such as tokenization and stemming, plus spell and grammar checking.
  • May not recognize out-of-vocabulary words. Use robust models (for example, RNNs with dropout) that can handle noisy and erroneous data.

Cultural and linguistic barriers

Advantages

  • Considering cultural differences makes systems accessible to broader audiences.
  • Accounting for linguistic nuances reduces errors and misinterpretation.
  • Enables expansion into new markets.

Challenges

  • May not handle language and cultural differences well. Provide multilingual support using language-specific or multilingual models.
  • Struggles with idioms and colloquialisms that vary across cultures. Adapt models to local customs and norms, and use human evaluation in different cultural contexts.

For speech recognition–specific sociocultural challenges, see the Speech Recognition section above.

Scalability and efficiency

Advantages

  • Scalable systems can process and analyze large volumes of text data.
  • Efficient processing improves accuracy and reduces errors.
  • Reduces costs by minimizing additional hardware, software, and personnel.

Challenges

  • NLP tasks can require significant computational resources. Use distributed computing to process data in parallel.
  • Requires large amounts of training data that can be difficult to obtain. Use model pruning and knowledge distillation to reduce model size and computational requirements without sacrificing accuracy.

For explainability, adversarial attacks, and ethical considerations as they apply to NLP, see the Large Language Models (LLMs) section above.

Dos and don’ts

The following checklist complements the Design Principles on the Voice page. Use it as a quick reference when reviewing your voice experience:
Do

  • Ensure Transparent Activation. Provide clear mic activation awareness through earcons or conversational cues to ensure users are aware of when their voice is being captured.
  • Provide Clear Privacy Controls. Offer users the option to opt-in to voice experiences where their voice will be captured, respecting their privacy.
  • Focus on the happy path with robust Error Handling and Repair paths.
  • Understand the common user intents and utterances, and the blocks that commonly occur when users attempt to use the system.
  • Establish consistent, systemic repair attempts that keep the context of the experience in mind. Conversation Design standards expect two attempts at repair and then a handoff to a live person or giving the user another solution.
  • Make sure to triage the error - is it a soft error or a hard error? Soft errors are issues in the design logic or data, while hard errors are backend service failures.
  • Provide users with visual or auditory cues when the system detects high levels of background noise, prompting them to move to a quieter location or adjust their microphone.
  • Focus on the most important aspect of speech within the technical limitations.
  • Provide accurate details such as names, dates, locations, and quantities for effective speech recognition interactions.
  • Offer explanations and alternatives to mitigate errors and maintain user satisfaction.
  • Use Part of Speech (POS) tagging to enable a more accurate interpretation.
  • Have the system clearly announce the available menu options and allow users to select a specific option using voice commands.
  • Provide shortcut options for frequently used actions.
  • Provide alternative input methods during multimodal interactions.
  • Ensure that the system is accessible and usable for all users, including those with disabilities.
  • Consider the visual components and audio-only interactions.
  • Use the POUR guiding principle of accessible design: Perceivable, Operable, Understandable, and Robust.
  • Use robust New User Experience to educate users.
  • Prioritize user understanding and resolution in the design of the error handling system.

Don't

  • Activate the microphone without user consent or by default. Default microphone activation can lead to unintended audio capture, compromising user privacy.
  • Underestimate the special needs of Voice experiences concerning Cognitive Load. Two to three choices are ideal, and be sure to design so that the user has a mental model with a clear path to understand where they are in the design.
  • Be ambiguous when designing using NLP models as the system may struggle if the information is not clear.
  • Overwhelm users with information but instead use breadcrumb techniques for clear and consistent navigation.
  • Limit the control the user has over their biometric data and how it's used.
  • Obscure or limit information on how the user's audio data is being used and protected.

Next steps

More design resources on voice

Designing experiences

Explore more design guidelines and learn how to design great experiences for your app users:
  • Input Modalities: Discover all the various input modalities.
  • Hands: Examine hands-based input methods.
  • Head: Examine head-based input methods.
  • Controllers: Examine controller-based input methods.
  • Peripherals: Learn how to design experiences that leverage peripherals.