Check in and pick up your name badge and swag for the Sawtooth Research Conference.
Kick off SRC with our opening reception at Muriel’s Jackson Square. Enjoy a full dinner featuring contemporary Creole cuisine.
We outline a practical framework for improving online survey data quality by integrating behavioral, device, and historical feedback signals. Traditional checks like RLH risk discarding genuine respondents or retaining fraudulent ones, undermining insights. The research demonstrates how combining in-survey behavioral indicators, device fraud detection, and cross-platform feedback loops creates a dynamic, adaptive quality management system. By leveraging Data Quality Co-op�s API within Lighthouse Studio, the study highlights holistic, iterative methods for identifying fraud and preserving authentic consumer voices.
We outline a practical framework for improving online survey data quality by integrating behavioral, device, and historical feedback signals. Traditional checks like RLH risk discarding genuine respondents or retaining fraudulent ones, undermining insights. The research demonstrates how combining in-survey behavioral indicators, device fraud detection, and cross-platform feedback loops creates a dynamic, adaptive quality management system. By leveraging Data Quality Co-op�s API within Lighthouse Studio, the study highlights holistic, iterative methods for identifying fraud and preserving authentic consumer voices.
Synthetic data promises faster, cheaper, and more flexible research � but it comes with risks. Using frameworks such as Truth, Transparency, and Trust, and blending Human and Artificial Intelligence, safe outcomes can be achieved. This paper explores the do�s and don�ts of synthetic augmentation, then presents an empirical study with a Digital Twin Panel, where AI-generated �clones� of real respondents are tested in choice studies, examining whether digital twins can represent the next generation of research automation.
Synthetic data promises faster, cheaper, and more flexible research � but it comes with risks. Using frameworks such as Truth, Transparency, and Trust, and blending Human and Artificial Intelligence, safe outcomes can be achieved. This paper explores the do�s and don�ts of synthetic augmentation, then presents an empirical study with a Digital Twin Panel, where AI-generated �clones� of real respondents are tested in choice studies, examining whether digital twins can represent the next generation of research automation.
Synthetic data is a polarizing topic, simultaneously touted as a solution for hard-to-reach populations and small-sample studies, and condemned as confidently misrepresenting the voice of the customer. In this session, we present a validation framework along with findings from our ongoing evaluation of synthetic data across common tasks including cross tabs, key driver analysis, and conjoint. Attendees will leave with practical guidance for evaluating synthetic data outcomes along with validated use cases.
Synthetic data is a polarizing topic, simultaneously touted as a solution for hard-to-reach populations and small-sample studies, and condemned as confidently misrepresenting the voice of the customer. In this session, we present a validation framework along with findings from our ongoing evaluation of synthetic data across common tasks including cross tabs, key driver analysis, and conjoint. Attendees will leave with practical guidance for evaluating synthetic data outcomes along with validated use cases.
In Venezuela�s uniquely complex market and economic environment, Empresas Polar adapted choice-based studies to generate fast, cost-effective insights under extreme volatility. By developing in-house capabilities, the team refined segmentation, tracked price sensitivity, forecasted sales and informed pack-size, pricing, and innovation strategies. Attendees will learn how abandoning traditional paradigms and tailoring conjoint designs to local constraints yielded actionable results. The case provides replicable lessons for organizations navigating unstable markets, offering practical guidance for agile, resilient research.
In Venezuela�s uniquely complex market and economic environment, Empresas Polar adapted choice-based studies to generate fast, cost-effective insights under extreme volatility. By developing in-house capabilities, the team refined segmentation, tracked price sensitivity, forecasted sales and informed pack-size, pricing, and innovation strategies. Attendees will learn how abandoning traditional paradigms and tailoring conjoint designs to local constraints yielded actionable results. The case provides replicable lessons for organizations navigating unstable markets, offering practical guidance for agile, resilient research.
Within the last few years Bank Millennium has carried out a series of research projects each with an objective to refresh one of its products to boost the acquisition of new clients in key segments. We will tell the story of both turning business questions into research tools and turning consumer insights into a new successful bank product.
Within the last few years Bank Millennium has carried out a series of research projects each with an objective to refresh one of its products to boost the acquisition of new clients in key segments. We will tell the story of both turning business questions into research tools and turning consumer insights into a new successful bank product.
While MaxDiff excels at selecting the best product claims, it falls short for evaluating longer technical narratives. We hypothesized that limited consumer awareness of our composite lumber�s advantages over industry standards was hindering adoption. Using a vignette-based survey design, this study tested behavioral shifts after exposure to a technical product story. Results demonstrated that incorporating educational content into our marketing strategy boosts conversion 80%. Vignettes are a promising addition to incorporate into mixed-method surveys.
While MaxDiff excels at selecting the best product claims, it falls short for evaluating longer technical narratives. We hypothesized that limited consumer awareness of our composite lumber�s advantages over industry standards was hindering adoption. Using a vignette-based survey design, this study tested behavioral shifts after exposure to a technical product story. Results demonstrated that incorporating educational content into our marketing strategy boosts conversion 80%. Vignettes are a promising addition to incorporate into mixed-method surveys.
For B2B SaaS companies, growth requires moving beyond intuition-based pricing. This paper presents an integrated case study with Docusign, showcasing a multi-stage conjoint research program that tackled distinct strategic challenges: launching a new product line, managing cannibalization with license types, creating tiered service offerings,�and channel expansion. Attendees will gain a practical framework for using advanced choice modeling to drive customer-centric monetization and translate complex research into measurable revenue growth and clear go-to-market guidance.
For B2B SaaS companies, growth requires moving beyond intuition-based pricing. This paper presents an integrated case study with Docusign, showcasing a multi-stage conjoint research program that tackled distinct strategic challenges: launching a new product line, managing cannibalization with license types, creating tiered service offerings,�and channel expansion. Attendees will gain a practical framework for using advanced choice modeling to drive customer-centric monetization and translate complex research into measurable revenue growth and clear go-to-market guidance.
Can large language models replace manual coding of open-end survey responses without sacrificing quality? We outline a four-phase approach: (1) testing AI-generated codeframes against human baselines; (2) evaluating assignment accuracy via interrater reliability and agreement subsets; (3) weighing cost-benefits of additional training data; and (4) sharing lessons from operationalizing at KS&R. Attendees will leave with methods for quantitatively assessing AI�s utility in open-end coding and practical tools to begin experimenting within their own organizations.
Can large language models replace manual coding of open-end survey responses without sacrificing quality? We outline a four-phase approach: (1) testing AI-generated codeframes against human baselines; (2) evaluating assignment accuracy via interrater reliability and agreement subsets; (3) weighing cost-benefits of additional training data; and (4) sharing lessons from operationalizing at KS&R. Attendees will leave with methods for quantitatively assessing AI�s utility in open-end coding and practical tools to begin experimenting within their own organizations.
Shapley Values (SVs) are widely used in summarizing item contributions and importance, but become computationally impossible for, say, 30-plus items. We will briefly introduce SVs and a trick for computation in variants of TURF, but focus on a new class of experimental designs that dramatically improve the accuracy of SV sampling approaches and make even 200 items feasible, for all problem types. We will also compare SVs to Johnson�s Relative Weighting for key driver regressions.
Shapley Values (SVs) are widely used in summarizing item contributions and importance, but become computationally impossible for, say, 30-plus items. We will briefly introduce SVs and a trick for computation in variants of TURF, but focus on a new class of experimental designs that dramatically improve the accuracy of SV sampling approaches and make even 200 items feasible, for all problem types. We will also compare SVs to Johnson�s Relative Weighting for key driver regressions.
To reduce survey length, researchers have respondents answer only a subset of questions. The missing data is then imputed using advanced data imputation techniques. We evaluate the effectiveness of LLMs� imputation capabilities against traditional methods, using real human responses as the benchmark. Our findings offer valuable insights into how and under what conditions LLMs can enhance data quality and efficiency in survey analysis.
To reduce survey length, researchers have respondents answer only a subset of questions. The missing data is then imputed using advanced data imputation techniques. We evaluate the effectiveness of LLMs� imputation capabilities against traditional methods, using real human responses as the benchmark. Our findings offer valuable insights into how and under what conditions LLMs can enhance data quality and efficiency in survey analysis.
Come see how we employed AI technology within Sawtooth�s Discover survey platform to measure the effectiveness of AI-driven hyper-personalization of marketing messages. We created 8 diverse, high-quality marketing messages following best practices in copywriting and communications. We also used AI to generate hyper-personalized messages for respondents, based on responses to multiple closed-end and open-ended questions in the survey. We�ll report on what we learned and how well the AI-generated messages performed on appeal and purchase intent compared to the pre-generated marketing messages.
Come see how we employed AI technology within Sawtooth�s Discover survey platform to measure the effectiveness of AI-driven hyper-personalization of marketing messages. We created 8 diverse, high-quality marketing messages following best practices in copywriting and communications. We also used AI to generate hyper-personalized messages for respondents, based on responses to multiple closed-end and open-ended questions in the survey. We�ll report on what we learned and how well the AI-generated messages performed on appeal and purchase intent compared to the pre-generated marketing messages.
Mingle with fellow conference attendees over hors d'oeuvres and drinks at the Loews New Orleans Hotel.
Conjoint analysis often begins with static feature glossaries, but are respondents really reading them? This research-on-research study compares glossary-based warm-ups with a Kano-inspired warm-up, with and without the Kurz-Binner Priming questions. We test impacts on comprehension, ANA, engagement, LOI, and utilities. Attendees will walk away with evidence-based, practical guidance for designing conjoint warm-ups that improve both the respondent experience and the quality of the data.
Conjoint analysis often begins with static feature glossaries, but are respondents really reading them? This research-on-research study compares glossary-based warm-ups with a Kano-inspired warm-up, with and without the Kurz-Binner Priming questions. We test impacts on comprehension, ANA, engagement, LOI, and utilities. Attendees will walk away with evidence-based, practical guidance for designing conjoint warm-ups that improve both the respondent experience and the quality of the data.
This study examines whether differentiating outside good choices in discrete choice models, into channel switching, category substitution, and category exit, yield richer insights than treating it as a single no-choice option. Using instant coffee as a case study, we evaluate improvements in model fit, predictive validity, and stability of willingness-to-pay across consumer subgroups. By clarifying how consumers opt out, the approach enhances market realism and equips researchers with strategies to navigate demand shifts in competitive markets.
This study examines whether differentiating outside good choices in discrete choice models, into channel switching, category substitution, and category exit, yield richer insights than treating it as a single no-choice option. Using instant coffee as a case study, we evaluate improvements in model fit, predictive validity, and stability of willingness-to-pay across consumer subgroups. By clarifying how consumers opt out, the approach enhances market realism and equips researchers with strategies to navigate demand shifts in competitive markets.
Conjoint analysis often includes a no-choice option allowing respondents to opt out and not make a purchase. We evaluate a dual-response format in which respondents first select their preferred option excluding the no-choice alternative, then decide whether they would actually purchase it. Results show that respondents apply budget constraints only in the second response, separating affordability from preference. The dual-response model improves parameter estimation efficiency by 30% and corrects overestimated equilibrium prices in competitive market simulations.
Conjoint analysis often includes a no-choice option allowing respondents to opt out and not make a purchase. We evaluate a dual-response format in which respondents first select their preferred option excluding the no-choice alternative, then decide whether they would actually purchase it. Results show that respondents apply budget constraints only in the second response, separating affordability from preference. The dual-response model improves parameter estimation efficiency by 30% and corrects overestimated equilibrium prices in competitive market simulations.
Clients often request feature-level willingness to pay (WTP) insights, but additive methods face scaling and feasibility issues. Sawtooth�s Sampling of Scenarios (SOS) improves realism through scenario averaging yet still overstates values. SKIM is testing alternative extrapolation functions (quadratic, exponential, and piecewise) to better capture price responses, with preliminary results showing a 16% WTP reduction using quadratic extrapolation versus linear. Future work also includes broader scaling approaches comparing �all-features-on� versus �all-features-off� simulations.
Clients often request feature-level willingness to pay (WTP) insights, but additive methods face scaling and feasibility issues. Sawtooth�s Sampling of Scenarios (SOS) improves realism through scenario averaging yet still overstates values. SKIM is testing alternative extrapolation functions (quadratic, exponential, and piecewise) to better capture price responses, with preliminary results showing a 16% WTP reduction using quadratic extrapolation versus linear. Future work also includes broader scaling approaches comparing �all-features-on� versus �all-features-off� simulations.
In Filter-CBC, respondents often fail to apply filters sufficiently, leaving them overloaded with irrelevant alternatives and increasing position bias. This study introduces pre-filtering modules before the conjoint exercise to reduce these effects. Two approaches are evaluated: an adaptive CBC pre-module and a simplified attribute-ranking with must-have probing. Both identify which attributes and levels to prefilter, thereby reducing task complexity and cognitive load, and ultimately leading to more reliable and predictive utility estimates.
In Filter-CBC, respondents often fail to apply filters sufficiently, leaving them overloaded with irrelevant alternatives and increasing position bias. This study introduces pre-filtering modules before the conjoint exercise to reduce these effects. Two approaches are evaluated: an adaptive CBC pre-module and a simplified attribute-ranking with must-have probing. Both identify which attributes and levels to prefilter, thereby reducing task complexity and cognitive load, and ultimately leading to more reliable and predictive utility estimates.
At the 2021 Turbo Choice Modeling Event, Peitz and Lerner introduced Bespoke CBC, where respondents choose which attributes matter to them, and only those appear in their CBC tasks. This personalization improved engagement, reduced dropout, and increased predictive validity�but was hard to program. We�ve developed a simple coding method that removes this barrier. Our next steps include empirically testing Bespoke CBC against other CBC types, examining attribute non-attendance, and assessing robustness when respondents misidentify key attributes.
At the 2021 Turbo Choice Modeling Event, Peitz and Lerner introduced Bespoke CBC, where respondents choose which attributes matter to them, and only those appear in their CBC tasks. This personalization improved engagement, reduced dropout, and increased predictive validity�but was hard to program. We�ve developed a simple coding method that removes this barrier. Our next steps include empirically testing Bespoke CBC against other CBC types, examining attribute non-attendance, and assessing robustness when respondents misidentify key attributes.
Markets like video streaming services in which buyers purchase more than one good at the same time pose substantial problems to Conjoint practitioners. We present and discuss ways to properly model these choices. Based on real projects we show advantages and disadvantages of different approaches and guide users on how to make the most of the choice data that they collected.
Markets like video streaming services in which buyers purchase more than one good at the same time pose substantial problems to Conjoint practitioners. We present and discuss ways to properly model these choices. Based on real projects we show advantages and disadvantages of different approaches and guide users on how to make the most of the choice data that they collected.
Traditional product development relies on expert-led ideation and evaluation, which is time-intensive and may overlook valuable concepts. We present an AI-augmented approach engaging consumers directly in co-creation using an AI-powered conversational survey agent. Participants generate benefit-feature ideas with chatbot assistance, followed by rolling MaxDiff evaluation. This method produces human-created ideas with quantitative scores, AI-enabled analysis, and rich qualitative insights, demonstrating AI's potential for augmenting human creativity in product development.
Traditional product development relies on expert-led ideation and evaluation, which is time-intensive and may overlook valuable concepts. We present an AI-augmented approach engaging consumers directly in co-creation using an AI-powered conversational survey agent. Participants generate benefit-feature ideas with chatbot assistance, followed by rolling MaxDiff evaluation. This method produces human-created ideas with quantitative scores, AI-enabled analysis, and rich qualitative insights, demonstrating AI's potential for augmenting human creativity in product development.
Generative AI offers a new way to improve conjoint analysis: by supplying utility priors for design and estimation. In this session, we show how large language models (LLMs) can generate realistic prior utilities that enhance predictive accuracy in CBC studies. Using a job search platform experiment with ~1,000 participants, we compare standard designs with GenAI-augmented ones and share practical lessons and pitfalls.
Generative AI offers a new way to improve conjoint analysis: by supplying utility priors for design and estimation. In this session, we show how large language models (LLMs) can generate realistic prior utilities that enhance predictive accuracy in CBC studies. Using a job search platform experiment with ~1,000 participants, we compare standard designs with GenAI-augmented ones and share practical lessons and pitfalls.
Companies possess a significant amount of relevant data, but due to a lack of resources and time, they often experience difficulties in utilising it. We aim to demonstrate our LLM-based approach solution, in which LLM acts as a�smart interface�between human intent and complex and unorganized data enabling�automated, scalable, and user-friendly�processing of complex product characteristics. The presentation will show how we transformed a technical bottleneck into a flexible, collaborative workflow.
Companies possess a significant amount of relevant data, but due to a lack of resources and time, they often experience difficulties in utilising it. We aim to demonstrate our LLM-based approach solution, in which LLM acts as a�smart interface�between human intent and complex and unorganized data enabling�automated, scalable, and user-friendly�processing of complex product characteristics. The presentation will show how we transformed a technical bottleneck into a flexible, collaborative workflow.
Exercise-related biases in stated choice experiments, such as position effects and screen fatigue, can reduce model accuracy by introducing skewed responses. We evaluate two mitigation strategies: adding contextual information to utility models and adjusting pooling based on respondent behavior. Meta-analytic findings will assess their effectiveness across studies. Attendees will gain practical tools for identifying quasi-straightliners, cleaning noisy data, and applying recommendations to model, simulate, and improve the overall quality of choice experiment results.
Driver-CBC is an innovative approach that integrates brand association analysis with choice-based conjoint models to inform market decisions. This method captures the impact of brand perceptions on consumer choices, enabling simulations of various brand perceptions, price and product feature scenarios and their consequent economic outcomes. By combining tactical and strategic elements, Driver-CBC provides a unified framework for optimising marketing strategies and brand development, offering actionable insights to maximise market share and revenue growth.
Automated Machine Learning (AutoML) identifies and then executes the best algorithm fitting to each specific data set automatically with little or no human intervention. Expanding AutoML to clustering is gaining attention but facing additional hurdles due to the lack of ground-truth for model training. Making it work for market research is even more challenging because of the art part. But the value can be big. This research explores an automated system for market segmentation, shares learnings and novel solutions for future expansion.
In this talk I will describe several techniques for aligning questions from multiple surveys so that they can be integrated and used in models to predict segment membership.� I will demonstrate the techniques and describe the necessary assumptions behind them. No AI is required.