Nov
“Humanizing” Market Research with AI
jerry97890 comments artificial intelligence, Brand Surveys and Testing, Burning Questions
The Boon and Bane of AI
The increasing and widespread utilization and demand for Artificial Intelligence have been met with both excitement and reservation. Excitement for the possibilities AI’s implementation unlocks, oftentimes steps ahead of the curve or beyond expectations; reservations not only stemming from the risks over its unethical and unchecked use, but also the ramifications for human involvement now that intelligent machines represent an optimized and economical choice for completing tasks and processes. But can there be a middle ground somewhere where AI and human engagement coexist and collaborate?
The “Humanization” of Market Research
“Capturing the Human Element in an Artificial World” by Eric Tayce (Quirk’s Marketing Research Review, Sep-Oct 2024) posits that such a midground is possible, especially in market research. An industry that’s all aware of its excessive dependence on technology to necessitate a push to “humanize” research data, it saw a dramatic shift from “data-intense tomes and clinical-sounding slide titles” to “streamlined, narrative-style reporting” focused on “the unique motivations and experiences that drive customer behaviors.” The latter “humanized” approach is able to communicate business goals while connecting and engaging on an emotional level. However, generative large language models (LLMs) cast a shadow on this “humanized” approach by offering synthetic outputs and progressive algorithms.
But combining both AI and efforts to “humanize” research can result in the whole being greater than the sum of its parts. The article shared that AI can help collect more unstructured data from survey research by employing conversational chatbots to create a natural, richer experience for the respondent. That unstructured data in turn can potentially provide more organic, more human insights with AI-powered algorithms, an undertaking that was once considered too complex or time-consuming. AI can also build multifaceted perspectives through context by linking survey records with a broad range of data sources. And in lieu of traditional static deliverables, data and insights can be presented in a vibrant and interactive narrative by an AI-powered persona.
The “Human” Element
All these interesting prospects can only be achieved when AI is tempered by high-quality human input and thoughtful implementation considerate of ethical and moral implications. Aside from AI mistakes and hallucinations existing, AI has been observed to be too helpful and excitable. Human oversight and input remain key in ensuring AI models are trained, fine-tuned and grounded with quality and relevant datasets while having enough flexibility to engage appropriately in open-ended interactions.
There’s no denying just how transformative AI is in reshaping industries today, including market research. Despite concerns of machines taking over jobs, one can look at it with the perspective of roles changing and adapting. AI with its generative and synthetic capabilities can elevate the “humanization” of market research, but to get to that point we simply can’t forget that humans are indispensable to the whole process.
Featured Image Copyright: GrumpyBeere
Top Image Copyright: Darlene Anderson
Aug
Can Synthetic Respondents Take Over Surveys?
jerry97890 comments artificial intelligence, Burning Questions
What Are Synthetic Respondents?
AI has increased operational efficiency by streamlining knowledge bases and shortcutting processes so it’s no surprise people and companies are looking for more ways for its application. For market research, one curious consideration is whether it could take over surveys, essentially by replacing actual respondents with synthetic respondents.
Also known as virtual respondents, digital personas, and Virtual Audiences, synthetic respondents are individual profiles constructed by Large Language Models (LLMs) from real or simulated data. Ideally, the data or descriptions used to generate these profiles come from previously conducted surveys and are combined with individual-level demographics, attitudes and behaviors.
Using these synthetic respondents over real respondents could benefit your research with speed, accuracy and cost savings, at least according to their advocates. Basically, you just need to conduct one survey and from the profile description or data you gathered from the actual respondents, you’re able to generate results from the constructed individuals over and over for succeeding studies and research.
Testing Synthetic Respondents
While synthetic respondents could accurately represent real respondents, relying exclusively on the results from these AI-based individuals may not be entirely beneficial. A webinar hosted by Radius Global took a closer look at the potential of AI-generated synthetic respondents through three case studies of quantitative concept testing, quantitative communications research, and qualitative communications research.
Aggregate results for the concept tests involving game controllers indicate somewhat strong similarities between the results of the real and synthetic respondents. This extends to the results from the quantitative communications research when it comes to the believability of statements on the benefits of milk, although there were some differences. The differences were much more pronounced though when it comes to surprise over the same statements, and there was incongruence when considering how each statement could possibly increase milk consumption.
The qualitative communications research was seeking in-depth insights into women’s needs, perceptions, and preferences for running a race or marathon, with the feedback gathered meant to be used for creative content. Personas were constructed from the profiles of six women aged between 18 and 64 years old who ran at least once in an average week. They had an LLM assume each persona to allow a comparison between findings from real participants to synthetic respondents.
They found that while both real and synthetic respondents have somewhat similar responses when it comes to functional aspects as goals for women in general pursuing fitness, the AI responses lacked emotional expressions. There are also little differences in the synthetic respondents’s responses despite having different profiles, and there was even a lack of subtle differences.
As for concerns among women who are aspirational marathon runners, the synthetic personas were consistent in their responses while the real respondents provided more nuances, variety, and perspectives more prevalent among women.
Synthetic Respondents Vs. Real Respondents
Synthetic respondents appear to be useful if you’re evaluating existing ideas and concepts; however, if you’re looking for “breakthroughs” or essentially new insights you would’ve never arrived at had you not performed the case study or research, you would need to engage with real respondents, relying exclusively on their results or combining them with that of synthetic respondents. Yes, there could be cases where synthetic respondents could be used, but the results must be extensively validated. It would also require increasing the efficiency of how data used to construct these individuals are analyzed in addition to enhancing the quality of the data and information gathered for these profiles through thorough screening, intelligent probing, and smart choice models.
There is a place for synthetic respondents in market research, but as another tool in a researcher’s disposal. They won’t be taking over surveys nor replacing actual respondents wholesale anytime soon it seems, as that elusive “Eureka” moment researchers seek are inherently tied with the nuances and perspectives of human emotion and experience you just simply couldn’t construct.
Photo courtesy of Pavel Danilyuk
Jul
AI In Market Research: The Story So Far – Chapter 3: A Glimpse Into A Future with AI
jerry97890 comments artificial intelligence, Burning Questions
This is the third installment in our series on AI webinars. The inspiration for this series is a simple question about what these AI seminars are saying. There are hundreds of these seminars floating about, all based on the premise that AI technology is here to stay, people are curious about it, and they want to know how it will affect their lives.
We asked one of our staff members to attend several of these AI webinars in pursuit of the answer to this question: what are the AI webinars really saying? What are the common themes, if any? While the first and second chapters focused on AI’s ubiquity and limitations respectively, this third installment focuses on the replacement of humans and human work.
Will Jobs Be Lost Because of AI?
AI is a threat to most jobs including those in the market research industry but this is most especially true with any repetitive or routine work grounded by established knowledge or processes. AI provides the advantages of streamlining your knowledge base and shortcutting processes. If you’re on the process side and you fail to embrace AI, clients might find you costlier and less optimized.
The market research industry had already learned this lesson in the early 2000s when the big companies didn’t take online research seriously. They subsequently found themselves trying to catch up some years later after the widespread acceptance and adoption of online research. Whoever waits too long or neglects to embrace the newest tool would most likely fall behind as the industry shifts towards AI-driven processes.
What are the AI webinars saying about Human Replacement?
This doesn’t necessarily mean that humans would be fully replaced and displaced by AI in market research. AI is, after all, a tool. All tools revolutionize optimization but optimization by definition doesn’t make things better. AI will revolutionize things, but it is not the big revolution that will make everything different.
To illustrate that last point, a question was raised in one of the webinars about the possibility of a data collection tool that can replace surveys. There wasn’t a definitive answer given by the panelists since it’s more of a question of what surveys would actually look like. It would depend on what is wanted to be accomplished, the type of information sought, and how they would engage and elicit reactions.
The subject of AI-powered market research alone attracts investors. Embracing AI would not only optimize your market research processes but it would also add value to your insights. Having said that, AI places everybody on the same playing field, except those who recognize and seize the opportunity to experiment with AI are able to gain an edge. The key would be to build on experience rather than purely on the thirst for innovation; try to be at the front of things, but don’t try to be the first one. Try to find a good balance by going with the flow while making smart moves and decisions.
It’s been noted that the profile of the researcher of the future is a little bit more techy and into IT integration. New business intelligence leaders today have IT backgrounds, and this is different from two decades ago. Even in a world with AI-based market research, there would be room for the human factor that adds value from experience — something AI won’t be able to replace.
Jun
AI In Market Research: The Story So Far – Chapter 2: Limitations of AI
jerry97890 comments artificial intelligence, Burning Questions
Despite AI’s expanding popularity in market research, experts are fully aware that there is still a lot of ground to cover regarding their effectiveness and optimization for use cases, along with understanding and mitigating their risks and limitations.
These limitations reveal themselves most especially in efforts to replicate human behavior. One research paper on a survey employing Large Language Models observed how effective these LLMs are in understanding consumer preferences with their behaviors consistent with four economic theories, but noted that there were demonstrations of extra sensitivity to the prompts they were given. In addition, there were indications of positional bias wherein the first concept was selected more often than the others that were also presented.
AI has also been found to be too optimistic, tech-forward, and self-interested. For example, ChatGPT is inherently focusing on maximizing expected payoffs, whereas a person would often act in a risk-averse way for gains and risk-seeking for losses. AI also exhibits a generally higher level of brand association than humans, resulting in higher brand scores. However, it struggles with lesser-known topics, notably in scenarios where new commercial products are tested and targeted toward a specific audience.
While it can be addressed by cautious prompt engineering, AI hallucinations are an unintended effect of the helpful aspect of these models where they generate unnecessary output stemming from patterns or elements they perceived but are nonexistent or imperceptible to human observers.
And while more on the side of risks than limitations, there is an understandably and famously increasing concern from artists over how text-to-image generators threaten to replace them and their work, just as there are certain roles in the market research sector that are in danger of being taken over by AI.
Perhaps the ideal recommendation for utilizing AI while keeping in mind its limitations is to use it in cases where it’s most effective and productive with the understanding that it might excel in one scenario, but it doesn’t mean it will be just as effective in another situation.
May
AI In Market Research: The Story So Far – Chapter 1: Adapt or Get Left Behind
jerry97890 comments artificial intelligence, Burning Questions
Whether you like it or not, AI is here to stay. Yes, AI is a threat to most jobs, including those in the market research industry, since it shortcuts processes while optimizing operational efficiency. While market research technology didn’t develop as fast as other industries in the early to mid-2000s, the advent and subsequent mainstream appeal of AI has forced market research to get with the times. You’re in trouble if you fail to embrace it but if you do, you get to be on the winning side.
Experts expressed that we’re still in the early exploratory stages of AI but there is already depth in its application in market research. Take, for example, the humanization of surveys. An interactive and dynamically probing AI improved overall data quality in more than one experiment due to an increased engagement from respondents resulting from a sense of appreciation over the perceived but simulated attention paid to them and their responses during the survey. In the same vein, employing a conversational AI voice has been shown to dramatically drive engagement for better data.
That latter effort to humanize surveys has created an influx of voice responses and content, leading to the new question of what we should now do with all those resources, which would be a byproduct of AI-based solutions. Of course, LLMs and other existing AI models would be employed to help find the answer to this question.
Aside from solving dark data, AI has also displayed impressive capabilities to answer choice tasks, especially performing well with well-known topics and products, even outdoing humans in some surveys where humans get confused or find it hard to render a judgment. It’s also been considered for AI to adapt existing survey data for a new topic to save time. AI’s role in market research might still be experimental at this point, but it has grown to the point where it’s being utilized and adapted to take on one challenge after another.
Our second entry in this four-part blog series highlights some of AI’s risks and limitations, and how understanding and mitigating these factors can lead to their effective and optimal utilization.
May
AI Webinars Are Everywhere – What Are They Really Saying?
jerry97890 comments artificial intelligence, Burning Questions
With AI becoming more ubiquitous each passing year, it’s no surprise that webinars dedicated to the subject have been springing up everywhere. Amid the hype, people are either curious, interested, or to some degree invested in what AI’s increasing popularity means for them as well as the industries they’re part of. These webinars serve as the perfect platform for industry experts to share their experiences, thoughts, and opinions on AI’s current and future implications.
What are these AI webinars really saying? We sent one of our staff members in search of the answers. Each webinar talks about the impact of AI on the economy, society, and culture, but they must share some common themes or overarching ideas. What are these common ideas? To get at the answers, we asked our staff member (Emil Deverala) to focus on the impacts on an industry we truly understand: market research.
During April and May 2024, Emil attended three AI webinars: “Market Research in an AI World,” “AI in Marketing Research: Expert Panel Discussion,” and “Building New Business: Five Ways Firms Are Driving New Revenue with Automation And AI.” After each webinar, Emil was asked to not only summarize the items that were discussed, but also share his larger thoughts about what the webinar was really trying to say to the world about what we can expect from AI.
Emil eventually boiled things down to four main ideas or themes. In this series of blogs, we’ll be exploring each of those themes. Here’s the first in our series. It focuses on why people in the market research industry need to pay attention to AI in the first place.
Sep
What It Means to Choose or Decide In The Age of AI
jerry97890 comments artificial intelligence, Burning Questions
Longstanding Concerns Over AI
From an open letter endorsed by tech leaders like Elon Musk and Steve Wozniak which proposed a six-month pause on AI development to Henry Kissinger co-writing a book on the pitfalls of unchecked, self-learning machines, it may come as no surprise that AI’s mainstream rise comes with its own share of caution and warnings. But these worries didn’t pop up with the sudden popularity of AI apps like ChatGPT; rather, concerns over AI’s influence have existed decades long before, expressed even by one of its early researchers, Joseph Weizenbaum.
ELIZA
In his book Computer Power and Human Reason: From Judgment to Calculation (1976), Weizenbaum recounted how he gradually transitioned from exalting the advancement of computer technology to a cautionary, philosophical outlook on machines imitating human behavior. As encapsulated in a 1996 review of his book by Amy Stout, Weizenbaum created a natural-language processing system he called ELIZA which is capable of conversing in a human-like fashion. When ELIZA began to be considered by psychiatrists for human therapy and his own secretary interacted with it too personally for Weizenbaum’s comfort, it led him to start pondering philosophically on what would be lost when aspects of humanity are compromised for production and efficiency.
Copyright chenspec (Pixabay)
The Importance of Human Intelligence
Weizenbaum posits that human intelligence can’t be simply measured nor can it be restricted by rationality. Human intelligence isn’t just scientific as it is also artistic and creative. He remarked with the following on what a monopoly of scientific approach would stand for, “We can count, but we are rapidly forgetting how to say what is worth counting and why.”
Weizenbaum’s ambivalence towards computer technology is further supported by the distinction he made between deciding and choosing; a computer can make decisions based on its calculation and programming but it can not ultimately choose since that requires judgment which is capable of factoring in emotions, values, and experience. Choice fundamentally is a human quality. Thus, we shouldn’t leave the most important decisions to be made for us by machines but rather, resolve matters from a perspective of choice and human understanding.
AI and Human Intelligence in Market Research
In the field of market research, AI is being utilized to analyze a multitude of data to produce accurate and actionable results or insights. One such example is deep learning models which, as Health IT Analytics explains, filter data through a cascade of multiple layers. Each successive layer improves its result by using or “learning” from the output of the previous one. This means the more data deep learning models process, the more accurate the results they provide thanks to the continuing refinement of their ability to correlate and connect information.
While you can depend on the accuracy of AI-generated results, Cascade Strategies takes it one step further by applying a high level of human thinking. This allows Cascade Strategies to interpret and unravel insights a machine would’ve otherwise missed because it can only decide, not choose.
Take a look at the market research project we performed for HP to help create a new marketing campaign. As part of our efforts, we chose to employ very perceptive researchers to spend time with worldwide HP engineers as well as engineers from other companies.
This resulted in our researchers discovering that HP engineers showed greater qualities of “mentorship” than other engineers. Yes, conducting their own technical work was important but just as significant for them was the opportunity to impart to others, especially younger people, what they were doing and why what they were doing was important. This deeper level of understanding led the way for a different approach to expressing the meaning of the HP brand for people and ultimately resulted in the award-winning and profitable “Mentor” campaign.
If you’re tired of the hype about AI-generated market research results and would like more thoughtful and original solutions for your brand, choose the high level of intuitive, interpretive, and synthesis-building thinking Cascade Strategies brings to the table. Please visit https://cascadestrategies.com/ to learn more about Cascade Strategies and more examples of our better thinking for clients.
Aug
The Future Is Here
“Mitigating the risk of extinction from AI should be a global priority alongside other societal-scale risks such as pandemics and nuclear war.”
With just 22 words, we are ushered into a future once heralded in science fiction movies and literature of the past, a future our collective consciousness anticipated but has now taken us by surprise upon the realization of our unreadiness. It is a future where machines are intelligent enough to replicate a growing number of significant and specialized tasks. A future where machines are intelligent enough to not only threaten to replace the human workforce but humanity itself.
Published by the San Francisco-based Center for AI Safety, this 22-word statement was co-signed by leading tech figures such as Google DeepMind CEO Demis Hassabis and OpenAI CEO Sam Altman. Both have also expressed calls for caution before, joining the ranks of other tech specialists and executives like Elon Musk and Steve Wozniak.
Earlier in the year, Musk, Wozniak, and other tech leaders and experts endorsed an open letter proposing a six-month halt on AI research and development. The suggested pause is presumed to allow for time to determine and implement AI safety standards and protocols.
Max Tegmark, physicist and AI researcher at the Massachusetts Institute of Technology and co-founder of the Future of Life Institute, once held an optimistic view of the possibilities granted by AI but has now recently issued a warning. He remarked that humanity is failing this new technology’s challenge by rushing into the development and release of AI systems that aren’t fully understood or completely regulated.
Henry Kissinger himself co-wrote a book on the topic. In The Age of AI, Kissinger warned us about AI eventually becoming capable of making conclusions and decisions no human is able to consider or understand. This is a notion made more unsettling when taken into the context of everyday life and warfare.
Working With AI
We at Cascade Strategies wholeheartedly agree with this now emerging consensus and additionally, we believe that we’ve been obedient in upholding the responsible and conscientious use of AI. Not only have we long been advocating for the “Appropriate Use” of AI, but we’ve also made it a hallmark of how we find solutions for our client’s needs with market research and brand management.
Just consider the work we’ve done with the Expedia Group. For years, they’ve utilized a segmentation model to engage with their lodging partners by offering advice that could lead to the partner winning a booking over a competitor. AI filters through the thousands of possible recommendations to arrive at a shortlist of the best selections optimized for revenue.
With the continued growth and diversification of their partners, they then needed a more effective approach in engaging and appealing to them, something that focuses more on that associate’s behavior and motivations. We came up with two things for Expedia: a psychographic segmentation formed into subgroups based on patterns of thinking, feeling, and perceiving to explain and predict behavior, and more importantly, a Scenario Analyzer that utilizes the underlying AI model but now delivers recommendations in very action-oriented and compelling messaging tailor-fit for that specific partner.
The best part about the Scenario Analyzer is whether the partner follows any of the advice recommended or does nothing, Expedia still stands to make a profit while maintaining an image of personalized attentiveness to their partner’s needs. And ultimately, it’s the partner who gets to decide, not the AI.
Copyright Tara Winstead
Our Future With AI
This is how we view and approach AI- it’s not the end-all, be-all solution but rather an essential tool in increasing productivity and efficiency in tandem with excellent human thinking, judgment, and creativity. Yes, it is going to be part of our future but in line with the new consensus, we believe that AI shaped by human values and experience is the way to go with this emerging and exciting technology.
Aug
How Can Healthcare Companies Identify Who Needs Remediation Programs?
jerry97890 comments artificial intelligence, Burning Questions
What Is Remediation?
The Cambridge Dictionary defines remediation as “the process of improving or correcting a situation.” Remediation programs are commonly employed in teaching and education wherein they address learning gaps by reteaching basic skills with a focus on core areas like reading and math. And as pointed out in an understood.org article, remedial programs are expanding in many places in our post-COVID 19 world.
In healthcare, there’s a wide range of remediation programs, or “remedial care,” diversified based on their end goal which may include smoking cessation, anti-obesity, weight reduction, diet improvement, exercise, heart-healthy living, alcoholism treatment, drug treatment, and more. But how do you identify the people who need remedial care the most?
Who Needs Remediation?
You might say you can tell who needs remedial care by just looking at the physical aspect of the prospective patient, but this is a shortsighted answer to the question. And what about those who need remedial care for a heart-healthy lifestyle? Surely you can’t tell a likely candidate for this remediation program with just one look alone.
It goes deeper than that. What if you, a healthcare representative, could only devote remedial care to a select few individuals given limited resources and time but you want to make sure that the whole remediation program is successful by achieving its intended goals? Just imagine all that time, effort and resources spent only for the patient to relapse back into their old ways not too long after program completion- or even in the middle of the remediation process itself.
Deep Learning and Remediation
This is where deep learning comes in. Also known as hierarchical learning or deep structured learning, Health IT Analytics defines deep learning as a type of machine learning that uses a layered algorithmic architecture to analyze data. In deep learning models, data is filtered through a cascade of multiple layers, with each successive layer using the output from the previous one to inform its results. Deep learning models can become more and more accurate as they process more data, essentially learning from previous results to refine their ability to make correlations and connections.
Deep learning models handle and process huge volumes of complex data through multi-layered analytics to provide fast, accurate, and actionable results or insights. When applied to the scenario we mentioned beforehand, deep learning filters through that multitude of patient data and prioritizes those who need remedial care the most.
You can also align its findings to effectively identify individuals who will not only return monetary value to your healthcare brand, but at the same time are most likely to “engage” or participate in programs offered by your company, such as wellness, diet, fitness or exercise. They can also be the best people to commit to avoiding poor lifestyle choices, such as overeating, smoking, and alcohol, helping guarantee the success of the remediation program.
With a combination of three decades of market research experience and conscientious use of AI, Cascade Strategies has been helping healthcare organizations develop advanced models to handle, filter and identify the likeliest of candidates for their program purposes. Cascade Strategies helps industry professionals not only recognize their ideal customers but also reach out to them with some of the most effective and award-winning marketing campaigns, thanks to our array of services such as Brand Development Research and Segmentation Studies. To see more examples of how we help leading worldwide companies achieve their goals, please visit our website.
Here are some of our suggestions for further reading on deep learning and healthcare:
https://builtin.com/artificial-intelligence/machine-learning-healthcare
https://research.aimultiple.com/deep-learning-in-healthcare/
https://healthitanalytics.com/features/types-of-deep-learning-their-uses-in-healthcare