How content analytics can be brought into research. The presentation was given as a webinar for the IE Business school where John Griffiths is a visiting professor.It features examples of the use of Purefold transmedia as a research methodology and the use of demographic replicator research bots. Part of the Cloud of Knowing project
1. The Cloud of Knowing content analytics and the future of market research John Griffiths Planning Above and Beyond Jan 20 th
2.
3.
4. Redrawing the lines around research Will research advance to include content? Or will research retreat and become a specialism? the guys who do Interviews..
9. So can online research help? Online survey panels Online focus groups Face to face/CATI surveys Offline discussion/ Focus groups Research Communities { } { } Open platforms Eg FB group
12. Purefold: open source co-created transmedia film director story â seedsâ RSS feeds gather raw content Friendfeed hopper Visitors /linkers grade and link to other Web content storylines amplified Brand owners sponsor & create characters storylines Output put into production Ridley Scott Assoc/Ag 8
15. Demographic replicators: Research bots â Wefeelfineâ emotional wrapper DGR Felix text bot twitter posterous friendfeed network Analysis of others who share bot tastes Comments/ retweets of those who dialogue with the bot Findings aggregated with other bots
16.
17. What is Felix? A projective device for researchers to reframe research questions? A dynamic sample/panel which can be observed and with which the researcher can interact A lure/decoy for drawing out customer response
18.
19. Past the sample/content bottleneck Sample Content Quant & Qual research keep sample and content questions separate Inability to control sample is what makes web content analysis problematic Sampling is literal: 20 male, works in local govt, lives in London
20. Computing has a similar bottleneck Instructions/data Memory/CPU
Online Research is a horseless carriage which looks a lot like offline research. It hasnât really started to take on unique forms which are suited to the internet.
Ridley Scott Associates â directors came up with short film stories â the concept being that these would be co-created with web content â funded by brand sponsorship and would be a collaboration between commercial storytellers the film industry, the content of the web, and brands creating entertainment which allowed them to dialogue with their customers. A trans media concept â using the creative commons â copyright free content anyone could write stories, make films or hyper documents as part of this collaboration. Take script ideas and put them into a wiki Using RSS feeds to collect story seeds and to feed them into a friendfeed forum where visitors would grade them, link to other content or bring in their own. Extended to storyline characters, set design. The shaped stories were then presented to brand sponsors who could work branded entertainment story lines, characters and themes into them.
Data gathered by RSS feed â research participants who go out tagging relevant content. So finding more relevant data. Other research participants who grade what comes through the friend feed/wiki. Â Analysis and grading carried out by research professionals. How does this look as a hybrid research methodology? Sampling? Yes getting selection and explanation from respondents? Yes - This broadens out the ORC online research community to something much richer where participants play a much more active part. And the richness comes from the data flowing through â less pressure on moderating and surveys of the internal community. Recognisably a research hybrid â with a similar cost structure as ORCs.
Going deeper. Demographic replicators â living samples who represent a quantitative sample whose blogging tweeting is gathered in real time. Who respond to live questions. Who have been aggregated to research sample criteria. But who may not even know that they are a subject for research. Â Demographic replicators also incorporate other forms of data GPS, potentially spending data, media consumption. So represent an aggregation of actual engagement and behavioural data. Â So still fulfilling sampling criteria but now a weird combination of machine and human. Still recognisable as research but becoming more abstract.
Research needs to establish 2 things âwho and what â what they are and what they think, what they feel, how they behave In quant sampling questions at the end of surveys- used for analysis but not for content. For qual sampling questions are asked during recruitment â using profile questionnaire. Sampling not only distinct but kept separate. Rule is that you canât use the same piece of data to identify the participant (who) and as evidence of their opinion (what) â the strain of keeping them separate is preventing research from using data when it is not possible to identify who it is who has produced the material. Secondly sampling is to be literal. As far as possible year, demographics and usage are to be verified. By the word of the research respondent.
This is analogous to the development of the computer when instructions and data used to be on 2 different media card based knitting machine. Alan Turing whose papers conceptualised the computer and Van Neuman who separated out the storing of data from the processing of data â turingâs concept allowed instructions and data to be in the same medium. Van Neuman introduced a bottleneck because processors cost so much more than memory. So programmes were stored and put through a CPU. That bottleneck is still with us today. I suggest that the separate identification of the sample from the content is a similar bottleneck which is preventing research from going any further.
Sampling â first probabilistic and using a tagging system to tag content either sample or thematic content tags. Cookies already in usage for identifying webusers. Once we have overcome the bottleneck. There is no reason why we shouldnât identify content first and then use cookies/sampling tags to eliminate all the content which is off sample. Now this starts to make available to us the riches of the web in realtime searchengines and google profile.
Will research go this route? At present this is not allowed. Vested interest in not going this route makes it unlikely. Qualitative researchers functioning rather like portrait painters in the 19 th century â not immediately displaced by the invention of photography but increasingly irrelevant as it became possible for people to take their own photos â not as good but much quicker and much more cost effective. Quantitative research has colonised the internet with a close analogue of what offline surveys have offered as a successful industrial model of data collection. The fact that online panels have become discredited so quickly is a sign the drawbacks of offline surveys were simply imported without being addressed â artificial, overly structured, measuring themselves rather than capturing reliable insightful claims of customer behaviour.
I hope it does for the sake of research. Because research has a great future following customers. This is where CRM is going without the data warehouses. Research methodologies badly needed where customer data is being aggregated and marketing decisions being made client side. If research doesnât make this jump then it risks being demoted to a specialism. If you donât know ask the customer. When more and more customer decision making is made without the customer.
Parting shot. As lines of demarcation are redrawn around companies it is possible that customers become just another stakeholder group with whom client companies are continually in dialogue and they wonât call it research any more. When I started my career within a week I was attending meetings at which the most senior marketing people were present. If we donât broaden research to include content analytics it wonât stop clients anlysisng content as part of decision support. They may just not ask us to do it for them. So a junior researcher in 2013 (30 years after I started) may see rather less of the marketing director.