To gather data on characteristics of CA facilitated ideas from real users, we initiated an open call during a research project involving partners from research and practice in the field of public administration. The call on the topic of “Mobility of the Future” was distributed via different university and city mailing lists, social media, and student groups to invite a wide group of participants to generate ideas. Guided to a website via link in the open call, participants were provided with information about the subsequent task and the possibility of winning vouchers. The topic was presented in the form and length of an abstract describing advantages and disadvantages of current mobility solutions. The participants were asked to propose ideas for a change of mobility at the national level. The idea generation process with the CA could be started by clicking a designated button. In total 40 ideas could be collected and served as data for a two-fold idea evaluation, reported in the following. First, interviews with domain experts were conducted to qualitatively assess the collected ideas. Second, computerized text-based analyses of the submitted ideas were performed to examine textual features of the ideas and establish links between idea contributors’ social behaviors and cognitive processes.
7.1.1 Expert Evaluation of Ideas and CA Facilitation
To allow an in-depth evaluation of the subject matter, the ideas and the utilized approach were assessed by four experts with different backgrounds of relevant experience in the domain of open innovation and ideation (see Table
2). Based on established idea evaluation dimensions in literature (Dean et al.,
2006), we conducted semi-structured interviews via video call that lasted between 41 and 53 min. The interviews were conducted with an interview guide comprising open-ended questions. Questions about the general impression of the ideas and the approach of CA facilitation were followed by questions about the completeness, level of detail, comprehensibility, originality, acceptability, and relevance of the submitted ideas. Prior to the interviews, each expert was provided with context information regarding the conceptual approach, process, and topic, as well as a random subset of ten idea submission. The interviews were recorded and transcribed by paraphrasing and noting verbatim statements.
Table 2
Interviewees for evaluation of the idea generation approach and idea characteristics
E1 | Software | Business development manager | 5–8 years | Male |
E2 | Real estate | Innovation project consultant | 5–8 years | Male |
E3 | Mobility | Technology manager for innovation projects | 2–4 years | Female |
E4 | Logistics | Startup software developer | > 8 years | Male |
The experts understood the CA facilitator approach to gather external ideas and considered it useful, even if CA technology is currently applied for different use cases in their organization (i.e., all experts were familiar with CA technology). In particular, the dialogue-based interaction was judged to be promising to receive ideas from external contributors as part of an open innovation initiative (“It is easier for contributors, because you receive feedback from the CA.”). Regarding the presented ideas, the experts emphasized the formulated ideas to be an extension of their own perspective. In this respect, some ideas particularly stood out, which were considered surprisingly unusual and novel (e.g., “I wouldn't have thought of such a thing.”). However, the experts noted that some ideas might be too radical from their point of view to be generally accepted. Nevertheless, one interviewee mentioned that radical approaches are a good sign, as they show an open process (e.g., “These are good food for thought and you don't want to see them stalled either.”).
The ideas were judged to be well elaborated and understandable. Regarding the level of detail, however, it was noted that even more idea-specific input would have been desirable. This would have allowed to go even deeper into the minds of the idea providers. It was suggested that the CA could have been even more proactive about specific terms used by the contributors, such as ridesharing, and asked specific questions (e.g., “What exactly do you mean by this?”). This would allow to obtain even more contextual knowledge. For example, the CA could also actively, i.e., without being addressed, have provided suitable suggestions from a database as an additional stimulus for the contributors to elaborate their idea (“It would be useful if there was a kick-start to trigger participation”). In relation to the assumed goal of the CA facilitation, i.e., collecting a large number of ideas, the experts mentioned that the ideas were already very well elaborated for a first idea collection step. “More detail is always possible, but it was enough for understanding” and an even more detailed level of elaboration could also complicate the idea screening and selection (e.g., “Who is meant to read through all that?”). Whether more content would be advantageous for a (partially) automated evaluation could not be conclusively assessed by the experts. The advantage of a more intensive dialogue should be weighed against the possible tendency of idea contributors to abort the process and a declining motivation to finish the idea generation (e.g., “They might get bored despite the engaging conversation at some point.”). Despite this, the experts expressed that the clear structure of ideas is certainly an advantage for the subsequent evaluation and selection, regardless of the method applied. Looking at the entire subset of ideas, the content was judged to be mostly consistent in terms of the different idea attributes. No obvious extreme deviations were noted by the interviewees.
When asked to what extent the provided ideas solve a problem in the context of the subject matter, it was stated that “the ideas address and comprehensively include the problem” and that very promising ideas had been proposed. However, further details would have been desirable and useful in some cases. Nevertheless, these ideas were a suitable starting point to identify one visionary idea among many in order to enter an in-depth exchange with this individual about his or her idea for solving the problem. Regarding the advantages of using a CA facilitator, the overall adaptability, and the possibility of accessing a current and large database that can be incorporated into the process of idea generation were highlighted. In the same context, the need for a large amount of data and its preparation for CA training was considered critical. One advantage that one expert emphasized was that a dialogical CA facilitation is suitable to involve all users regardless of their individual prerequisites, i.e., from a cognitive perspective, who can also have very useful ideas. In this regard, automatic adaptation of CA’s behavior and utterances based on personal characteristics of the idea contributor was considered potentially valuable and could be leveraged with technological advances (e.g., “Especially when you think about the future possibilities that you don't want to miss, this is a great playground.”).
7.1.2 Text-based Evaluation of Ideas
To link idea contributors’ written language style in the gathered texts during the idea generation process to affective (e.g., negative and positive emotions) and cognitive processes (e.g., problem-solving), we conducted linguistic analyses with computerized text analysis. This form of text analysis has been used to study social networking sites, online discussion forums, group dynamics, and interactions between individuals (Kacewicz et al.,
2014; Oliver et al.,
2021; van Swol & Kane,
2019) and yields reliable psychological insights about individuals’ thought processes, emotional states, intentions, and motivations (Boyd & Pennebaker,
2015; Tausczik & Pennebaker,
2010).
We examined the collected idea texts by applying a dictionary approach. We used the program Linguistic Inquiry and Word Count (LIWC) (Pennebaker et al.,
2015a,
b). LIWC utilizes over 90 pre-defined categories, analyzes and classifies words within these categories, which allows for a consistent measurement of words, leading to concordant validity (Humphreys & Wang,
2018; Moore et al.,
2021; Pilny et al.,
2019). The fundamental power of the LIWC dictionary lies in the fact that it was thoroughly developed using established and standardized psychometric procedures that ensure external validity and high internal reliability (Boyd,
2017; Pennebaker et al.,
2015a,
b). Given the German text corpora, we rely on the translated German LIWC2015 dictionary (Meier et al.,
2019), which captures an average of 83 percent of the words people use in written and spoken language.
To prepare the linguistic analysis, we followed the guidelines for German text samples to preprocess the texts before analysis (Meier et al.,
2019). For the linguistic analysis (see Table
3), we use five general descriptive dimensions: word count (WC), words per idea (WPI), words per sentence (WPS), percent of words in the text that are longer than six letters (Sixltr), and percent of target words captured by the dictionary (Dic). In addition, we utilized four summary variables: analytical thinking (Analytic), clout, authenticity (Authentic), and emotional tone (Tone).
Table 3
Results of the linguistic analysis of the submitted ideas
Idea text | 1295 | 32.38 | 15.42 | 35.98 | 79.31 | 97.06 | 68.78 | 33.31 | 83.54 |
Elaboration questions | 1456 | 36.40 | 8.88 | 37.50 | 76.30 | 96.61 | 60.02 | 41.20 | 39.82 |
Title | 133 | 3.33 | 3.80 | 54.14 | 54.89 | 99.00 | 69.14 | 57.71 | 88.32 |
Mean | 961.33 | 24.04 | 9.37 | 42.54 | 70.17 | 97.56 | 65.98 | 44.07 | 70.56 |
SD | 589.40 | 14.73 | 4.76 | 8.23 | 10.87 | 1.04 | 4.22 | 10.17 | 21.82 |
Grand meana | 5429.44 | – | 20.18 | 22.90 | 82.72 | 49.53 | 60.63 | 48.34 | 61.20 |
SDa | 9245.24 | – | 119.94 | 4.15 | 6.93 | 20.62 | 14.86 | 24.41 | 27.69 |
The four summary measures each reflect a 100-point scale ranging from 0 to 100 with standardized scores. The underlying complex algorithms are proprietary. The variables are constructed from various LIWC variables based on previous language research (Boyd & Pennebaker,
2015). The scale values reliably reflect the following psychological dimensions (Boyd & Pennebaker,
2015, pp. 21–22):
-
Analytical thinking: a high number reflects formal, logical, and hierarchical thinking; lower numbers reflect more informal, personal, here-and-now, and narrative thinking.
-
Clout: a high number suggests that the author is speaking from the perspective of high expertise and is confident; low Clout numbers suggest a more tentative, humble, even anxious style.
-
Authentic: higher numbers are associated with a more honest, personal, and disclosing text; lower numbers suggest a more guarded, distanced form of discourse.
-
Emotional tone: a high number is associated with a more positive, upbeat style; a low number reveals greater anxiety, sadness, or hostility. A number around 50 suggests either a lack of emotionality or different levels of ambivalence.
The level of analysis refers to the gathered texts during the idea generation process steps, i.e., idea text, elaboration questions (which problem is solved, idea novelty, target audience), and title, as these are sufficiently self-contained and distinct from each other to allow meaningful intra-process comparison. Keywords were not included, since the analysis of individual keywords based on the LIWC categories appears to make little sense. A total of 54 keywords, mostly one to two per idea and compound words (see explanation below), were assigned for identification purposes from the idea contributors. A single idea was submitted without any keyword.
Although idea titles were relatively short on average (3.33 words per idea), they were included in the LIWC analysis because they are potentially informative covering a range from concise and descriptive to bold and lurid in a wording continuum. 54.14% of words in the title text were longer than six letters, which is notably higher than the respective percentages for the idea texts (35.98%) and question answers (37.50%). The result for the titles is related to the frequent utilization of compound words. Compound words consist of several nouns attached to each other and their extensive use is a peculiarity of the German language. While some of the most common compound words are included in the German LIWC dictionary, less common compound words are not recognized (Meier et al.,
2019). This was reflected in the title texts with 54.89% of the target words identified.
The percentage of words longer than six letters were fairly at the same level regarding the idea texts and answers to the elaboration questions at 35.98% and 37.50%, indicating more active, i.e., less frequent use of long compound words, and consistent language across the process steps. Accordingly, the percentage of target words captured by the dictionary for the idea texts and answers were higher than for titles, at 79.31% and 76.30%, respectively. This puts them at about the same level as the fundamental German LIWC dictionary capture rate of 83%.
Considering the idea texts and the answers to the elaboration questions, the phrased sentences were almost one-half shorter at 15.42 to 8.88 words. This discrepancy is associated to the mixture of key phrases and rather short sentences in the answers to the questions. Remarkably, though, answers to the elaboration questions with 36.40 words per idea were longer than the idea texts with an average of 32.38 words. Thus, the elaboration questions contributed substantially to the overall idea generation process.
The idea texts, answers to questions, and titles are characterized by strong analytical thinking (opposed to narrative thinking) with each over 97-scale points. Accordingly, during the idea generation process, the idea contributors predominantly used a formal, categorical style of textual language that is associated with increased abstract thinking and a logical, complex way of cognitive processing. Individuals with such a predisposition in processing information tend to analyze, break down problems and are more likely to weigh facts (Boyd & Pennebaker,
2015; Pennebaker et al.,
2014).
The texts of the ideas with 68.78 points and the titles with 69.14 points on the clout scale were almost on par. The answers to the questions were somewhat lower with 60.02 points. Compared to the grand mean clout score of 60.63 (
SD = 14.86) points from the German LIWC dictionary (Meier et al.,
2019), these scores reflect a somewhat greater level of contributors’ competence and confidence in the text. In addition, individuals who score high on the clout dimension usually use more outward words and are more focused on the people they interact with than on themselves. This type of interaction has been found to be conducive in the context of online discussion forums supporting the type of interaction and engagement required to build knowledge (Adaji & Olakanmi,
2019; Kacewicz et al.,
2014; Moore et al.,
2021).
Authenticity scores for the text segments ranged from 33.31 (idea texts), and 41.20 (answers to elaboration questions, to 57.71 (title). Compared to the grand mean value of 48.34 (
SD = 24.41) (Meier et al.,
2019), the value for the idea texts was relatively low and the value for the titles was relatively high. In order to understand these values, it is helpful to look at base rates of word usage from which the grand mean was calculated. The data sets of “Expressive writing” (76.73 points) and German-speaking “Reddit” (35.09 points) formed the ends of the authenticity continuum. Reddit is a social media platform where individuals discuss and exchange ideas on various subject matters (e.g., sports, politics, and leisure) in the form of threads and forums. Expressive writing, on the other hand, comprised samples from cross-sectional and longitudinal studies in which individuals wrote about profoundly personal issues in stream of consciousness mode (Meier et al.,
2019). This put the idea texts at about the same level as social media which reflects informal, netspeak language (Meier et al.,
2019). Nevertheless, the relatively low values are related to a rather reserved and distanced form of communication.
Looking at the scale for emotional tone, it is noticeable that the answers to the elaboration questions reflected a lack of emotional terms (39.82 points). In comparison, the scores for the idea texts (83.54) and the titles (88.32) showed a rather high occurrence of positive verbal signs of emotion on the emotion scale, suggesting that the idea contributors were more emotionally involved during these steps in the idea generation process.