Hostname: page-component-586b7cd67f-2brh9 Total loading time: 0 Render date: 2024-11-23T19:34:51.193Z Has data issue: false hasContentIssue false

The generation of problem-focussed patent clusters: a comparative analysis of crowd intelligence with algorithmic and expert approaches

Published online by Cambridge University Press:  11 October 2017

Rights & Permissions [Opens in a new window]

Abstract

This paper presents a new crowdsourcing approach to the construction of patent clusters, and systematically benchmarks it against previous expert and algorithmic approaches. Patent databases should be rich sources of inspiration which could lead engineering designers to novel solutions for creative problems. However, the sheer volume and complexity of patent information means that this potential is rarely realised. Rather than the keyword driven searches common in commercial systems, designers need tools that help them to understand patents in the context of the problem they are considering. This paper presents an approach to address this problem by using crowd intelligence for effective generation of patent clusters at lower cost and with greater rationale. A systematic study was carried out to compare the crowd’s efficiency with both expert and algorithmic patent clusters, with the results indicating that the crowd was able to create 80% more patent pairs with appropriate rationale.

Type
Research Article
Creative Commons
Creative Common License - CCCreative Common License - BY
Distributed as Open Access under a CC-BY 4.0 license (http://creativecommons.org/licenses/by/4.0/)
Copyright
Copyright © The Author(s) 2017

1 Introduction

Patents not only provide intellectual property (IP) protection, but are an extremely valuable source of information. They can be freely used by designers to review solutions to comparable technical problems, draw inspiration for new areas of research, identify trends and development in the field, examine competitors and their activity, acquire technologies to licence or use freely, and avoid unintentional duplication of research. While they relate to all aspects of the innovation ecosystem (Luo Reference Luo2015), in reality, for designers, they have most often become a checkpoint in the design process rather than a resource to be utilised in support of design activity.

This is partly attributable to the size and complexity of the database. Given that there are now over 50 million online patent records instantly available, and with the global growth in patent applications at more than 7.5% (World Intellectual Property Organization 2016), effective navigation of these has become increasingly difficult. Additionally, the language and presentation of patents has evolved to suit the legislative and bureaucratic requirements of the patent system – they are not engaging documents written and presented in a way that is useful to a designer. The abstract and main image are the main areas for interpretation, but nevertheless the pertinent inventive detail can be difficult to infer from these abbreviations. There is a lot of variation in writing style which is often impenetrable to the average reader. Finally, the aim of the patent system – to encourage innovation by providing a period of protection for a fixed length of time, and thereafter sharing the information across society for its greater benefit – has been hijacked by companies utilising patents purely for commercial purposes. Practices such as ‘thicketing’ (deliberately stifling innovation in an area by filing a series of blocking patents) and ‘trolling’ (attempting to enforce rights against infringers beyond a patent’s value or contribution) have distorted the innovation economy to the point where they have become synonymous with the widely reported ‘patent wars’ (Moser Reference Moser2013). In essence, the translation of patent documents into practical design ideas or requirements is not easy.

We have identified crowdsourcing as a way to potentially address some of these concerns and to provide an economic, scalable way of collating and applying appropriate taxonomic information that could reclaim the patent database as a source of engineering design inspiration. Crowdsourcing is the process of obtaining information or input from a large number of people to complete a task. Key components in crowdsourcing workflows are repletion (i.e. multiple, parallel tasks to generate sets of ‘answers’), peer review and merger, iteration, and the linkage of payment to quality assessments.

In assessing the viability of its use in the design context, the key question is whether crowdsourcing offers any benefits when compared with other approaches. To this end, we have focussed on the use of the crowd in patent clustering, as this is an aspect of analysis where the need for insight and interpretation has proven challenging for computational approaches. We have performed systematic benchmarking, including comparative analyses of the crowd with computer algorithms, design experts and commercial landscaping software to assess the benefits. This incorporates and builds on data from previous work on computationally structured databases, our aim being to provide a clear basis for moving forward with this approach as well as to establish continuity in this line of investigation in the literature.

1.1 Foundational research

In previous work, we have reported on two experiments to establish basic crowd capability. These included a patent textual information categorisation task and an exercise that required the interpretation of patent drawing information (Vasantha et al. Reference Vasantha, Corney, Maclachlan and Wodehouse2016). The key conclusions included the following:

  • a large crowd of people from many countries is available on demand to undertake posted tasks;

  • completion time is quick, although it is dependent on the inclusion of test questions;

  • judgement works best by aggregating all responses rather than relying on individual responses;

  • clear task instructions, test questions, and payment for each task can play a role in getting acceptable responses;

  • the best workers should be chosen through initial tests, then nurtured for higher-reasoning patent analysis tasks.

While there are clear challenges in designing task protocols and managing effective workflows for the crowd, these initial findings have provided us with sufficient encouragement to progress to more formal analyses. Our research has identified three key themes in relation to patents in design: finding relevant patents (searching), organising groupings in relation to particular problems (clustering), and applying these in design activities (utilising), as illustrated in Figure 1. Clustering of patents relies on insight and judgement to create meaningful interpretations, and as such is challenging to achieve using computational approaches. This has therefore been the theme that we have focussed on through the implementation of a crowd-based approach.

Figure 1. Overview of project themes for patent analysis.

2 Patent clustering

Patents are often used to assist with technological forecasting (Jin et al. Reference Jin, Jeong and Yoon2015; Kim & Jun Reference Kim and Jun2015) and the identification of opportunities for development (Trappey et al. Reference Trappey, Trappey, Peng, Lin and Wang2014). However, there is limited research on how patent clusters can be presented to support design activity. Kitamura et al. (Reference Kitamura, Kashiwase, Fuse and Mizoguchi2004) outlined the possibility of creating patent maps of functional decomposition to be used in design reviews as a means to link design functions to patents. A limitation is that the proposed software requires the user to detail function decomposition trees on a graphical user interface. Vandevenne et al. (Reference Vandevenne, Verhaegen, Dewulf and Duflou2016) proposed a scalable search for systematic biologically inspired design. This utilised functional characteristics to identify candidate products for design by analogy, thereby increasing the variety and novelty of ideas subsequently generated (Verhaegen et al. Reference Verhaegen, Peeters, Vandevenne, Dewulf and Duflou2011). They outline challenges in interactive result filtering and classification in successfully extracting knowledge from the patents. Fu et al. (Reference Fu, Murphy, Yang, Otto, Jensen and Wood2014) presented a method for finding analogies based on a functional vector space representation (Murphy et al. Reference Murphy, Fu, Otto, Yang, Jensen and Wood2014) from patent databases to assist designers. The results demonstrated improved novelty (but no change in quantity) in design solutions subsequently generated. Together, these results indicate that there is potential within the patent database to support enhanced ideation. The rest of this section addresses in more detail the principles and issues that can be applied to clustering, with a view to applying these in the most appropriate way to the design context.

2.1 Clustering principles

Patent clustering aims to develop meaningful patterns for identifying dominant technologies, forecasting future technological trends, detecting patent infringement, and identifying technological vacuums. It is carried out using structured (e.g. inventors, granted date, etc.) and/or unstructured patent data (free text; e.g. claims, description, etc.). It is clear that although computational algorithms are maturing, human intervention remains essential to bring insight to patent clusters: several authors have emphasised that human expert judgement and outcomes are safest and most accurate on patent analysis and clustering (Fantoni et al. Reference Fantoni, Apreda, Dell’Orletta and Monge2013; Park et al. Reference Park, Kim, Choi and Yoon2013). Moreover, given its importance in the visualisation of patents, the current reliance of computational approaches on the International Patent Classification (IPC) classes is unsatisfactory and can result in landscapes that lack meaning or require further interpretation to be useful (Widodo & Budi Reference Widodo and Budi2011).

The major processes involved in algorithmic clustering are text mining (to extract useful information from the patent corpus) and grouping (to identify similar patents based on structured information received from the text mining step). The key objective of text mining is to extract and classify information that provides some kind of structure for further analysis. For example, function, behaviour and structure (FBS) is one such classification used for text mining purposes (Yan Reference Yan1993). Key requirements for patent text mining are that all patent texts are coded appropriately using a chosen classification mechanism, with steps taken to ensure consistency in applying the established definitions and relationships. Patent grouping is concerned with maximising the similarity within clusters (i.e. homogeneous, with minimal intracluster distance) and the diversity of patents between clusters (i.e. heterogeneous, with maximum intercluster distance) (Yoon & Park Reference Yoon and Park2004). There are various measures for likeness, such as Euclidean distance (Widodo & Budi Reference Widodo and Budi2011), similarity measure (Widodo & Budi Reference Widodo and Budi2011), silhouette measure (Jun et al. Reference Jun, Park and Jang2014), cosine angle (Fu et al. Reference Fu, Cagan and Kotovsky2011) and Hamming distance (Xu et al. Reference Xu, Zhiyong and Bin2013). Table 1 summarises various algorithmic approaches used in the text mining and grouping processes.

Table 1. Summary of text mining and grouping approaches to patent clustering, with highlighted references

Keyword and K-means are the popular approaches used in text mining and grouping respectively. The frequency-inverse document frequency (tf-idf) is often used for determining the importance of words and keyword extraction in text mining (Kukolj et al. Reference Kukolj, Tekic, Nikolic, Panjkov, Pokric, Drazic and Nemet2012). Commonly, a binary patent-keyword/similarity matrix/noun–verb matrix is provided as an input to the K-means clustering approach (Suh & Park Reference Suh and Park2006; Park et al. Reference Park, Kim, Choi and Yoon2013). Indeed, various combinations of the text mining and grouping approaches highlighted in Table 1 can be utilised to find an appropriate clustering process and outcome. Even in doing so, however, there are persistent issues that have been highlighted in the literature.

2.2 Clustering issues

Patent textual analysis is challenging due to variations in writing style, usage of abstract terminologies, and confusion in classifying terms (e.g. technology or effect). When two patents are very different, this leads to ineffective word overlap for text-based analysis methods and causes difficulties in measuring semantic similarities (Xu et al. Reference Xu, Zhiyong and Bin2013; Li et al. Reference Li, Atherton and Harrison2014). The WordNet database can be used to expand patent terms for synonyms, hypernyms, hyponyms and polysemy (one word having multiple meanings) matching to create a level of generalisation. However, this does not contain all of the domain-specific technical terms (Choi et al. Reference Choi, Park, Kang, Lee and Kim2012; Park et al. Reference Park, Kim, Choi and Yoon2013; Abbas et al. Reference Abbas, Zhang and Khan2014). The implicit domain knowledge that is not mentioned in patent documents is missed in cluster analysis procedures (Fu et al. Reference Fu, Chan, Schunn, Cagan and Kotovsky2013b ). Since the importance (weightage) of words is given to frequently occurring terms (Tekic et al. Reference Tekic, Kukolj, Nikolic, Drazic, Pokric, Vitas and Nemet2012), there is a high risk that less-frequent but important terms are not given due attention in the analysis process (Park et al. Reference Park, Kim, Choi and Yoon2013). It has been emphasised that frequencies and co-occurrences of the patterns may not reflect substantial characteristics of inventions (Yoon et al. Reference Yoon, Park and Kim2012), and are unable to discriminate noise from information (Ryley et al. Reference Ryley, Saffer and Gibbs2008), and that keyword vectors cannot reflect structural relationships among technological components (Trappey et al. Reference Trappey, Trappey, Hsu and Hsiao2009; Park et al. Reference Park, Yoon and Kim2011).

In terms of grouping, it is difficult to choose the appropriate technique for a specific task; evaluations have pointed out that the efficiency varies widely for different data sets and approaches (Kukolj et al. Reference Kukolj, Tekic, Nikolic, Panjkov, Pokric, Drazic and Nemet2012; Segev & Kantola Reference Segev and Kantola2012), and that tested cluster similarity scores are not high (Widodo & Budi Reference Widodo and Budi2011). Typically, computational grouping is carried out by the relative distance between text mined data from compared patents, so outcomes do not provide clear categorical clustering (Trappey et al. Reference Trappey, Trappey and Wu2010), which is required to navigate effectively among patents. Most of the clustered outcomes do not clearly establish cluster meaning or provide valuable labels for each cluster (Yoon & Park Reference Yoon and Park2004; Spangler et al. Reference Spangler, Ying, Kreulen, Boyer, Griffin, Alba and Yan2011; Xu et al. Reference Xu, Zhiyong and Bin2013). Additionally, the loss of semantic information through the translation of high-level data in patent documents into specific design functions or requirements can lead to inaccurate clustering (Fan et al. Reference Fan, Lai, Huang and Huang2011; Tekic et al. Reference Tekic, Kukolj, Nikolic, Drazic, Pokric, Vitas and Nemet2012; Abbas et al. Reference Abbas, Zhang and Khan2014) and uncertainty in establishing the distance between patents (Yoon et al. Reference Yoon, Park and Kim2012). Therefore, clustering techniques normally require human expert input at various stages of execution, including semantic annotation (Xu et al. Reference Xu, Zhiyong and Bin2013), extracting key phrases (Kim et al. Reference Kim, Suh and Park2008; Trappey et al. Reference Trappey, Wu, Taghaboni-Dutta and Trappey2011; Choi et al. Reference Choi, Park, Kang, Lee and Kim2012; Park et al. Reference Park, Kim, Choi and Yoon2013), removing irrelevant SAO structures (Choi et al. Reference Choi, Park, Kang, Lee and Kim2012), defining clusters at each iteration (Jun et al. Reference Jun, Park and Jang2014), validating cluster labels (Fu et al. Reference Fu, Chan, Schunn, Cagan and Kotovsky2013b ), feedback for semantic recognition of keywords (Atzmüller & Landl Reference Atzmüller and Landl2009), and verification (Trappey et al. Reference Trappey, Trappey, Peng, Lin and Wang2014). Atzmüller & Landl (Reference Atzmüller and Landl2009) have suggested that it could take up to ‘…6–8 h of the client’s time to develop a thorough understanding of how the map can be interpreted’. This interpretive effort should ideally be carried out when the map is constructed in order to present visualisations that are clear, understandable and useful.

3 Method

We therefore performed comparative analyses of a crowd-based approach to clustering with computer algorithms, design experts and commercial landscaping software to assess its benefits. This included the use of data – a 45 patent set and different clusterings of it – from research previously reported by Fu et al. (Reference Fu, Cagan, Kotovsky and Wood2013a ,Reference Fu, Chan, Schunn, Cagan and Kotovsky b ,Reference Fu, Dilmore, Cagan and Dougherty c ) on computationally structured databases. Its aim was to compare the algorithm developed by the authors with the interpretation of four experts. Use of these findings has allowed us to extend the benchmarking for comparison with the crowd and with the PatentInspiration landscaping software.

3.1 Fu et al.’s data set

The design scenario used in Fu et al.’s work was to develop a human-powered energy generator for use in rural communities, and the set of 45 patents was selected by the authors. The patents fell across 96 cooperative patent classification (CPC) codes and 113 International Patent Classification (IPC) codes, and were in the date range of 1976–2007. All were owned by different applicants, and all were filed in the USA. A full list of the patents has been included in Appendix A.

3.1.1 Algorithm clustering

To assist designers in visualising the patent database appropriately, Fu et al. (Reference Fu, Cagan, Kotovsky and Wood2013a ,Reference Fu, Chan, Schunn, Cagan and Kotovsky b ) proposed a Bayesian model combined with latent semantic analysis for discovering structural form in a dataset. The algorithm output is in the form of labelled clusters based on functional (verbs) and surface (nouns) patent similarity. The cluster structures take different forms (such as a ring, chain or tree) to yield different insights and interrelationships between the patents. They argued that these clusters provide cross-domain associations and transfer of knowledge based on functional and surface content similarity. The algorithm could place the design problem in a space within the generated cluster to highlight the most relevant patents. They used two different methods for cluster label generation, namely the highest average rank labelling method (top five highest average ranked words) and the highest cosine similarity labelling method (top five words with the highest cosine similarity values in the pool of words). The clustering outputs generated for the highest average rank labelling are shown in Figure 2. The black filled circle in the top left corner represents the design problem; the farthest patent relevant to the design problem is patent number 20 in the bottom right corner.

Figure 2. Patent clusters labelled using the highest average rank method (Fu et al. Reference Fu, Chan, Schunn, Cagan and Kotovsky2013b ).

3.1.2 Expert clustering

Fu et al. (Reference Fu, Chan, Schunn, Cagan and Kotovsky2013b ) presented results of patent clusters generated by four design experts for this dataset. The experts had at least 10 years of experience in the field of product design and an educational background in engineering. It was noticeable that the clusters created by the experts were all fairly distinct, with limited correlation across them. The resulting patent space generated by Expert 3 is illustrated in Figure 3.

Figure 3. The patent space defined by Expert 3 (Fu et al. Reference Fu, Chan, Schunn, Cagan and Kotovsky2013b ).

3.2 PatentInspiration landscaping software

PatentInspiration™ is an online patent analysis software, which intends to provide a quick and easy way to analyse text, and yield information on trends and technology across industries and domains. It provides 50 standardised presentation templates, the most widely used of which are ‘landscapes’, graphical representations of the dataset which borrow characteristics of cartography. Like most applications, this utilises the classification codes issued to the patents during the filing process to group ‘islands’ according to their area of application – an example using the CPC codes for the 45 patent set is shown in Figure 4. Additional semantic information for each cluster is then applied in the form of labels that help to identify the nature of the patents contained in each cluster. However, there is no information to understand the significance of the distance between the clusters, or the relevance to a particular design problem. Additionally, the fact that clusters are constrained to specific CPC or IPC classification codes limits the emergence of cross-classification patent groups.

Figure 4. A patent landscape based on CPC classification codes (generated with PatentInspiration, http://www.patentinspiration.com).

3.3 Distribution of clustering task to the crowd

During the course of the research, we have developed a bespoke software platform to manage the distribution of tasks to the crowd. This was a significant undertaking but was deemed necessary, as existing crowdsourcing platform templates proved insufficient for the complexity of patent clustering tasks. As well as providing a flexible, graphic interface for workers to interact with patent sets in a more intuitive way, it allows control over backend workflows in delegating tasks to the crowd. We were able to manage recruitment and payment of workers by integrating with existing systems (i.e. Crowdflower and mTurk), as illustrated in Figure 5. In working with the 45 patent set, the main characteristics of the interface presented to the crowd are the following:

  • a split screen allows for scrolling through individual patents’ abstract, title and main figure (left side), and through patent clusters (right side);

  • it allows each patent to be maximised to full screen view for visualising the patent abstract and image in detail;

  • generated cluster labels follow the right side scroll bar to remain visible and facilitate easy patent grouping;

  • it enables patents in both grouped and ungrouped clusters to be searched, with the number of matches highlighted;

  • there are options for editing labels, adding comments and deleting individual patents or patent clusters.

The workflow for distributing patent clustering tasks is depicted in Figure 6. The platform facilitates the presentation of the patent abstract and image in such a way that workers can easily group, label and rank individual patent groups in relation to the given design problem. This task design was deemed more efficient and effective than setting each patent pair as a discrete crowdsourcing task then synthesising to form clusters. Moreover, while Fu’s data set and other computational approaches generally use full-text analysis, it was determined that the abstract and image provided a sufficient level of detail for the crowd to review and assess patent content. The distribution of the tasks and review of the results is managed in the system backend, before being used to create clusters for use by designers. In terms of the system architecture, a variety of open-source tools were utilised. The platform was compiled using a Node.js package and a Meteor Up client deployed through Amazon Web Services’ cloud database storage. MongoDB (a document-oriented database program) was used to collect, structure and store securely all cluster information generated by the crowd. While we have been working with an initial 45 patent set, the system developed is scalable to increase the number of patents to be clustered. Since the focus of this work is to create a knowledge base that designers can browse and utilise with reference to a particular design problem, we anticipate patent sets scaling into the hundreds rather than thousands. Sets larger than this are suited to landscaping and scoping rather than interactive use. The crowdsourcing approach as detailed can comfortably scale to patent sets in the hundreds. However, if it was desired to scale into larger sets of thousands of patents, the most effective way would be to partition the task and synthesise the results from each. This is still possible but would require more time and cost.

Figure 5. Articulation between established crowdsourcing platforms and the customised patent cluster platform.

Figure 6. The workflow for assignation, gathering and presentation of crowd results.

4 Results – performance of the crowd

In comparing the performance of the crowd against other approaches, we utilised the same design problem and clustering procedure as set out by Fu et al.. The challenge for the crowd was to identify patents that would help with the development of a low-cost device to collect energy from human motion for use in developing and impoverished communities – the full briefing text is documented in Appendix B. Three aspects of clustering were assessed across two discrete tasks. In the first task, the crowd was asked to create and name patent groups. These lend a patent map its ‘shape’, and rely on patent similarity. When the basic groupings were established, the crowd was then asked to generate appropriate labelling for each group to form a cluster. This followed the format of the open clustering exercise undertaken by experts in Fu et al.’s benchmarking work. The second task was to assess the applicability of patents. To do this, we asked the crowd to evaluate the relevance of each patent group relative to the design problem in order to orient the clusters effectively. Fixed amounts were paid to the crowd for the patent clustering ($3) and relevance ranking ($0.50) tasks. On top of these, we provided a bonus payment of up to $3 to incentivise workers to provide elaboration of their rationale for labels. Table 2 summarises the statistics of crowd participation using the two crowdsourcing platforms.

Table 2. Participation statistics for the two crowdsourcing platforms

On posting the tasks, 181 responses were received, with nearly 95% of these coming within 48 hours of posting. The acceptable output percentage (all patents grouped, all groups labelled) of 45% was encouraging considering that the crowd consisted of untrained workers participating for the first time. While the rate of response to the tasks suggests that this percentage would be workable, with a curated crowd returning to complete tasks consistently, a significantly higher acceptable output percentage, and more economically efficient crowd, would be anticipated. Furthermore, nearly 80% of participants involved in Task 1 (similarity and labelling) went on to complete Task 2 (relevance rating), a positive retention rate and indicator of crowd satisfaction with the task/reward structuring. On average, crowd workers spent 48 minutes on Task 1 and 17 minutes on Task 2. This is a considerable amount of time when compared with the low-complexity tasks generally advertised, such as image tagging, collecting information from websites or categorising tweets. These generally take a couple of minutes to complete and use the standard templates available on commercial crowdsourcing platforms. Our results suggest that the crowd is adaptable towards more involved and time-consuming tasks with appropriate task formatting and reward. In our case, payment for successful completion was US$5 on average, a figure that included web platform administration charges. Figure 7 shows the location of participants during the tasks, and highlights the global reach of crowdsourcing platforms. While we were unable to track personal information on our platform, a recent demographic survey shows that the crowd population is becoming increasingly international, with an equal gender distribution, average age of 32 years, and almost half having graduate education (Ipeirotis Reference Ipeirotis2010; Ross et al. Reference Ross, Irani, Silberman, Zaldivar and Tomlinson2010).

Figure 7. Location of workers participating in patent tasks (generated with BatchGeo, https://batchgeo.com).

4.1 Task 1: similarity and labelling

4.1.1 Similarity

In the first part of the grouping task, the crowd was asked to create functionally similar groups from the 45 patent set, and if two patents resided in the same group – having a discernible link in functional content – they were designated as a pair. For a set of $N$ patents, the number of possible pairs ( $P$ ) is $P=N\times (N-1)/2$ , so for the 45 patent set, 990 pairs were possible. Table 3 lists the average number of clusters and patent pairs generated through the different approaches. We only used the results of two of the experts (E3 and E4) reported in Fu et al.’s work (refer to Appendix B in Fu et al. (Reference Fu, Chan, Schunn, Cagan and Kotovsky2013b )), as the demarcations between patent groups were unclear in the clusters generated by the other two experts. We found that the average number of clusters generated by the crowd was equivalent to experts’ results. The average number of pairs identified by the crowd (164) was, however, higher than that generated by the two experts ( $69+128/2=99$ ). If a single group has a large number of patents, this will tend to increase the overall number of pairings. For example, the maximum number of patent pairs for a clustered outcome is 513. This is due to the fact that 32 patents are grouped in a single cluster, creating a large number of pairings (496). Thus, spreading the patents more evenly across clusters will reduce the number of pairs. While one of the experts eliminated some patents as irrelevant to the design problem, reducing the experts’ average number of pairs, this practice appears to be equivalent to the crowd workers’ tendency to organise many patents into a single group. For the algorithms, it was found that Fu et al.’s identified 125 pairs and the PatentInspiration software 253. This suggests that Fu’s algorithm is more discerning than the PatentInspiration software, which relies purely on CPC/IPC classifications.

Figure 8 shows the correlation between the number of clusters and the number of connected patent pairs using the different approaches. The average for the crowd has been drawn with an inverse exponential curve, and the other approaches, bar one of the experts, follow this reasonably closely. The overall results show significant variety in patent clusters and labels, which is valuable in providing a range of insights for sifting and prioritisation.

Figure 8. Number of clusters versus number of connected patent pairs for the crowd, experts and algorithms.

Table 3. Composition of clusters and connected patent pairs for different approaches

Examining the performance of the crowd in more detail, Figure 9 shows the frequency of agreement in patent pairings across the 83 workers, i.e. the number of patent pairs. The left of the peak can be characterised as largely speculative pairings between patents with little in common. The peak itself between eight and 12 workers indicates the greatest degree of consensus on patent pairs that are less explicitly linked. The aggregated intellectual interpretation in this region provides a degree of working confidence in the crowd insight. Moreover, to the right of the peak, we see increasingly explicit connections, with a maximum of 58 workers agreeing to a patent pair. At least one crowd worker identified a link in all 989 patent pairs (i.e. there was only one unconnected pair).

Figure 9. The number of crowd workers identifying identical patent pairs.

Table 4 describes the five patent pairs with the highest number of workers (i.e. towards the right of Figure 9) in agreement in more detail. This shows that patent pairs are linked and extensively labelled by the crowd for both explicit (e.g. different valve types) and implicit (e.g. joint assembly and steering device) functional relationships. It is in the less obvious pairings that the crowd provides a distinct advantage over computational approaches. Further analysis to compare expert and algorithm outcomes with aggregated crowd results indicates that only unlinked patent pairs can be predicted with a high degree of confidence (refer to Appendix C for Phi nominal correlation and regression results).

Table 4. Patent pairs with highest generated labels by the crowd compared with algorithmic and expert assessment

We conducted a further analysis with the number of the crowd who agreed on a pairing compared with the percentage of agreement with Expert 1, Expert 2, Fu et al.’s algorithm and the PatentInspiration software (Figure 10). With reference to a particular number of crowd workers, the agreement percentage was calculated by comparing the frequency of crowd pair agreement with the frequency of pair agreement of the various other approaches. This allows us to examine the degree of consensus in identifying patents that should be grouped, are of equal importance, and those that should not. For patent pairs where few of the crowd agreed on a link, we can see a similarly low proportion of proposed pairs by the experts and algorithms. A fairly stable direct relationship holds until we reach patent pairs where the level of crowd agreement is 30 and above. At this point, the relationships break down. For instance, 45 crowd workers agreed on a link between a pair (1 and 43 from Table 4), but experts and algorithmic approaches did not link that particular patent pair. This highlights that both algorithmic and expert approaches have overlooked patent pairs that the crowd workers identified, and supported with meaningful rationale, as relevant.

Figure 10. Percentage of benchmark pairings identified by numbers of crowd workers.

4.1.2 Labelling

The second part of the grouping task was for the crowd to create appropriate labels with rationale. Clusters require appropriate labelling to summarise their content and assist with positioning. Both algorithmic outputs (Fu et al. algorithm and PatentInspiration) produced a list of words as labels, whereas both the crowd and experts can add more context and rationale. Therefore, while a higher number of patents can be processed computationally to provide an overview of a sector or field, there is a potential advantage to be gained from human intervention in the creation of more meaningful cluster labels at a functional level. Table 5 sets out the number of distinct cluster labels generated in both algorithmic and human approaches, and shows that a greater number of labels were generated by the crowd. Given the size of the crowd over the expert group, it would appear that there is convergence in the cluster labelling undertaken by the individual crowd members.

Table 5. Distinct cluster labels generated in various approaches

Fu et al. used two methods to generate labels for each cluster. They selected the top five highest average words generated by the highest average rank labelling method (which ranks each word in the latent semantic analysis space for each patent within a cluster) and the highest cosine similarity labelling method (which generates the highest cosine similarity values in the pool of words for the patents within that cluster). Raising the cut-off values of rank and cosine similarity scores (i.e. choosing the top 10 rather than the top five words with the highest scores) to increase the number of labels generated through the algorithmic approaches may dilute the quality of the generated labels.

Table 6. Comparison of cosine similarity scores for labels across different clustering approaches

Table 7. Grammar percentages for labels generated by various approaches

We went on to compare the labelling of the different methods using the cosine similarity across common label sets (Table 6), with the vector representing the frequency of a particular label presented in each approach. In this case, scores range from 0–1, with a 1 indicating that they are identical. The results show a strong cosine similarity correlation between the experts and the crowd workers, with a much lower match between the experts and the algorithmic approaches.

The reason for this non-convergence is that the algorithmic approaches mostly generated object- (i.e. noun) oriented labels rather than functionally oriented labels. The part-of-speech tagger from NLTK (http://www.nltk.org) was used to categorise labels generated with various approaches with respect to English grammar. Table 7 summarises the grammar percentages for labels generated by various approaches. The grammar percentages demonstrate clear differences between algorithmic and human labelling approaches. The algorithmic approaches mostly chose noun labels, whereas humans provide more context to labels by using a wide variety of English grammars.

4.2 Task 2: Relevance

The third aspect of cluster analysis we explored was the relevance of patents to the design problem. Fu’s clustering algorithm situates the design problem in the dark circle at the top left of Figure 2. This allows us to identify clusters in close proximity, which might provide practical design information to assist in embodiment design, and more distant clusters, which might provide more analogous ideas and principles that could be brought to bear. Most software landscaping tools, however, do not have the capability of locating the design problem in relation to clusters (providing a ‘centre’ to the map), so instead these become visual aids to browse and understand but are more difficult to apply. One of the main advantages in asking experts to review patents in relation to a design problem is the facility for this kind of interpretation: they can find the shallow and deep links for different patent types, consider clustering depending on the nature of the problem and the perspective this brings, and prioritise cluster groups based on this information. The question for us was whether this kind of expert analysis, so dependent on experience and insight, could be replicated in a crowd that has a greater number but less expertise.

A task was therefore set which asked the crowd workers to rate the level of relevance of patents to the design brief of designing a device to collect energy from human motion for use in developing and impoverished rural communities. A 1–6 rating was chosen to match the distance rankings generated by Fu’s algorithm – Figure 2 shows that the patents occupying the top left node are most relevant (a rating of 1), with those five nodes removed in the bottom right (a rating of 6). Figure 11 compares the relevance ranking generated by Fu’s algorithm and the highest cumulative rank frequency marked by the crowd workers. This shows that the crowd rankings concentrated at the extremes of the ranking scale, whereas Fu’s algorithm populated mostly the middle rankings.

Figure 11. Comparison between the crowd and Fu’s algorithm ranking of patent relevance to the given problem.

This indicates a significant variance between the crowd and the computational approach with respect to patent relevance for a particular design problem. While it may have been useful to correlate with relevance as identified by experts, only one of the four in Fu’s analysis (Expert 2) was deemed to have explicitly incorporated distance from the design problem in their clustering. Even for that expert, distance was measured roughly from the design problem circle to the centre of each cluster rather than each patent, so an accurate correlation was not deemed appropriate.

4.3 Evaluation of crowd responses

To assess the validity of the crowd results, evaluation was conducted on the 989 patent pairs and their labels by three evaluators. Evaluator 1 was an RA with eight years of design research experience, Evaluator 2 was a post-doc with a design-related PhD and Evaluator 3 was a senior product design undergraduate with prior experience in patent research. The experts reviewed every patent pair with associated labels, and for each identified whether the crowd had created suitable similarity labels, generating any additional labels not already present, or indicated that there was no link between the patent pair. It took the evaluators an average of approximately 20 hours to complete the analysis of crowd labels. The Cronbach alpha inter-coding reliability scores were 0.811 (Evaluators 1 and 2), 0.878 (Evaluators 1 and 3) and 0.735 (Evaluators 2 and 3), indicating a high level of internal consistency. Table 8 summarises patent pair agreement between the evaluator and crowd results. On average, the evaluators agreed with 59% of the crowd’s patent pairs and labels, and with 6% of the crowd’s patent pairs (but not the labels). The evaluators generated additional labels for 18% of agreed patent pairs.

Table 8. Patent pair agreement between evaluator and crowd results

Figure 12 portrays box plots showing the correlation between the number of evaluators and the number of crowd workers agreeing with a patent pair. A one-way ANOVA was conducted to determine whether the number of crowd workers who agreed with a patent pair varied in relation to the number of evaluators who agreed with a patent pair. The mean $\pm$ standard deviation for zero, one, two and three evaluators in agreement against the crowd results is shown. The number of crowd workers agreeing pairs increased from $9.40\pm 4.71$ for no evaluators agreeing pairs, to $10.87\pm 5.19$ for one evaluator agreeing, to $13.94\pm 5.81$ for two evaluators agreeing, to $17.33\pm 7.88$ for all three evaluators agreeing. The relationship between these groups was statistically significant, ( $F(3,985)=96.392$ , $p<0.001$ ), showing that the likelihood of all evaluators agreeing with a patent pair is high if there are more crowd workers in agreement for that pair.

Figure 12. Box plot illustrating the relationship between the number of agreed evaluators and the number of crowd workers agreed with a patent pair.

4.4 Final cluster outcome

The generation and evaluation of patent pairs, cluster labelling and ranking of relevance to the design problem led to the production of a patent matrix to aid visualisation and understanding of the groupings created. We used ClustergrammerFootnote 1 to generate a dynamic and interactive matrix visualisation rather than the static patent clusters represented in Section 2. Figure 13 illustrates the partial patent cluster matrix between patents ( $x$ -axis) and generated labels ( $y$ -axis). This choropleth visualisation uses colour variation to represent relevance to the design problem (a darker shade represents stronger relevance to the problem). For example, the ‘material handling’ label is associated with Patents 3, 9 and 14, and Patents 9 and 14 are more relevant to the design problem than Patent 3. The order of the labels and patents is augmented by the software to optimise the clusters; however, this can be controlled by the user if desired.

Figure 13. Visualisation of clustering between patents and patent labels for a cropped patent cluster zone.

The visualisations are highly interactive: they can be reordered according to hierarchical clustering, rank by sum, and rank by variance approaches, and the right top and bottom left sliders can be changed to increase or decrease the number of patent clusters with the slider. This level of control allows designers to dynamically choose and change the number of clusters according to the level of detail or granularity required for the task in hand. The generation of different views in this way can help to identify unique patent to label relationships, and stimulate new insights in relation to the design problem. Furthermore, matrices can be generated for patent to patent and label to label clusters in addition to the patent to label clusters shown in Figure 13. This can help to identify relationships between them and their importance to the design problem. Important labels can be automatically chosen by varying the relevance sum or variance score. This function represents dimensionality reduction performed by various algorithms.

The visualisation communicates the specific relationship between each patent pair with appropriate labels, showing both dense patent clusters and also sparse links between patents. In other words, the map does not categorically isolate different clusters. It comprehensively represents the information generated from the evaluated crowd results, avoiding any loss of information that can occur in the algorithmic refinement process. Additionally, there is the opportunity for designers to navigate patent clusters dynamically according to their requirements: mouse hovering highlights a particular patent and connected label, and clicking a particular cell in the cluster bring up the relevance score. Moreover, the addition of IPC/CPC codes as categories can help engineers to understand how different domains interact with reference to crowd-generated similarity labels.

5 Discussion

Based on the experiments conducted, we can summarise what added value the crowd currently provides in terms of the clustering patents for design, and how we can optimise its response going forward.

In determining patent similarity, the aggregated and evaluated crowd results generated 83% and 79% more patent pairs in comparison with average pairs created by experts and Fu’s algorithm respectively. This result is substantial, suggesting a significant number of overlooked patent pairs by experts and algorithmic approaches. With regards to the experts, the limited amount of time undertaking the exercise may have been a factor. For Fu’s algorithmic approach, the important information generated about patent relationships was hidden in the process. The final patent cluster shown in Figure 2 presents only narrow information generated from the overall process. For example, the pairwise patent relationships created through latent semantic analysis were not presented in the published paper. Designers may benefit from exploring the results after each of the algorithmic steps, allowing the exploration of specific pair relationships – in particular those dismissed as unlinked. Rather than simply presenting the final outcome (as shown in Figure 2), a more accessible resource could help to highlight the depth of relationships required for leverage in design practice. The crowd’s ability to identify implicit functional patent pairs could potentially be useful to train and develop advanced clustering algorithms. Another characteristic of the crowd was to place more patents in a single group when identifying similarity. This could reflect a clear perception of how patents are related, but could also indicate a lack of full consideration of the implications for each patent. Identification of optimal grouping and pairing numbers that emerged through our evaluation could prove valuable in the self-evaluation and monitoring of results. When reviewing the patent pairs in detail, an advantage of the crowd is that labels are generated both for explicitly linked patents (e.g. functionally similar valves that lie in the same product category) as well as less obvious connections (e.g. steering devices and universal joints that could be cross-applicable). These insights can be aggregated and evaluated to provide a fuller overview of patents that may be relevant to a technology or problem.

In the generation of patent labels, both the experts’ and the crowd’s differed greatly with algorithmic approaches. Algorithmic extraction through ‘word’ analysis such as frequency and cosine similarity is not rich enough to represent meaningful coherent information of a patent group. The results show that people generate high-level functional labels, with extended descriptions that show flexibility in adapting to different domains and technical properties. Human heuristics to generate labels according to these criteria, and the adaptive behaviour necessary to do so, is still far from being replicated by algorithmic approaches. Given that the crowd was untrained, with unknown levels of prior experience or knowledge, the strong overlap with labels generated by experts is a promising result. With higher levels of curation, it could be expected that the level of convergence among the crowd and the overlap with expert perception would rise. The size of the crowd to some extent mitigates the lack of expertise of the workers, as through the evaluation of aggregate responses we can find the meaningful and valuable insights that might not emerge when relying on computational approaches alone.

In evaluating patents’ relevance to a design problem, the results show a clear difference between algorithmic and human reasoning. The crowd tended to rate patents using the full scale available, whereas the Fu algorithm had less distinct results. This suggests that the crowd may offer more definitive insights than computational approaches with respect to patent relevance. While potentially useful in highlighting creative triggers or unusual insights from patents that might otherwise have been dismissed, strong opinions need to be aggregated to identify those that are meaningful and insightful. By associating and consolidating the rationale provided by a large number of people to different relevance ratings, the identification and demarcation of patent usefulness becomes much clearer for designers.

The discrepancies between the algorithm and crowd results suggest that it may be advisable to utilise a combined approach. The algorithm reliance on classification and explicit noun terms allows for the rapid grouping and categorisation of functionally similar patents using full-text analysis; the crowd can be utilised to verify these groupings and provide enhanced understanding and rationale for why these groupings emerge through comparison of images and abstracts. In addition, in the pairings where there are less obvious but potentially valuable connections, the crowd results can be aggregated and evaluated to provide additional insights. The patent cluster matrix produced in this work can act as a new way to create clusters based on individual labels rather than clubbing labels to represent groups of patents. In doing so, it provides clusters based on relationships between labels, and between patents themselves.

Table 9. Practical considerations of crowdsourcing clustering with reference to other approaches

5.1 Application

In practical terms, the crowd is highly appropriate for utilisation in patent cluster analysis. Table 9 sets out time, cost and basic performance characteristics for the different approaches to patent clustering. It can be seen that the crowd cost is significantly less than expert time and software applications. We have no information on the potential cost of the algorithm associated with Fu’s work, but based on its sophistication, we can assume that if it was integrated in a software package it would be at the higher end of the market. We chose to use PatentInspiration, as it offers a good level of functionality and a user-friendly interface at a lower price point than the majority of its competitors, making it suitable for comparison with the very low cost of the crowdsourcing approach. However, we should be aware of the other potential clustering outcomes offered by more expensive packages with significant development resources behind them. For example, Thompson’s Themescape™ is one of the most recognised in the field and as well as employing a sophisticated algorithm, it utilises additional classification by ‘expert teams’ to try to ensure the accuracy and relevance of the results. As organisations seek to make such classifications more time and cost efficient, and employ outsourcing facilities to assist and augment their core operation, the process begins to move into the domain of crowdsourcing. The crowd is also extremely quick. The majority of responses to tasks were received within a matter of hours, with a small number of responses increasing the overall task response time. While this will never match the speed of computational approaches, it is not a critical factor in the preparation of a cluster analysis. We can consider that increasing the payments to the crowd with a corresponding increase in task response time could significantly increase the quality of responses. A key part of our recommendation is that a curated crowd, which has been primed to complete patent analysis tasks, will still be an order of magnitude cheaper than alternatives, comparatively quick, and produce useful results. Even though we conducted the study based on a 45 patent set, the availability of global crowd resources means that it is possible to scale the workflows as necessary for larger patent sets.

5.2 Limitations

It is important to note that the use of crowdsourcing has been found to be challenging in the design context. For example, Burnap et al. (Reference Burnap, Ren, Gerth, Papazoglou, Gonzalez and Papalambros2015) found the use of crowd consensus for design evaluation problematic due to challenges in establishing appropriate levels of evaluator expertise among workers. We have tried to address the issues of technical knowledge through the nature of the workflow – the crowd is asked to pair and label patents, so their competence is demonstrated by the successful completion of these. Moreover, while an individual’s result is not reliable, aggregation of these results across larger numbers can provide workable results. Additionally, we have relied on evaluators to check the pairings identified by the crowd. An aspect of future work is to reconfigure this evaluation as a crowdsourced task to eliminate this part of the workflow.

Another issue in the early experimentation reported here was the degree of variability in the quality of the work submitted despite the use of test questions and reward payments. When these are in place and a core of responsive workers have been secured, the crowd can be expected to mature and the overall quality of results can be expected to rise over a period of engagement. Furthermore, a well-structured task instruction page and intuitive user interface are essential to facilitate good crowd responses, and this will generally require some iteration through pilot testing. While our software platform facilitated the design of optimised task workflows, there are limits to what can be achieved using the most common crowdsourcing platforms such as CrowdFlower and mTurk. Since neither of these platforms gives the task provider control over the choice of individuals when allocating work, it is not easy to develop the best workers over a period of time. Ideally, this would be done by initial screening tests and then progressively harder jobs in preparation for patent analysis tasks which require high levels of reasoning and judgement.

The nature of the patent set is something that should also be considered in relation to the findings. Depending on the design context – scoping, generation, embodiment or testing – we would expect the patent set to vary in size and scope. While the 45 patent set used here was not particularly large, it was very diverse. We found that the crowd was capable of identifying relationships between patents from very different contexts. It would be desirable to run further tests with a more focussed patent set, perhaps centred on a particular technical topic, to understand whether the crowd is able to discern between patents at this more detailed level. We anticipate that with appropriate guidelines and task design for workers, this should be viable, but the effect on completion time and acceptance rate would be revealing.

As we have outlined above, we anticipate the patent sets to be clustered and accessed in the application of this approach to be in the hundreds rather than the thousands. However, a clear benefit of the algorithmic approach is the ability to scale into thousands of patents at no additional cost. To achieve a larger data set using the crowdsourcing approach, the most practical approach would be to run the task a number of times and synthesise the results for analysis. The global source of workers suggests that there would be no issues in terms of the execution of this; however, there is an additional cost in terms of time (if running the tasks sequentially) and cost (on a worker-by-worker basis). A combination of algorithmic and crowd approaches may have advantages in, for example, first acquiring and refining a patent set for a particular design problem through computational approaches and then utilising the crowd to further parse and cluster. This focusses the human effort where it is most valuable – in interpreting and synthesising design information.

6 Conclusion

The research reported here provides a new approach for the utilisation of patents in design contexts. While a significant amount of material has been written on how patents relate to innovation at a macroeconomic or strategic level, we have focussed on how value and insights can be extracted from functional groupings to support design problem solving. Specifically, we have demonstrated the advantages and limitations of utilising crowdsourcing to review, cluster and label patents. The construction of a bespoke software platform meant that we had an architecture flexible enough to design visually orientated tasks where information was clearly presented to workers. The design of the clustering workflow, implementing similarity, labelling and relevance ranking tasks and aggregating them for analysis, has allowed us to establish the viability of the crowd in comparison with computational and expert approaches. In practice, different stages of the design process require different types of patent sets. For example, a scoping exercise requires a large and diverse patent set to allow for effective horizon scanning, whereas embodiment design may require a more focussed and technical patent set to solve a particular mechanical issue. In future work, we therefore intend to work with patent sets of varying size and diversity, developing search strategies to define these. Using these, we will monitor the effect on crowd performance, and evaluate the effect of the generated visualisations on the work produced by designers.

Acknowledgments

This work was supported by the Engineering and Physical Sciences Research Council (EPSRC) grant reference EP/N005880/1. All data supporting this research are openly available from http://dx.doi.org/10.15129/eae5f958-4135-4caf-9dc4-df6d5de19c9f.

Appendix A. Patent set

The 45 patent set table used in crowdsourcing studies.

Appendix B. Clustering tasks set to the crowd

Task 1: Group patents

You are going to support an engineer who is currently solving the following design problem.

‘Design a device to collect energy from human motion for use in developing and impoverished rural communities in places like India and many African countries. Our goal is to build a low-cost, easy to manufacture device targeted at individuals and small households to provide energy to be stored in a rechargeable battery with approximately 80% efficiency. The energy is intended to be used by small, low-power-draw electrical devices, such as a radio or lighting device, hopefully leading to an increase in the quality of life of the communities by increasing productivity, connection to the outside world, etc. The target energy production is 1 kW-hour per day, roughly enough to power eight 25 Watt compact florescent light bulbs for 5 hours each per day, or enough to power a CB radio for the entire day.’

The engineer identified 45 patents that could help in this task. However, these patents need to be grouped to help the engineer to easily navigate among them. Patents should be grouped based on ‘functional similarity’ to facilitate searching through the collection of patents for potential inspiration to solve the abovementioned design problem. The group of patents should capture functional similarity, and not necessarily problem or technological domain similarity. The example for functional similarity is a patent for ‘a car suspension system’ is grouped with ‘a running shoe’ based on the fact that ‘shock absorbing’ is a function that is common in both patents.

Your task is to structure these 45 patents into smaller groups based on functional similarity, and provide a suitable functional group title.

Task 2: Patent relevance to the design problem

Your task is to rate your patent groups on a scale of 1 to 6 based on importance to the given problem.

Table 10. Phi nominal correlations between crowd workers and other approaches

Appendix C. Predicting expert and algorithm outcomes using the crowd

Table 10 presents the positive Phi nominal correlation values with statistical significance between the crowd workers’ pairs and the other clustering approaches, as well as the maximum Phi nominal correlation values achieved. The results show that the average and maximum number of crowd-generated clusters are correlated significantly with the PatentInspiration software and Expert 1 outputs. However, the highest significant correlation achieved is only 0.286, which provides moderate confidence in predicting experts’ outcomes from individual crowd cluster results.

This conclusion is further confirmed in logistic regression of the sum of patent pairs generated by all of the crowd workers concerning other approaches (Table 11). The logistic regression establishes that there is a statistically significant prediction possibility of Fu’s algorithm, experts and PatentInspiration outcomes from the aggregated sum of all of the crowd workers’ results. However, the prediction percentage is higher only for unlinked patent pairs (i.e. ungrouped patent pair) in all cases. This result shows that patent clustering is a complex activity, in which each and every approach and person generates different results, which makes prediction difficult.

Table 11. Chi-square values and percentage of correct prediction of unlinked and linked pairs through crowd workers’ results of logistic regression

Footnotes

References

Abbas, A., Zhang, L. & Khan, S. U. 2014 A literature review on the state-of-the-art in patent analysis. World Patent Information 37, 313.CrossRefGoogle Scholar
Atzmüller, P. & Landl, G. 2009 Semantic enrichment and added metadata – examples of efficient usage in an industrial environment. World Patent Information 31 (2), 8996.Google Scholar
Burnap, A., Ren, Y., Gerth, R., Papazoglou, G., Gonzalez, R. & Papalambros, P. Y. 2015 When crowdsourcing fails: a study of expertise on crowdsourced design evaluation. Journal of Mechanical Design, Transactions of the ASME 137 (3), doi:10.1115/1.4029065.CrossRefGoogle Scholar
Choi, S., Park, H., Kang, D., Lee, J. Y. & Kim, K. 2012 An SAO-based text mining approach to building a technology tree for technology planning. Expert Systems with Applications 39 (13), 1144311455.Google Scholar
Dereli, T. & Durmuşoğlu, A. 2009 Classifying technology patents to identify trends: applying a fuzzy-based clustering approach in the Turkish textile industry. Technology in Society 31 (3), 263272.Google Scholar
Dražić, M., Kukolj, D., Vitas, M., Pokrić, M., Manojlović, S. & Tekić, Z.2013 Effectiveness of Text Processing in Patent Documents Visualization. Paper presented at the 2013 IEEE 11th International Symposium on Intelligent Systems and Informatics (SISY).Google Scholar
Fan, C. Y., Lai, M., Huang, T. & Huang, C.2011 Applying K-means Clustering and Technology Map in Asia Pacific-Semiconductors Industry Analysis. Paper presented at the Industrial Engineering and Engineering Management (IEEM), 2011 IEEE International Conference on.Google Scholar
Fantoni, G., Apreda, R., Dell’Orletta, F. & Monge, M. 2013 Automatic extraction of function–behaviour–state information from patents. Advanced Engineering Informatics 27 (3), 317334.Google Scholar
Fu, K., Cagan, J. & Kotovsky, K.2011 A Methodology for Discovering Structure in Design Databases. Paper presented at the DS 68-6: Proceedings of the 18th International Conference on Engineering Design (ICED 11), Impacting Society through Engineering Design, Vol. 6: Design Information and Knowledge, Lyngby/Copenhagen, Denmark, 15–19.08. 2011.Google Scholar
Fu, K., Cagan, J., Kotovsky, K. & Wood, K. 2013a Discovering structure in design databases through functional and surface based mapping. Journal of Mechanical Design 135 (3), 031006.CrossRefGoogle Scholar
Fu, K., Chan, J., Schunn, C., Cagan, J. & Kotovsky, K. 2013b Expert representation of design repository space: a comparison to and validation of algorithmic output. Design Studies 34 (6), 729762.Google Scholar
Fu, K., Dilmore, J., Cagan, J. & Dougherty, C. H. Jr. 2013c Using Design Database Structures to Characterize Freedom-to-Operate in a Design Space: A Legal Case Study. Paper presented at the DS 75-1: Proceedings of the 19th International Conference on Engineering Design (ICED13), Design for Harmonies, Vol. 1: Design Processes, Seoul, Korea, 19–22.08. 2013.Google Scholar
Fu, K., Murphy, J., Yang, M., Otto, K., Jensen, D. & Wood, K. 2014 Design-by-analogy: experimental evaluation of a functional analogy search methodology for concept generation improvement. Research in Engineering Design 26 (1), 7795.Google Scholar
Huang, C.-Y., Lin, Y.-H. & Tsai, P.-F. 2015 Developing a rework process for underfilled electronics components via integration of TRIZ and cluster analysis. IEEE Transactions on Components, Packaging and Manufacturing Technology 5 (3), 422438.Google Scholar
Igami, M. 2008 Exploration of the evolution of nanotechnology via mapping of patent applications. Scientometrics 77 (2), 289308.Google Scholar
Ipeirotis, P.2010 Demographics of Mechanical Turk. NYU Center for Digital Economy Research Working Paper CeDER-10-01, New York University.Google Scholar
Jeong, C. & Kim, K. 2014 Creating patents on the new technology using analogy-based patent mining. Expert Systems with Applications 41 (8), 36053614.Google Scholar
Jin, G., Jeong, Y. & Yoon, B. 2015 Technology-driven roadmaps for identifying new product/market opportunities: use of text mining and quality function deployment. Advanced Engineering Informatics 29 (1), 126138.Google Scholar
Jun, S., Park, S.-S. & Jang, D.-S. 2014 Document clustering method using dimension reduction and support vector clustering to overcome sparseness. Expert Systems with Applications 41 (7), 32043212.Google Scholar
Kim, J.-M. & Jun, S. 2015 Graphical causal inference and copula regression model for apple keywords by text mining. Advanced Engineering Informatics 29 (4), 918929.Google Scholar
Kim, Y. G., Suh, J. H. & Park, S. C. 2008 Visualization of patent analysis for emerging technology. Expert Systems with Applications 34 (3), 18041812.CrossRefGoogle Scholar
Kitamura, Y., Kashiwase, M., Fuse, M. & Mizoguchi, R. 2004 Deployment of an ontological framework of functional design knowledge. Advanced Engineering Informatics 18 (2), 115127.Google Scholar
Kukolj, D., Tekic, Z., Nikolic, L., Panjkov, Z., Pokric, M., Drazic, M. & Nemet, D.2012 Comparison of Algorithms for Patent Documents Clusterization. Paper presented at the MIPRO, 2012 Proceedings of the 35th International Convention.Google Scholar
Lee, C., Kang, B. & Shin, J. 2015 Novelty-focused patent mapping for technology opportunity analysis. Technological Forecasting and Social Change 90, 355365.CrossRefGoogle Scholar
Lee, S., Yoon, B. & Park, Y. 2009 An approach to discovering new technology opportunities: keyword-based patent map approach. Technovation 29 (6–7), 481497.CrossRefGoogle Scholar
Li, Z., Atherton, M. & Harrison, D. 2014 Identifying patent conflicts: TRIZ-led patent mapping. World Patent Information 39, 1123.CrossRefGoogle Scholar
Luo, J. 2015 The united innovation process: integrating science, design, and entrepreneurship as sub-processes. Design Science 1, doi:10.1017/dsj.2015.2.CrossRefGoogle Scholar
Moehrle, M. G. & Passing, F. 2016 Applying an anchor based patent mapping approach: basic conception and the case of carbon fiber reinforcements. World Patent Information 45, 19.CrossRefGoogle Scholar
Moser, P. 2013 Patents and innovation: evidence from economic history. The Journal of Economic Perspectives 27 (1), 2344.Google Scholar
Murphy, J., Fu, K., Otto, K., Yang, M., Jensen, D. & Wood, K. 2014 Function based design-by-analogy: a functional vector approach to analogical search. Journal of Mechanical Design 136 (10), 101102.Google Scholar
Park, H., Kim, K., Choi, S. & Yoon, J. 2013 A patent intelligence system for strategic technology planning. Expert Systems with Applications 40 (7), 23732390.Google Scholar
Park, H., Yoon, J. & Kim, K. 2011 Identifying patent infringement using SAO based semantic technological similarities. Scientometrics 90 (2), 515529.CrossRefGoogle Scholar
Ross, J., Irani, L., Silberman, M. S., Zaldivar, A. & Tomlinson, B.2010 Who are the Crowdworkers?: Shifting Demographics in Mechanical Turk. Paper presented at the CHI ’10 Extended Abstracts on Human Factors in Computing Systems, Atlanta, Georgia, USA.CrossRefGoogle Scholar
Ryley, J. F., Saffer, J. & Gibbs, A. 2008 Advanced document retrieval techniques for patent research. World Patent Information 30 (3), 238243.Google Scholar
Segev, A. & Kantola, J. 2012 Identification of trends from patents using self-organizing maps. Expert Systems with Applications 39 (18), 1323513242.Google Scholar
Spangler, S., Ying, C., Kreulen, J., Boyer, S., Griffin, T., Alba, A. & Yan, S. 2011 Exploratory analytics on patent data sets using the SIMPLE platform. World Patent Information 33 (4), 328339.CrossRefGoogle Scholar
Suh, J. H. & Park, S. C.2006 A New Visualization Method for Patent Map: Application to Ubiquitous Computing Technology. Paper presented at the International Conference on Advanced Data Mining and Applications.CrossRefGoogle Scholar
Suh, J. H. & Park, S. C. 2009 Service-oriented Technology Roadmap (SoTRM) using patent map for R&D strategy of service industry. Expert Systems with Applications 36 (3), 67546772.Google Scholar
Tekic, Z., Kukolj, D., Nikolic, L., Drazic, M., Pokric, M., Vitas, M. & Nemet, D.2012 PSALM-Tool for Business Intelligence. Paper presented at the MIPRO, 2012 Proceedings of the 35th International Convention.Google Scholar
Trappey, A. J., Trappey, C. V., Hsu, F.-C. & Hsiao, D. W. 2009 A fuzzy ontological knowledge document clustering methodology. IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics) 39 (3), 806814.Google Scholar
Trappey, C. V., Trappey, A. J. C., Peng, H.-Y., Lin, L.-D. & Wang, T.-M. 2014 A knowledge centric methodology for dental implant technology assessment using ontology based patent analysis and clinical meta-analysis. Advanced Engineering Informatics 28 (2), 153165.CrossRefGoogle Scholar
Trappey, C. V., Trappey, A. J. C. & Wu, C.-Y. 2010 Clustering patents using non-exhaustive overlaps. Journal of Systems Science and Systems Engineering 19 (2), 162181.Google Scholar
Trappey, C. V., Wu, H.-Y., Taghaboni-Dutta, F. & Trappey, A. J. C. 2011 Using patent data for technology forecasting: China RFID patent analysis. Advanced Engineering Informatics 25 (1), 5364.CrossRefGoogle Scholar
Vandevenne, D., Verhaegen, P., Dewulf, S. & Duflou, J. 2016 SEABIRD: scalable search for systematic biologically inspired design. Artificial Intelligence for Engineering Design, Analysis and Manufacturing 30 (1), 7895.Google Scholar
Vasantha, G., Corney, J., Maclachlan, R. & Wodehouse, A.2016 The Analysis and Presentation of Patents to Support Engineering Design. Paper presented at the 7th International Conference on Design Computing and Cognition (DCC16), Chicago, USA.CrossRefGoogle Scholar
Verhaegen, P.-A., Peeters, J., Vandevenne, D., Dewulf, S. & Duflou, J. R. 2011 Effectiveness of the PAnDA ideation tool. Procedia Engineering 9, 6376.Google Scholar
Widodo, A. & Budi, I.2011 Clustering Patent Document in the Field of ICT (Information & Communication Technology). Paper presented at the 2011 International Conference on Semantic Technology and Information Retrieval.Google Scholar
World Intellectual Property Organization2016 WIPO IP Facts and Figures 2016. Retrieved from Geneva, Switzerland. Available at: http://www.wipo.int/edocs/pubdocs/en/wipo_pub_943_2016.pdf.Google Scholar
Xu, C., Zhiyong, P. & Bin, L.2013 Technology and Effect Matrix for Patent Clustering. Paper presented at the Web Information System and Application Conference (WISA), 2013 10th.CrossRefGoogle Scholar
Yan, M. 1993 Representing design knowledge as a network of function, behaviour and structure. Design Studies 14 (3), 314329.Google Scholar
Yoon, B. & Park, Y. 2004 A text-mining-based patent network: analytical tool for high-technology trend. The Journal of High Technology Management Research 15 (1), 3750.Google Scholar
Yoon, B. U., Yoon, C. B. & Park, Y. T. 2002 On the development and application of a self-organizing feature map-based patent map. R&D Management 32 (4), 291300.Google Scholar
Yoon, J., Park, H. & Kim, K. 2012 Identifying technological competition trends for R&D planning using dynamic patent maps: SAO-based content analysis. Scientometrics 94 (1), 313331.Google Scholar
Zhang, L.-w. & Zhu, D.-h.2010 Research of Technical Development Trend and Hot Points Based on Text Mining. Paper presented at the 2nd International Conference on Information Engineering and Computer Science (ICIECS).Google Scholar
Figure 0

Figure 1. Overview of project themes for patent analysis.

Figure 1

Table 1. Summary of text mining and grouping approaches to patent clustering, with highlighted references

Figure 2

Figure 2. Patent clusters labelled using the highest average rank method (Fu et al.2013b).

Figure 3

Figure 3. The patent space defined by Expert 3 (Fu et al.2013b).

Figure 4

Figure 4. A patent landscape based on CPC classification codes (generated with PatentInspiration, http://www.patentinspiration.com).

Figure 5

Figure 5. Articulation between established crowdsourcing platforms and the customised patent cluster platform.

Figure 6

Figure 6. The workflow for assignation, gathering and presentation of crowd results.

Figure 7

Table 2. Participation statistics for the two crowdsourcing platforms

Figure 8

Figure 7. Location of workers participating in patent tasks (generated with BatchGeo, https://batchgeo.com).

Figure 9

Figure 8. Number of clusters versus number of connected patent pairs for the crowd, experts and algorithms.

Figure 10

Table 3. Composition of clusters and connected patent pairs for different approaches

Figure 11

Figure 9. The number of crowd workers identifying identical patent pairs.

Figure 12

Table 4. Patent pairs with highest generated labels by the crowd compared with algorithmic and expert assessment

Figure 13

Figure 10. Percentage of benchmark pairings identified by numbers of crowd workers.

Figure 14

Table 5. Distinct cluster labels generated in various approaches

Figure 15

Table 6. Comparison of cosine similarity scores for labels across different clustering approaches

Figure 16

Table 7. Grammar percentages for labels generated by various approaches

Figure 17

Figure 11. Comparison between the crowd and Fu’s algorithm ranking of patent relevance to the given problem.

Figure 18

Table 8. Patent pair agreement between evaluator and crowd results

Figure 19

Figure 12. Box plot illustrating the relationship between the number of agreed evaluators and the number of crowd workers agreed with a patent pair.

Figure 20

Figure 13. Visualisation of clustering between patents and patent labels for a cropped patent cluster zone.

Figure 21

Table 9. Practical considerations of crowdsourcing clustering with reference to other approaches

Figure 22

Table 10. Phi nominal correlations between crowd workers and other approaches

Figure 23

Table 11. Chi-square values and percentage of correct prediction of unlinked and linked pairs through crowd workers’ results of logistic regression